hash
stringlengths 40
40
| authorName
stringclasses 42
values | authorEmail
stringclasses 41
values | date
timestamp[ms]date 2021-07-26 09:52:55
2025-07-18 10:19:56
| subject
stringlengths 11
116
| diff
stringlengths 0
987k
|
---|---|---|---|---|---|
673fd5b2cdce29644027cea1c3e60d75f52a48cb
|
Sylvain Lesage
| 2023-03-17T10:43:15 |
feat: 🎸 reduce log level to INFO in prod (#954)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index abbcf823..77bd813b 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -71 +71 @@ common:
- logLevel: "DEBUG"
+ logLevel: "INFO"
|
|
d597db6aaa417647681f1991ebef03189570760a
|
Sylvain Lesage
| 2023-03-17T10:32:39 |
feat: 🎸 increase the timeout to Hub from 200ms to 1.5s (#952)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 66b202df..abbcf823 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -146,0 +147,2 @@ admin:
+ # the timeout in seconds for the requests to the Hugging Face Hub.
+ hfTimeoutSeconds: "1.5"
@@ -165,0 +168,2 @@ api:
+ # the timeout in seconds for the requests to the Hugging Face Hub.
+ hfTimeoutSeconds: "1.5"
|
|
9b7f5ab80771a1979dae0086fd02e5b4abba2bbd
|
Sylvain Lesage
| 2023-03-17T10:23:44 |
feat: 🎸 add logs in case of error in /webhook (#951)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 6ef19d95..66b202df 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -71 +71 @@ common:
- logLevel: "INFO"
+ logLevel: "DEBUG"
diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py
index caa9b17e..8c5498b6 100644
--- a/services/api/src/api/routes/webhook.py
+++ b/services/api/src/api/routes/webhook.py
@@ -111,0 +112 @@ def create_webhook_endpoint(
+ logging.info("/webhook: the body could not be parsed as a JSON.")
@@ -117 +118 @@ def create_webhook_endpoint(
- except ValidationError:
+ except ValidationError as e:
@@ -118,0 +120 @@ def create_webhook_endpoint(
+ logging.info(f"/webhook: the JSON body is invalid. JSON: {json}. Error: {e}")
@@ -122,0 +125 @@ def create_webhook_endpoint(
+ logging.warning(f"/webhook: unexpected error while parsing the JSON body is invalid. Error: {e}")
@@ -134 +137 @@ def create_webhook_endpoint(
- except DatasetError:
+ except DatasetError as e:
@@ -135,0 +139,2 @@ def create_webhook_endpoint(
+ dataset = payload["repo"]["name"]
+ logging.debug(f"/webhook: the dataset {dataset} is not supported. JSON: {json}. Error: {e}")
|
|
a2b67d629c0e7ba85851c7e7cea05af76d4e52c6
|
Sylvain Lesage
| 2023-03-16T15:25:40 |
Try to fix the e2e tests in the CI (#944)
|
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml
index 4112de64..90852c97 100644
--- a/.github/workflows/_e2e_tests.yml
+++ b/.github/workflows/_e2e_tests.yml
@@ -27,2 +27,9 @@ jobs:
- WORKER_SLEEP_TIME: "1"
- # hard coded, see e2e/tests/fixtures/hub.py
+ ADMIN_UVICORN_NUM_WORKERS: "2"
+ ADMIN_UVICORN_PORT: "8081"
+ API_HF_JWT_ALGORITHM: "EdDSA"
+ API_HF_JWT_PUBLIC_KEY_URL: "https://hub-ci.huggingface.co/api/keys/jwt"
+ # ^ it's not tested in the e2e tests, but it's good to ensure we're able to fetch it at least.
+ API_HF_TIMEOUT_SECONDS: "10"
+ # ^ allow hub-ci to be slow to answer
+ API_UVICORN_NUM_WORKERS: "2"
+ API_UVICORN_PORT: "8080"
@@ -31 +38,2 @@ jobs:
- PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: "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"
+ # ^ hard coded, see e2e/tests/fixtures/hub.py
+ COMMON_LOG_LEVEL: "DEBUG"
@@ -32,0 +41 @@ jobs:
+ PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: "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"
@@ -35,5 +44 @@ jobs:
- API_UVICORN_NUM_WORKERS: "2"
- API_UVICORN_PORT: "8080"
- ADMIN_UVICORN_NUM_WORKERS: "2"
- ADMIN_UVICORN_PORT: "8081"
- COMMON_LOG_LEVEL: "DEBUG"
+ WORKER_SLEEP_SECONDS: "1"
@@ -61,2 +66,9 @@ jobs:
- QUEUE_SLEEP_TIME: "1"
- # hard coded, see e2e/tests/fixtures/hub.py
+ ADMIN_UVICORN_NUM_WORKERS: "2"
+ ADMIN_UVICORN_PORT: "8081"
+ API_HF_JWT_ALGORITHM: "EdDSA"
+ API_HF_JWT_PUBLIC_KEY_URL: "https://hub-ci.huggingface.co/api/keys/jwt"
+ # ^ it's not tested in the e2e tests, but it's good to ensure we're able to fetch it at least.
+ API_HF_TIMEOUT_SECONDS: "10"
+ # ^ allow hub-ci to be slow to answer
+ API_UVICORN_NUM_WORKERS: "2"
+ API_UVICORN_PORT: "8080"
@@ -65 +77,2 @@ jobs:
- PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: "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"
+ # ^ hard coded, see e2e/tests/fixtures/hub.py
+ COMMON_LOG_LEVEL: "DEBUG"
@@ -66,0 +80 @@ jobs:
+ PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: "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"
@@ -69,9 +83 @@ jobs:
- API_UVICORN_NUM_WORKERS: "2"
- API_UVICORN_PORT: "8080"
- ADMIN_UVICORN_NUM_WORKERS: "2"
- ADMIN_UVICORN_PORT: "8081"
- COMMON_LOG_LEVEL: "DEBUG"
- API_HF_TIMEOUT_SECONDS: "10"
- # JWT: it's not tested in the e2e tests, but it's good to ensure we're able to fetch it at least.
- API_HF_JWT_PUBLIC_KEY_URL: "https://hub-ci.huggingface.co/api/keys/jwt"
- API_HF_JWT_ALGORITHM: "EdDSA"
+ WORKER_SLEEP_SECONDS : "1"
diff --git a/e2e/Makefile b/e2e/Makefile
index 5322ef53..d26328ae 100644
--- a/e2e/Makefile
+++ b/e2e/Makefile
@@ -2,4 +2,9 @@
-export COMPOSE_PROJECT_NAME := e2e
-export PORT_REVERSE_PROXY := 9000
-export MONGO_PORT := 27050
-export QUEUE_SLEEP_TIME := 1
+export ADMIN_UVICORN_NUM_WORKERS := 2
+export ADMIN_UVICORN_PORT := 9081
+export API_HF_JWT_ALGORITHM := EdDSA
+export API_HF_JWT_PUBLIC_KEY_URL := https://hub-ci.huggingface.co/api/keys/jwt
+# ^ it's not tested in the e2e tests, but it's good to ensure we're able to fetch it at least.
+export API_HF_TIMEOUT_SECONDS := 10
+# ^ allow hub-ci to be slow to answer
+export API_UVICORN_NUM_WORKERS := 2
+export API_UVICORN_PORT := 9080
@@ -8,3 +13,2 @@ export COMMON_HF_TOKEN := hf_app_datasets-server_token
-export PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN := 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
-export API_UVICORN_PORT := 9080
-export ADMIN_UVICORN_PORT := 9081
+# ^ hard coded, see e2e/tests/fixtures/hub.py
+export COMMON_LOG_LEVEL := DEBUG
@@ -11,0 +16,3 @@ export FIRST_ROWS_MAX_NUMBER := 4
+export MONGO_PORT := 27050
+export PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN := 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
+export PORT_REVERSE_PROXY := 9000
@@ -13,5 +20,3 @@ export PROMETHEUS_MULTIPROC_DIR := /tmp
-export ADMIN_UVICORN_NUM_WORKERS := 2
-export API_UVICORN_NUM_WORKERS := 2
-export COMMON_LOG_LEVEL := DEBUG
-export API_HF_TIMEOUT_SECONDS := 10
-# ^ allow hub-ci to be slow to answer
+export WORKER_SLEEP_SECONDS := 1
+
+export COMPOSE_PROJECT_NAME := e2e
|
|
1d917ebbc3e895103a7fb04bb1ca8c515baf7b4a
|
Andrea Francis Soria Jimenez
| 2023-03-16T14:34:41 |
dataset-level split names from dataset info (#937)
|
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 583bf2c1..a511b792 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -117,0 +118,4 @@ class ProcessingGraphConfig:
+ "dataset-split-names-from-dataset-info": {
+ "input_type": "dataset",
+ "requires": "/split-names-from-dataset-info",
+ },
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index 4e4693cc..30c0be9b 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -23,0 +24 @@ def test_default_graph() -> None:
+ dataset_split_names_from_dataset_info = graph.get_step("dataset-split-names-from-dataset-info")
@@ -67 +68 @@ def test_default_graph() -> None:
- assert split_names_from_dataset_info.children == []
+ assert split_names_from_dataset_info.children == [dataset_split_names_from_dataset_info]
@@ -84,0 +86,9 @@ def test_default_graph() -> None:
+ assert dataset_split_names_from_dataset_info is not None
+ assert dataset_split_names_from_dataset_info.parent is split_names_from_dataset_info
+ assert dataset_split_names_from_dataset_info.children == []
+ assert dataset_split_names_from_dataset_info.get_ancestors() == [
+ parquet_and_dataset_info,
+ dataset_info,
+ split_names_from_dataset_info,
+ ]
+
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index adac6782..346cc3a6 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -113 +113 @@ class EndpointConfig:
- "dataset": ["/splits", "dataset-split-names-from-streaming"],
+ "dataset": ["/splits", "dataset-split-names-from-streaming", "dataset-split-names-from-dataset-info"],
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index 6737a5f5..2fa834f5 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -36 +36 @@ def test_endpoints_definition() -> None:
- assert len(splits["dataset"]) == 2 # Has two processing steps
+ assert len(splits["dataset"]) == 3 # Has three processing steps
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 633916f9..c9ad8081 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -22,0 +23,3 @@ from worker.job_runners.config_size import ConfigSizeJobRunner
+from worker.job_runners.dataset.split_names_from_dataset_info import (
+ DatasetSplitNamesFromDatasetInfoJobRunner,
+)
@@ -154,0 +158,7 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ if job_type == DatasetSplitNamesFromDatasetInfoJobRunner.get_job_type():
+ return DatasetSplitNamesFromDatasetInfoJobRunner(
+ job_info=job_info,
+ processing_step=processing_step,
+ common_config=self.app_config.common,
+ worker_config=self.app_config.worker,
+ )
@@ -167,0 +178 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ DatasetSplitNamesFromDatasetInfoJobRunner.get_job_type(),
diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py
new file mode 100644
index 00000000..700fb262
--- /dev/null
+++ b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py
@@ -0,0 +1,158 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+from http import HTTPStatus
+from typing import Any, List, Literal, Mapping, Optional, Tuple
+
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
+
+from worker.job_runner import JobResult, JobRunner, JobRunnerError
+from worker.utils import (
+ ConfigItem,
+ DatasetSplitNamesResponse,
+ FailedConfigItem,
+ SplitItem,
+)
+
+DatasetSplitNamesFromDatasetInfoErrorCode = Literal[
+ "PreviousStepStatusError",
+ "PreviousStepFormatError",
+]
+
+
+class DatasetSplitNamesFromDatasetInfoJobRunnerError(JobRunnerError):
+ """Base class for exceptions in this module."""
+
+ def __init__(
+ self,
+ message: str,
+ status_code: HTTPStatus,
+ code: DatasetSplitNamesFromDatasetInfoErrorCode,
+ cause: Optional[BaseException] = None,
+ disclose_cause: bool = False,
+ ):
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
+
+
+class PreviousStepStatusError(DatasetSplitNamesFromDatasetInfoJobRunnerError):
+ """Raised when the previous step gave an error. The job should not have been created."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False)
+
+
+class PreviousStepFormatError(DatasetSplitNamesFromDatasetInfoJobRunnerError):
+ """Raised when the content of the previous step has not the expected format."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False)
+
+
+def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tuple[DatasetSplitNamesResponse, float]:
+ """
+ Get the response of /splits for one specific dataset on huggingface.co
+ computed from responses cached in /split-names-from-dataset-info step.
+ Args:
+ dataset (`str`):
+ A namespace (user or an organization) and a repo name separated by a `/`.
+ Returns:
+ `DatasetSplitNamesResponse`: An object with a list of split names for the dataset [splits],
+ a list of pending configs to be processed [pending] and the list of errors [failed] by config.
+ <Tip>
+ Raises the following errors:
+ - [`~job_runners.dataset_split_names_from_dataset_info.PreviousStepStatusError`]
+ If the the previous step gave an error.
+ - [`~job_runners.dataset_split_names_from_dataset_info.PreviousStepFormatError`]
+ If the content of the previous step has not the expected format
+ - [`~libcommon.dataset.DatasetNotFoundError`]
+ If previous step content was not found for the dataset
+ </Tip>
+ """
+ logging.info(f"get dataset split names from dataset info for dataset={dataset}")
+
+ try:
+ response = get_response(kind="/dataset-info", dataset=dataset)
+ dataset_info_content = response["content"]["dataset_info"]
+ except DoesNotExist as e:
+ raise DatasetNotFoundError("No response found in previous step for this dataset: '/dataset-info'.", e) from e
+ except KeyError as e:
+ raise PreviousStepFormatError("Previous step '/dataset-info' did not return the expected content.") from e
+
+ if response["http_status"] != HTTPStatus.OK:
+ raise PreviousStepStatusError(
+ f"Previous step gave an error: {response['http_status']}. This job should not have been created."
+ )
+
+ try:
+ splits: List[SplitItem] = []
+ pending: List[ConfigItem] = []
+ failed: List[FailedConfigItem] = []
+ total = 0
+ for config in dataset_info_content.keys():
+ total += 1
+ try:
+ response = get_response(kind="/split-names-from-dataset-info", dataset=dataset, config=config)
+ except DoesNotExist:
+ logging.debug("No response found in previous step '/split-names-from-dataset-info' for this dataset.")
+ pending.append(ConfigItem({"dataset": dataset, "config": config}))
+ continue
+ if response["http_status"] != HTTPStatus.OK:
+ logging.debug(f"Previous step gave an error: {response['http_status']}.")
+ failed.append(
+ FailedConfigItem(
+ {
+ "dataset": dataset,
+ "config": config,
+ "error": response["content"],
+ }
+ )
+ )
+ continue
+ splits.extend(
+ [
+ SplitItem({"dataset": dataset, "config": config, "split": split_content["split"]})
+ for split_content in response["content"]["splits"]
+ ]
+ )
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e
+
+ progress = (total - len(pending)) / total if total else 1.0
+
+ return (
+ DatasetSplitNamesResponse(
+ {
+ "splits": splits,
+ "pending": pending,
+ "failed": failed,
+ }
+ ),
+ progress,
+ )
+
+
+class DatasetSplitNamesFromDatasetInfoJobRunner(JobRunner):
+ @staticmethod
+ def get_job_type() -> str:
+ return "dataset-split-names-from-dataset-info"
+
+ @staticmethod
+ def get_job_runner_version() -> int:
+ return 1
+
+ def compute(self) -> JobResult:
+ if self.dataset is None:
+ raise ValueError("dataset is required")
+ response_content, progress = compute_dataset_split_names_from_dataset_info_response(dataset=self.dataset)
+ return JobResult(response_content, progress=progress)
+
+ def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]:
+ """Get the set of new splits, from the content created by the compute."""
+ return {
+ SplitFullName(dataset=split_item["dataset"], config=split_item["config"], split=split_item["split"])
+ for split_item in content["splits"]
+ }
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index 35e6ac92..906f7c91 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -4 +4 @@
-from typing import List, Optional, TypedDict
+from typing import Any, List, Mapping, Optional, TypedDict
@@ -20,0 +21,10 @@ class SplitsList(TypedDict):
+
+
+class FailedConfigItem(ConfigItem):
+ error: Mapping[str, Any]
+
+
+class DatasetSplitNamesResponse(TypedDict):
+ splits: List[SplitItem]
+ pending: List[ConfigItem]
+ failed: List[FailedConfigItem]
diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py
new file mode 100644
index 00000000..01adbf29
--- /dev/null
+++ b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py
@@ -0,0 +1,246 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from http import HTTPStatus
+from typing import Any, Callable
+
+import pytest
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import SplitFullName, upsert_response
+
+from worker.config import AppConfig
+from worker.job_runners.dataset.split_names_from_dataset_info import (
+ DatasetSplitNamesFromDatasetInfoJobRunner,
+ PreviousStepFormatError,
+)
+
+GetJobRunner = Callable[[str, AppConfig, bool], DatasetSplitNamesFromDatasetInfoJobRunner]
+
+
[email protected]
+def get_job_runner(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> DatasetSplitNamesFromDatasetInfoJobRunner:
+ return DatasetSplitNamesFromDatasetInfoJobRunner(
+ job_info={
+ "type": DatasetSplitNamesFromDatasetInfoJobRunner.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ common_config=app_config.common,
+ worker_config=app_config.worker,
+ processing_step=ProcessingStep(
+ name=DatasetSplitNamesFromDatasetInfoJobRunner.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ )
+
+ return _get_job_runner
+
+
[email protected](
+ "dataset,split_names_from_dataset_info,expected_content,progress",
+ [
+ (
+ "pending_response",
+ [
+ {
+ "config": "config_a",
+ "response": {
+ "splits": [
+ {
+ "dataset": "pending_response",
+ "config": "config_a",
+ "split": "split_a",
+ }
+ ]
+ },
+ }
+ ],
+ {
+ "splits": [
+ {
+ "dataset": "pending_response",
+ "config": "config_a",
+ "split": "split_a",
+ },
+ ],
+ "pending": [{"dataset": "pending_response", "config": "config_b"}],
+ "failed": [],
+ },
+ 0.5,
+ ),
+ (
+ "complete",
+ [
+ {
+ "config": "config_a",
+ "response": {
+ "splits": [
+ {
+ "dataset": "complete",
+ "config": "config_a",
+ "split": "split_a",
+ }
+ ]
+ },
+ },
+ {
+ "config": "config_b",
+ "response": {
+ "splits": [
+ {
+ "dataset": "complete",
+ "config": "config_b",
+ "split": "split_b",
+ }
+ ]
+ },
+ },
+ ],
+ {
+ "splits": [
+ {
+ "dataset": "complete",
+ "config": "config_a",
+ "split": "split_a",
+ },
+ {
+ "dataset": "complete",
+ "config": "config_b",
+ "split": "split_b",
+ },
+ ],
+ "pending": [],
+ "failed": [],
+ },
+ 1,
+ ),
+ ],
+)
+def test_compute_progress(
+ app_config: AppConfig,
+ get_job_runner: GetJobRunner,
+ dataset: str,
+ split_names_from_dataset_info: Any,
+ expected_content: Any,
+ progress: float,
+) -> None:
+ upsert_response(
+ kind="/dataset-info",
+ dataset=dataset,
+ content={"dataset_info": {"config_a": {"splits": {}}, "config_b": {"splits": {}}}},
+ http_status=HTTPStatus.OK,
+ )
+ for config in split_names_from_dataset_info:
+ upsert_response(
+ kind="/split-names-from-dataset-info",
+ dataset=dataset,
+ config=config["config"],
+ content=config["response"],
+ http_status=HTTPStatus.OK,
+ )
+ job_runner = get_job_runner(dataset, app_config, False)
+ response = job_runner.compute()
+ assert response.content == expected_content
+ assert response.progress == progress
+
+
+def test_compute_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "error"
+ config = "error"
+ upsert_response(
+ kind="/dataset-info",
+ dataset=dataset,
+ content={"dataset_info": {config: {"splits": {}}}},
+ http_status=HTTPStatus.OK,
+ )
+ upsert_response(
+ kind="/split-names-from-dataset-info",
+ dataset=dataset,
+ config=config,
+ content={},
+ http_status=HTTPStatus.INTERNAL_SERVER_ERROR,
+ )
+ job_runner = get_job_runner(dataset, app_config, False)
+ response = job_runner.compute()
+ assert response.content == {
+ "splits": [],
+ "failed": [{"dataset": dataset, "config": config, "error": {}}],
+ "pending": [],
+ }
+ assert response.progress == 1.0
+
+
+def test_compute_format_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "error"
+ config = "error"
+ upsert_response(
+ kind="/dataset-info",
+ dataset=dataset,
+ content={"dataset_info": {config: {"splits": {}}}},
+ http_status=HTTPStatus.OK,
+ )
+ upsert_response(
+ kind="/split-names-from-dataset-info",
+ dataset=dataset,
+ config=config,
+ content={"wrong_format": []},
+ http_status=HTTPStatus.OK,
+ )
+ job_runner = get_job_runner(dataset, app_config, False)
+ with pytest.raises(PreviousStepFormatError):
+ job_runner.compute()
+
+
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "doesnotexist"
+ job_runner = get_job_runner(dataset, app_config, False)
+ with pytest.raises(DatasetNotFoundError):
+ job_runner.compute()
+
+
+def test_get_new_splits(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "dataset"
+ job_runner = get_job_runner(dataset, app_config, False)
+ content = {
+ "splits": [
+ {
+ "dataset": dataset,
+ "config": "config_a",
+ "split": "split_a",
+ },
+ {
+ "dataset": dataset,
+ "config": "config_b",
+ "split": "split_b",
+ },
+ ],
+ "pending": [],
+ "failed": [],
+ }
+ expected = {
+ SplitFullName(dataset=dataset, config="config_a", split="split_a"),
+ SplitFullName(dataset=dataset, config="config_b", split="split_b"),
+ }
+ new_splits = job_runner.get_new_splits(content=content)
+ assert new_splits
+ assert new_splits == expected
|
|
648f5114df5fc7879d13e60f088ef6839dab304a
|
Andrea Francis Soria Jimenez
| 2023-03-16T14:01:57 |
dataset-level split names from streaming (#936)
|
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index ca702637..583bf2c1 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -113,0 +114,4 @@ class ProcessingGraphConfig:
+ "dataset-split-names-from-streaming": {
+ "input_type": "dataset",
+ "requires": "/split-names-from-streaming",
+ },
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index 620622be..4e4693cc 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -22,0 +23 @@ def test_default_graph() -> None:
+ dataset_split_names_from_streaming = graph.get_step("dataset-split-names-from-streaming")
@@ -31 +32 @@ def test_default_graph() -> None:
- assert split_names_from_streaming.children == [first_rows]
+ assert split_names_from_streaming.children == [first_rows, dataset_split_names_from_streaming]
@@ -78,0 +80,5 @@ def test_default_graph() -> None:
+ assert dataset_split_names_from_streaming is not None
+ assert dataset_split_names_from_streaming.parent is split_names_from_streaming
+ assert dataset_split_names_from_streaming.children == []
+ assert dataset_split_names_from_streaming.get_ancestors() == [config_names, split_names_from_streaming]
+
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 920b7146..adac6782 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -113 +113 @@ class EndpointConfig:
- "dataset": ["/splits"],
+ "dataset": ["/splits", "dataset-split-names-from-streaming"],
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index aef6da7a..6737a5f5 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -36,2 +36,2 @@ def test_endpoints_definition() -> None:
- assert len(splits["dataset"]) == 1 # Only has one processing step
- assert len(splits["config"]) == 2 # Has two processing step
+ assert len(splits["dataset"]) == 2 # Has two processing steps
+ assert len(splits["config"]) == 2 # Has two processing steps
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index cdb1d944..633916f9 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -13,0 +14,6 @@ from worker.job_runner import JobRunner
+from worker.job_runners.config.split_names_from_dataset_info import (
+ SplitNamesFromDatasetInfoJobRunner,
+)
+from worker.job_runners.config.split_names_from_streaming import (
+ SplitNamesFromStreamingJobRunner,
+)
@@ -16,0 +23,3 @@ from worker.job_runners.config_size import ConfigSizeJobRunner
+from worker.job_runners.dataset.split_names_from_streaming import (
+ DatasetSplitNamesFromStreamingJobRunner,
+)
@@ -22,6 +30,0 @@ from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoJob
-from worker.job_runners.split_names_from_dataset_info import (
- SplitNamesFromDatasetInfoJobRunner,
-)
-from worker.job_runners.split_names_from_streaming import (
- SplitNamesFromStreamingJobRunner,
-)
@@ -144,0 +148,7 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ if job_type == DatasetSplitNamesFromStreamingJobRunner.get_job_type():
+ return DatasetSplitNamesFromStreamingJobRunner(
+ job_info=job_info,
+ processing_step=processing_step,
+ common_config=self.app_config.common,
+ worker_config=self.app_config.worker,
+ )
@@ -156,0 +167 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ DatasetSplitNamesFromStreamingJobRunner.get_job_type(),
diff --git a/services/worker/src/worker/job_runners/config/__init__.py b/services/worker/src/worker/job_runners/config/__init__.py
new file mode 100644
index 00000000..1e9d0c5a
--- /dev/null
+++ b/services/worker/src/worker/job_runners/config/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py
similarity index 90%
rename from services/worker/src/worker/job_runners/split_names_from_dataset_info.py
rename to services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py
index 975e7e1e..a94762b7 100644
--- a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
+++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Any, List, Literal, Mapping, Optional, TypedDict
+from typing import Any, List, Literal, Mapping, Optional
@@ -12,0 +13 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.utils import SplitItem, SplitsList
@@ -50,10 +50,0 @@ class PreviousStepFormatError(SplitNamesFromDatasetInfoJobRunnerError):
-class SplitNameItem(TypedDict):
- dataset: str
- config: str
- split: str
-
-
-class SplitNamesFromDatasetInfoResponseContent(TypedDict):
- splits: List[SplitNameItem]
-
-
@@ -63 +54 @@ def compute_split_names_from_dataset_info_response(
-) -> SplitNamesFromDatasetInfoResponseContent:
+) -> SplitsList:
@@ -78 +69 @@ def compute_split_names_from_dataset_info_response(
- `SplitNamesFromDatasetInfoResponseContent`: An object with the list of split names for the dataset and config.
+ `SplitsList`: An object with the list of split names for the dataset and config.
@@ -104 +95 @@ def compute_split_names_from_dataset_info_response(
- split_name_items: List[SplitNameItem] = [
+ split_name_items: List[SplitItem] = [
@@ -108 +99 @@ def compute_split_names_from_dataset_info_response(
- return {"splits": split_name_items}
+ return SplitsList({"splits": split_name_items})
diff --git a/services/worker/src/worker/job_runners/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
similarity index 90%
rename from services/worker/src/worker/job_runners/split_names_from_streaming.py
rename to services/worker/src/worker/job_runners/config/split_names_from_streaming.py
index b89e6eaf..0d2c7008 100644
--- a/services/worker/src/worker/job_runners/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Any, List, Literal, Mapping, Optional, TypedDict, Union
+from typing import Any, List, Literal, Mapping, Optional, Union
@@ -13,0 +14 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.utils import SplitItem, SplitsList
@@ -51,10 +51,0 @@ class EmptyDatasetError(SplitNamesFromStreamingJobRunnerError):
-class SplitNameItem(TypedDict):
- dataset: str
- config: str
- split: str
-
-
-class SplitNamesFromStreamingResponseContent(TypedDict):
- splits: List[SplitNameItem]
-
-
@@ -65 +56 @@ def compute_split_names_from_streaming_response(
-) -> SplitNamesFromStreamingResponseContent:
+) -> SplitsList:
@@ -88 +79 @@ def compute_split_names_from_streaming_response(
- `SplitNamesFromStreamingResponseContent`: An object with the list of split names for the dataset and config.
+ `SplitsList`: An object with the list of split names for the dataset and config.
@@ -101 +92 @@ def compute_split_names_from_streaming_response(
- split_name_items: List[SplitNameItem] = [
+ split_name_items: List[SplitItem] = [
@@ -112 +103 @@ def compute_split_names_from_streaming_response(
- return {"splits": split_name_items}
+ return SplitsList({"splits": split_name_items})
diff --git a/services/worker/src/worker/job_runners/dataset/__init__.py b/services/worker/src/worker/job_runners/dataset/__init__.py
new file mode 100644
index 00000000..1e9d0c5a
--- /dev/null
+++ b/services/worker/src/worker/job_runners/dataset/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py
new file mode 100644
index 00000000..ac9ea187
--- /dev/null
+++ b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py
@@ -0,0 +1,168 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+from http import HTTPStatus
+from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict
+
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
+
+from worker.job_runner import JobResult, JobRunner, JobRunnerError
+from worker.utils import ConfigItem, SplitItem
+
+DatasetSplitNamesFromStreamingJobRunnerErrorCode = Literal[
+ "PreviousStepStatusError", "PreviousStepFormatError", "ResponseNotReady"
+]
+
+
+class DatasetSplitNamesFromStreamingJobRunnerError(JobRunnerError):
+ """Base class for dataset split from streaming names job runner exceptions."""
+
+ def __init__(
+ self,
+ message: str,
+ status_code: HTTPStatus,
+ code: DatasetSplitNamesFromStreamingJobRunnerErrorCode,
+ cause: Optional[BaseException] = None,
+ disclose_cause: bool = False,
+ ):
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
+
+
+class PreviousStepStatusError(DatasetSplitNamesFromStreamingJobRunnerError):
+ """Raised when the previous step gave an error. The job should not have been created."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False)
+
+
+class PreviousStepFormatError(DatasetSplitNamesFromStreamingJobRunnerError):
+ """Raised when the content of the previous step has not the expected format."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False)
+
+
+class ResponseNotReadyError(DatasetSplitNamesFromStreamingJobRunnerError):
+ """Raised when the response has not been processed yet from any source."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseNotReady")
+
+
+class FailedConfigItem(ConfigItem):
+ error: Mapping[str, Any]
+
+
+class DatasetSplitNamesFromStreamingResponse(TypedDict):
+ splits: List[SplitItem]
+ pending: List[ConfigItem]
+ failed: List[FailedConfigItem]
+
+
+def compute_dataset_split_names_from_streaming_response(
+ dataset: str,
+) -> Tuple[DatasetSplitNamesFromStreamingResponse, float]:
+ """
+ Get the response of /splits for one specific dataset on huggingface.co
+ computed from responses cached in /split-names-from-streaming step.
+
+ Args:
+ dataset (`str`):
+ A namespace (user or an organization) and a repo name separated by a `/`.
+ Returns:
+ `DatasetSplitNamesFromStreamingResponse`: An object with a list of split names for the dataset [splits],
+ a list of pending configs to be processed [pending] and the list of errors [failed] by config.
+ <Tip>
+ Raises the following errors:
+ - [`~job_runners.dataset_split_names_from_streaming.PreviousStepStatusError`]
+ If the the previous step gave an error.
+ - [`~job_runners.dataset_split_names_from_streaming.PreviousStepFormatError`]
+ If the content of the previous step has not the expected format
+ - [`~libcommon.dataset.DatasetNotFoundError`]
+ If previous step content was not found for the dataset
+ </Tip>
+ """
+ logging.info(f"get dataset split names from dataset info for dataset={dataset}")
+ try:
+ config_names = get_response(kind="/config-names", dataset=dataset)
+ config_content = config_names["content"]["config_names"]
+ except DoesNotExist as e:
+ raise DatasetNotFoundError("No response found in previous step '/config-names' for this dataset.", e) from e
+ except KeyError as e:
+ raise PreviousStepFormatError("Previous step '/config-names' did not return the expected content.") from e
+
+ if config_names["http_status"] != HTTPStatus.OK:
+ raise PreviousStepStatusError(
+ f"Previous step gave an error: {config_names['http_status']}. This job should not have been created."
+ )
+ try:
+ splits: List[SplitItem] = []
+ pending: List[ConfigItem] = []
+ failed: List[FailedConfigItem] = []
+ total = 0
+ for config_item in config_content:
+ config = config_item["config"]
+ total += 1
+ try:
+ response = get_response(kind="/split-names-from-streaming", dataset=dataset, config=config)
+ except DoesNotExist:
+ logging.debug("No response found in previous step '/split-names-from-streaming' for this dataset.")
+ pending.append(ConfigItem({"dataset": dataset, "config": config}))
+ continue
+ if response["http_status"] != HTTPStatus.OK:
+ logging.debug(f"Previous step gave an error: {response['http_status']}.")
+ failed.append(
+ FailedConfigItem(
+ {
+ "dataset": dataset,
+ "config": config,
+ "error": response["content"],
+ }
+ )
+ )
+ continue
+ splits.extend(
+ [
+ SplitItem({"dataset": dataset, "config": config, "split": split_content["split"]})
+ for split_content in response["content"]["splits"]
+ ]
+ )
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e
+
+ progress = (total - len(pending)) / total if total else 1.0
+
+ return (
+ DatasetSplitNamesFromStreamingResponse(
+ {
+ "splits": splits,
+ "pending": pending,
+ "failed": failed,
+ }
+ ),
+ progress,
+ )
+
+
+class DatasetSplitNamesFromStreamingJobRunner(JobRunner):
+ @staticmethod
+ def get_job_type() -> str:
+ return "dataset-split-names-from-streaming"
+
+ @staticmethod
+ def get_job_runner_version() -> int:
+ return 1
+
+ def compute(self) -> JobResult:
+ if self.dataset is None:
+ raise ValueError("dataset is required")
+ response_content, progress = compute_dataset_split_names_from_streaming_response(dataset=self.dataset)
+ return JobResult(response_content, progress=progress)
+
+ def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]:
+ """Get the set of new splits, from the content created by the compute."""
+ return {SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["splits"]}
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
new file mode 100644
index 00000000..35e6ac92
--- /dev/null
+++ b/services/worker/src/worker/utils.py
@@ -0,0 +1,20 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from typing import List, Optional, TypedDict
+
+
+class DatasetItem(TypedDict):
+ dataset: str
+
+
+class ConfigItem(DatasetItem):
+ config: Optional[str]
+
+
+class SplitItem(ConfigItem):
+ split: Optional[str]
+
+
+class SplitsList(TypedDict):
+ splits: List[SplitItem]
diff --git a/services/worker/tests/job_runners/config/__init__.py b/services/worker/tests/job_runners/config/__init__.py
new file mode 100644
index 00000000..1e9d0c5a
--- /dev/null
+++ b/services/worker/tests/job_runners/config/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
diff --git a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py
similarity index 98%
rename from services/worker/tests/job_runners/test_split_names_from_dataset_info.py
rename to services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py
index 9c68913a..bfd6dbfa 100644
--- a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
+++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py
@@ -16 +16 @@ from worker.config import AppConfig
-from worker.job_runners.split_names_from_dataset_info import (
+from worker.job_runners.config.split_names_from_dataset_info import (
diff --git a/services/worker/tests/job_runners/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
similarity index 97%
rename from services/worker/tests/job_runners/test_split_names_from_streaming.py
rename to services/worker/tests/job_runners/config/test_split_names_from_streaming.py
index 80410d76..e457e9d6 100644
--- a/services/worker/tests/job_runners/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
@@ -16 +16 @@ from worker.config import AppConfig
-from worker.job_runners.split_names_from_streaming import (
+from worker.job_runners.config.split_names_from_streaming import (
@@ -21 +21 @@ from worker.resources import LibrariesResource
-from ..fixtures.hub import HubDatasets, get_default_config_split
+from ...fixtures.hub import HubDatasets, get_default_config_split
diff --git a/services/worker/tests/job_runners/dataset/__init__.py b/services/worker/tests/job_runners/dataset/__init__.py
new file mode 100644
index 00000000..1e9d0c5a
--- /dev/null
+++ b/services/worker/tests/job_runners/dataset/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py
new file mode 100644
index 00000000..48490bd3
--- /dev/null
+++ b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py
@@ -0,0 +1,268 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from http import HTTPStatus
+from typing import Any, Callable
+
+import pytest
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import SplitFullName, upsert_response
+
+from worker.config import AppConfig
+from worker.job_runners.dataset.split_names_from_streaming import (
+ DatasetSplitNamesFromStreamingJobRunner,
+ PreviousStepFormatError,
+)
+
+GetJobRunner = Callable[[str, AppConfig, bool], DatasetSplitNamesFromStreamingJobRunner]
+
+
[email protected]
+def get_job_runner(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> DatasetSplitNamesFromStreamingJobRunner:
+ return DatasetSplitNamesFromStreamingJobRunner(
+ job_info={
+ "type": DatasetSplitNamesFromStreamingJobRunner.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ common_config=app_config.common,
+ worker_config=app_config.worker,
+ processing_step=ProcessingStep(
+ name=DatasetSplitNamesFromStreamingJobRunner.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ )
+
+ return _get_job_runner
+
+
[email protected](
+ "dataset,split_names_from_streaming,expected_content,progress",
+ [
+ (
+ "pending_response",
+ [
+ {
+ "config": "config_a",
+ "response": {
+ "splits": [
+ {
+ "dataset": "pending_response",
+ "config": "config_a",
+ "split": "split_a",
+ }
+ ]
+ },
+ }
+ ],
+ {
+ "splits": [
+ {
+ "dataset": "pending_response",
+ "config": "config_a",
+ "split": "split_a",
+ },
+ ],
+ "pending": [{"dataset": "pending_response", "config": "config_b"}],
+ "failed": [],
+ },
+ 0.5,
+ ),
+ (
+ "complete",
+ [
+ {
+ "config": "config_a",
+ "response": {
+ "splits": [
+ {
+ "dataset": "complete",
+ "config": "config_a",
+ "split": "split_a",
+ }
+ ]
+ },
+ },
+ {
+ "config": "config_b",
+ "response": {
+ "splits": [
+ {
+ "dataset": "complete",
+ "config": "config_b",
+ "split": "split_b",
+ }
+ ]
+ },
+ },
+ ],
+ {
+ "splits": [
+ {
+ "dataset": "complete",
+ "config": "config_a",
+ "split": "split_a",
+ },
+ {
+ "dataset": "complete",
+ "config": "config_b",
+ "split": "split_b",
+ },
+ ],
+ "pending": [],
+ "failed": [],
+ },
+ 1,
+ ),
+ ],
+)
+def test_compute_progress(
+ app_config: AppConfig,
+ get_job_runner: GetJobRunner,
+ dataset: str,
+ split_names_from_streaming: Any,
+ expected_content: Any,
+ progress: float,
+) -> None:
+ upsert_response(
+ kind="/config-names",
+ dataset=dataset,
+ content={
+ "config_names": [
+ {
+ "dataset": dataset,
+ "config": "config_a",
+ },
+ {"dataset": dataset, "config": "config_b"},
+ ]
+ },
+ http_status=HTTPStatus.OK,
+ )
+ for config in split_names_from_streaming:
+ upsert_response(
+ kind="/split-names-from-streaming",
+ dataset=dataset,
+ config=config["config"],
+ content=config["response"],
+ http_status=HTTPStatus.OK,
+ )
+ job_runner = get_job_runner(dataset, app_config, False)
+ response = job_runner.compute()
+ assert response.content == expected_content
+ assert response.progress == progress
+
+
+def test_compute_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "error"
+ config = "error"
+ upsert_response(
+ kind="/config-names",
+ dataset=dataset,
+ content={
+ "config_names": [
+ {
+ "dataset": dataset,
+ "config": config,
+ }
+ ]
+ },
+ http_status=HTTPStatus.OK,
+ )
+ upsert_response(
+ kind="/split-names-from-streaming",
+ dataset=dataset,
+ config=config,
+ content={},
+ http_status=HTTPStatus.INTERNAL_SERVER_ERROR,
+ )
+ job_runner = get_job_runner(dataset, app_config, False)
+ response = job_runner.compute()
+ assert response.content == {
+ "splits": [],
+ "failed": [{"dataset": dataset, "config": config, "error": {}}],
+ "pending": [],
+ }
+ assert response.progress == 1.0
+
+
+def test_compute_format_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "error"
+ config = "error"
+ upsert_response(
+ kind="/config-names",
+ dataset=dataset,
+ content={
+ "config_names": [
+ {
+ "dataset": dataset,
+ "config": config,
+ }
+ ]
+ },
+ http_status=HTTPStatus.OK,
+ )
+ upsert_response(
+ kind="/split-names-from-streaming",
+ dataset=dataset,
+ config=config,
+ content={"wrong_format": []},
+ http_status=HTTPStatus.OK,
+ )
+ job_runner = get_job_runner(dataset, app_config, False)
+ with pytest.raises(PreviousStepFormatError):
+ job_runner.compute()
+
+
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "doesnotexist"
+ job_runner = get_job_runner(dataset, app_config, False)
+ with pytest.raises(DatasetNotFoundError):
+ job_runner.compute()
+
+
+def test_get_new_splits(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "dataset"
+ job_runner = get_job_runner(dataset, app_config, False)
+ content = {
+ "splits": [
+ {
+ "dataset": dataset,
+ "config": "config_a",
+ "split": "split_a",
+ },
+ {
+ "dataset": dataset,
+ "config": "config_b",
+ "split": "split_b",
+ },
+ ],
+ "pending": [],
+ "failed": [],
+ }
+ expected = {
+ SplitFullName(dataset=dataset, config="config_a", split="split_a"),
+ SplitFullName(dataset=dataset, config="config_b", split="split_b"),
+ }
+ new_splits = job_runner.get_new_splits(content=content)
+ assert new_splits
+ assert new_splits == expected
|
|
e1e602f66c1d31d2438b145c60a05ea92ecec1d2
|
Sylvain Lesage
| 2023-03-16T13:35:22 |
ci: 🎡 use the public key of the CI Hub (#941)
|
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml
index 75e4bf11..4112de64 100644
--- a/.github/workflows/_e2e_tests.yml
+++ b/.github/workflows/_e2e_tests.yml
@@ -75 +75,2 @@ jobs:
- API_HF_JWT_PUBLIC_KEY_URL: "https://huggingface.co/api/keys/jwt"
+ # JWT: it's not tested in the e2e tests, but it's good to ensure we're able to fetch it at least.
+ API_HF_JWT_PUBLIC_KEY_URL: "https://hub-ci.huggingface.co/api/keys/jwt"
|
|
caca20fbc0dbe799dd83f0c89f73a6175990aa93
|
Sylvain Lesage
| 2023-03-16T10:17:05 |
feat: 🎸 revert log level to INFO (#940)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 66b202df..6ef19d95 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -71 +71 @@ common:
- logLevel: "DEBUG"
+ logLevel: "INFO"
|
|
d6d5a64996230880d45c5d34fd19ad2578ff1b69
|
Sylvain Lesage
| 2023-03-16T09:39:13 |
Increase logs to debug jwt issue (#939)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 6ef19d95..66b202df 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -71 +71 @@ common:
- logLevel: "INFO"
+ logLevel: "DEBUG"
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index 029276ca..fadabe51 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -73,0 +74,3 @@ def auth_check(
+ logging.debug(
+ f"Looking if jwt is passed in request headers {request.headers} to bypass authentication."
+ )
@@ -75,4 +78,9 @@ def auth_check(
- token = request.headers.get(HEADER)
- if token and is_jwt_valid(
- dataset=dataset, token=token, public_key=hf_jwt_public_key, algorithm=hf_jwt_algorithm
- ):
+ if token := request.headers.get(HEADER):
+ if is_jwt_valid(
+ dataset=dataset, token=token, public_key=hf_jwt_public_key, algorithm=hf_jwt_algorithm
+ ):
+ logging.debug(
+ f"By-passing the authentication step, because a valid JWT was passed in header: '{HEADER}'"
+ f" for dataset {dataset}. JWT was: {token}"
+ )
+ return True
@@ -80,2 +88,7 @@ def auth_check(
- f"By-passing the authentication step, because a valid JWT was passed in header: '{HEADER}' for"
- f" dataset {dataset}. JWT was: {token}"
+ f"Error while validating the JWT passed in header: '{HEADER}' for dataset {dataset}. Trying"
+ f" with the following authentication mechanisms. JWT was: {token}"
+ )
+ else:
+ logging.debug(
+ f"No JWT was passed in header: '{HEADER}' for dataset {dataset}. Trying with the following"
+ " authentication mechanisms."
@@ -83,5 +95,0 @@ def auth_check(
- return True
- logging.debug(
- f"Error while validating the JWT passed in header: '{HEADER}' for dataset {dataset}. Trying with"
- f" the following authentication mechanisms. JWT was: {token}"
- )
diff --git a/services/api/src/api/jwt_token.py b/services/api/src/api/jwt_token.py
index dd4469fc..cfc3b8fe 100644
--- a/services/api/src/api/jwt_token.py
+++ b/services/api/src/api/jwt_token.py
@@ -3,0 +4 @@
+import logging
@@ -137,0 +139,4 @@ def is_jwt_valid(
+ logging.debug(
+ f"Missing public key '{public_key}' or algorithm '{algorithm}' to decode JWT token. Skipping JWT"
+ " validation."
+ )
@@ -145,0 +151 @@ def is_jwt_valid(
+ logging.debug(f"Decoded JWT is: '{public_key}'.")
@@ -146,0 +153,4 @@ def is_jwt_valid(
+ logging.debug(
+ f"Missing public key '{public_key}' or algorithm '{algorithm}' to decode JWT token. Skipping JWT"
+ " validation."
+ )
|
|
cd58e8e26ce51016258eb030f1b7437219b2f3cb
|
Quentin Lhoest
| 2023-03-15T11:03:47 |
Config-level parquet (#931)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 7c5571e4..6ef19d95 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -278 +278 @@ workers:
- workerOnlyJobTypes: "/parquet"
+ workerOnlyJobTypes: "dataset-parquet,config-parquet"
diff --git a/chart/values.yaml b/chart/values.yaml
index 0deb1192..744663c0 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -387 +387 @@ workers:
- workerOnlyJobTypes: "/parquet"
+ workerOnlyJobTypes: "dataset-parquet,config-parquet"
diff --git a/e2e/tests/test_31_admin_metrics.py b/e2e/tests/test_31_admin_metrics.py
index 83677edb..84f548b8 100644
--- a/e2e/tests/test_31_admin_metrics.py
+++ b/e2e/tests/test_31_admin_metrics.py
@@ -33 +33 @@ def test_metrics() -> None:
- for queue in ["/splits", "/first-rows", "/parquet"]:
+ for queue in ["/splits", "/first-rows", "dataset-parquet"]:
@@ -40 +40 @@ def test_metrics() -> None:
- for cache_kind in ["/splits", "/first-rows", "/parquet"]:
+ for cache_kind in ["/splits", "/first-rows", "dataset-parquet"]:
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 9041e263..ca702637 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -108 +108,2 @@ class ProcessingGraphConfig:
- "/parquet": {"input_type": "dataset", "requires": "/parquet-and-dataset-info"},
+ "config-parquet": {"input_type": "config", "requires": "/parquet-and-dataset-info"},
+ "dataset-parquet": {"input_type": "dataset", "requires": "config-parquet"},
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index 8689d804..620622be 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -17 +17,2 @@ def test_default_graph() -> None:
- parquet = graph.get_step("/parquet")
+ config_parquet = graph.get_step("config-parquet")
+ dataset_parquet = graph.get_step("dataset-parquet")
@@ -45 +46 @@ def test_default_graph() -> None:
- assert parquet_and_dataset_info.children == [parquet, dataset_info, config_size]
+ assert parquet_and_dataset_info.children == [config_parquet, dataset_info, config_size]
@@ -48,4 +49,9 @@ def test_default_graph() -> None:
- assert parquet is not None
- assert parquet.parent is parquet_and_dataset_info
- assert parquet.children == []
- assert parquet.get_ancestors() == [parquet_and_dataset_info]
+ assert config_parquet is not None
+ assert config_parquet.parent is parquet_and_dataset_info
+ assert config_parquet.children == [dataset_parquet]
+ assert config_parquet.get_ancestors() == [parquet_and_dataset_info]
+
+ assert dataset_parquet is not None
+ assert dataset_parquet.parent is config_parquet
+ assert dataset_parquet.children == []
+ assert dataset_parquet.get_ancestors() == [parquet_and_dataset_info, config_parquet]
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 45e31c0e..920b7146 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -118 +118,4 @@ class EndpointConfig:
- "/parquet": {"dataset": ["/parquet"]},
+ "/parquet": {
+ "dataset": ["dataset-parquet"],
+ "config": ["config-parquet"],
+ },
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 37cb76ee..cdb1d944 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -14,0 +15 @@ from worker.job_runners.config_names import ConfigNamesJobRunner
+from worker.job_runners.config_parquet import ConfigParquetJobRunner
@@ -16,0 +18 @@ from worker.job_runners.dataset_info import DatasetInfoJobRunner
+from worker.job_runners.dataset_parquet import DatasetParquetJobRunner
@@ -19 +20,0 @@ from worker.job_runners.first_rows import FirstRowsJobRunner
-from worker.job_runners.parquet import ParquetJobRunner
@@ -102,2 +103,9 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- if job_type == ParquetJobRunner.get_job_type():
- return ParquetJobRunner(
+ if job_type == ConfigParquetJobRunner.get_job_type():
+ return ConfigParquetJobRunner(
+ job_info=job_info,
+ common_config=self.app_config.common,
+ worker_config=self.app_config.worker,
+ processing_step=processing_step,
+ )
+ if job_type == DatasetParquetJobRunner.get_job_type():
+ return DatasetParquetJobRunner(
@@ -143 +151,2 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- ParquetJobRunner.get_job_type(),
+ ConfigParquetJobRunner.get_job_type(),
+ DatasetParquetJobRunner.get_job_type(),
diff --git a/services/worker/src/worker/job_runners/parquet.py b/services/worker/src/worker/job_runners/config_parquet.py
similarity index 50%
rename from services/worker/src/worker/job_runners/parquet.py
rename to services/worker/src/worker/job_runners/config_parquet.py
index 5b2aeab4..5f0b3030 100644
--- a/services/worker/src/worker/job_runners/parquet.py
+++ b/services/worker/src/worker/job_runners/config_parquet.py
@@ -12 +12,4 @@ from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError
-from worker.job_runners.parquet_and_dataset_info import ParquetFileItem
+from worker.job_runners.parquet_and_dataset_info import (
+ ParquetAndDatasetInfoResponse,
+ ParquetFileItem,
+)
@@ -14 +17 @@ from worker.job_runners.parquet_and_dataset_info import ParquetFileItem
-ParquetJobRunnerErrorCode = Literal[
+ConfigParquetJobRunnerErrorCode = Literal[
@@ -16,0 +20 @@ ParquetJobRunnerErrorCode = Literal[
+ "MissingInfoForConfigError",
@@ -20 +24 @@ ParquetJobRunnerErrorCode = Literal[
-class ParquetResponse(TypedDict):
+class ConfigParquetResponse(TypedDict):
@@ -24 +28 @@ class ParquetResponse(TypedDict):
-class ParquetJobRunnerError(JobRunnerError):
+class ConfigParquetJobRunnerError(JobRunnerError):
@@ -31 +35 @@ class ParquetJobRunnerError(JobRunnerError):
- code: ParquetJobRunnerErrorCode,
+ code: ConfigParquetJobRunnerErrorCode,
@@ -40 +44 @@ class ParquetJobRunnerError(JobRunnerError):
-class PreviousStepStatusError(ParquetJobRunnerError):
+class PreviousStepStatusError(ConfigParquetJobRunnerError):
@@ -47 +51 @@ class PreviousStepStatusError(ParquetJobRunnerError):
-class PreviousStepFormatError(ParquetJobRunnerError):
+class PreviousStepFormatError(ConfigParquetJobRunnerError):
@@ -54 +58,8 @@ class PreviousStepFormatError(ParquetJobRunnerError):
-def compute_parquet_response(dataset: str) -> ParquetResponse:
+class MissingInfoForConfigError(ConfigParquetJobRunnerError):
+ """Raised when the dataset info from the parquet export is missing the requested dataset configuration."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "MissingInfoForConfigError", cause, False)
+
+
+def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse:
@@ -60,0 +72,2 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
+ config (`str`):
+ A configuration name.
@@ -62 +75 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- `ParquetResponse`: An object with the parquet_response (list of parquet files).
+ `ConfigParquetResponse`: An object with the parquet_response (list of parquet files).
@@ -68,0 +82,5 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
+ - [`~job_runners.config_size.MissingInfoForConfigError`]
+ If the dataset info from the parquet export is missing the requested dataset configuration
+ - [`~libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the
+ token does not give the sufficient access to the dataset, or if the dataset is private
+ (private datasets are not supported by the datasets server)
@@ -71 +89 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- logging.info(f"get parquet files for dataset={dataset}")
+ logging.info(f"get parquet files for dataset={dataset}, config={config}")
@@ -73,2 +90,0 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- # TODO: we should move this dependency to the JobRunner class: defining which are the inputs, and just getting
- # their value here
@@ -82,2 +98,20 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- raise PreviousStepStatusError(
- f"Previous step gave an error: {response['http_status']}. This job should not have been created."
+ raise PreviousStepStatusError(f"Previous step gave an error: {response['http_status']}.")
+ try:
+ content = ParquetAndDatasetInfoResponse(
+ parquet_files=response["content"]["parquet_files"], dataset_info=response["content"]["dataset_info"]
+ )
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e
+
+ if config not in content["dataset_info"]:
+ if not isinstance(content["dataset_info"], dict):
+ raise PreviousStepFormatError(
+ "Previous step did not return the expected content.",
+ TypeError(f"dataset_info should be a dict, but got {type(content['dataset_info'])}"),
+ )
+ raise MissingInfoForConfigError(
+ f"Dataset configuration '{config}' is missing in the dataset info from the parquet export. "
+ f"Available configurations: {', '.join(list(content['dataset_info'])[:10])}"
+ + f"... ({len(content['dataset_info']) - 10})"
+ if len(content["dataset_info"]) > 10
+ else ""
@@ -85 +119 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- content = response["content"]
+
@@ -88,3 +122,2 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- return {
- "parquet_files": content["parquet_files"],
- }
+ parquet_files = [parquet_file for parquet_file in content["parquet_files"] if parquet_file.get("config") == config]
+ return ConfigParquetResponse(parquet_files=parquet_files)
@@ -93 +126 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
-class ParquetJobRunner(JobRunner):
+class ConfigParquetJobRunner(JobRunner):
@@ -96 +129 @@ class ParquetJobRunner(JobRunner):
- return "/parquet"
+ return "config-parquet"
@@ -103 +136,3 @@ class ParquetJobRunner(JobRunner):
- return CompleteJobResult(compute_parquet_response(dataset=self.dataset))
+ if self.config is None:
+ raise ValueError("config is required")
+ return CompleteJobResult(compute_parquet_response(dataset=self.dataset, config=self.config))
diff --git a/services/worker/src/worker/job_runners/dataset_parquet.py b/services/worker/src/worker/job_runners/dataset_parquet.py
new file mode 100644
index 00000000..aa69cce8
--- /dev/null
+++ b/services/worker/src/worker/job_runners/dataset_parquet.py
@@ -0,0 +1,164 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+from http import HTTPStatus
+from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict
+
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
+
+from worker.job_runner import JobResult, JobRunner, JobRunnerError
+from worker.job_runners.config_parquet import ConfigParquetResponse
+from worker.job_runners.parquet_and_dataset_info import ParquetFileItem
+
+SizesJobRunnerErrorCode = Literal[
+ "PreviousStepStatusError",
+ "PreviousStepFormatError",
+]
+
+
+class PreviousJob(TypedDict):
+ kind: str
+ dataset: str
+ config: Optional[str]
+ split: Optional[str]
+
+
+class DatasetParquetResponse(TypedDict):
+ parquet_files: List[ParquetFileItem]
+ pending: list[PreviousJob]
+ failed: list[PreviousJob]
+
+
+class DatasetParquetJobRunnerError(JobRunnerError):
+ """Base class for exceptions in this module."""
+
+ def __init__(
+ self,
+ message: str,
+ status_code: HTTPStatus,
+ code: SizesJobRunnerErrorCode,
+ cause: Optional[BaseException] = None,
+ disclose_cause: bool = False,
+ ):
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
+
+
+class PreviousStepStatusError(DatasetParquetJobRunnerError):
+ """Raised when the previous step gave an error. The job should not have been created."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False)
+
+
+class PreviousStepFormatError(DatasetParquetJobRunnerError):
+ """Raised when the content of the previous step has not the expected format."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False)
+
+
+def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float]:
+ """
+ Get the response of dataset-parquet for one specific dataset on huggingface.co.
+ Args:
+ dataset (`str`):
+ A namespace (user or an organization) and a repo name separated
+ by a `/`.
+ Returns:
+ `DatasetParquetResponse`: An object with the parquet_response (list of parquet files).
+ <Tip>
+ Raises the following errors:
+ - [`~job_runners.dataset_size.PreviousStepStatusError`]
+ If the previous step gave an error.
+ - [`~job_runners.dataset_size.PreviousStepFormatError`]
+ If the content of the previous step has not the expected format
+ </Tip>
+ """
+ logging.info(f"get parquet files for dataset={dataset}")
+
+ try:
+ response = get_response(kind="/parquet-and-dataset-info", dataset=dataset)
+ except DoesNotExist as e:
+ raise DatasetNotFoundError(
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e
+ ) from e
+ if response["http_status"] != HTTPStatus.OK:
+ raise PreviousStepStatusError(
+ f"Previous step gave an error: {response['http_status']}. This job should not have been created."
+ )
+ content = response["content"]
+ try:
+ parquet_files: list[ParquetFileItem] = []
+ total = 0
+ pending = []
+ failed = []
+ for config in content["dataset_info"].keys():
+ total += 1
+ try:
+ response = get_response(kind="config-parquet", dataset=dataset, config=config)
+ except DoesNotExist:
+ logging.debug("No response found in previous step for this dataset: 'config-parquet' endpoint.")
+ pending.append(
+ PreviousJob(
+ {
+ "kind": "config-parquet",
+ "dataset": dataset,
+ "config": config,
+ "split": None,
+ }
+ )
+ )
+ continue
+ if response["http_status"] != HTTPStatus.OK:
+ logging.debug(f"Previous step gave an error: {response['http_status']}.")
+ failed.append(
+ PreviousJob(
+ {
+ "kind": "config-parquet",
+ "dataset": dataset,
+ "config": config,
+ "split": None,
+ }
+ )
+ )
+ continue
+ config_parquet_content = ConfigParquetResponse(parquet_files=response["content"]["parquet_files"])
+ parquet_files.extend(config_parquet_content["parquet_files"])
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e
+
+ progress = (total - len(pending)) / total if total else 1.0
+
+ return (
+ DatasetParquetResponse(
+ parquet_files=parquet_files,
+ pending=pending,
+ failed=failed,
+ ),
+ progress,
+ )
+
+
+class DatasetParquetJobRunner(JobRunner):
+ @staticmethod
+ def get_job_type() -> str:
+ return "dataset-parquet"
+
+ @staticmethod
+ def get_job_runner_version() -> int:
+ return 1
+
+ def compute(self) -> JobResult:
+ response_content, progress = compute_sizes_response(dataset=self.dataset)
+ return JobResult(response_content, progress=progress)
+
+ def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]:
+ """Get the set of new splits, from the content created by the compute."""
+ return {
+ SplitFullName(dataset=parquet_file["dataset"], config=parquet_file["config"], split=parquet_file["split"])
+ for parquet_file in content["parquet_files"]
+ }
diff --git a/services/worker/tests/job_runners/test_parquet.py b/services/worker/tests/job_runners/test_config_parquet.py
similarity index 58%
rename from services/worker/tests/job_runners/test_parquet.py
rename to services/worker/tests/job_runners/test_config_parquet.py
index ddf4adb2..6850a899 100644
--- a/services/worker/tests/job_runners/test_parquet.py
+++ b/services/worker/tests/job_runners/test_config_parquet.py
@@ -15,2 +15,3 @@ from worker.config import AppConfig
-from worker.job_runners.parquet import (
- ParquetJobRunner,
+from worker.job_runners.config_parquet import (
+ ConfigParquetJobRunner,
+ ConfigParquetResponse,
@@ -19,0 +21,4 @@ from worker.job_runners.parquet import (
+from worker.job_runners.parquet_and_dataset_info import (
+ ParquetAndDatasetInfoResponse,
+ ParquetFileItem,
+)
@@ -28 +33 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-GetJobRunner = Callable[[str, AppConfig, bool], ParquetJobRunner]
+GetJobRunner = Callable[[str, str, AppConfig, bool], ConfigParquetJobRunner]
@@ -37,0 +43 @@ def get_job_runner(
+ config: str,
@@ -40,2 +46,2 @@ def get_job_runner(
- ) -> ParquetJobRunner:
- return ParquetJobRunner(
+ ) -> ConfigParquetJobRunner:
+ return ConfigParquetJobRunner(
@@ -43 +49 @@ def get_job_runner(
- "type": ParquetJobRunner.get_job_type(),
+ "type": ConfigParquetJobRunner.get_job_type(),
@@ -45 +51 @@ def get_job_runner(
- "config": None,
+ "config": config,
@@ -54,2 +60,2 @@ def get_job_runner(
- name=ParquetJobRunner.get_job_type(),
- input_type="dataset",
+ name=ConfigParquetJobRunner.get_job_type(),
+ input_type="config",
@@ -68 +74 @@ def get_job_runner(
- "dataset,upstream_status,upstream_content,expected_error_code,expected_content,should_raise",
+ "dataset,config,upstream_status,upstream_content,expected_error_code,expected_content,should_raise",
@@ -71,0 +78 @@ def get_job_runner(
+ "config_1",
@@ -73 +80,11 @@ def get_job_runner(
- {"parquet_files": [{"key": "value"}], "dataset_info": {"key": "value"}},
+ ParquetAndDatasetInfoResponse(
+ parquet_files=[
+ ParquetFileItem(
+ dataset="ok", config="config_1", split="train", url="url1", filename="filename1", size=0
+ ),
+ ParquetFileItem(
+ dataset="ok", config="config_2", split="train", url="url2", filename="filename2", size=0
+ ),
+ ],
+ dataset_info={"config_1": "value", "config_2": "value"},
+ ),
@@ -75 +92,7 @@ def get_job_runner(
- {"parquet_files": [{"key": "value"}]},
+ ConfigParquetResponse(
+ parquet_files=[
+ ParquetFileItem(
+ dataset="ok", config="config_1", split="train", url="url1", filename="filename1", size=0
+ )
+ ]
+ ),
@@ -78 +101,9 @@ def get_job_runner(
- ("status_error", HTTPStatus.NOT_FOUND, {"error": "error"}, PreviousStepStatusError.__name__, None, True),
+ (
+ "status_error",
+ "config_1",
+ HTTPStatus.NOT_FOUND,
+ {"error": "error"},
+ PreviousStepStatusError.__name__,
+ None,
+ True,
+ ),
@@ -80,0 +112 @@ def get_job_runner(
+ "config_1",
@@ -92,0 +125 @@ def test_compute(
+ config: str,
@@ -102 +135 @@ def test_compute(
- job_runner = get_job_runner(dataset, app_config, False)
+ job_runner = get_job_runner(dataset, config, app_config, False)
@@ -113 +146,2 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No
- job_runner = get_job_runner(dataset, app_config, False)
+ config = "doesnotexist"
+ job_runner = get_job_runner(dataset, config, app_config, False)
diff --git a/services/worker/tests/job_runners/test_dataset_parquet.py b/services/worker/tests/job_runners/test_dataset_parquet.py
new file mode 100644
index 00000000..ad435715
--- /dev/null
+++ b/services/worker/tests/job_runners/test_dataset_parquet.py
@@ -0,0 +1,222 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from http import HTTPStatus
+from typing import Any, Callable, Mapping, Optional, TypedDict
+
+import pytest
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import upsert_response
+
+from worker.config import AppConfig
+from worker.job_runners.config_parquet import ConfigParquetResponse
+from worker.job_runners.dataset_parquet import (
+ DatasetParquetJobRunner,
+ DatasetParquetResponse,
+ PreviousStepFormatError,
+ PreviousStepStatusError,
+)
+from worker.job_runners.parquet_and_dataset_info import (
+ ParquetAndDatasetInfoResponse,
+ ParquetFileItem,
+)
+
+
[email protected](autouse=True)
+def prepare_and_clean_mongo(app_config: AppConfig) -> None:
+ # prepare the database before each test, and clean it afterwards
+ pass
+
+
+GetJobRunner = Callable[[str, AppConfig, bool], DatasetParquetJobRunner]
+
+
+class UpstreamResponse(TypedDict):
+ kind: str
+ dataset: str
+ config: Optional[str]
+ http_status: HTTPStatus
+ content: Mapping[str, Any]
+
+
[email protected]
+def get_job_runner(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> DatasetParquetJobRunner:
+ return DatasetParquetJobRunner(
+ job_info={
+ "type": DatasetParquetJobRunner.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ common_config=app_config.common,
+ worker_config=app_config.worker,
+ processing_step=ProcessingStep(
+ name=DatasetParquetJobRunner.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ )
+
+ return _get_job_runner
+
+
[email protected](
+ "dataset,upstream_responses,expected_error_code,expected_content,should_raise",
+ [
+ (
+ "ok",
+ [
+ UpstreamResponse(
+ kind="/parquet-and-dataset-info",
+ dataset="ok",
+ config=None,
+ http_status=HTTPStatus.OK,
+ content=ParquetAndDatasetInfoResponse(
+ parquet_files=[
+ ParquetFileItem(
+ dataset="ok",
+ config="config_1",
+ split="train",
+ url="url1",
+ filename="filename1",
+ size=0,
+ ),
+ ParquetFileItem(
+ dataset="ok",
+ config="config_2",
+ split="train",
+ url="url2",
+ filename="filename2",
+ size=0,
+ ),
+ ],
+ dataset_info={"config_1": "value", "config_2": "value"},
+ ),
+ ),
+ UpstreamResponse(
+ kind="config-parquet",
+ dataset="ok",
+ config="config_1",
+ http_status=HTTPStatus.OK,
+ content=ConfigParquetResponse(
+ parquet_files=[
+ ParquetFileItem(
+ dataset="ok",
+ config="config_1",
+ split="train",
+ url="url1",
+ filename="filename1",
+ size=0,
+ ),
+ ]
+ ),
+ ),
+ UpstreamResponse(
+ kind="config-parquet",
+ dataset="ok",
+ config="config_2",
+ http_status=HTTPStatus.OK,
+ content=ConfigParquetResponse(
+ parquet_files=[
+ ParquetFileItem(
+ dataset="ok",
+ config="config_2",
+ split="train",
+ url="url2",
+ filename="filename2",
+ size=0,
+ ),
+ ]
+ ),
+ ),
+ ],
+ None,
+ DatasetParquetResponse(
+ parquet_files=[
+ ParquetFileItem(
+ dataset="ok", config="config_1", split="train", url="url1", filename="filename1", size=0
+ ),
+ ParquetFileItem(
+ dataset="ok", config="config_2", split="train", url="url2", filename="filename2", size=0
+ ),
+ ],
+ pending=[],
+ failed=[],
+ ),
+ False,
+ ),
+ (
+ "status_error",
+ [
+ UpstreamResponse(
+ kind="/parquet-and-dataset-info",
+ dataset="status_error",
+ config=None,
+ http_status=HTTPStatus.NOT_FOUND,
+ content={"error": "error"},
+ )
+ ],
+ PreviousStepStatusError.__name__,
+ None,
+ True,
+ ),
+ (
+ "format_error",
+ [
+ UpstreamResponse(
+ kind="/parquet-and-dataset-info",
+ dataset="format_error",
+ config=None,
+ http_status=HTTPStatus.OK,
+ content={"not_parquet_files": "wrong_format"},
+ )
+ ],
+ PreviousStepFormatError.__name__,
+ None,
+ True,
+ ),
+ ],
+)
+def test_compute(
+ app_config: AppConfig,
+ get_job_runner: GetJobRunner,
+ dataset: str,
+ upstream_responses: list[UpstreamResponse],
+ expected_error_code: str,
+ expected_content: Any,
+ should_raise: bool,
+) -> None:
+ for upstream_response in upstream_responses:
+ upsert_response(**upstream_response)
+ job_runner = get_job_runner(dataset, app_config, False)
+ if should_raise:
+ with pytest.raises(Exception) as e:
+ job_runner.compute()
+ assert e.type.__name__ == expected_error_code
+ else:
+ assert job_runner.compute().content == expected_content
+
+
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "doesnotexist"
+ job_runner = get_job_runner(dataset, app_config, False)
+ with pytest.raises(DatasetNotFoundError):
+ job_runner.compute()
diff --git a/services/worker/tests/job_runners/test_dataset_size.py b/services/worker/tests/job_runners/test_dataset_size.py
index 9c88487a..72b53aa3 100644
--- a/services/worker/tests/job_runners/test_dataset_size.py
+++ b/services/worker/tests/job_runners/test_dataset_size.py
@@ -377 +376,0 @@ def test_compute(
- # raise ArithmeticError(str(job_runner.compute().content))
diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py
index 5f7d8d2e..cadd352d 100644
--- a/services/worker/tests/test_job_runner_factory.py
+++ b/services/worker/tests/test_job_runner_factory.py
@@ -28 +28,2 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph:
- ("/parquet", "ParquetJobRunner"),
+ ("config-parquet", "ConfigParquetJobRunner"),
+ ("dataset-parquet", "DatasetParquetJobRunner"),
|
|
2ead70d394f785db4977a19a6d1dff11df3285c6
|
Sylvain Lesage
| 2023-03-15T09:54:15 |
fix: 🐛 take the algorithm into account in config (#933)
|
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 1ebfb379..45e31c0e 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -66,0 +67 @@ class ApiConfig:
+ hf_jwt_algorithm=env.str(name="HF_JWT_ALGORITHM", default=API_HF_JWT_ALGORITHM),
|
|
e3fb17cf1b00e42acd6d8dc32fc5b33b8c56a95d
|
Sylvain Lesage
| 2023-03-15T09:46:10 |
refactor: 💡 fix type of the public key: str, not Any (#932)
|
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index 3ab735e3..029276ca 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -5 +5 @@ import logging
-from typing import Any, Literal, Optional
+from typing import Literal, Optional
@@ -45 +45 @@ def auth_check(
- hf_jwt_public_key: Optional[Any] = None,
+ hf_jwt_public_key: Optional[str] = None,
@@ -63 +63 @@ def auth_check(
- hf_jwt_public_key (Any|None): the public key to use to decode the JWT token
+ hf_jwt_public_key (str|None): the public key to use to decode the JWT token
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 561b4c05..17eaf93f 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -7 +7 @@ from http import HTTPStatus
-from typing import Any, List, Mapping, Optional, Tuple
+from typing import List, Mapping, Optional, Tuple
@@ -205 +205 @@ def create_endpoint(
- hf_jwt_public_key: Optional[Any] = None,
+ hf_jwt_public_key: Optional[str] = None,
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 48005992..1321dc60 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -5 +5 @@ import logging
-from typing import Any, List, Optional, Set
+from typing import List, Optional, Set
@@ -72 +72 @@ def create_is_valid_endpoint(
- hf_jwt_public_key: Optional[Any] = None,
+ hf_jwt_public_key: Optional[str] = None,
|
|
b1f7d0003bbad349bd6def7acf53b5c4eaeeb036
|
Andrea Francis Soria Jimenez
| 2023-03-14T15:37:32 |
Remove worker version (#927)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 5a6feeb7..ec2d79f7 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -28,0 +29,3 @@ from mongodb_migration.migrations._20230309141600_cache_add_job_runner_version i
+from mongodb_migration.migrations._20230313164200_cache_remove_worker_version import (
+ MigrationRemoveWorkerVersionFromCachedResponse,
+)
@@ -63,0 +67,3 @@ class MigrationsCollector:
+ MigrationRemoveWorkerVersionFromCachedResponse(
+ version="20230313164200", description="remove 'worker_version' field from cache"
+ ),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py
new file mode 100644
index 00000000..bf25e6b2
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py
@@ -0,0 +1,26 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.simple_cache import CachedResponse
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import IrreversibleMigration, Migration
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationRemoveWorkerVersionFromCachedResponse(Migration):
+ def up(self) -> None:
+ logging.info("Removing 'worker_version' field.")
+ db = get_db("cache")
+ db["cachedResponsesBlue"].update_many({}, {"$unset": {"worker_version": ""}})
+
+ def down(self) -> None:
+ raise IrreversibleMigration("This migration does not support rollback")
+
+ def validate(self) -> None:
+ logging.info("Ensure that a random selection of cached results don't have 'worker_version' field")
+
+ check_documents(DocCls=CachedResponse, sample_size=10)
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py b/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py
index 8bcfa1ee..5d3563bb 100644
--- a/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py
+++ b/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py
@@ -17 +16,0 @@ def test_cache_add_job_runner_version_without_worker_version(mongo_host: str) ->
- db["cachedResponsesBlue"].delete_many({})
@@ -27,0 +27 @@ def test_cache_add_job_runner_version_without_worker_version(mongo_host: str) ->
+ db["cachedResponsesBlue"].drop()
@@ -42 +41,0 @@ def test_cache_add_job_runner_version(mongo_host: str, worker_version: str, expe
- db["cachedResponsesBlue"].delete_many({})
@@ -52,0 +52 @@ def test_cache_add_job_runner_version(mongo_host: str, worker_version: str, expe
+ db["cachedResponsesBlue"].drop()
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py
new file mode 100644
index 00000000..d8078f00
--- /dev/null
+++ b/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py
@@ -0,0 +1,38 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from libcommon.resources import MongoResource
+from mongoengine.connection import get_db
+from pytest import raises
+
+from mongodb_migration.migration import IrreversibleMigration
+from mongodb_migration.migrations._20230313164200_cache_remove_worker_version import (
+ MigrationRemoveWorkerVersionFromCachedResponse,
+)
+
+
+def test_cache_remove_worker_version(mongo_host: str) -> None:
+ with MongoResource(database="test_cache_remove_worker_version", host=mongo_host, mongoengine_alias="cache"):
+ db = get_db("cache")
+ db["cachedResponsesBlue"].delete_many({})
+ db["cachedResponsesBlue"].insert_many(
+ [
+ {
+ "kind": "/splits",
+ "dataset": "dataset_without_worker_version",
+ "http_status": 200,
+ "worker_version": "1.0.0",
+ }
+ ]
+ )
+ migration = MigrationRemoveWorkerVersionFromCachedResponse(
+ version="20230313164200", description="remove 'worker_version' field from cache"
+ )
+ migration.up()
+ result = db["cachedResponsesBlue"].find_one({"dataset": "dataset_without_worker_version"})
+ assert result
+ assert "worker_version" not in result
+
+ with raises(IrreversibleMigration):
+ migration.down()
+ db["cachedResponsesBlue"].drop()
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index f63a530d..b54933e8 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -82 +81,0 @@ class CachedResponse(Document):
- worker_version (`str`): The semver version of the worker that cached the response.
@@ -98 +96,0 @@ class CachedResponse(Document):
- worker_version = StringField()
|
|
e58240944787657439dadeb256fcf2a8eb7134fc
|
Quentin Lhoest
| 2023-03-14T15:36:10 |
Config-level size (#926)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 54f1b731..7c5571e4 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -323 +323 @@ workers:
- workerOnlyJobTypes: "/sizes"
+ workerOnlyJobTypes: "config-size,dataset-size"
diff --git a/chart/values.yaml b/chart/values.yaml
index cdc95256..0deb1192 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -432 +432 @@ workers:
- workerOnlyJobTypes: "/sizes"
+ workerOnlyJobTypes: "config-size,dataset-size"
diff --git a/e2e/tests/test_11_auth.py b/e2e/tests/test_11_auth.py
index 8d38fab2..6c693ddf 100644
--- a/e2e/tests/test_11_auth.py
+++ b/e2e/tests/test_11_auth.py
@@ -48 +48 @@ def test_auth_e2e(
- f"/sizes?dataset={dataset}",
+ f"/size?dataset={dataset}",
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 75a99799..9041e263 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -111 +111,2 @@ class ProcessingGraphConfig:
- "/sizes": {"input_type": "dataset", "requires": "/parquet-and-dataset-info"},
+ "config-size": {"input_type": "config", "requires": "/parquet-and-dataset-info"},
+ "dataset-size": {"input_type": "dataset", "requires": "config-size"},
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index 02974524..8689d804 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -19 +19,2 @@ def test_default_graph() -> None:
- sizes = graph.get_step("/sizes")
+ config_size = graph.get_step("config-size")
+ dataset_size = graph.get_step("dataset-size")
@@ -44 +45 @@ def test_default_graph() -> None:
- assert parquet_and_dataset_info.children == [parquet, dataset_info, sizes]
+ assert parquet_and_dataset_info.children == [parquet, dataset_info, config_size]
@@ -62,4 +63,9 @@ def test_default_graph() -> None:
- assert sizes is not None
- assert sizes.parent is parquet_and_dataset_info
- assert sizes.children == []
- assert sizes.get_ancestors() == [parquet_and_dataset_info]
+ assert config_size is not None
+ assert config_size.parent is parquet_and_dataset_info
+ assert config_size.children == [dataset_size]
+ assert config_size.get_ancestors() == [parquet_and_dataset_info]
+
+ assert dataset_size is not None
+ assert dataset_size.parent is config_size
+ assert dataset_size.children == []
+ assert dataset_size.get_ancestors() == [parquet_and_dataset_info, config_size]
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index cb325cea..1ebfb379 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -119 +119,4 @@ class EndpointConfig:
- "/sizes": {"dataset": ["/sizes"]},
+ "/size": {
+ "dataset": ["dataset-size"],
+ "config": ["config-size"],
+ },
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index aefa84d7..aef6da7a 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -59,4 +59,6 @@ def test_endpoints_definition() -> None:
- sizes = definition["/sizes"]
- assert sizes is not None
- assert sizes["dataset"] is not None
- assert len(sizes["dataset"]) == 1 # Only has one processing step
+ size = definition["/size"]
+ assert size is not None
+ assert size["dataset"] is not None
+ assert size["config"] is not None
+ assert len(size["dataset"]) == 1 # Only has one processing step
+ assert len(size["config"]) == 1 # Only has one processing step
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 1207e9de..37cb76ee 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -14,0 +15 @@ from worker.job_runners.config_names import ConfigNamesJobRunner
+from worker.job_runners.config_size import ConfigSizeJobRunner
@@ -15,0 +17 @@ from worker.job_runners.dataset_info import DatasetInfoJobRunner
+from worker.job_runners.dataset_size import DatasetSizeJobRunner
@@ -19 +20,0 @@ from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoJob
-from worker.job_runners.sizes import SizesJobRunner
@@ -115,2 +116,9 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- if job_type == SizesJobRunner.get_job_type():
- return SizesJobRunner(
+ if job_type == DatasetSizeJobRunner.get_job_type():
+ return DatasetSizeJobRunner(
+ job_info=job_info,
+ common_config=self.app_config.common,
+ worker_config=self.app_config.worker,
+ processing_step=processing_step,
+ )
+ if job_type == ConfigSizeJobRunner.get_job_type():
+ return ConfigSizeJobRunner(
@@ -137 +145,2 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- SizesJobRunner.get_job_type(),
+ DatasetSizeJobRunner.get_job_type(),
+ ConfigSizeJobRunner.get_job_type(),
diff --git a/services/worker/src/worker/job_runners/config_size.py b/services/worker/src/worker/job_runners/config_size.py
new file mode 100644
index 00000000..933de199
--- /dev/null
+++ b/services/worker/src/worker/job_runners/config_size.py
@@ -0,0 +1,206 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+from http import HTTPStatus
+from typing import Any, Literal, Mapping, Optional, TypedDict
+
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
+
+from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError
+from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoResponse
+
+ConfigSizeJobRunnerErrorCode = Literal[
+ "PreviousStepStatusError",
+ "PreviousStepFormatError",
+ "MissingInfoForConfigError",
+]
+
+
+class ConfigSize(TypedDict):
+ dataset: str
+ config: str
+ num_bytes_original_files: int
+ num_bytes_parquet_files: int
+ num_bytes_memory: int
+ num_rows: int
+ num_columns: int
+
+
+class SplitSize(TypedDict):
+ dataset: str
+ config: str
+ split: str
+ num_bytes_parquet_files: int
+ num_bytes_memory: int
+ num_rows: int
+ num_columns: int
+
+
+class ConfigSizeContent(TypedDict):
+ config: ConfigSize
+ splits: list[SplitSize]
+
+
+class ConfigSizeResponse(TypedDict):
+ size: ConfigSizeContent
+
+
+class ConfigSizeJobRunnerError(JobRunnerError):
+ """Base class for exceptions in this module."""
+
+ def __init__(
+ self,
+ message: str,
+ status_code: HTTPStatus,
+ code: ConfigSizeJobRunnerErrorCode,
+ cause: Optional[BaseException] = None,
+ disclose_cause: bool = False,
+ ):
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
+
+
+class PreviousStepStatusError(ConfigSizeJobRunnerError):
+ """Raised when the previous step gave an error. The job should not have been created."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False)
+
+
+class PreviousStepFormatError(ConfigSizeJobRunnerError):
+ """Raised when the content of the previous step has not the expected format."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False)
+
+
+class MissingInfoForConfigError(ConfigSizeJobRunnerError):
+ """Raised when the dataset info from the parquet export is missing the requested dataset configuration."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "MissingInfoForConfigError", cause, False)
+
+
+def compute_config_size_response(dataset: str, config: str) -> ConfigSizeResponse:
+ """
+ Get the response of config-size for one specific dataset and config on huggingface.co.
+ Args:
+ dataset (`str`):
+ A namespace (user or an organization) and a repo name separated
+ by a `/`.
+ config (`str`):
+ A configuration name.
+ Returns:
+ `ConfigSizeResponse`: An object with the size_response.
+ <Tip>
+ Raises the following errors:
+ - [`~job_runners.config_size.PreviousStepStatusError`]
+ If the previous step gave an error.
+ - [`~job_runners.config_size.PreviousStepFormatError`]
+ If the content of the previous step has not the expected format
+ - [`~job_runners.config_size.MissingInfoForConfigError`]
+ If the dataset info from the parquet export is missing the requested dataset configuration
+ - [`~libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the
+ token does not give the sufficient access to the dataset, or if the dataset is private
+ (private datasets are not supported by the datasets server)
+ </Tip>
+ """
+ logging.info(f"get size for dataset={dataset}, config={config}")
+
+ try:
+ response = get_response(kind="/parquet-and-dataset-info", dataset=dataset)
+ except DoesNotExist as e:
+ raise DatasetNotFoundError(
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e
+ ) from e
+ if response["http_status"] != HTTPStatus.OK:
+ raise PreviousStepStatusError(f"Previous step gave an error: {response['http_status']}..")
+
+ try:
+ content = ParquetAndDatasetInfoResponse(
+ parquet_files=response["content"]["parquet_files"], dataset_info=response["content"]["dataset_info"]
+ )
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e
+
+ if config not in content["dataset_info"]:
+ if not isinstance(content["dataset_info"], dict):
+ raise PreviousStepFormatError(
+ "Previous step did not return the expected content.",
+ TypeError(f"dataset_info should be a dict, but got {type(content['dataset_info'])}"),
+ )
+ raise MissingInfoForConfigError(
+ f"Dataset configuration '{config}' is missing in the dataset info from the parquet export. "
+ f"Available configurations: {', '.join(list(content['dataset_info'])[:10])}"
+ + f"... ({len(content['dataset_info']) - 10})"
+ if len(content["dataset_info"]) > 10
+ else ""
+ )
+ try:
+ config_dataset_info = content["dataset_info"][config]
+ num_columns = len(config_dataset_info["features"])
+ split_sizes: list[SplitSize] = [
+ {
+ "dataset": dataset,
+ "config": config,
+ "split": split_info["name"],
+ "num_bytes_parquet_files": sum(
+ x["size"]
+ for x in content["parquet_files"]
+ if x["config"] == config and x["split"] == split_info["name"]
+ ),
+ "num_bytes_memory": split_info["num_bytes"],
+ "num_rows": split_info["num_examples"],
+ "num_columns": num_columns,
+ }
+ for split_info in config_dataset_info["splits"].values()
+ ]
+ config_size = ConfigSize(
+ {
+ "dataset": dataset,
+ "config": config,
+ "num_bytes_original_files": config_dataset_info["download_size"],
+ "num_bytes_parquet_files": sum(split_size["num_bytes_parquet_files"] for split_size in split_sizes),
+ "num_bytes_memory": sum(
+ split_size["num_bytes_memory"] for split_size in split_sizes
+ ), # or "num_bytes_memory": config_dataset_info["dataset_size"],
+ "num_rows": sum(split_size["num_rows"] for split_size in split_sizes),
+ "num_columns": num_columns,
+ }
+ )
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e
+
+ return ConfigSizeResponse(
+ {
+ "size": {
+ "config": config_size,
+ "splits": split_sizes,
+ }
+ }
+ )
+
+
+class ConfigSizeJobRunner(JobRunner):
+ @staticmethod
+ def get_job_type() -> str:
+ return "config-size"
+
+ @staticmethod
+ def get_job_runner_version() -> int:
+ return 1
+
+ def compute(self) -> CompleteJobResult:
+ if self.config is None:
+ raise ValueError("config is required")
+ return CompleteJobResult(compute_config_size_response(dataset=self.dataset, config=self.config))
+
+ def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]:
+ """Get the set of new splits, from the content created by the compute."""
+ return {
+ SplitFullName(dataset=split_size["dataset"], config=split_size["config"], split=split_size["split"])
+ for split_size in content["size"]["splits"]
+ }
diff --git a/services/worker/src/worker/job_runners/dataset_info.py b/services/worker/src/worker/job_runners/dataset_info.py
index 6b839ebb..0d45dc81 100644
--- a/services/worker/src/worker/job_runners/dataset_info.py
+++ b/services/worker/src/worker/job_runners/dataset_info.py
@@ -76 +76 @@ def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse:
- "No response found in previous step for this dataset: '/parquet-and-dataset-info' endpoint.", e
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e
diff --git a/services/worker/src/worker/job_runners/sizes.py b/services/worker/src/worker/job_runners/dataset_size.py
similarity index 53%
rename from services/worker/src/worker/job_runners/sizes.py
rename to services/worker/src/worker/job_runners/dataset_size.py
index 62f18941..49d3b495 100644
--- a/services/worker/src/worker/job_runners/sizes.py
+++ b/services/worker/src/worker/job_runners/dataset_size.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Any, Literal, Mapping, Optional, TypedDict
+from typing import Any, Literal, Mapping, Optional, Tuple, TypedDict
@@ -11 +11,2 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError
+from worker.job_runner import JobResult, JobRunner, JobRunnerError
+from worker.job_runners.config_size import ConfigSize, ConfigSizeResponse, SplitSize
@@ -27,21 +28 @@ class DatasetSize(TypedDict):
-class ConfigSize(TypedDict):
- dataset: str
- config: str
- num_bytes_original_files: int
- num_bytes_parquet_files: int
- num_bytes_memory: int
- num_rows: int
- num_columns: int
-
-
-class SplitSize(TypedDict):
- dataset: str
- config: str
- split: str
- num_bytes_parquet_files: int
- num_bytes_memory: int
- num_rows: int
- num_columns: int
-
-
-class SizesContent(TypedDict):
+class DatasetSizeContent(TypedDict):
@@ -53,2 +34,6 @@ class SizesContent(TypedDict):
-class SizesResponse(TypedDict):
- sizes: SizesContent
+class PreviousJob(TypedDict):
+ kind: str
+ dataset: str
+ config: Optional[str]
+ split: Optional[str]
+
@@ -55,0 +41,4 @@ class SizesResponse(TypedDict):
+class DatasetSizeResponse(TypedDict):
+ size: DatasetSizeContent
+ pending: list[PreviousJob]
+ failed: list[PreviousJob]
@@ -57 +46,2 @@ class SizesResponse(TypedDict):
-class SizesJobRunnerError(JobRunnerError):
+
+class DatasetSizeJobRunnerError(JobRunnerError):
@@ -73 +63 @@ class SizesJobRunnerError(JobRunnerError):
-class PreviousStepStatusError(SizesJobRunnerError):
+class PreviousStepStatusError(DatasetSizeJobRunnerError):
@@ -80 +70 @@ class PreviousStepStatusError(SizesJobRunnerError):
-class PreviousStepFormatError(SizesJobRunnerError):
+class PreviousStepFormatError(DatasetSizeJobRunnerError):
@@ -87 +77 @@ class PreviousStepFormatError(SizesJobRunnerError):
-def compute_sizes_response(dataset: str) -> SizesResponse:
+def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
@@ -89 +79 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- Get the response of /sizes for one specific dataset on huggingface.co.
+ Get the response of dataset-size for one specific dataset on huggingface.co.
@@ -95 +85 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- `SizesResponse`: An object with the sizes_response.
+ `DatasetSizeResponse`: An object with the sizes_response.
@@ -98 +88 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- - [`~job_runners.sizes.PreviousStepStatusError`]
+ - [`~job_runners.dataset_size.PreviousStepStatusError`]
@@ -100 +90 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- - [`~job_runners.sizes.PreviousStepFormatError`]
+ - [`~job_runners.dataset_size.PreviousStepFormatError`]
@@ -110 +100 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- "No response found in previous step for this dataset: '/parquet-and-dataset-info' endpoint.", e
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e
@@ -119,0 +110,3 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
+ total = 0
+ pending = []
+ failed = []
@@ -121,34 +114,32 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- config_dataset_info = content["dataset_info"][config]
- num_columns = len(config_dataset_info["features"])
- config_split_sizes: list[SplitSize] = [
- {
- "dataset": dataset,
- "config": config,
- "split": split_info["name"],
- "num_bytes_parquet_files": sum(
- x["size"]
- for x in content["parquet_files"]
- if x["config"] == config and x["split"] == split_info["name"]
- ),
- "num_bytes_memory": split_info["num_bytes"],
- "num_rows": split_info["num_examples"],
- "num_columns": num_columns,
- }
- for split_info in config_dataset_info["splits"].values()
- ]
- config_sizes.append(
- {
- "dataset": dataset,
- "config": config,
- "num_bytes_original_files": config_dataset_info["download_size"],
- "num_bytes_parquet_files": sum(
- split_size["num_bytes_parquet_files"] for split_size in config_split_sizes
- ),
- "num_bytes_memory": sum(
- split_size["num_bytes_memory"] for split_size in config_split_sizes
- ), # or "num_bytes_memory": config_dataset_info["dataset_size"],
- "num_rows": sum(split_size["num_rows"] for split_size in config_split_sizes),
- "num_columns": len(config_dataset_info["features"]),
- }
- )
- split_sizes.extend(config_split_sizes)
+ total += 1
+ try:
+ response = get_response(kind="config-size", dataset=dataset, config=config)
+ except DoesNotExist:
+ logging.debug("No response found in previous step for this dataset: 'config-size' endpoint.")
+ pending.append(
+ PreviousJob(
+ {
+ "kind": "config-size",
+ "dataset": dataset,
+ "config": config,
+ "split": None,
+ }
+ )
+ )
+ continue
+ if response["http_status"] != HTTPStatus.OK:
+ logging.debug(f"Previous step gave an error: {response['http_status']}.")
+ failed.append(
+ PreviousJob(
+ {
+ "kind": "config-size",
+ "dataset": dataset,
+ "config": config,
+ "split": None,
+ }
+ )
+ )
+ continue
+ config_size_content = ConfigSizeResponse(size=response["content"]["size"])
+ config_sizes.append(config_size_content["size"]["config"])
+ split_sizes.extend(config_size_content["size"]["splits"])
@@ -165,10 +156,19 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- return {
- "sizes": {
- "dataset": dataset_size,
- "configs": config_sizes,
- "splits": split_sizes,
- }
- }
-
-
-class SizesJobRunner(JobRunner):
+ progress = (total - len(pending)) / total if total else 1.0
+
+ return (
+ DatasetSizeResponse(
+ {
+ "size": {
+ "dataset": dataset_size,
+ "configs": config_sizes,
+ "splits": split_sizes,
+ },
+ "pending": pending,
+ "failed": failed,
+ }
+ ),
+ progress,
+ )
+
+
+class DatasetSizeJobRunner(JobRunner):
@@ -177 +177 @@ class SizesJobRunner(JobRunner):
- return "/sizes"
+ return "dataset-size"
@@ -183,2 +183,3 @@ class SizesJobRunner(JobRunner):
- def compute(self) -> CompleteJobResult:
- return CompleteJobResult(compute_sizes_response(dataset=self.dataset))
+ def compute(self) -> JobResult:
+ response_content, progress = compute_sizes_response(dataset=self.dataset)
+ return JobResult(response_content, progress=progress)
@@ -190 +191 @@ class SizesJobRunner(JobRunner):
- for split_size in content["sizes"]["splits"]
+ for split_size in content["size"]["splits"]
diff --git a/services/worker/src/worker/job_runners/parquet.py b/services/worker/src/worker/job_runners/parquet.py
index 09e7b336..5b2aeab4 100644
--- a/services/worker/src/worker/job_runners/parquet.py
+++ b/services/worker/src/worker/job_runners/parquet.py
@@ -79 +79 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- "No response found in previous step for this dataset: '/parquet-and-dataset-info' endpoint.", e
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
index 3b0198ca..975e7e1e 100644
--- a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
+++ b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
@@ -69 +69 @@ def compute_split_names_from_dataset_info_response(
- like the size or number of samples. See /dataset-info or /sizes for that.
+ like the size or number of samples. See /dataset-info or dataset-size for that.
diff --git a/services/worker/src/worker/job_runners/split_names_from_streaming.py b/services/worker/src/worker/job_runners/split_names_from_streaming.py
index c55f20d7..b89e6eaf 100644
--- a/services/worker/src/worker/job_runners/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split_names_from_streaming.py
@@ -77 +77 @@ def compute_split_names_from_streaming_response(
- like the size or number of samples. See /dataset-info or /sizes for that.
+ like the size or number of samples. See /dataset-info or dataset-size for that.
diff --git a/services/worker/src/worker/job_runners/splits.py b/services/worker/src/worker/job_runners/splits.py
index 4e15d1c1..7751a912 100644
--- a/services/worker/src/worker/job_runners/splits.py
+++ b/services/worker/src/worker/job_runners/splits.py
@@ -97 +97 @@ def compute_splits_response(
- of samples. See /dataset-info or /sizes for that.
+ of samples. See /dataset-info or dataset-size for that.
diff --git a/services/worker/tests/job_runners/test_sizes.py b/services/worker/tests/job_runners/test_config_size.py
similarity index 75%
rename from services/worker/tests/job_runners/test_sizes.py
rename to services/worker/tests/job_runners/test_config_size.py
index 1895211e..35add26e 100644
--- a/services/worker/tests/job_runners/test_sizes.py
+++ b/services/worker/tests/job_runners/test_config_size.py
@@ -15 +15,2 @@ from worker.config import AppConfig
-from worker.job_runners.sizes import (
+from worker.job_runners.config_size import (
+ ConfigSizeJobRunner,
@@ -18 +18,0 @@ from worker.job_runners.sizes import (
- SizesJobRunner,
@@ -28 +28 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-GetJobRunner = Callable[[str, AppConfig, bool], SizesJobRunner]
+GetJobRunner = Callable[[str, str, AppConfig, bool], ConfigSizeJobRunner]
@@ -37,0 +38 @@ def get_job_runner(
+ config: str,
@@ -40,2 +41,2 @@ def get_job_runner(
- ) -> SizesJobRunner:
- return SizesJobRunner(
+ ) -> ConfigSizeJobRunner:
+ return ConfigSizeJobRunner(
@@ -43 +44 @@ def get_job_runner(
- "type": SizesJobRunner.get_job_type(),
+ "type": ConfigSizeJobRunner.get_job_type(),
@@ -45 +46 @@ def get_job_runner(
- "config": None,
+ "config": config,
@@ -54,2 +55,2 @@ def get_job_runner(
- name=SizesJobRunner.get_job_type(),
- input_type="dataset",
+ name=ConfigSizeJobRunner.get_job_type(),
+ input_type="config",
@@ -68 +69 @@ def get_job_runner(
- "dataset,upstream_status,upstream_content,expected_error_code,expected_content,should_raise",
+ "dataset,config,upstream_status,upstream_content,expected_error_code,expected_content,should_raise",
@@ -71,0 +73 @@ def get_job_runner(
+ "config_1",
@@ -163,2 +165,2 @@ def get_job_runner(
- "sizes": {
- "dataset": {
+ "size": {
+ "config": {
@@ -166,4 +168,6 @@ def get_job_runner(
- "num_bytes_original_files": 21507144,
- "num_bytes_parquet_files": 19057017,
- "num_bytes_memory": 20394144,
- "num_rows": 74000,
+ "config": "config_1",
+ "num_bytes_original_files": 11594722,
+ "num_bytes_parquet_files": 16665091,
+ "num_bytes_memory": 20387232,
+ "num_rows": 70000,
+ "num_columns": 2,
@@ -171,20 +174,0 @@ def get_job_runner(
- "configs": [
- {
- "dataset": "dataset_ok",
- "config": "config_1",
- "num_bytes_original_files": 11594722,
- "num_bytes_parquet_files": 16665091,
- "num_bytes_memory": 20387232,
- "num_rows": 70000,
- "num_columns": 2,
- },
- {
- "dataset": "dataset_ok",
- "config": "config_2",
- "num_bytes_original_files": 9912422,
- "num_bytes_parquet_files": 2391926,
- "num_bytes_memory": 6912,
- "num_rows": 4000,
- "num_columns": 3,
- },
- ],
@@ -210,18 +193,0 @@ def get_job_runner(
- {
- "dataset": "dataset_ok",
- "config": "config_2",
- "split": "train",
- "num_bytes_parquet_files": 8023,
- "num_bytes_memory": 5678,
- "num_rows": 3000,
- "num_columns": 3,
- },
- {
- "dataset": "dataset_ok",
- "config": "config_2",
- "split": "test",
- "num_bytes_parquet_files": 2383903,
- "num_bytes_memory": 1234,
- "num_rows": 1000,
- "num_columns": 3,
- },
@@ -233 +199,9 @@ def get_job_runner(
- ("status_error", HTTPStatus.NOT_FOUND, {"error": "error"}, PreviousStepStatusError.__name__, None, True),
+ (
+ "status_error",
+ "config_1",
+ HTTPStatus.NOT_FOUND,
+ {"error": "error"},
+ PreviousStepStatusError.__name__,
+ None,
+ True,
+ ),
@@ -235,0 +210 @@ def get_job_runner(
+ "config_1",
@@ -247,0 +223 @@ def test_compute(
+ config: str,
@@ -257 +233 @@ def test_compute(
- job_runner = get_job_runner(dataset, app_config, False)
+ job_runner = get_job_runner(dataset, config, app_config, False)
@@ -267,2 +243,2 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No
- dataset = "doesnotexist"
- job_runner = get_job_runner(dataset, app_config, False)
+ dataset = config = "doesnotexist"
+ job_runner = get_job_runner(dataset, config, app_config, False)
diff --git a/services/worker/tests/job_runners/test_dataset_size.py b/services/worker/tests/job_runners/test_dataset_size.py
new file mode 100644
index 00000000..9c88487a
--- /dev/null
+++ b/services/worker/tests/job_runners/test_dataset_size.py
@@ -0,0 +1,385 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from http import HTTPStatus
+from typing import Any, Callable, Mapping, Optional, TypedDict
+
+import pytest
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import upsert_response
+
+from worker.config import AppConfig
+from worker.job_runners.dataset_size import (
+ DatasetSizeJobRunner,
+ PreviousStepFormatError,
+ PreviousStepStatusError,
+)
+
+
[email protected](autouse=True)
+def prepare_and_clean_mongo(app_config: AppConfig) -> None:
+ # prepare the database before each test, and clean it afterwards
+ pass
+
+
+GetJobRunner = Callable[[str, AppConfig, bool], DatasetSizeJobRunner]
+
+
+class UpstreamResponse(TypedDict):
+ kind: str
+ dataset: str
+ config: Optional[str]
+ http_status: HTTPStatus
+ content: Mapping[str, Any]
+
+
[email protected]
+def get_job_runner(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> DatasetSizeJobRunner:
+ return DatasetSizeJobRunner(
+ job_info={
+ "type": DatasetSizeJobRunner.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ common_config=app_config.common,
+ worker_config=app_config.worker,
+ processing_step=ProcessingStep(
+ name=DatasetSizeJobRunner.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ )
+
+ return _get_job_runner
+
+
[email protected](
+ "dataset,upstream_responses,expected_error_code,expected_content,should_raise",
+ [
+ (
+ "dataset_ok",
+ [
+ UpstreamResponse(
+ kind="/parquet-and-dataset-info",
+ dataset="dataset_ok",
+ config=None,
+ http_status=HTTPStatus.OK,
+ content={
+ "parquet_files": [
+ {"dataset": "dataset_ok", "config": "config_1", "split": "train", "size": 14281188},
+ {"dataset": "dataset_ok", "config": "config_1", "split": "test", "size": 2383903},
+ {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 1234},
+ {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 6789},
+ {"dataset": "dataset_ok", "config": "config_2", "split": "test", "size": 2383903},
+ ],
+ "dataset_info": {
+ "config_1": {
+ "features": {
+ "image": {"_type": "Image"},
+ "label": {
+ "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"],
+ "_type": "ClassLabel",
+ },
+ },
+ "splits": {
+ "train": {
+ "name": "train",
+ "num_bytes": 17470800,
+ "num_examples": 60000,
+ "dataset_name": "dataset_ok",
+ },
+ "test": {
+ "name": "test",
+ "num_bytes": 2916432,
+ "num_examples": 10000,
+ "dataset_name": "dataset_ok",
+ },
+ },
+ "download_checksums": {
+ "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": {
+ "num_bytes": 9912422,
+ "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609",
+ },
+ "https://storage.googleapis.com/cvdf-datasets/mnist/train-labels-idx1-ubyte.gz": {
+ "num_bytes": 28881,
+ "checksum": "3552534a0a558bbed6aed32b30c495cca23d567ec52cac8be1a0730e8010255c",
+ },
+ "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-images-idx3-ubyte.gz": {
+ "num_bytes": 1648877,
+ "checksum": "8d422c7b0a1c1c79245a5bcf07fe86e33eeafee792b84584aec276f5a2dbc4e6",
+ },
+ "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-labels-idx1-ubyte.gz": {
+ "num_bytes": 4542,
+ "checksum": "f7ae60f92e00ec6debd23a6088c31dbd2371eca3ffa0defaefb259924204aec6",
+ },
+ },
+ "download_size": 11594722,
+ "dataset_size": 20387232,
+ "size_in_bytes": 31981954,
+ },
+ "config_2": {
+ "features": {
+ "image": {"_type": "Image"},
+ "image2": {"_type": "Image"},
+ "label": {
+ "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"],
+ "_type": "ClassLabel",
+ },
+ },
+ "splits": {
+ "train": {
+ "name": "train",
+ "num_bytes": 5678,
+ "num_examples": 3000,
+ "dataset_name": "dataset_ok",
+ },
+ "test": {
+ "name": "test",
+ "num_bytes": 1234,
+ "num_examples": 1000,
+ "dataset_name": "dataset_ok",
+ },
+ },
+ "download_checksums": {
+ "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": {
+ "num_bytes": 9912422,
+ "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609",
+ },
+ },
+ "download_size": 9912422,
+ "dataset_size": 6912,
+ "size_in_bytes": 9919334,
+ },
+ },
+ },
+ ),
+ UpstreamResponse(
+ kind="config-size",
+ dataset="dataset_ok",
+ config="config_1",
+ http_status=HTTPStatus.OK,
+ content={
+ "size": {
+ "config": {
+ "dataset": "dataset_ok",
+ "config": "config_1",
+ "num_bytes_original_files": 11594722,
+ "num_bytes_parquet_files": 16665091,
+ "num_bytes_memory": 20387232,
+ "num_rows": 70000,
+ "num_columns": 2,
+ },
+ "splits": [
+ {
+ "dataset": "dataset_ok",
+ "config": "config_1",
+ "split": "train",
+ "num_bytes_parquet_files": 14281188,
+ "num_bytes_memory": 17470800,
+ "num_rows": 60000,
+ "num_columns": 2,
+ },
+ {
+ "dataset": "dataset_ok",
+ "config": "config_1",
+ "split": "test",
+ "num_bytes_parquet_files": 2383903,
+ "num_bytes_memory": 2916432,
+ "num_rows": 10000,
+ "num_columns": 2,
+ },
+ ],
+ }
+ },
+ ),
+ UpstreamResponse(
+ kind="config-size",
+ dataset="dataset_ok",
+ config="config_2",
+ http_status=HTTPStatus.OK,
+ content={
+ "size": {
+ "config": {
+ "dataset": "dataset_ok",
+ "config": "config_2",
+ "num_bytes_original_files": 9912422,
+ "num_bytes_parquet_files": 2391926,
+ "num_bytes_memory": 6912,
+ "num_rows": 4000,
+ "num_columns": 3,
+ },
+ "splits": [
+ {
+ "dataset": "dataset_ok",
+ "config": "config_2",
+ "split": "train",
+ "num_bytes_parquet_files": 8023,
+ "num_bytes_memory": 5678,
+ "num_rows": 3000,
+ "num_columns": 3,
+ },
+ {
+ "dataset": "dataset_ok",
+ "config": "config_2",
+ "split": "test",
+ "num_bytes_parquet_files": 2383903,
+ "num_bytes_memory": 1234,
+ "num_rows": 1000,
+ "num_columns": 3,
+ },
+ ],
+ }
+ },
+ ),
+ ],
+ None,
+ {
+ "size": {
+ "dataset": {
+ "dataset": "dataset_ok",
+ "num_bytes_original_files": 21507144,
+ "num_bytes_parquet_files": 19057017,
+ "num_bytes_memory": 20394144,
+ "num_rows": 74000,
+ },
+ "configs": [
+ {
+ "dataset": "dataset_ok",
+ "config": "config_1",
+ "num_bytes_original_files": 11594722,
+ "num_bytes_parquet_files": 16665091,
+ "num_bytes_memory": 20387232,
+ "num_rows": 70000,
+ "num_columns": 2,
+ },
+ {
+ "dataset": "dataset_ok",
+ "config": "config_2",
+ "num_bytes_original_files": 9912422,
+ "num_bytes_parquet_files": 2391926,
+ "num_bytes_memory": 6912,
+ "num_rows": 4000,
+ "num_columns": 3,
+ },
+ ],
+ "splits": [
+ {
+ "dataset": "dataset_ok",
+ "config": "config_1",
+ "split": "train",
+ "num_bytes_parquet_files": 14281188,
+ "num_bytes_memory": 17470800,
+ "num_rows": 60000,
+ "num_columns": 2,
+ },
+ {
+ "dataset": "dataset_ok",
+ "config": "config_1",
+ "split": "test",
+ "num_bytes_parquet_files": 2383903,
+ "num_bytes_memory": 2916432,
+ "num_rows": 10000,
+ "num_columns": 2,
+ },
+ {
+ "dataset": "dataset_ok",
+ "config": "config_2",
+ "split": "train",
+ "num_bytes_parquet_files": 8023,
+ "num_bytes_memory": 5678,
+ "num_rows": 3000,
+ "num_columns": 3,
+ },
+ {
+ "dataset": "dataset_ok",
+ "config": "config_2",
+ "split": "test",
+ "num_bytes_parquet_files": 2383903,
+ "num_bytes_memory": 1234,
+ "num_rows": 1000,
+ "num_columns": 3,
+ },
+ ],
+ },
+ "failed": [],
+ "pending": [],
+ },
+ False,
+ ),
+ (
+ "status_error",
+ [
+ UpstreamResponse(
+ kind="/parquet-and-dataset-info",
+ dataset="status_error",
+ config=None,
+ http_status=HTTPStatus.NOT_FOUND,
+ content={"error": "error"},
+ )
+ ],
+ PreviousStepStatusError.__name__,
+ None,
+ True,
+ ),
+ (
+ "format_error",
+ [
+ UpstreamResponse(
+ kind="/parquet-and-dataset-info",
+ dataset="format_error",
+ config=None,
+ http_status=HTTPStatus.OK,
+ content={"not_dataset_info": "wrong_format"},
+ )
+ ],
+ PreviousStepFormatError.__name__,
+ None,
+ True,
+ ),
+ ],
+)
+def test_compute(
+ app_config: AppConfig,
+ get_job_runner: GetJobRunner,
+ dataset: str,
+ upstream_responses: list[UpstreamResponse],
+ expected_error_code: str,
+ expected_content: Any,
+ should_raise: bool,
+) -> None:
+ for upstream_response in upstream_responses:
+ upsert_response(**upstream_response)
+ job_runner = get_job_runner(dataset, app_config, False)
+ if should_raise:
+ with pytest.raises(Exception) as e:
+ job_runner.compute()
+ assert e.type.__name__ == expected_error_code
+ else:
+ # raise ArithmeticError(str(job_runner.compute().content))
+ assert job_runner.compute().content == expected_content
+
+
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "doesnotexist"
+ job_runner = get_job_runner(dataset, app_config, False)
+ with pytest.raises(DatasetNotFoundError):
+ job_runner.compute()
diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py
index 7034927b..5f7d8d2e 100644
--- a/services/worker/tests/test_job_runner_factory.py
+++ b/services/worker/tests/test_job_runner_factory.py
@@ -30 +30,2 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph:
- ("/sizes", "SizesJobRunner"),
+ ("config-size", "ConfigSizeJobRunner"),
+ ("dataset-size", "DatasetSizeJobRunner"),
|
|
57b49e14f051ca0ef108a28f83554b4c405973ff
|
Sylvain Lesage
| 2023-03-14T14:48:52 |
Fix jwt decoding (#929)
|
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 96ccf8d8..5d594000 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -75,0 +76 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ hf_jwt_algorithm=app_config.api.hf_jwt_algorithm,
@@ -95,0 +97 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ hf_jwt_algorithm=app_config.api.hf_jwt_algorithm,
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index 3ab521cd..3ab735e3 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -74 +74,2 @@ def auth_check(
- token = request.headers.get("x-api-key")
+ HEADER = "x-api-key"
+ token = request.headers.get(HEADER)
@@ -77,0 +79,4 @@ def auth_check(
+ logging.debug(
+ f"By-passing the authentication step, because a valid JWT was passed in header: '{HEADER}' for"
+ f" dataset {dataset}. JWT was: {token}"
+ )
@@ -78,0 +84,4 @@ def auth_check(
+ logging.debug(
+ f"Error while validating the JWT passed in header: '{HEADER}' for dataset {dataset}. Trying with"
+ f" the following authentication mechanisms. JWT was: {token}"
+ )
diff --git a/services/api/src/api/jwt_token.py b/services/api/src/api/jwt_token.py
index c0a673a9..dd4469fc 100644
--- a/services/api/src/api/jwt_token.py
+++ b/services/api/src/api/jwt_token.py
@@ -115 +115,3 @@ def fetch_jwt_public_key(
-def is_jwt_valid(dataset: str, token: Any, public_key: Optional[str], algorithm: Optional[str]) -> bool:
+def is_jwt_valid(
+ dataset: str, token: Any, public_key: Optional[str], algorithm: Optional[str], verify_exp: Optional[bool] = True
+) -> bool:
@@ -129,0 +132 @@ def is_jwt_valid(dataset: str, token: Any, public_key: Optional[str], algorithm:
+ verify_exp (bool|None): whether to verify the expiration of the JWT token. Default to True.
@@ -138 +141,4 @@ def is_jwt_valid(dataset: str, token: Any, public_key: Optional[str], algorithm:
- jwt=token, key=public_key, algorithms=[algorithm], options={"require": ["exp", "sub", "read"]}
+ jwt=token,
+ key=public_key,
+ algorithms=[algorithm],
+ options={"require": ["exp", "sub", "read"], "verify_exp": verify_exp},
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index f620f368..561b4c05 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -205,0 +206 @@ def create_endpoint(
+ hf_jwt_algorithm: Optional[str] = None,
@@ -249,0 +251 @@ def create_endpoint(
+ hf_jwt_algorithm=hf_jwt_algorithm,
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 218ef36b..48005992 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -72,0 +73 @@ def create_is_valid_endpoint(
+ hf_jwt_algorithm: Optional[str] = None,
@@ -93,0 +95 @@ def create_is_valid_endpoint(
+ hf_jwt_algorithm=hf_jwt_algorithm,
diff --git a/services/api/tests/test_jwt_token.py b/services/api/tests/test_jwt_token.py
index 4330881a..4a9f6835 100644
--- a/services/api/tests/test_jwt_token.py
+++ b/services/api/tests/test_jwt_token.py
@@ -13,0 +14,5 @@ HUB_JWT_KEYS = [{"crv": "Ed25519", "x": "-RBhgyNluwaIL5KFJb6ZOL2H1nmyI8mW4Z2EHGD
+HUB_JWT_ALGORITHM = "EdDSA"
+HUB_JWT_PUBLIC_KEY = """-----BEGIN PUBLIC KEY-----
+MCowBQYDK2VwAyEA+RBhgyNluwaIL5KFJb6ZOL2H1nmyI8mW4Z2EHGDGCXM=
+-----END PUBLIC KEY-----
+"""
@@ -39 +44,23 @@ def test_parse_jwk(
- parse_jwt_public_key(keys=keys, hf_jwt_algorithm="EdDSA")
+ key = parse_jwt_public_key(keys=keys, hf_jwt_algorithm=HUB_JWT_ALGORITHM)
+ assert key == HUB_JWT_PUBLIC_KEY
+
+
+HUB_JWT_TOKEN_FOR_SEVERO_GLUE = (
+ "eyJhbGciOiJFZERTQSJ9.eyJyZWFkIjp0cnVlLCJzdWIiOiJkYXRhc2V0cy9zZXZlcm8vZ2x1ZSIsImV4cCI6MTY3ODgwMjk0NH0"
+ ".nIi1ZKinMBpYi4kKtirW-cQEt1cGnAziTGmJsZeN5UpE62jz4DcPaIPlSI5P5ciGOlTxy4SEhD1WITkQzpo3Aw"
+)
+DATASET_SEVERO_GLUE = "severo/glue"
+
+
+def test_is_jwt_valid_with_ec() -> None:
+ assert (
+ is_jwt_valid(
+ dataset=DATASET_SEVERO_GLUE,
+ token=HUB_JWT_TOKEN_FOR_SEVERO_GLUE,
+ public_key=HUB_JWT_PUBLIC_KEY,
+ algorithm=HUB_JWT_ALGORITHM,
+ verify_exp=False,
+ # This is a test token generated on 2023/03/14, so we don't want to verify the exp.
+ )
+ is True
+ )
|
|
e24942d637cd6d46116caf73c67af670ba1ba756
|
Andrea Francis Soria Jimenez
| 2023-03-13T20:34:14 |
Change worker_version to int job_runner_version (#916)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index fa9fe261..5a6feeb7 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -25,0 +26,3 @@ from mongodb_migration.migrations._20230309123100_cache_add_progress import (
+from mongodb_migration.migrations._20230309141600_cache_add_job_runner_version import (
+ MigrationAddJobRunnerVersionToCacheResponse,
+)
@@ -57,0 +61,3 @@ class MigrationsCollector:
+ MigrationAddJobRunnerVersionToCacheResponse(
+ version="20230309141600", description="add 'job_runner_version' field based on 'worker_version' value"
+ ),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py
new file mode 100644
index 00000000..bcbf4326
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py
@@ -0,0 +1,45 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.simple_cache import CachedResponse
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import Migration
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationAddJobRunnerVersionToCacheResponse(Migration):
+ def up(self) -> None:
+ # See https://docs.mongoengine.org/guide/migration.html#example-1-addition-of-a-field
+ logging.info("If missing, add 'job_runner_version' field based on 'worker_version' value")
+ db = get_db("cache")
+ db["cachedResponsesBlue"].update_many(
+ {"job_runner_version": {"$exists": False}},
+ [
+ {
+ "$set": {
+ "job_runner_version": {
+ "$convert": {
+ "input": {"$first": {"$split": ["$worker_version", "."]}},
+ "to": "int",
+ "onError": None,
+ "onNull": None,
+ }
+ }
+ }
+ }
+ ], # type: ignore
+ )
+
+ def down(self) -> None:
+ logging.info("Remove 'job_runner_version' field from all the cached results")
+ db = get_db("cache")
+ db["cachedResponsesBlue"].update_many({}, {"$unset": {"job_runner_version": ""}})
+
+ def validate(self) -> None:
+ logging.info("Ensure that a random selection of cached results have the 'job_runner_version' field")
+
+ check_documents(DocCls=CachedResponse, sample_size=10)
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py b/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py
new file mode 100644
index 00000000..8bcfa1ee
--- /dev/null
+++ b/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py
@@ -0,0 +1,52 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+from typing import Optional
+
+import pytest
+from libcommon.resources import MongoResource
+from mongoengine.connection import get_db
+
+from mongodb_migration.migrations._20230309141600_cache_add_job_runner_version import (
+ MigrationAddJobRunnerVersionToCacheResponse,
+)
+
+
+def test_cache_add_job_runner_version_without_worker_version(mongo_host: str) -> None:
+ with MongoResource(database="test_cache_add_job_runner_version", host=mongo_host, mongoengine_alias="cache"):
+ db = get_db("cache")
+ db["cachedResponsesBlue"].delete_many({})
+ db["cachedResponsesBlue"].insert_many(
+ [{"kind": "/splits", "dataset": "dataset_without_worker_version", "http_status": 200}]
+ )
+ migration = MigrationAddJobRunnerVersionToCacheResponse(
+ version="20230309141600", description="add 'job_runner_version' field based on 'worker_version' value"
+ )
+ migration.up()
+ result = db["cachedResponsesBlue"].find_one({"dataset": "dataset_without_worker_version"})
+ assert result
+ assert not result["job_runner_version"]
+
+
[email protected](
+ "worker_version,expected",
+ [
+ ("2.0.0", 2),
+ ("1.5.0", 1),
+ ("WrongFormat", None),
+ (None, None),
+ ],
+)
+def test_cache_add_job_runner_version(mongo_host: str, worker_version: str, expected: Optional[int]) -> None:
+ with MongoResource(database="test_cache_add_job_runner_version", host=mongo_host, mongoengine_alias="cache"):
+ db = get_db("cache")
+ db["cachedResponsesBlue"].delete_many({})
+ db["cachedResponsesBlue"].insert_many(
+ [{"kind": "/splits", "dataset": "dataset", "http_status": 200, "worker_version": worker_version}]
+ )
+ migration = MigrationAddJobRunnerVersionToCacheResponse(
+ version="20230309141600", description="add 'job_runner_version' field based on 'worker_version' value"
+ )
+ migration.up()
+ result = db["cachedResponsesBlue"].find_one({"dataset": "dataset"})
+ assert result
+ assert result["job_runner_version"] == expected
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index aaa1c5fe..f63a530d 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -27,0 +28 @@ from mongoengine.fields import (
+ IntField,
@@ -81,0 +83 @@ class CachedResponse(Document):
+ job_runner_version (`int`): The version of the job runner that cached the response.
@@ -98,0 +101 @@ class CachedResponse(Document):
+ job_runner_version = IntField()
@@ -135 +138 @@ def upsert_response(
- worker_version: Optional[str] = None,
+ job_runner_version: Optional[int] = None,
@@ -144 +146,0 @@ def upsert_response(
- worker_version=worker_version,
@@ -147,0 +150 @@ def upsert_response(
+ job_runner_version=job_runner_version,
@@ -164 +166,0 @@ class CacheEntryWithoutContent(TypedDict):
- worker_version: Optional[str]
@@ -166,0 +169 @@ class CacheEntryWithoutContent(TypedDict):
+ job_runner_version: Optional[int]
@@ -175 +178 @@ def get_response_without_content(
- .only("http_status", "error_code", "worker_version", "dataset_git_revision", "progress")
+ .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress")
@@ -181 +183,0 @@ def get_response_without_content(
- "worker_version": response.worker_version,
@@ -182,0 +185 @@ def get_response_without_content(
+ "job_runner_version": response.job_runner_version,
@@ -189 +192 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L
- "http_status", "error_code", "worker_version", "dataset_git_revision", "progress"
+ "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress"
@@ -195 +198 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L
- "worker_version": response.worker_version,
+ "job_runner_version": response.job_runner_version,
@@ -211 +214 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O
- .only("content", "http_status", "error_code", "worker_version", "dataset_git_revision", "progress")
+ .only("content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress")
@@ -218 +221 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O
- "worker_version": response.worker_version,
+ "job_runner_version": response.job_runner_version,
@@ -283 +286 @@ class CacheReport(TypedDict):
- worker_version: Optional[str]
+ job_runner_version: Optional[int]
@@ -345 +348 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo
- "worker_version",
+ "job_runner_version",
@@ -359 +362 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo
- "worker_version": object.worker_version,
+ "job_runner_version": object.job_runner_version,
@@ -425 +428 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int)
- "worker_version",
+ "job_runner_version",
@@ -442 +445 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int)
- "worker_version": object.worker_version,
+ "job_runner_version": object.job_runner_version,
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index e151c0f1..addace5b 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -116 +116 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
- "worker_version": None,
+ "job_runner_version": None,
@@ -126 +126 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
- "worker_version": None,
+ "job_runner_version": None,
@@ -147 +147 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
- worker_version = "0.1.2"
+ job_runner_version = 0
@@ -157 +157 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
- worker_version=worker_version,
+ job_runner_version=job_runner_version,
@@ -166 +166 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
- "worker_version": worker_version,
+ "job_runner_version": job_runner_version,
@@ -421 +421 @@ def test_get_cache_reports() -> None:
- worker_version_b = "0.1.2"
+ job_runner_version_b = 0
@@ -431 +431 @@ def test_get_cache_reports() -> None:
- worker_version=worker_version_b,
+ job_runner_version=job_runner_version_b,
@@ -483 +483 @@ def test_get_cache_reports() -> None:
- "worker_version": None,
+ "job_runner_version": None,
@@ -493 +493 @@ def test_get_cache_reports() -> None:
- "worker_version": worker_version_b,
+ "job_runner_version": job_runner_version_b,
@@ -509 +509 @@ def test_get_cache_reports() -> None:
- "worker_version": None,
+ "job_runner_version": None,
@@ -530 +530 @@ def test_get_cache_reports() -> None:
- "worker_version": None,
+ "job_runner_version": None,
@@ -543 +543 @@ def test_get_cache_reports() -> None:
- "worker_version": worker_version_b,
+ "job_runner_version": job_runner_version_b,
@@ -564 +564 @@ def test_get_cache_reports() -> None:
- "worker_version": None,
+ "job_runner_version": None,
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index 960b09dd..d4e5e53e 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -25 +24,0 @@ from libcommon.utils import orjson_dumps
-from packaging import version
@@ -202 +201 @@ class JobRunner(ABC):
- def get_version() -> str:
+ def get_job_runner_version() -> int:
@@ -272,19 +270,0 @@ class JobRunner(ABC):
- def compare_major_version(self, other_version: str) -> int:
- """
- Compare the major version of job runner's self version and the other version's.
-
- Args:
- other_version (:obj:`str`): the other semantic version
-
- Returns:
- :obj:`int`: the difference between the major version of both versions.
- 0 if they are equal. Negative if job runner's major version is lower than other_version, positive
- otherwise.
- Raises:
- :obj:`ValueError`: if job runner's version or other_version is not a valid semantic version.
- """
- try:
- return version.parse(self.get_version()).major - version.parse(other_version).major
- except Exception as err:
- raise RuntimeError(f"Could not get major versions: {err}") from err
-
@@ -327,2 +307,2 @@ class JobRunner(ABC):
- cached_response["worker_version"] is None
- or self.compare_major_version(cached_response["worker_version"]) != 0
+ cached_response["job_runner_version"] is None
+ or self.get_job_runner_version() > cached_response["job_runner_version"]
@@ -330,4 +309,0 @@ class JobRunner(ABC):
- # no job runner version in the cache, or the job runner has been updated - we process the job to update
- # the cache
- # note: the collection field is named "worker_version" for historical reasons, it might be renamed
- # "job_runner_version" in the future.
@@ -378 +354 @@ class JobRunner(ABC):
- worker_version=self.get_version(),
+ job_runner_version=self.get_job_runner_version(),
@@ -406 +382 @@ class JobRunner(ABC):
- worker_version=self.get_version(),
+ job_runner_version=self.get_job_runner_version(),
@@ -513 +489 @@ class JobRunner(ABC):
- worker_version=self.get_version(),
+ job_runner_version=self.get_job_runner_version(),
diff --git a/services/worker/src/worker/job_runners/config_names.py b/services/worker/src/worker/job_runners/config_names.py
index 052f7c6a..3ecccf11 100644
--- a/services/worker/src/worker/job_runners/config_names.py
+++ b/services/worker/src/worker/job_runners/config_names.py
@@ -104,2 +104,2 @@ class ConfigNamesJobRunner(DatasetsBasedJobRunner):
- def get_version() -> str:
- return "1.0.0"
+ def get_job_runner_version() -> int:
+ return 1
diff --git a/services/worker/src/worker/job_runners/dataset_info.py b/services/worker/src/worker/job_runners/dataset_info.py
index 9f4019fa..6b839ebb 100644
--- a/services/worker/src/worker/job_runners/dataset_info.py
+++ b/services/worker/src/worker/job_runners/dataset_info.py
@@ -96,2 +96,2 @@ class DatasetInfoJobRunner(JobRunner):
- def get_version() -> str:
- return "1.0.0"
+ def get_job_runner_version() -> int:
+ return 1
diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py
index c594f5ba..b21202c0 100644
--- a/services/worker/src/worker/job_runners/first_rows.py
+++ b/services/worker/src/worker/job_runners/first_rows.py
@@ -648,2 +648,2 @@ class FirstRowsJobRunner(DatasetsBasedJobRunner):
- def get_version() -> str:
- return "2.0.0"
+ def get_job_runner_version() -> int:
+ return 2
diff --git a/services/worker/src/worker/job_runners/parquet.py b/services/worker/src/worker/job_runners/parquet.py
index b76bb5d6..09e7b336 100644
--- a/services/worker/src/worker/job_runners/parquet.py
+++ b/services/worker/src/worker/job_runners/parquet.py
@@ -99,2 +99,2 @@ class ParquetJobRunner(JobRunner):
- def get_version() -> str:
- return "3.0.0"
+ def get_job_runner_version() -> int:
+ return 3
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index a1d6e416..0cd5189f 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -872,2 +872,2 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner):
- def get_version() -> str:
- return "1.1.0"
+ def get_job_runner_version() -> int:
+ return 1
diff --git a/services/worker/src/worker/job_runners/sizes.py b/services/worker/src/worker/job_runners/sizes.py
index c8ae0a6a..62f18941 100644
--- a/services/worker/src/worker/job_runners/sizes.py
+++ b/services/worker/src/worker/job_runners/sizes.py
@@ -180,2 +180,2 @@ class SizesJobRunner(JobRunner):
- def get_version() -> str:
- return "1.0.0"
+ def get_job_runner_version() -> int:
+ return 1
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
index a5fa8f46..3b0198ca 100644
--- a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
+++ b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
@@ -117,2 +117,2 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner):
- def get_version() -> str:
- return "2.0.0"
+ def get_job_runner_version() -> int:
+ return 2
diff --git a/services/worker/src/worker/job_runners/split_names_from_streaming.py b/services/worker/src/worker/job_runners/split_names_from_streaming.py
index 56a7477f..c55f20d7 100644
--- a/services/worker/src/worker/job_runners/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split_names_from_streaming.py
@@ -121,2 +121,2 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner):
- def get_version() -> str:
- return "2.0.0"
+ def get_job_runner_version() -> int:
+ return 2
diff --git a/services/worker/src/worker/job_runners/splits.py b/services/worker/src/worker/job_runners/splits.py
index d276cc2d..4e15d1c1 100644
--- a/services/worker/src/worker/job_runners/splits.py
+++ b/services/worker/src/worker/job_runners/splits.py
@@ -133,2 +133,2 @@ class SplitsJobRunner(DatasetsBasedJobRunner):
- def get_version() -> str:
- return "2.0.0"
+ def get_job_runner_version() -> int:
+ return 2
diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py
index 44affd1e..38059247 100644
--- a/services/worker/tests/job_runners/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__datasets_based_worker.py
@@ -33,2 +33,2 @@ class DummyJobRunner(DatasetsBasedJobRunner):
- def get_version() -> str:
- return "1.0.0"
+ def get_job_runner_version() -> int:
+ return 1
@@ -85,8 +84,0 @@ def get_job_runner(
-def test_version(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
- dataset, config, split = get_default_config_split("dataset")
- job_runner = get_job_runner(dataset, config, split, app_config, False)
- assert len(job_runner.get_version().split(".")) == 3
- assert job_runner.compare_major_version(other_version="0.0.0") > 0
- assert job_runner.compare_major_version(other_version="1000.0.0") < 0
-
-
diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py
index e4aa763f..29c53611 100644
--- a/services/worker/tests/job_runners/test_config_names.py
+++ b/services/worker/tests/job_runners/test_config_names.py
@@ -79 +79 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner: Get
- assert cached_response["worker_version"] == job_runner.get_version()
+ assert cached_response["job_runner_version"] == job_runner.get_job_runner_version()
diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py
index 683d61de..9fae0ae7 100644
--- a/services/worker/tests/job_runners/test_first_rows.py
+++ b/services/worker/tests/job_runners/test_first_rows.py
@@ -93 +93 @@ def test_compute(
- assert cached_response["worker_version"] == job_runner.get_version()
+ assert cached_response["job_runner_version"] == job_runner.get_job_runner_version()
diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index 68825b9b..39f55cd6 100644
--- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -133 +133 @@ def test_compute(
- assert cached_response["worker_version"] == job_runner.get_version()
+ assert cached_response["job_runner_version"] == job_runner.get_job_runner_version()
diff --git a/services/worker/tests/job_runners/test_split_names_from_streaming.py b/services/worker/tests/job_runners/test_split_names_from_streaming.py
index 9b677562..80410d76 100644
--- a/services/worker/tests/job_runners/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/test_split_names_from_streaming.py
@@ -71 +71 @@ def test_process(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public
- assert cached_response["worker_version"] == job_runner.get_version()
+ assert cached_response["job_runner_version"] == job_runner.get_job_runner_version()
diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py
index dafd920a..b68f8800 100644
--- a/services/worker/tests/job_runners/test_splits.py
+++ b/services/worker/tests/job_runners/test_splits.py
@@ -79 +79 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner: Get
- assert cached_response["worker_version"] == job_runner.get_version()
+ assert cached_response["job_runner_version"] == job_runner.get_job_runner_version()
diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py
index 36716405..af0cee4e 100644
--- a/services/worker/tests/test_job_runner.py
+++ b/services/worker/tests/test_job_runner.py
@@ -35,2 +35,2 @@ class DummyJobRunner(JobRunner):
- def get_job_type() -> str:
- return "/dummy"
+ def get_job_runner_version() -> int:
+ return 1
@@ -39,2 +39,2 @@ class DummyJobRunner(JobRunner):
- def get_version() -> str:
- return "1.0.1"
+ def get_job_type() -> str:
+ return "/dummy"
@@ -49,41 +48,0 @@ class DummyJobRunner(JobRunner):
[email protected](
- "other_version, expected, should_raise",
- [
- ("1.0.0", 0, False),
- ("0.1.0", 1, False),
- ("2.0.0", -1, False),
- ("not a version", None, True),
- ],
-)
-def test_compare_major_version(
- test_processing_step: ProcessingStep,
- other_version: str,
- expected: int,
- should_raise: bool,
-) -> None:
- job_id = "job_id"
- dataset = "dataset"
- config = "config"
- split = "split"
- force = False
- job_runner = DummyJobRunner(
- job_info={
- "job_id": job_id,
- "type": test_processing_step.job_type,
- "dataset": dataset,
- "config": config,
- "split": split,
- "force": force,
- "priority": Priority.NORMAL,
- },
- processing_step=test_processing_step,
- common_config=CommonConfig(),
- worker_config=WorkerConfig(),
- )
- if should_raise:
- with pytest.raises(Exception):
- job_runner.compare_major_version(other_version)
- else:
- assert job_runner.compare_major_version(other_version) == expected
-
-
@@ -93 +52 @@ class CacheEntry:
- worker_version: Optional[str]
+ job_runner_version: Optional[int]
@@ -98 +56,0 @@ class CacheEntry:
-# .get_version()
@@ -106 +64 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -115 +73 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -124 +82 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -138 +96 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -147 +105 @@ class CacheEntry:
- worker_version=None, # no version
+ job_runner_version=None, # no version
@@ -156 +114 @@ class CacheEntry:
- worker_version="0.0.1", # a different version
+ job_runner_version=0, # a different version
@@ -165 +123 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -174 +132 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -183 +141 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -193 +151 @@ class CacheEntry:
- worker_version=DummyJobRunner.get_version(),
+ job_runner_version=DummyJobRunner.get_job_runner_version(),
@@ -232 +190 @@ def test_should_skip_job(
- worker_version=cache_entry.worker_version,
+ job_runner_version=cache_entry.job_runner_version,
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 11d755e7..3ef192d4 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -25,2 +25,2 @@ class DummyJobRunner(JobRunner):
- def get_version() -> str:
- return "1.0.1"
+ def get_job_runner_version() -> int:
+ return 1
|
|
958ef26a19991b01c09dba78e2456780a09eccd0
|
Sylvain Lesage
| 2023-03-13T19:48:43 |
Fix access to gated datasets, and forbid access to private datasets! (#925)
|
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py
index 362bc3bd..83d3e642 100644
--- a/libs/libcommon/src/libcommon/dataset.py
+++ b/libs/libcommon/src/libcommon/dataset.py
@@ -103,0 +104,5 @@ class GatedExtraFieldsError(DatasetError):
+DOES_NOT_EXIST_OR_PRIVATE_DATASET_ERROR_MESSAGE = (
+ "The dataset does not exist on the Hub, or is private. Private datasets are not yet supported."
+)
+
+
@@ -142 +147 @@ def ask_access(
- err,
+ cause=err,
@@ -151 +156 @@ def ask_access(
- "The dataset is gated with extra fields: not supported at the moment."
+ "The dataset is gated with extra fields: not supported at the moment.", cause=err
@@ -154 +159 @@ def ask_access(
- raise GatedDisabledError("The dataset is gated and access is disabled.") from err
+ raise GatedDisabledError("The dataset is gated and access is disabled.", cause=err) from err
@@ -156,3 +161 @@ def ask_access(
- raise DatasetNotFoundError(
- "The dataset does not exist on the Hub, or is private. Private datasets are not yet supported."
- ) from err
+ raise DatasetNotFoundError(DOES_NOT_EXIST_OR_PRIVATE_DATASET_ERROR_MESSAGE, cause=err) from err
@@ -200,5 +203,15 @@ def get_dataset_info_for_supported_datasets(
- dataset_info = HfApi(endpoint=hf_endpoint).dataset_info(
- repo_id=dataset, token=hf_token, timeout=hf_timeout_seconds
- )
- except RepositoryNotFoundError:
- ask_access(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds)
+ try:
+ dataset_info = HfApi(endpoint=hf_endpoint).dataset_info(
+ repo_id=dataset, token=hf_token, timeout=hf_timeout_seconds
+ )
+ except RepositoryNotFoundError:
+ ask_access(
+ dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds
+ )
+ dataset_info = HfApi(endpoint=hf_endpoint).dataset_info(
+ repo_id=dataset, token=hf_token, timeout=hf_timeout_seconds
+ )
+ except DatasetError as err:
+ raise err
+ except RepositoryNotFoundError as err:
+ raise DatasetNotFoundError(DOES_NOT_EXIST_OR_PRIVATE_DATASET_ERROR_MESSAGE, cause=err) from err
@@ -215 +228 @@ def get_dataset_info_for_supported_datasets(
- err,
+ cause=err,
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index bd2f1ed9..a1d6e416 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -268,0 +269,2 @@ def get_dataset_info_or_raise(
+ if dataset_info.private:
+ raise DatasetNotFoundError("The dataset does not exist on the Hub.")
diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py
index 1a329ac0..e4aa763f 100644
--- a/services/worker/tests/job_runners/test_config_names.py
+++ b/services/worker/tests/job_runners/test_config_names.py
@@ -133,6 +133,2 @@ def test_compute_splits_response_simple_csv(
- if cause is None:
- assert not exc_info.value.disclose_cause
- assert exc_info.value.cause_exception is None
- else:
- assert exc_info.value.disclose_cause
- assert exc_info.value.cause_exception == cause
+ assert exc_info.value.cause_exception == cause
+ if exc_info.value.disclose_cause:
diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py
index 1d281884..683d61de 100644
--- a/services/worker/tests/job_runners/test_first_rows.py
+++ b/services/worker/tests/job_runners/test_first_rows.py
@@ -165,6 +165,2 @@ def test_number_rows(
- if cause is None:
- assert not exc_info.value.disclose_cause
- assert exc_info.value.cause_exception is None
- else:
- assert exc_info.value.disclose_cause
- assert exc_info.value.cause_exception == cause
+ assert exc_info.value.cause_exception == cause
+ if exc_info.value.disclose_cause:
diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index 1b911c27..68825b9b 100644
--- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -434,2 +434,2 @@ def test_compute_splits_response_simple_csv_ok(
- ("does_not_exist", "DatasetNotFoundError", None),
- ("gated_extra_fields", "GatedExtraFieldsError", None),
+ ("does_not_exist", "DatasetNotFoundError", "HTTPError"),
+ ("gated_extra_fields", "GatedExtraFieldsError", "HTTPError"),
@@ -453,6 +453,2 @@ def test_compute_splits_response_simple_csv_error(
- if cause is None:
- assert not exc_info.value.disclose_cause
- assert exc_info.value.cause_exception is None
- else:
- assert exc_info.value.disclose_cause
- assert exc_info.value.cause_exception == cause
+ assert exc_info.value.cause_exception == cause
+ if exc_info.value.disclose_cause:
diff --git a/services/worker/tests/job_runners/test_split_names_from_streaming.py b/services/worker/tests/job_runners/test_split_names_from_streaming.py
index 19228d75..9b677562 100644
--- a/services/worker/tests/job_runners/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/test_split_names_from_streaming.py
@@ -127,6 +127,2 @@ def test_compute_split_names_from_streaming_response(
- if cause is None:
- assert not exc_info.value.disclose_cause
- assert exc_info.value.cause_exception is None
- else:
- assert exc_info.value.disclose_cause
- assert exc_info.value.cause_exception == cause
+ assert exc_info.value.cause_exception == cause
+ if exc_info.value.disclose_cause:
diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py
index a987b48f..dafd920a 100644
--- a/services/worker/tests/job_runners/test_splits.py
+++ b/services/worker/tests/job_runners/test_splits.py
@@ -134,6 +134,2 @@ def test_compute_splits_response_simple_csv(
- if cause is None:
- assert not exc_info.value.disclose_cause
- assert exc_info.value.cause_exception is None
- else:
- assert exc_info.value.disclose_cause
- assert exc_info.value.cause_exception == cause
+ assert exc_info.value.cause_exception == cause
+ if exc_info.value.disclose_cause:
|
|
1cb8049f1a4f3bd2ba1a0a00c90d50c3ffd66734
|
Sylvain Lesage
| 2023-03-13T19:36:01 |
Add jwt to bypass auth (#898)
|
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml
index 952bf02c..75e4bf11 100644
--- a/.github/workflows/_e2e_tests.yml
+++ b/.github/workflows/_e2e_tests.yml
@@ -74,0 +75,2 @@ jobs:
+ API_HF_JWT_PUBLIC_KEY_URL: "https://huggingface.co/api/keys/jwt"
+ API_HF_JWT_ALGORITHM: "EdDSA"
diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl
index 40214e51..08a088cf 100644
--- a/chart/templates/services/api/_container.tpl
+++ b/chart/templates/services/api/_container.tpl
@@ -14,0 +15,4 @@
+ - name: API_HF_JWT_PUBLIC_KEY_URL
+ value: {{ .Values.api.hfJwtPublicKeyUrl | quote }}
+ - name: API_HF_JWT_ALGORITHM
+ value: {{ .Values.api.hfJwtAlgorithm | quote }}
diff --git a/chart/values.yaml b/chart/values.yaml
index c621c6a2..cdc95256 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -255,0 +256,6 @@ api:
+ # the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format.
+ # It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication
+ # check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored.
+ hfJwtPublicKeyUrl: "https://huggingface.co/api/keys/jwt"
+ # the algorithm used to encode the JWT.
+ hfJwtAlgorithm: "EdDSA"
diff --git a/services/api/README.md b/services/api/README.md
index 49320e7f..2b9d5b8b 100644
--- a/services/api/README.md
+++ b/services/api/README.md
@@ -13,0 +14,2 @@ Set environment variables to configure the application (`API_` prefix):
+- `API_HF_JWT_PUBLIC_KEY_URL`: the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format. It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored. Defaults to empty.
+- `API_HF_JWT_ALGORITHM`: the algorithm used to encode the JWT. Defaults to `"EdDSA"`.
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index 264f58cc..6b1a8955 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -146,0 +147,77 @@ files = [
+[[package]]
+name = "cffi"
+version = "1.15.1"
+description = "Foreign Function Interface for Python calling C code."
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"},
+ {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"},
+ {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"},
+ {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"},
+ {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"},
+ {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"},
+ {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"},
+ {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"},
+ {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"},
+ {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"},
+ {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"},
+ {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"},
+ {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"},
+ {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"},
+ {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"},
+ {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"},
+ {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"},
+ {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"},
+ {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"},
+ {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"},
+ {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"},
+ {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"},
+ {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"},
+ {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"},
+ {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"},
+ {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"},
+ {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"},
+ {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"},
+ {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"},
+ {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"},
+ {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"},
+ {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"},
+ {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"},
+ {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"},
+ {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"},
+]
+
+[package.dependencies]
+pycparser = "*"
+
@@ -335,0 +413,46 @@ toml = ["tomli"]
+[[package]]
+name = "cryptography"
+version = "39.0.2"
+description = "cryptography is a package which provides cryptographic recipes and primitives to Python developers."
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "cryptography-39.0.2-cp36-abi3-macosx_10_12_universal2.whl", hash = "sha256:2725672bb53bb92dc7b4150d233cd4b8c59615cd8288d495eaa86db00d4e5c06"},
+ {file = "cryptography-39.0.2-cp36-abi3-macosx_10_12_x86_64.whl", hash = "sha256:23df8ca3f24699167daf3e23e51f7ba7334d504af63a94af468f468b975b7dd7"},
+ {file = "cryptography-39.0.2-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.manylinux_2_24_aarch64.whl", hash = "sha256:eb40fe69cfc6f5cdab9a5ebd022131ba21453cf7b8a7fd3631f45bbf52bed612"},
+ {file = "cryptography-39.0.2-cp36-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc0521cce2c1d541634b19f3ac661d7a64f9555135e9d8af3980965be717fd4a"},
+ {file = "cryptography-39.0.2-cp36-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ffd394c7896ed7821a6d13b24657c6a34b6e2650bd84ae063cf11ccffa4f1a97"},
+ {file = "cryptography-39.0.2-cp36-abi3-manylinux_2_24_x86_64.whl", hash = "sha256:e8a0772016feeb106efd28d4a328e77dc2edae84dfbac06061319fdb669ff828"},
+ {file = "cryptography-39.0.2-cp36-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:8f35c17bd4faed2bc7797d2a66cbb4f986242ce2e30340ab832e5d99ae60e011"},
+ {file = "cryptography-39.0.2-cp36-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:b49a88ff802e1993b7f749b1eeb31134f03c8d5c956e3c125c75558955cda536"},
+ {file = "cryptography-39.0.2-cp36-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:5f8c682e736513db7d04349b4f6693690170f95aac449c56f97415c6980edef5"},
+ {file = "cryptography-39.0.2-cp36-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:d7d84a512a59f4412ca8549b01f94be4161c94efc598bf09d027d67826beddc0"},
+ {file = "cryptography-39.0.2-cp36-abi3-win32.whl", hash = "sha256:c43ac224aabcbf83a947eeb8b17eaf1547bce3767ee2d70093b461f31729a480"},
+ {file = "cryptography-39.0.2-cp36-abi3-win_amd64.whl", hash = "sha256:788b3921d763ee35dfdb04248d0e3de11e3ca8eb22e2e48fef880c42e1f3c8f9"},
+ {file = "cryptography-39.0.2-pp38-pypy38_pp73-macosx_10_12_x86_64.whl", hash = "sha256:d15809e0dbdad486f4ad0979753518f47980020b7a34e9fc56e8be4f60702fac"},
+ {file = "cryptography-39.0.2-pp38-pypy38_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:50cadb9b2f961757e712a9737ef33d89b8190c3ea34d0fb6675e00edbe35d074"},
+ {file = "cryptography-39.0.2-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:103e8f7155f3ce2ffa0049fe60169878d47a4364b277906386f8de21c9234aa1"},
+ {file = "cryptography-39.0.2-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:6236a9610c912b129610eb1a274bdc1350b5df834d124fa84729ebeaf7da42c3"},
+ {file = "cryptography-39.0.2-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:e944fe07b6f229f4c1a06a7ef906a19652bdd9fd54c761b0ff87e83ae7a30354"},
+ {file = "cryptography-39.0.2-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.manylinux_2_24_aarch64.whl", hash = "sha256:35d658536b0a4117c885728d1a7032bdc9a5974722ae298d6c533755a6ee3915"},
+ {file = "cryptography-39.0.2-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:30b1d1bfd00f6fc80d11300a29f1d8ab2b8d9febb6ed4a38a76880ec564fae84"},
+ {file = "cryptography-39.0.2-pp39-pypy39_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:e029b844c21116564b8b61216befabca4b500e6816fa9f0ba49527653cae2108"},
+ {file = "cryptography-39.0.2-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:fa507318e427169ade4e9eccef39e9011cdc19534f55ca2f36ec3f388c1f70f3"},
+ {file = "cryptography-39.0.2-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:8bc0008ef798231fac03fe7d26e82d601d15bd16f3afaad1c6113771566570f3"},
+ {file = "cryptography-39.0.2.tar.gz", hash = "sha256:bc5b871e977c8ee5a1bbc42fa8d19bcc08baf0c51cbf1586b0e87a2694dde42f"},
+]
+
+[package.dependencies]
+cffi = ">=1.12"
+
+[package.extras]
+docs = ["sphinx (>=5.3.0)", "sphinx-rtd-theme (>=1.1.1)"]
+docstest = ["pyenchant (>=1.6.11)", "sphinxcontrib-spelling (>=4.0.1)", "twine (>=1.12.0)"]
+pep8test = ["black", "check-manifest", "mypy", "ruff", "types-pytz", "types-requests"]
+sdist = ["setuptools-rust (>=0.11.4)"]
+ssh = ["bcrypt (>=3.1.5)"]
+test = ["hypothesis (>=1.11.4,!=3.79.2)", "iso8601", "pretend", "pytest (>=6.2.0)", "pytest-benchmark", "pytest-cov", "pytest-shard (>=0.1.2)", "pytest-subtests", "pytest-xdist", "pytz"]
+test-randomorder = ["pytest-randomly"]
+tox = ["tox"]
+
@@ -1226,0 +1350,12 @@ files = [
+[[package]]
+name = "pycparser"
+version = "2.21"
+description = "C parser in Python"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"},
+ {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"},
+]
+
@@ -1253,0 +1389,21 @@ plugins = ["importlib-metadata"]
+[[package]]
+name = "pyjwt"
+version = "2.6.0"
+description = "JSON Web Token implementation in Python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "PyJWT-2.6.0-py3-none-any.whl", hash = "sha256:d83c3d892a77bbb74d3e1a2cfa90afaadb60945205d1095d9221f04466f64c14"},
+ {file = "PyJWT-2.6.0.tar.gz", hash = "sha256:69285c7e31fc44f68a1feb309e948e0df53259d579295e6cfe2b1792329f05fd"},
+]
+
+[package.dependencies]
+cryptography = {version = ">=3.4.0", optional = true, markers = "extra == \"crypto\""}
+
+[package.extras]
+crypto = ["cryptography (>=3.4.0)"]
+dev = ["coverage[toml] (==5.0.4)", "cryptography (>=3.4.0)", "pre-commit", "pytest (>=6.0.0,<7.0.0)", "sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"]
+docs = ["sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"]
+tests = ["coverage[toml] (==5.0.4)", "pytest (>=6.0.0,<7.0.0)"]
+
@@ -1987 +2143 @@ python-versions = "3.9.15"
-content-hash = "7b9462e0e596249181d714cb2a687ade42839f6ea4e972aa97ed86fb44640b66"
+content-hash = "9f3e741a9858013b84eef040c1c7d4e02cf9843e3a6766d537147784bd05605d"
diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml
index 20a9ba25..00f5afea 100644
--- a/services/api/pyproject.toml
+++ b/services/api/pyproject.toml
@@ -11,0 +12 @@ libcommon = {path = "../../libs/libcommon", develop = true}
+pyjwt = { extras = ["crypto"], version = "^2.6.0" }
@@ -56,0 +58,2 @@ strict = true
+disallow_untyped_calls = false
+# ^ call to expected_algorithm.from_jwk force to set this to false
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 1514a2e3..96ccf8d8 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -15,0 +16 @@ from api.config import AppConfig, EndpointConfig, UvicornConfig
+from api.jwt_token import fetch_jwt_public_key
@@ -38,0 +40,9 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ hf_jwt_public_key = (
+ fetch_jwt_public_key(
+ url=app_config.api.hf_jwt_public_key_url,
+ hf_jwt_algorithm=app_config.api.hf_jwt_algorithm,
+ hf_timeout_seconds=app_config.api.hf_timeout_seconds,
+ )
+ if app_config.api.hf_jwt_public_key_url and app_config.api.hf_jwt_algorithm
+ else None
+ )
@@ -64,0 +75 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ hf_jwt_public_key=hf_jwt_public_key,
@@ -83,0 +95 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ hf_jwt_public_key=hf_jwt_public_key,
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index e2b06d19..3ab521cd 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -5 +5 @@ import logging
-from typing import Literal, Optional
+from typing import Any, Literal, Optional
@@ -11,0 +12 @@ from starlette.requests import Request
+from api.jwt_token import is_jwt_valid
@@ -43,0 +45,2 @@ def auth_check(
+ hf_jwt_public_key: Optional[Any] = None,
+ hf_jwt_algorithm: Optional[str] = None,
@@ -58,2 +61,4 @@ def auth_check(
- request (Request | None): the request which optionally bears authentication headers: "cookie" or
- "authorization"
+ request (Request | None): the request which optionally bears authentication headers: "cookie",
+ "authorization" or "X-Api-Key"
+ hf_jwt_public_key (Any|None): the public key to use to decode the JWT token
+ hf_jwt_algorithm (str): the algorithm to use to decode the JWT token
@@ -67,0 +73,6 @@ def auth_check(
+ if request:
+ token = request.headers.get("x-api-key")
+ if token and is_jwt_valid(
+ dataset=dataset, token=token, public_key=hf_jwt_public_key, algorithm=hf_jwt_algorithm
+ ):
+ return True
@@ -95,15 +106,15 @@ def auth_check(
- with StepProfiler(method="auth_check", step="return or raise"):
- if response.status_code == 200:
- return True
- elif response.status_code == 401:
- raise ExternalUnauthenticatedError(
- "The dataset does not exist, or is not accessible without authentication (private or gated)."
- " Please check the spelling of the dataset name or retry with authentication."
- )
- elif response.status_code in [403, 404]:
- raise ExternalAuthenticatedError(
- "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
- " Please check the spelling of the dataset name or retry with other authentication credentials."
- )
- else:
- raise ValueError(f"Unexpected status code {response.status_code}")
+ with StepProfiler(method="auth_check", step="return or raise"):
+ if response.status_code == 200:
+ return True
+ elif response.status_code == 401:
+ raise ExternalUnauthenticatedError(
+ "The dataset does not exist, or is not accessible without authentication (private or gated). Please"
+ " check the spelling of the dataset name or retry with authentication."
+ )
+ elif response.status_code in [403, 404]:
+ raise ExternalAuthenticatedError(
+ "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ " Please check the spelling of the dataset name or retry with other authentication credentials."
+ )
+ else:
+ raise ValueError(f"Unexpected status code {response.status_code}")
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 8265621e..cb325cea 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -39,0 +40,2 @@ API_HF_AUTH_PATH = "/api/datasets/%s/auth-check"
+API_HF_JWT_PUBLIC_KEY_URL = None
+API_HF_JWT_ALGORITHM = "EdDSA"
@@ -48,0 +51,2 @@ class ApiConfig:
+ hf_jwt_public_key_url: Optional[str] = API_HF_JWT_PUBLIC_KEY_URL
+ hf_jwt_algorithm: Optional[str] = API_HF_JWT_ALGORITHM
@@ -61,0 +66 @@ class ApiConfig:
+ hf_jwt_public_key_url=env.str(name="HF_JWT_PUBLIC_KEY_URL", default=API_HF_JWT_PUBLIC_KEY_URL),
diff --git a/services/api/src/api/jwt_token.py b/services/api/src/api/jwt_token.py
new file mode 100644
index 00000000..c0a673a9
--- /dev/null
+++ b/services/api/src/api/jwt_token.py
@@ -0,0 +1,146 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from typing import Any, Optional, Union
+
+import jwt
+import requests
+from cryptography.hazmat.primitives import serialization
+from cryptography.hazmat.primitives.asymmetric.ec import (
+ EllipticCurvePrivateKey,
+ EllipticCurvePublicKey,
+)
+from cryptography.hazmat.primitives.asymmetric.ed448 import (
+ Ed448PrivateKey,
+ Ed448PublicKey,
+)
+from cryptography.hazmat.primitives.asymmetric.ed25519 import (
+ Ed25519PrivateKey,
+ Ed25519PublicKey,
+)
+from cryptography.hazmat.primitives.asymmetric.rsa import RSAPrivateKey, RSAPublicKey
+from jwt.algorithms import (
+ ECAlgorithm,
+ HMACAlgorithm,
+ OKPAlgorithm,
+ RSAAlgorithm,
+ RSAPSSAlgorithm,
+)
+
+from api.utils import JWKError
+
+ASYMMETRIC_ALGORITHMS = (ECAlgorithm, OKPAlgorithm, RSAAlgorithm, RSAPSSAlgorithm)
+SYMMETRIC_ALGORITHMS = (HMACAlgorithm,)
+
+
+def is_public_key(
+ key: Union[
+ EllipticCurvePublicKey,
+ EllipticCurvePrivateKey,
+ Ed25519PublicKey,
+ Ed448PublicKey,
+ Ed25519PrivateKey,
+ Ed448PrivateKey,
+ RSAPrivateKey,
+ RSAPublicKey,
+ ]
+) -> bool:
+ return hasattr(key, "public_bytes")
+
+
+def parse_jwt_public_key(keys: Any, hf_jwt_algorithm: str) -> str:
+ """parse the input JSON to extract the public key
+
+ Note that the return type is Any in order not to enter in too much details. See
+ https://github.com/jpadilla/pyjwt/blob/777efa2f51249f63b0f95804230117723eca5d09/jwt/algorithms.py#L629-L651
+ In our case, the type should be cryptography.hazmat.backends.openssl.ed25519._Ed25519PublicKey
+
+ Args:
+ keys (Any): the JSON to parse
+ hf_jwt_algorithm (str): the JWT algorithm to use.
+
+ Returns:
+ str: the public key
+ """
+ try:
+ expected_algorithm = jwt.get_algorithm_by_name(hf_jwt_algorithm)
+ if not isinstance(expected_algorithm, (*ASYMMETRIC_ALGORITHMS, *SYMMETRIC_ALGORITHMS)):
+ raise NotImplementedError()
+ except NotImplementedError as err:
+ raise RuntimeError(f"Invalid algorithm for JWT verification: {hf_jwt_algorithm} is not supported") from err
+
+ if not isinstance(keys, list):
+ raise ValueError("Payload from moon must be a list of JWK formatted keys.")
+ try:
+ key = expected_algorithm.from_jwk(keys[0])
+ if not isinstance(expected_algorithm, ASYMMETRIC_ALGORITHMS):
+ return key.decode("utf-8") # type: ignore[no-any-return]
+ if not is_public_key(key):
+ raise RuntimeError("Failed to parse JWT key: the provided key is a private key")
+ return key.public_bytes( # type: ignore[no-any-return]
+ encoding=serialization.Encoding.PEM,
+ format=serialization.PublicFormat.SubjectPublicKeyInfo,
+ ).decode("utf-8")
+ # ^ we assume that the key contain UTF-8 encoded bytes, which is why we use type ignore for mypy
+ except (jwt.InvalidKeyError, KeyError) as err:
+ raise RuntimeError(f"Failed to parse JWT key: {err.args[0]}") from err
+
+
+def fetch_jwt_public_key(
+ url: str,
+ hf_jwt_algorithm: str,
+ hf_timeout_seconds: Optional[float] = None,
+) -> str:
+ """fetch the public key to decode the JWT token from the input URL
+
+ See https://huggingface.co/api/keys/jwt
+
+ Args:
+ url (str): the URL to fetch the public key from
+ hf_jwt_algorithm (str): the JWT algorithm to use.
+ hf_timeout_seconds (float|None): the timeout in seconds for the external authentication service. It
+ is used both for the connection timeout and the read timeout. If None, the request never timeouts.
+
+ Returns:
+ str: the public key
+ """
+ try:
+ response = requests.get(url, timeout=hf_timeout_seconds)
+ response.raise_for_status()
+ return parse_jwt_public_key(keys=response.json(), hf_jwt_algorithm=hf_jwt_algorithm)
+ except Exception as err:
+ raise JWKError(f"Failed to fetch or parse the JWT public key from {url}. ", cause=err) from err
+
+
+def is_jwt_valid(dataset: str, token: Any, public_key: Optional[str], algorithm: Optional[str]) -> bool:
+ """
+ Check if the JWT is valid for the dataset.
+
+ The JWT is decoded with the public key, and the "sub" claim must be:
+ {"repoName": <...>, "repoType": "dataset", "read": true}
+ where <...> is the dataset identifier.
+
+ Returns True only if all the conditions are met. Else, it returns False.
+
+ Args:
+ dataset (str): the dataset identifier
+ token (Any): the JWT token to decode
+ public_key (str|None): the public key to use to decode the JWT token
+ algorithm (str|None): the algorithm to use to decode the JWT token
+
+ Returns:
+ bool: True if the JWT is valid for the input dataset, else False
+ """
+ if not public_key or not algorithm:
+ return False
+ try:
+ decoded = jwt.decode(
+ jwt=token, key=public_key, algorithms=[algorithm], options={"require": ["exp", "sub", "read"]}
+ )
+ except Exception:
+ return False
+ sub = decoded.get("sub")
+ if not isinstance(sub, str) or not sub.startswith("datasets/") or sub.removeprefix("datasets/") != dataset:
+ return False
+ read = decoded.get("read")
+ return read is True
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index e8ab75d6..f620f368 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -7 +7 @@ from http import HTTPStatus
-from typing import List, Mapping, Optional, Tuple
+from typing import Any, List, Mapping, Optional, Tuple
@@ -204,0 +205 @@ def create_endpoint(
+ hf_jwt_public_key: Optional[Any] = None,
@@ -247,0 +249 @@ def create_endpoint(
+ hf_jwt_public_key=hf_jwt_public_key,
@@ -250 +251,0 @@ def create_endpoint(
-
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 65f54177..218ef36b 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -5 +5 @@ import logging
-from typing import List, Optional, Set
+from typing import Any, List, Optional, Set
@@ -71,0 +72 @@ def create_is_valid_endpoint(
+ hf_jwt_public_key: Optional[Any] = None,
@@ -91,0 +93 @@ def create_is_valid_endpoint(
+ hf_jwt_public_key=hf_jwt_public_key,
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index 52a04024..09eb4a24 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -20,0 +21 @@ ApiErrorCode = Literal[
+ "JWKError",
@@ -84,0 +86,7 @@ class ExternalAuthenticatedError(ApiCustomError):
+class JWKError(ApiCustomError):
+ """Raised when the JWT key (JWK) could not be fetched or parsed."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "JWKError", cause=cause, disclose_cause=False)
+
+
diff --git a/services/api/tests/test_authentication.py b/services/api/tests/test_authentication.py
index 4ffabefc..ab32c1dd 100644
--- a/services/api/tests/test_authentication.py
+++ b/services/api/tests/test_authentication.py
@@ -6 +6 @@ from contextlib import nullcontext as does_not_raise
-from typing import Any, Mapping, Optional
+from typing import Any, Dict, Mapping, Optional
@@ -7,0 +8 @@ from typing import Any, Mapping, Optional
+import jwt
@@ -12,0 +14,2 @@ from starlette.requests import Request
+from werkzeug.wrappers import Request as WerkzeugRequest
+from werkzeug.wrappers import Response as WerkzeugResponse
@@ -20,0 +24,7 @@ from api.utils import (
+from .test_jwt_token import (
+ algorithm_rs256,
+ dataset_ok,
+ payload_ok,
+ private_key,
+ public_key,
+)
@@ -130,0 +141,37 @@ def test_valid_responses_with_request(
+
+
+def raise_value_error(request: WerkzeugRequest) -> WerkzeugResponse:
+ return WerkzeugResponse(status=500) # <- will raise ValueError in auth_check
+
+
[email protected](
+ "hf_jwt_public_key,header,payload,expectation",
+ [
+ (None, None, payload_ok, pytest.raises(ValueError)),
+ (None, "X-Api-Key", payload_ok, pytest.raises(ValueError)),
+ (public_key, None, payload_ok, pytest.raises(ValueError)),
+ (public_key, "X-Api-Key", {}, pytest.raises(ValueError)),
+ (public_key, "X-Api-Key", payload_ok, does_not_raise()),
+ (public_key, "x-api-key", payload_ok, does_not_raise()),
+ ],
+)
+def test_bypass_auth_public_key(
+ httpserver: HTTPServer,
+ hf_endpoint: str,
+ hf_auth_path: str,
+ hf_jwt_public_key: Optional[str],
+ header: Optional[str],
+ payload: Dict[str, str],
+ expectation: Any,
+) -> None:
+ external_auth_url = hf_endpoint + hf_auth_path
+ httpserver.expect_request(hf_auth_path % dataset_ok).respond_with_handler(raise_value_error)
+ headers = {header: jwt.encode(payload, private_key, algorithm=algorithm_rs256)} if header else {}
+ with expectation:
+ auth_check(
+ dataset=dataset_ok,
+ external_auth_url=external_auth_url,
+ request=create_request(headers=headers),
+ hf_jwt_public_key=hf_jwt_public_key,
+ hf_jwt_algorithm=algorithm_rs256,
+ )
diff --git a/services/api/tests/test_jwt_token.py b/services/api/tests/test_jwt_token.py
new file mode 100644
index 00000000..4330881a
--- /dev/null
+++ b/services/api/tests/test_jwt_token.py
@@ -0,0 +1,98 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import datetime
+from contextlib import nullcontext as does_not_raise
+from typing import Any, Dict, Optional
+
+import jwt
+import pytest
+
+from api.jwt_token import is_jwt_valid, parse_jwt_public_key
+
+HUB_JWT_KEYS = [{"crv": "Ed25519", "x": "-RBhgyNluwaIL5KFJb6ZOL2H1nmyI8mW4Z2EHGDGCXM", "kty": "OKP"}]
+UNSUPPORTED_ALGORITHM_JWT_KEYS = [
+ {
+ "alg": "EC",
+ "crv": "P-256",
+ "x": "MKBCTNIcKUSDii11ySs3526iDZ8AiTo7Tu6KPAqv7D4",
+ "y": "4Etl6SRW2YiLUrN5vfvVHuhp7x8PxltmWWlbbM4IFyM",
+ "use": "enc",
+ "kid": "1",
+ }
+]
+
+
[email protected](
+ "keys,expectation",
+ [
+ (HUB_JWT_KEYS, does_not_raise()),
+ ([], pytest.raises(Exception)),
+ (UNSUPPORTED_ALGORITHM_JWT_KEYS, pytest.raises(Exception)),
+ ],
+)
+def test_parse_jwk(
+ keys: Any,
+ expectation: Any,
+) -> None:
+ with expectation:
+ parse_jwt_public_key(keys=keys, hf_jwt_algorithm="EdDSA")
+
+
+private_key = """-----BEGIN RSA PRIVATE KEY-----
+MIIBOQIBAAJAZTmplhS/Jd73ycVut7TglMObheQqXM7RZYlwazLU4wpfIVIwOh9I
+sCZGSgLyFq42KWIikKLEs/yqx3pRGfq+rwIDAQABAkAMyF9WCICq86Eu5bO5lynV
+H26AVfPTjHp87AI6R00C7p9n8hO/DhHaHpc3InOSsXsw9d2hmz37jwwBFiwMHMMh
+AiEAtbttHlIO+yO29oXw4P6+yO11lMy1UpT1sPVTnR9TXbUCIQCOl7Zuyy2ZY9ZW
+pDhW91x/14uXjnLXPypgY9bcfggJUwIhAJQG1LzrzjQWRUPMmgZKuhBkC3BmxhM8
+LlwzmCXVjEw5AiA7JnAFEb9+q82T71d3q/DxD0bWvb6hz5ASoBfXK2jGBQIgbaQp
+h4Tk6UJuj1xgKNs75Pk3pG2tj8AQiuBk3l62vRU=
+-----END RSA PRIVATE KEY-----"""
+public_key = """-----BEGIN PUBLIC KEY-----
+MFswDQYJKoZIhvcNAQEBBQADSgAwRwJAZTmplhS/Jd73ycVut7TglMObheQqXM7R
+ZYlwazLU4wpfIVIwOh9IsCZGSgLyFq42KWIikKLEs/yqx3pRGfq+rwIDAQAB
+-----END PUBLIC KEY-----"""
+other_public_key = """-----BEGIN PUBLIC KEY-----
+MFswDQYJKoZIhvcNAQEBBQADSgAwRwJAecoNIHMXczWkzTp9ePEcx6vPibrZVz/z
+xYGX6G2jFcwFdsrO9nCecrtpSw5lwjW40aNVL9NL9yxPxDi2dyq4wQIDAQAB
+-----END PUBLIC KEY-----"""
+
+
+dataset_ok = "dataset"
+wrong_dataset = "wrong_dataset"
+exp_ok = datetime.datetime.now().timestamp() + 1000
+wrong_exp_1 = datetime.datetime.now().timestamp() - 1000
+wrong_exp_2 = 1
+sub_ok = f"datasets/{dataset_ok}"
+sub_wrong_1 = dataset_ok
+sub_wrong_2 = f"dataset/{dataset_ok}"
+sub_wrong_3 = f"models/{dataset_ok}"
+sub_wrong_4 = f"datasets/{wrong_dataset}"
+read_ok = True
+read_wrong_1 = False
+read_wrong_2 = "True"
+payload_ok = {"sub": sub_ok, "read": read_ok, "exp": exp_ok}
+algorithm_rs256 = "RS256"
+
+
[email protected](
+ "public_key,payload,expected",
+ [
+ (None, payload_ok, False),
+ (other_public_key, payload_ok, False),
+ (public_key, {}, False),
+ (public_key, {"sub": dataset_ok}, False),
+ (public_key, {"sub": sub_wrong_1, "read": read_ok, "exp": exp_ok}, False),
+ (public_key, {"sub": sub_wrong_2, "read": read_ok, "exp": exp_ok}, False),
+ (public_key, {"sub": sub_wrong_3, "read": read_ok, "exp": exp_ok}, False),
+ (public_key, {"sub": sub_wrong_4, "read": read_ok, "exp": exp_ok}, False),
+ (public_key, {"sub": sub_ok, "read": read_wrong_1, "exp": exp_ok}, False),
+ (public_key, {"sub": sub_ok, "read": read_wrong_2, "exp": exp_ok}, False),
+ (public_key, {"sub": sub_ok, "read": read_ok, "exp": wrong_exp_1}, False),
+ (public_key, {"sub": sub_ok, "read": read_ok, "exp": wrong_exp_2}, False),
+ (public_key, payload_ok, True),
+ ],
+)
+def test_is_jwt_valid(public_key: Optional[str], payload: Dict[str, str], expected: bool) -> None:
+ token = jwt.encode(payload, private_key, algorithm=algorithm_rs256)
+ assert is_jwt_valid(dataset=dataset_ok, token=token, public_key=public_key, algorithm=algorithm_rs256) is expected
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index 455a7021..bab6e8be 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -56,0 +57,2 @@ services:
+ API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://huggingface.co/api/keys/jwt}
+ API_HF_JWT_ALGORITHM: ${API_HF_JWT_ALGORITHM-EdDSA}
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 3797b83e..8c68791c 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -56,0 +57,2 @@ services:
+ API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://huggingface.co/api/keys/jwt}
+ API_HF_JWT_ALGORITHM: ${API_HF_JWT_ALGORITHM-EdDSA}
|
|
d9545543b881e0f1232f0c1ea6c2f6217272ac62
|
Quentin Lhoest
| 2023-03-13T18:00:41 |
Don't rebuild dev images (#923)
|
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml
index c6ab2b65..952bf02c 100644
--- a/.github/workflows/_e2e_tests.yml
+++ b/.github/workflows/_e2e_tests.yml
@@ -9 +9 @@ env:
- poetry-version: 1.3.2
+ poetry-version: 1.4.0
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml
index 1ad2c343..2b4e6b15 100644
--- a/.github/workflows/_quality-python.yml
+++ b/.github/workflows/_quality-python.yml
@@ -18 +18 @@ env:
- poetry-version: "1.3.2"
+ poetry-version: "1.4.0"
diff --git a/.github/workflows/_unit-tests-python.yml b/.github/workflows/_unit-tests-python.yml
index 446d3046..232c189b 100644
--- a/.github/workflows/_unit-tests-python.yml
+++ b/.github/workflows/_unit-tests-python.yml
@@ -19 +19 @@ env:
- poetry-version: "1.3.2"
+ poetry-version: "1.4.0"
diff --git a/.github/workflows/openapi-spec.yml b/.github/workflows/openapi-spec.yml
index 2106a43e..a8062cfc 100644
--- a/.github/workflows/openapi-spec.yml
+++ b/.github/workflows/openapi-spec.yml
@@ -19 +19 @@ env:
- poetry-version: 1.3.2
+ poetry-version: 1.4.0
diff --git a/jobs/mongodb_migration/Dockerfile b/jobs/mongodb_migration/Dockerfile
index 3b83aa7e..2dfb42a4 100644
--- a/jobs/mongodb_migration/Dockerfile
+++ b/jobs/mongodb_migration/Dockerfile
@@ -13 +13 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VERSION=1.3.2 \
+ POETRY_VERSION=1.4.0 \
@@ -19 +19 @@ RUN apt-get update \
- && apt-get install -y build-essential python3-dev make \
+ && apt-get install -y build-essential make \
diff --git a/services/admin/Dockerfile b/services/admin/Dockerfile
index 1ce73069..f63efeab 100644
--- a/services/admin/Dockerfile
+++ b/services/admin/Dockerfile
@@ -13 +13 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VERSION=1.3.2 \
+ POETRY_VERSION=1.4.0 \
@@ -19 +19 @@ RUN apt-get update \
- && apt-get install -y build-essential python3-dev make \
+ && apt-get install -y build-essential make \
diff --git a/services/admin/dev.Dockerfile b/services/admin/dev.Dockerfile
new file mode 100644
index 00000000..c471992e
--- /dev/null
+++ b/services/admin/dev.Dockerfile
@@ -0,0 +1,44 @@
+# build with
+# docker build -t some_tag_admin -f Dockerfile ../..
+FROM python:3.9.15-slim
+
+ENV PYTHONFAULTHANDLER=1 \
+ PYTHONUNBUFFERED=1 \
+ PYTHONHASHSEED=random \
+ PIP_NO_CACHE_DIR=1 \
+ PIP_DISABLE_PIP_VERSION_CHECK=on \
+ PIP_DEFAULT_TIMEOUT=100 \
+ POETRY_NO_INTERACTION=1 \
+ # Versions:
+ POETRY_VERSION=1.4.0 \
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
+
+# System deps:
+RUN apt-get update \
+ && apt-get install -y build-essential make \
+ && rm -rf /var/lib/apt/lists/*
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
+
+WORKDIR /src
+COPY libs/libcommon/poetry.lock ./libs/libcommon/poetry.lock
+COPY libs/libcommon/pyproject.toml ./libs/libcommon/pyproject.toml
+COPY services/admin/poetry.lock ./services/admin/poetry.lock
+COPY services/admin/pyproject.toml ./services/admin/pyproject.toml
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# Initialize an empty libcommon
+# Mapping a volume to ./libs/libcommon/src is required when running this image.
+RUN mkdir ./libs/libcommon/src && mkdir ./libs/libcommon/src/libcommon && touch ./libs/libcommon/src/libcommon/__init__.py
+
+# Install dependencies
+WORKDIR /src/services/admin/
+RUN --mount=type=cache,target=/home/.cache/pypoetry/cache \
+ --mount=type=cache,target=/home/.cache/pypoetry/artifacts \
+ poetry install --no-root
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# Install the admin package.
+# Mapping a volume to ./services/admin/src is required when running this image.
+ENTRYPOINT ["/bin/sh", "-c" , "poetry install --only-root && poetry run python src/admin/main.py"]
diff --git a/services/api/Dockerfile b/services/api/Dockerfile
index 52483230..3a77155e 100644
--- a/services/api/Dockerfile
+++ b/services/api/Dockerfile
@@ -13 +13 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VERSION=1.3.2 \
+ POETRY_VERSION=1.4.0 \
@@ -19 +19 @@ RUN apt-get update \
- && apt-get install -y build-essential unzip wget python3-dev \
+ && apt-get install -y build-essential unzip wget \
diff --git a/services/api/dev.Dockerfile b/services/api/dev.Dockerfile
new file mode 100644
index 00000000..340c5927
--- /dev/null
+++ b/services/api/dev.Dockerfile
@@ -0,0 +1,44 @@
+# build with
+# docker build -t some_tag_api -f Dockerfile ../..
+FROM python:3.9.15-slim
+
+ENV PYTHONFAULTHANDLER=1 \
+ PYTHONUNBUFFERED=1 \
+ PYTHONHASHSEED=random \
+ PIP_NO_CACHE_DIR=1 \
+ PIP_DISABLE_PIP_VERSION_CHECK=on \
+ PIP_DEFAULT_TIMEOUT=100 \
+ POETRY_NO_INTERACTION=1 \
+ # Versions:
+ POETRY_VERSION=1.4.0 \
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
+
+# System deps:
+RUN apt-get update \
+ && apt-get install -y build-essential unzip wget \
+ && rm -rf /var/lib/apt/lists/*
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
+
+WORKDIR /src
+COPY libs/libcommon/poetry.lock ./libs/libcommon/poetry.lock
+COPY libs/libcommon/pyproject.toml ./libs/libcommon/pyproject.toml
+COPY services/api/poetry.lock ./services/api/poetry.lock
+COPY services/api/pyproject.toml ./services/api/pyproject.toml
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# Initialize an empty libcommon
+# Mapping a volume to ./libs/libcommon/src is required when running this image.
+RUN mkdir ./libs/libcommon/src && mkdir ./libs/libcommon/src/libcommon && touch ./libs/libcommon/src/libcommon/__init__.py
+
+# Install dependencies
+WORKDIR /src/services/api/
+RUN --mount=type=cache,target=/home/.cache/pypoetry/cache \
+ --mount=type=cache,target=/home/.cache/pypoetry/artifacts \
+ poetry install --no-root
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# Install the api package.
+# Mapping a volume to ./services/api/src is required when running this image.
+ENTRYPOINT ["/bin/sh", "-c" , "poetry install --only-root && poetry run python src/api/main.py"]
diff --git a/services/worker/Dockerfile b/services/worker/Dockerfile
index 93882054..922ebb17 100644
--- a/services/worker/Dockerfile
+++ b/services/worker/Dockerfile
@@ -13 +13 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VERSION=1.3.2 \
+ POETRY_VERSION=1.4.0 \
@@ -19 +19 @@ RUN apt-get update \
- && apt-get install -y build-essential unzip wget python3-dev make \
+ && apt-get install -y build-essential unzip wget make \
diff --git a/services/worker/dev.Dockerfile b/services/worker/dev.Dockerfile
index 2468dd10..e014bb61 100644
--- a/services/worker/dev.Dockerfile
+++ b/services/worker/dev.Dockerfile
@@ -13 +13 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VERSION=1.3.2 \
+ POETRY_VERSION=1.4.0 \
@@ -19 +19 @@ RUN apt-get update \
- && apt-get install -y build-essential unzip wget python3-dev make \
+ && apt-get install -y build-essential unzip wget make \
@@ -27,0 +28,2 @@ WORKDIR /src
+COPY libs/libcommon/poetry.lock ./libs/libcommon/poetry.lock
+COPY libs/libcommon/pyproject.toml ./libs/libcommon/pyproject.toml
@@ -31 +33,7 @@ COPY services/worker/pyproject.toml ./services/worker/pyproject.toml
-COPY libs/libcommon ./libs/libcommon
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# Initialize an empty libcommon
+# Mapping a volume to ./libs/libcommon/src is required when running this image.
+RUN mkdir ./libs/libcommon/src && mkdir ./libs/libcommon/src/libcommon && touch ./libs/libcommon/src/libcommon/__init__.py
+
+# Install dependencies
@@ -38,5 +46,2 @@ RUN --mount=type=cache,target=/home/.cache/pypoetry/cache \
-# No need to copy the source code since we map a volume in docker-compose-base.yaml
-# Removed: COPY services/worker/src ./src
-# Removed: RUN poetry install --no-cache
-# However we need to install the package when the container starts
-# Added: poetry install
+# Install the worker package.
+# Mapping a volume to ./services/worker/src is required when running this image.
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 23b97f42..0934c65b 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -276 +276 @@ name = "bandit"
-version = "1.7.4"
+version = "1.7.5"
@@ -282,2 +282,2 @@ files = [
- {file = "bandit-1.7.4-py3-none-any.whl", hash = "sha256:412d3f259dab4077d0e7f0c11f50f650cc7d10db905d98f6520a95a18049658a"},
- {file = "bandit-1.7.4.tar.gz", hash = "sha256:2d63a8c573417bae338962d4b9b06fbc6080f74ecd955a092849e1e65c717bd2"},
+ {file = "bandit-1.7.5-py3-none-any.whl", hash = "sha256:75665181dc1e0096369112541a056c59d1c5f66f9bb74a8d686c3c362b83f549"},
+ {file = "bandit-1.7.5.tar.gz", hash = "sha256:bdfc739baa03b880c2d15d0431b31c658ffc348e907fe197e54e0389dd59e11e"},
@@ -289,0 +290 @@ PyYAML = ">=5.3.1"
+rich = "*"
@@ -293,2 +294,2 @@ stevedore = ">=1.20.0"
-test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "toml"]
-toml = ["toml"]
+test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "tomli (>=1.1.0)"]
+toml = ["tomli (>=1.1.0)"]
@@ -636 +637 @@ name = "charset-normalizer"
-version = "3.0.1"
+version = "3.1.0"
@@ -640 +641 @@ optional = false
-python-versions = "*"
+python-versions = ">=3.7.0"
@@ -642,88 +643,75 @@ files = [
- {file = "charset-normalizer-3.0.1.tar.gz", hash = "sha256:ebea339af930f8ca5d7a699b921106c6e29c617fe9606fa7baa043c1cdae326f"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:88600c72ef7587fe1708fd242b385b6ed4b8904976d5da0893e31df8b3480cb6"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c75ffc45f25324e68ab238cb4b5c0a38cd1c3d7f1fb1f72b5541de469e2247db"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:db72b07027db150f468fbada4d85b3b2729a3db39178abf5c543b784c1254539"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:62595ab75873d50d57323a91dd03e6966eb79c41fa834b7a1661ed043b2d404d"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ff6f3db31555657f3163b15a6b7c6938d08df7adbfc9dd13d9d19edad678f1e8"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:772b87914ff1152b92a197ef4ea40efe27a378606c39446ded52c8f80f79702e"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:70990b9c51340e4044cfc394a81f614f3f90d41397104d226f21e66de668730d"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:292d5e8ba896bbfd6334b096e34bffb56161c81408d6d036a7dfa6929cff8783"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:2edb64ee7bf1ed524a1da60cdcd2e1f6e2b4f66ef7c077680739f1641f62f555"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:31a9ddf4718d10ae04d9b18801bd776693487cbb57d74cc3458a7673f6f34639"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:44ba614de5361b3e5278e1241fda3dc1838deed864b50a10d7ce92983797fa76"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:12db3b2c533c23ab812c2b25934f60383361f8a376ae272665f8e48b88e8e1c6"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c512accbd6ff0270939b9ac214b84fb5ada5f0409c44298361b2f5e13f9aed9e"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-win32.whl", hash = "sha256:502218f52498a36d6bf5ea77081844017bf7982cdbe521ad85e64cabee1b608b"},
- {file = "charset_normalizer-3.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:601f36512f9e28f029d9481bdaf8e89e5148ac5d89cffd3b05cd533eeb423b59"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0298eafff88c99982a4cf66ba2efa1128e4ddaca0b05eec4c456bbc7db691d8d"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:a8d0fc946c784ff7f7c3742310cc8a57c5c6dc31631269876a88b809dbeff3d3"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:87701167f2a5c930b403e9756fab1d31d4d4da52856143b609e30a1ce7160f3c"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:14e76c0f23218b8f46c4d87018ca2e441535aed3632ca134b10239dfb6dadd6b"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0c0a590235ccd933d9892c627dec5bc7511ce6ad6c1011fdf5b11363022746c1"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8c7fe7afa480e3e82eed58e0ca89f751cd14d767638e2550c77a92a9e749c317"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:79909e27e8e4fcc9db4addea88aa63f6423ebb171db091fb4373e3312cb6d603"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8ac7b6a045b814cf0c47f3623d21ebd88b3e8cf216a14790b455ea7ff0135d18"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:72966d1b297c741541ca8cf1223ff262a6febe52481af742036a0b296e35fa5a"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f9d0c5c045a3ca9bedfc35dca8526798eb91a07aa7a2c0fee134c6c6f321cbd7"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:5995f0164fa7df59db4746112fec3f49c461dd6b31b841873443bdb077c13cfc"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:4a8fcf28c05c1f6d7e177a9a46a1c52798bfe2ad80681d275b10dcf317deaf0b"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:761e8904c07ad053d285670f36dd94e1b6ab7f16ce62b9805c475b7aa1cffde6"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-win32.whl", hash = "sha256:71140351489970dfe5e60fc621ada3e0f41104a5eddaca47a7acb3c1b851d6d3"},
- {file = "charset_normalizer-3.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:9ab77acb98eba3fd2a85cd160851816bfce6871d944d885febf012713f06659c"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:84c3990934bae40ea69a82034912ffe5a62c60bbf6ec5bc9691419641d7d5c9a"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:74292fc76c905c0ef095fe11e188a32ebd03bc38f3f3e9bcb85e4e6db177b7ea"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c95a03c79bbe30eec3ec2b7f076074f4281526724c8685a42872974ef4d36b72"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f4c39b0e3eac288fedc2b43055cfc2ca7a60362d0e5e87a637beac5d801ef478"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:df2c707231459e8a4028eabcd3cfc827befd635b3ef72eada84ab13b52e1574d"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:93ad6d87ac18e2a90b0fe89df7c65263b9a99a0eb98f0a3d2e079f12a0735837"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:59e5686dd847347e55dffcc191a96622f016bc0ad89105e24c14e0d6305acbc6"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:cd6056167405314a4dc3c173943f11249fa0f1b204f8b51ed4bde1a9cd1834dc"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:083c8d17153ecb403e5e1eb76a7ef4babfc2c48d58899c98fcaa04833e7a2f9a"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:f5057856d21e7586765171eac8b9fc3f7d44ef39425f85dbcccb13b3ebea806c"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:7eb33a30d75562222b64f569c642ff3dc6689e09adda43a082208397f016c39a"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-win32.whl", hash = "sha256:95dea361dd73757c6f1c0a1480ac499952c16ac83f7f5f4f84f0658a01b8ef41"},
- {file = "charset_normalizer-3.0.1-cp36-cp36m-win_amd64.whl", hash = "sha256:eaa379fcd227ca235d04152ca6704c7cb55564116f8bc52545ff357628e10602"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:3e45867f1f2ab0711d60c6c71746ac53537f1684baa699f4f668d4c6f6ce8e14"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cadaeaba78750d58d3cc6ac4d1fd867da6fc73c88156b7a3212a3cd4819d679d"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:911d8a40b2bef5b8bbae2e36a0b103f142ac53557ab421dc16ac4aafee6f53dc"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:503e65837c71b875ecdd733877d852adbc465bd82c768a067badd953bf1bc5a3"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a60332922359f920193b1d4826953c507a877b523b2395ad7bc716ddd386d866"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16a8663d6e281208d78806dbe14ee9903715361cf81f6d4309944e4d1e59ac5b"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:a16418ecf1329f71df119e8a65f3aa68004a3f9383821edcb20f0702934d8087"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:9d9153257a3f70d5f69edf2325357251ed20f772b12e593f3b3377b5f78e7ef8"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:02a51034802cbf38db3f89c66fb5d2ec57e6fe7ef2f4a44d070a593c3688667b"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:2e396d70bc4ef5325b72b593a72c8979999aa52fb8bcf03f701c1b03e1166918"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:11b53acf2411c3b09e6af37e4b9005cba376c872503c8f28218c7243582df45d"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-win32.whl", hash = "sha256:0bf2dae5291758b6f84cf923bfaa285632816007db0330002fa1de38bfcb7154"},
- {file = "charset_normalizer-3.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:2c03cc56021a4bd59be889c2b9257dae13bf55041a3372d3295416f86b295fb5"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:024e606be3ed92216e2b6952ed859d86b4cfa52cd5bc5f050e7dc28f9b43ec42"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:4b0d02d7102dd0f997580b51edc4cebcf2ab6397a7edf89f1c73b586c614272c"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:358a7c4cb8ba9b46c453b1dd8d9e431452d5249072e4f56cfda3149f6ab1405e"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81d6741ab457d14fdedc215516665050f3822d3e56508921cc7239f8c8e66a58"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8b8af03d2e37866d023ad0ddea594edefc31e827fee64f8de5611a1dbc373174"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9cf4e8ad252f7c38dd1f676b46514f92dc0ebeb0db5552f5f403509705e24753"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e696f0dd336161fca9adbb846875d40752e6eba585843c768935ba5c9960722b"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c22d3fe05ce11d3671297dc8973267daa0f938b93ec716e12e0f6dee81591dc1"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:109487860ef6a328f3eec66f2bf78b0b72400280d8f8ea05f69c51644ba6521a"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:37f8febc8ec50c14f3ec9637505f28e58d4f66752207ea177c1d67df25da5aed"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:f97e83fa6c25693c7a35de154681fcc257c1c41b38beb0304b9c4d2d9e164479"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:a152f5f33d64a6be73f1d30c9cc82dfc73cec6477ec268e7c6e4c7d23c2d2291"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:39049da0ffb96c8cbb65cbf5c5f3ca3168990adf3551bd1dee10c48fce8ae820"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-win32.whl", hash = "sha256:4457ea6774b5611f4bed5eaa5df55f70abde42364d498c5134b7ef4c6958e20e"},
- {file = "charset_normalizer-3.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:e62164b50f84e20601c1ff8eb55620d2ad25fb81b59e3cd776a1902527a788af"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:8eade758719add78ec36dc13201483f8e9b5d940329285edcd5f70c0a9edbd7f"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8499ca8f4502af841f68135133d8258f7b32a53a1d594aa98cc52013fff55678"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3fc1c4a2ffd64890aebdb3f97e1278b0cc72579a08ca4de8cd2c04799a3a22be"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:00d3ffdaafe92a5dc603cb9bd5111aaa36dfa187c8285c543be562e61b755f6b"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c2ac1b08635a8cd4e0cbeaf6f5e922085908d48eb05d44c5ae9eabab148512ca"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f6f45710b4459401609ebebdbcfb34515da4fc2aa886f95107f556ac69a9147e"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3ae1de54a77dc0d6d5fcf623290af4266412a7c4be0b1ff7444394f03f5c54e3"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3b590df687e3c5ee0deef9fc8c547d81986d9a1b56073d82de008744452d6541"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ab5de034a886f616a5668aa5d098af2b5385ed70142090e2a31bcbd0af0fdb3d"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:9cb3032517f1627cc012dbc80a8ec976ae76d93ea2b5feaa9d2a5b8882597579"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:608862a7bf6957f2333fc54ab4399e405baad0163dc9f8d99cb236816db169d4"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0f438ae3532723fb6ead77e7c604be7c8374094ef4ee2c5e03a3a17f1fca256c"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:356541bf4381fa35856dafa6a965916e54bed415ad8a24ee6de6e37deccf2786"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-win32.whl", hash = "sha256:39cf9ed17fe3b1bc81f33c9ceb6ce67683ee7526e65fde1447c772afc54a1bb8"},
- {file = "charset_normalizer-3.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:0a11e971ed097d24c534c037d298ad32c6ce81a45736d31e0ff0ad37ab437d59"},
- {file = "charset_normalizer-3.0.1-py3-none-any.whl", hash = "sha256:7e189e2e1d3ed2f4aebabd2d5b0f931e883676e51c7624826e0a4e5fe8a0bf24"},
+ {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"},
+ {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"},
@@ -966 +954 @@ name = "datasets"
-version = "2.10.0"
+version = "2.10.1"
@@ -972,2 +960,2 @@ files = [
- {file = "datasets-2.10.0-py3-none-any.whl", hash = "sha256:42968d0cc88608ebc4a92a2fc26b4cd65e38f3ce3f5bb90a468cae6a6f0de989"},
- {file = "datasets-2.10.0.tar.gz", hash = "sha256:c2ddab0899e17590fe39bc5783828ac9a7331c74508ab7a345ea0dca3112ead0"},
+ {file = "datasets-2.10.1-py3-none-any.whl", hash = "sha256:bfde7253b31abfd075f9ad55961b13c135a5a9295cea04ae86c47b9c0e0466fd"},
+ {file = "datasets-2.10.1.tar.gz", hash = "sha256:e2764c90aa3af96450a9747a934b8893b121f79f58d89e123cb1a7046bb8e81e"},
@@ -1099 +1087 @@ name = "exceptiongroup"
-version = "1.1.0"
+version = "1.1.1"
@@ -1105,2 +1093,2 @@ files = [
- {file = "exceptiongroup-1.1.0-py3-none-any.whl", hash = "sha256:327cbda3da756e2de031a3107b81ab7b3770a602c4d16ca618298c526f4bec1e"},
- {file = "exceptiongroup-1.1.0.tar.gz", hash = "sha256:bcb67d800a4497e1b404c2dd44fca47d3b7a5e5433dbab67f96c1a685cdfdf23"},
+ {file = "exceptiongroup-1.1.1-py3-none-any.whl", hash = "sha256:232c37c63e4f682982c8b6459f33a8981039e5fb8756b2074364e5055c498c9e"},
+ {file = "exceptiongroup-1.1.1.tar.gz", hash = "sha256:d484c3090ba2889ae2928419117447a14daf3c1231d5e30d0aae34f354f01785"},
@@ -1114 +1102 @@ name = "fastavro"
-version = "1.7.2"
+version = "1.7.3"
@@ -1120,21 +1108,31 @@ files = [
- {file = "fastavro-1.7.2-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:20a94269bc0910c41f06106424429289e5c8cdce65b006e48d600c9bb6c442a5"},
- {file = "fastavro-1.7.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b34e9cd755e4d9d12377fe151ad95b9bff08215f0a3756601b21ed31f4d127ab"},
- {file = "fastavro-1.7.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8993e2f34e9ae195f417105d495fdd3d7f6d1532bacf57dc5dea600082f9b1a0"},
- {file = "fastavro-1.7.2-cp310-cp310-win_amd64.whl", hash = "sha256:e83069f971f307175ab806e9929a15826e586ef05a620fcfc7768ceaa3ec19c8"},
- {file = "fastavro-1.7.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:4c44f72a1a7b2d28cc21c95686916b782354dcc1f1aa1c8ca044c66d5ada3e14"},
- {file = "fastavro-1.7.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:88005d5c6fe436e80f756399a52aaa6d0cdc506c1ee195277287a09c7617aec3"},
- {file = "fastavro-1.7.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a5c5d466631f455213b8bfd54888d2d8691cfc240cd6dbb40eb3694bf05e562e"},
- {file = "fastavro-1.7.2-cp311-cp311-win_amd64.whl", hash = "sha256:4c2523c5e2279debb5fa788bf665a5abbd16a1e889338b973e3b318d50fcce77"},
- {file = "fastavro-1.7.2-cp37-cp37m-macosx_10_15_x86_64.whl", hash = "sha256:db3d9b6237934be5001fc18c07d25c09d3008ce6ab3448daf4b30608b84488e0"},
- {file = "fastavro-1.7.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e82862cb7d3bf99fdca65048b8b09db295795d7fb1eb7c5dae29c0a4d7dcf1f2"},
- {file = "fastavro-1.7.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5417dfbf99ac5bf963a929f2541c643d35e1a5a22876ba5ae0dca7b86326d9c2"},
- {file = "fastavro-1.7.2-cp37-cp37m-win_amd64.whl", hash = "sha256:b8cd8bd9a8ad4b687bd8dbf12effc2996bdf41ec04960875c19a7a90c949e8a8"},
- {file = "fastavro-1.7.2-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:f07d1f2ebdcc18938908e710a79ab3731061936fbea27534e436f9e12342450b"},
- {file = "fastavro-1.7.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9919d866319b1e123374fced84afb4728802872bd5938ba7e12a691c54593542"},
- {file = "fastavro-1.7.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8da1a9d4f0b871e7c968c8413caef19e19002100036b2714e037543c36550b8d"},
- {file = "fastavro-1.7.2-cp38-cp38-win_amd64.whl", hash = "sha256:e842beff23b0b207c45f99fa79d50cfc0d84ac137d84a7e283e9e59a8368b115"},
- {file = "fastavro-1.7.2-cp39-cp39-macosx_11_0_x86_64.whl", hash = "sha256:c1d5805c50a9507c99b814cb014da2812e6e0ee2770f4b793ce56b82a5b0af21"},
- {file = "fastavro-1.7.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:22cb5d2bf87c1ed9a1dcc97b6b13dc6b7ee4395dbf1fd5903dac01cdde0291c2"},
- {file = "fastavro-1.7.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:06bd51962df9b3f28c7081c0a3c8663d840df05d964b7817d85bd37c1d101236"},
- {file = "fastavro-1.7.2-cp39-cp39-win_amd64.whl", hash = "sha256:d787e7739d93db59f378c0e44a4eca8e56ff979e5b43dea6378e87c48548818c"},
- {file = "fastavro-1.7.2.tar.gz", hash = "sha256:3c524c65a5d578c020997800fce992cd90cc88ef13148b70f01a3429de34c8a4"},
+ {file = "fastavro-1.7.3-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:10a5ac9d8c66d4ba24f25ad7313e2dab56d98ceebcf53ba9cfa88acdd135c794"},
+ {file = "fastavro-1.7.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2e6d8bb79e53dc39e620c777f14b5f7122f1bf21309a9fcf60085f8e062e49c"},
+ {file = "fastavro-1.7.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9a0ba2f43844eb784f8abf5324a0c10474287beaecb14fb736e47136464e3044"},
+ {file = "fastavro-1.7.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e20db96b092d7b6208f3063a424d35bb48c283e2d8b4e7ad4ee6541dc1fac2ed"},
+ {file = "fastavro-1.7.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:52ba6bb5525561df577ebd94819784626caac9d8ad2ed167030403ba1bf73159"},
+ {file = "fastavro-1.7.3-cp310-cp310-win_amd64.whl", hash = "sha256:22d6f3e73f471e2b4ba0785cb60df939792e8904db4ba93037ba6b7858f7d6f9"},
+ {file = "fastavro-1.7.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:5dd5299cbc5bc2aa15f1c619f4cc55c054c6fe9ccd614f93eb1d6ab22cf314dd"},
+ {file = "fastavro-1.7.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f4706a77038bf31ad2e8cc752a0c007894bd39ffb0b775c7824113743182c5f6"},
+ {file = "fastavro-1.7.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b6b0e58e7dd34906d21738c3461cddef760de3b7845779169a378b2757afa693"},
+ {file = "fastavro-1.7.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:04740e2dd27084b4155337d082f2a232cf1d801a1b009f772e50c8306a8f8aaf"},
+ {file = "fastavro-1.7.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:2f95c767bf97e896640f24d58931b3a19df3d84ccaf0606c92e603c79de60f16"},
+ {file = "fastavro-1.7.3-cp311-cp311-win_amd64.whl", hash = "sha256:a727e07007230267e25702d5f3738854cb315747fc58b84839699db30dedf490"},
+ {file = "fastavro-1.7.3-cp37-cp37m-macosx_10_15_x86_64.whl", hash = "sha256:da71b9db7718f4682cc11e0f25b5e395d5f3bc17ddaf0224f39be3bac5309cfa"},
+ {file = "fastavro-1.7.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d1a62c359f4c9472c3ebe2be478e203ff434cc1d6bebaf61181a4a121c0899a6"},
+ {file = "fastavro-1.7.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9e2ba001864607b46fc2f6124d690731b19db215a84751c4b3b155e70b615d05"},
+ {file = "fastavro-1.7.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:cd10bffc26457402da9727663de71c40dd717d90e8ab3d3b893bc227cad5e410"},
+ {file = "fastavro-1.7.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:e5d0f16b85104aa0e2899a47c186be1082a10cecf6b331571afa92a4b8e6061a"},
+ {file = "fastavro-1.7.3-cp37-cp37m-win_amd64.whl", hash = "sha256:b86d1c1188ec47aeb76d6195e36ab52665984e8e98f69a224ab550c82991fe07"},
+ {file = "fastavro-1.7.3-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:a007151cc2a08e61dd5ea5b48989849d056a8d63b04d7e6799c36fdf0b702bf4"},
+ {file = "fastavro-1.7.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7b176d5731f336c2c9c88d95225f71f862b2512c33ef917b1fe7f87379cc92fd"},
+ {file = "fastavro-1.7.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c52e4a1b05306f82916eacf83c732a4637a5be748bc2ef2ff6fed1506535d692"},
+ {file = "fastavro-1.7.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:207bc7663133ca766eaf9033806da4cf08071dacf2e9779aa9427df40815f846"},
+ {file = "fastavro-1.7.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:cc98be4ad3d8fb9000abeeae0ecb0f8e62ec7898b791da5ec2f6de81dd2a73e2"},
+ {file = "fastavro-1.7.3-cp38-cp38-win_amd64.whl", hash = "sha256:fb3879aaeb3b56ee5b3a22ffa11cbdf4ba65c04be4688ee8bd152aa6535a00ee"},
+ {file = "fastavro-1.7.3-cp39-cp39-macosx_11_0_x86_64.whl", hash = "sha256:8c45f7fdfab351431d106f5981fdc2313a8cbfdb82d2b1172b2a144bfba376b7"},
+ {file = "fastavro-1.7.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:127e928604753d845fa0f2ae758c1640215ff901a5ce20cdf7e9f154500c3212"},
+ {file = "fastavro-1.7.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b82a71a6c014ec5d03293d8dc8b698220380266d5503779fd3712a94e4497069"},
+ {file = "fastavro-1.7.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:08bfd749cce456f925203895d6732f6b68c973d63ff886733f27db3c2d3c0b9a"},
+ {file = "fastavro-1.7.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:dc1e88d5db17e7ebc3fc764a1091f6c05a42e3cb0e2c8eaf49126743c7ca1bb5"},
+ {file = "fastavro-1.7.3-cp39-cp39-win_amd64.whl", hash = "sha256:7d525f3f99cc49a5e245e08d7ab947195a18cbdd5c43af75c0989fbe14a32597"},
+ {file = "fastavro-1.7.3.tar.gz", hash = "sha256:8b08bd3cba45830b64adda32ccc5b027a71b6941a99cc39f90d7019a7986cc19"},
@@ -1184 +1182 @@ name = "flatbuffers"
-version = "23.1.21"
+version = "23.3.3"
@@ -1190,2 +1188,2 @@ files = [
- {file = "flatbuffers-23.1.21-py2.py3-none-any.whl", hash = "sha256:2e4101b291b14f21e87ea20b7bf7127b11563f6084e352d2d708bddd545c9265"},
- {file = "flatbuffers-23.1.21.tar.gz", hash = "sha256:a948913bbb5d83c43a1193d7943c90e6c0ab732e7f2983111104250aeb61ff85"},
+ {file = "flatbuffers-23.3.3-py2.py3-none-any.whl", hash = "sha256:5ad36d376240090757e8f0a2cfaf6abcc81c6536c0dc988060375fd0899121f8"},
+ {file = "flatbuffers-23.3.3.tar.gz", hash = "sha256:cabd87c4882f37840f6081f094b2c5bc28cefc2a6357732746936d055ab45c3d"},
@@ -1280 +1278 @@ name = "fsspec"
-version = "2023.1.0"
+version = "2023.3.0"
@@ -1284 +1282 @@ optional = false
-python-versions = ">=3.7"
+python-versions = ">=3.8"
@@ -1286,2 +1284,2 @@ files = [
- {file = "fsspec-2023.1.0-py3-none-any.whl", hash = "sha256:b833e2e541e9e8cde0ab549414187871243177feb3d344f9d27b25a93f5d8139"},
- {file = "fsspec-2023.1.0.tar.gz", hash = "sha256:fbae7f20ff801eb5f7d0bedf81f25c787c0dfac5e982d98fa3884a9cde2b5411"},
+ {file = "fsspec-2023.3.0-py3-none-any.whl", hash = "sha256:bf57215e19dbfa4fe7edae53040cc1deef825e3b1605cca9a8d2c2fadd2328a0"},
+ {file = "fsspec-2023.3.0.tar.gz", hash = "sha256:24e635549a590d74c6c18274ddd3ffab4753341753e923408b1904eaabafe04d"},
@@ -1300 +1297,0 @@ dropbox = ["dropbox", "dropboxdrivefs", "requests"]
-entrypoints = ["importlib-metadata"]
@@ -1380 +1377 @@ name = "google-auth"
-version = "2.16.1"
+version = "2.16.2"
@@ -1386,2 +1383,2 @@ files = [
- {file = "google-auth-2.16.1.tar.gz", hash = "sha256:5fd170986bce6bfd7bb5c845c4b8362edb1e0cba901e062196e83f8bb5d5d32c"},
- {file = "google_auth-2.16.1-py2.py3-none-any.whl", hash = "sha256:75d76ea857df65938e1f71dcbcd7d0cd48e3f80b34b8870ba229c9292081f7ef"},
+ {file = "google-auth-2.16.2.tar.gz", hash = "sha256:07e14f34ec288e3f33e00e2e3cc40c8942aa5d4ceac06256a28cd8e786591420"},
+ {file = "google_auth-2.16.2-py2.py3-none-any.whl", hash = "sha256:2fef3cf94876d1a0e204afece58bb4d83fb57228aaa366c64045039fda6770a2"},
@@ -1811 +1808 @@ reference = "master"
-resolved_reference = "9af679c38477b564c26917a5dcf52d2c86177fb9"
+resolved_reference = "716251e7cac9feebba1138639420089a73d008a5"
@@ -2279 +2276 @@ name = "msgpack"
-version = "1.0.4"
+version = "1.0.5"
@@ -2285,52 +2282,63 @@ files = [
- {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:4ab251d229d10498e9a2f3b1e68ef64cb393394ec477e3370c457f9430ce9250"},
- {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:112b0f93202d7c0fef0b7810d465fde23c746a2d482e1e2de2aafd2ce1492c88"},
- {file = "msgpack-1.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:002b5c72b6cd9b4bafd790f364b8480e859b4712e91f43014fe01e4f957b8467"},
- {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:35bc0faa494b0f1d851fd29129b2575b2e26d41d177caacd4206d81502d4c6a6"},
- {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4733359808c56d5d7756628736061c432ded018e7a1dff2d35a02439043321aa"},
- {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:eb514ad14edf07a1dbe63761fd30f89ae79b42625731e1ccf5e1f1092950eaa6"},
- {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c23080fdeec4716aede32b4e0ef7e213c7b1093eede9ee010949f2a418ced6ba"},
- {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:49565b0e3d7896d9ea71d9095df15b7f75a035c49be733051c34762ca95bbf7e"},
- {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:aca0f1644d6b5a73eb3e74d4d64d5d8c6c3d577e753a04c9e9c87d07692c58db"},
- {file = "msgpack-1.0.4-cp310-cp310-win32.whl", hash = "sha256:0dfe3947db5fb9ce52aaea6ca28112a170db9eae75adf9339a1aec434dc954ef"},
- {file = "msgpack-1.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dea20515f660aa6b7e964433b1808d098dcfcabbebeaaad240d11f909298075"},
- {file = "msgpack-1.0.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:e83f80a7fec1a62cf4e6c9a660e39c7f878f603737a0cdac8c13131d11d97f52"},
- {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3c11a48cf5e59026ad7cb0dc29e29a01b5a66a3e333dc11c04f7e991fc5510a9"},
- {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1276e8f34e139aeff1c77a3cefb295598b504ac5314d32c8c3d54d24fadb94c9"},
- {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6c9566f2c39ccced0a38d37c26cc3570983b97833c365a6044edef3574a00c08"},
- {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:fcb8a47f43acc113e24e910399376f7277cf8508b27e5b88499f053de6b115a8"},
- {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:76ee788122de3a68a02ed6f3a16bbcd97bc7c2e39bd4d94be2f1821e7c4a64e6"},
- {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0a68d3ac0104e2d3510de90a1091720157c319ceeb90d74f7b5295a6bee51bae"},
- {file = "msgpack-1.0.4-cp36-cp36m-win32.whl", hash = "sha256:85f279d88d8e833ec015650fd15ae5eddce0791e1e8a59165318f371158efec6"},
- {file = "msgpack-1.0.4-cp36-cp36m-win_amd64.whl", hash = "sha256:c1683841cd4fa45ac427c18854c3ec3cd9b681694caf5bff04edb9387602d661"},
- {file = "msgpack-1.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:a75dfb03f8b06f4ab093dafe3ddcc2d633259e6c3f74bb1b01996f5d8aa5868c"},
- {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9667bdfdf523c40d2511f0e98a6c9d3603be6b371ae9a238b7ef2dc4e7a427b0"},
- {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11184bc7e56fd74c00ead4f9cc9a3091d62ecb96e97653add7a879a14b003227"},
- {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ac5bd7901487c4a1dd51a8c58f2632b15d838d07ceedaa5e4c080f7190925bff"},
- {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1e91d641d2bfe91ba4c52039adc5bccf27c335356055825c7f88742c8bb900dd"},
- {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:2a2df1b55a78eb5f5b7d2a4bb221cd8363913830145fad05374a80bf0877cb1e"},
- {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:545e3cf0cf74f3e48b470f68ed19551ae6f9722814ea969305794645da091236"},
- {file = "msgpack-1.0.4-cp37-cp37m-win32.whl", hash = "sha256:2cc5ca2712ac0003bcb625c96368fd08a0f86bbc1a5578802512d87bc592fe44"},
- {file = "msgpack-1.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:eba96145051ccec0ec86611fe9cf693ce55f2a3ce89c06ed307de0e085730ec1"},
- {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:7760f85956c415578c17edb39eed99f9181a48375b0d4a94076d84148cf67b2d"},
- {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:449e57cc1ff18d3b444eb554e44613cffcccb32805d16726a5494038c3b93dab"},
- {file = "msgpack-1.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d603de2b8d2ea3f3bcb2efe286849aa7a81531abc52d8454da12f46235092bcb"},
- {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:48f5d88c99f64c456413d74a975bd605a9b0526293218a3b77220a2c15458ba9"},
- {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6916c78f33602ecf0509cc40379271ba0f9ab572b066bd4bdafd7434dee4bc6e"},
- {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:81fc7ba725464651190b196f3cd848e8553d4d510114a954681fd0b9c479d7e1"},
- {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:d5b5b962221fa2c5d3a7f8133f9abffc114fe218eb4365e40f17732ade576c8e"},
- {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:77ccd2af37f3db0ea59fb280fa2165bf1b096510ba9fe0cc2bf8fa92a22fdb43"},
- {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b17be2478b622939e39b816e0aa8242611cc8d3583d1cd8ec31b249f04623243"},
- {file = "msgpack-1.0.4-cp38-cp38-win32.whl", hash = "sha256:2bb8cdf50dd623392fa75525cce44a65a12a00c98e1e37bf0fb08ddce2ff60d2"},
- {file = "msgpack-1.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:26b8feaca40a90cbe031b03d82b2898bf560027160d3eae1423f4a67654ec5d6"},
- {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:462497af5fd4e0edbb1559c352ad84f6c577ffbbb708566a0abaaa84acd9f3ae"},
- {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2999623886c5c02deefe156e8f869c3b0aaeba14bfc50aa2486a0415178fce55"},
- {file = "msgpack-1.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f0029245c51fd9473dc1aede1160b0a29f4a912e6b1dd353fa6d317085b219da"},
- {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ed6f7b854a823ea44cf94919ba3f727e230da29feb4a99711433f25800cf747f"},
- {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0df96d6eaf45ceca04b3f3b4b111b86b33785683d682c655063ef8057d61fd92"},
- {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6a4192b1ab40f8dca3f2877b70e63799d95c62c068c84dc028b40a6cb03ccd0f"},
- {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:0e3590f9fb9f7fbc36df366267870e77269c03172d086fa76bb4eba8b2b46624"},
- {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:1576bd97527a93c44fa856770197dec00d223b0b9f36ef03f65bac60197cedf8"},
- {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:63e29d6e8c9ca22b21846234913c3466b7e4ee6e422f205a2988083de3b08cae"},
- {file = "msgpack-1.0.4-cp39-cp39-win32.whl", hash = "sha256:fb62ea4b62bfcb0b380d5680f9a4b3f9a2d166d9394e9bbd9666c0ee09a3645c"},
- {file = "msgpack-1.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:4d5834a2a48965a349da1c5a79760d94a1a0172fbb5ab6b5b33cbf8447e109ce"},
- {file = "msgpack-1.0.4.tar.gz", hash = "sha256:f5d869c18f030202eb412f08b28d2afeea553d6613aee89e200d7aca7ef01f5f"},
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:525228efd79bb831cf6830a732e2e80bc1b05436b086d4264814b4b2955b2fa9"},
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4f8d8b3bf1ff2672567d6b5c725a1b347fe838b912772aa8ae2bf70338d5a198"},
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cdc793c50be3f01106245a61b739328f7dccc2c648b501e237f0699fe1395b81"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb47c21a8a65b165ce29f2bec852790cbc04936f502966768e4aae9fa763cb7"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e42b9594cc3bf4d838d67d6ed62b9e59e201862a25e9a157019e171fbe672dd3"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:55b56a24893105dc52c1253649b60f475f36b3aa0fc66115bffafb624d7cb30b"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:1967f6129fc50a43bfe0951c35acbb729be89a55d849fab7686004da85103f1c"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:20a97bf595a232c3ee6d57ddaadd5453d174a52594bf9c21d10407e2a2d9b3bd"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d25dd59bbbbb996eacf7be6b4ad082ed7eacc4e8f3d2df1ba43822da9bfa122a"},
+ {file = "msgpack-1.0.5-cp310-cp310-win32.whl", hash = "sha256:382b2c77589331f2cb80b67cc058c00f225e19827dbc818d700f61513ab47bea"},
+ {file = "msgpack-1.0.5-cp310-cp310-win_amd64.whl", hash = "sha256:4867aa2df9e2a5fa5f76d7d5565d25ec76e84c106b55509e78c1ede0f152659a"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9f5ae84c5c8a857ec44dc180a8b0cc08238e021f57abdf51a8182e915e6299f0"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:9e6ca5d5699bcd89ae605c150aee83b5321f2115695e741b99618f4856c50898"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5494ea30d517a3576749cad32fa27f7585c65f5f38309c88c6d137877fa28a5a"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ab2f3331cb1b54165976a9d976cb251a83183631c88076613c6c780f0d6e45a"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:28592e20bbb1620848256ebc105fc420436af59515793ed27d5c77a217477705"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe5c63197c55bce6385d9aee16c4d0641684628f63ace85f73571e65ad1c1e8d"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:ed40e926fa2f297e8a653c954b732f125ef97bdd4c889f243182299de27e2aa9"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b2de4c1c0538dcb7010902a2b97f4e00fc4ddf2c8cda9749af0e594d3b7fa3d7"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf22a83f973b50f9d38e55c6aade04c41ddda19b00c4ebc558930d78eecc64ed"},
+ {file = "msgpack-1.0.5-cp311-cp311-win32.whl", hash = "sha256:c396e2cc213d12ce017b686e0f53497f94f8ba2b24799c25d913d46c08ec422c"},
+ {file = "msgpack-1.0.5-cp311-cp311-win_amd64.whl", hash = "sha256:6c4c68d87497f66f96d50142a2b73b97972130d93677ce930718f68828b382e2"},
+ {file = "msgpack-1.0.5-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:a2b031c2e9b9af485d5e3c4520f4220d74f4d222a5b8dc8c1a3ab9448ca79c57"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f837b93669ce4336e24d08286c38761132bc7ab29782727f8557e1eb21b2080"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1d46dfe3832660f53b13b925d4e0fa1432b00f5f7210eb3ad3bb9a13c6204a6"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:366c9a7b9057e1547f4ad51d8facad8b406bab69c7d72c0eb6f529cf76d4b85f"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4c075728a1095efd0634a7dccb06204919a2f67d1893b6aa8e00497258bf926c"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:f933bbda5a3ee63b8834179096923b094b76f0c7a73c1cfe8f07ad608c58844b"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:36961b0568c36027c76e2ae3ca1132e35123dcec0706c4b7992683cc26c1320c"},
+ {file = "msgpack-1.0.5-cp36-cp36m-win32.whl", hash = "sha256:b5ef2f015b95f912c2fcab19c36814963b5463f1fb9049846994b007962743e9"},
+ {file = "msgpack-1.0.5-cp36-cp36m-win_amd64.whl", hash = "sha256:288e32b47e67f7b171f86b030e527e302c91bd3f40fd9033483f2cacc37f327a"},
+ {file = "msgpack-1.0.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:137850656634abddfb88236008339fdaba3178f4751b28f270d2ebe77a563b6c"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c05a4a96585525916b109bb85f8cb6511db1c6f5b9d9cbcbc940dc6b4be944b"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:56a62ec00b636583e5cb6ad313bbed36bb7ead5fa3a3e38938503142c72cba4f"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef8108f8dedf204bb7b42994abf93882da1159728a2d4c5e82012edd92c9da9f"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1835c84d65f46900920b3708f5ba829fb19b1096c1800ad60bae8418652a951d"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:e57916ef1bd0fee4f21c4600e9d1da352d8816b52a599c46460e93a6e9f17086"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:17358523b85973e5f242ad74aa4712b7ee560715562554aa2134d96e7aa4cbbf"},
+ {file = "msgpack-1.0.5-cp37-cp37m-win32.whl", hash = "sha256:cb5aaa8c17760909ec6cb15e744c3ebc2ca8918e727216e79607b7bbce9c8f77"},
+ {file = "msgpack-1.0.5-cp37-cp37m-win_amd64.whl", hash = "sha256:ab31e908d8424d55601ad7075e471b7d0140d4d3dd3272daf39c5c19d936bd82"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:b72d0698f86e8d9ddf9442bdedec15b71df3598199ba33322d9711a19f08145c"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:379026812e49258016dd84ad79ac8446922234d498058ae1d415f04b522d5b2d"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:332360ff25469c346a1c5e47cbe2a725517919892eda5cfaffe6046656f0b7bb"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:476a8fe8fae289fdf273d6d2a6cb6e35b5a58541693e8f9f019bfe990a51e4ba"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a9985b214f33311df47e274eb788a5893a761d025e2b92c723ba4c63936b69b1"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:48296af57cdb1d885843afd73c4656be5c76c0c6328db3440c9601a98f303d87"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:addab7e2e1fcc04bd08e4eb631c2a90960c340e40dfc4a5e24d2ff0d5a3b3edb"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:916723458c25dfb77ff07f4c66aed34e47503b2eb3188b3adbec8d8aa6e00f48"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:821c7e677cc6acf0fd3f7ac664c98803827ae6de594a9f99563e48c5a2f27eb0"},
+ {file = "msgpack-1.0.5-cp38-cp38-win32.whl", hash = "sha256:1c0f7c47f0087ffda62961d425e4407961a7ffd2aa004c81b9c07d9269512f6e"},
+ {file = "msgpack-1.0.5-cp38-cp38-win_amd64.whl", hash = "sha256:bae7de2026cbfe3782c8b78b0db9cbfc5455e079f1937cb0ab8d133496ac55e1"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:20c784e66b613c7f16f632e7b5e8a1651aa5702463d61394671ba07b2fc9e025"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:266fa4202c0eb94d26822d9bfd7af25d1e2c088927fe8de9033d929dd5ba24c5"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:18334484eafc2b1aa47a6d42427da7fa8f2ab3d60b674120bce7a895a0a85bdd"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57e1f3528bd95cc44684beda696f74d3aaa8a5e58c816214b9046512240ef437"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:586d0d636f9a628ddc6a17bfd45aa5b5efaf1606d2b60fa5d87b8986326e933f"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a740fa0e4087a734455f0fc3abf5e746004c9da72fbd541e9b113013c8dc3282"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3055b0455e45810820db1f29d900bf39466df96ddca11dfa6d074fa47054376d"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:a61215eac016f391129a013c9e46f3ab308db5f5ec9f25811e811f96962599a8"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:362d9655cd369b08fda06b6657a303eb7172d5279997abe094512e919cf74b11"},
+ {file = "msgpack-1.0.5-cp39-cp39-win32.whl", hash = "sha256:ac9dd47af78cae935901a9a500104e2dea2e253207c924cc95de149606dc43cc"},
+ {file = "msgpack-1.0.5-cp39-cp39-win_amd64.whl", hash = "sha256:06f5174b5f8ed0ed919da0e62cbd4ffde676a374aba4020034da05fab67b9164"},
+ {file = "msgpack-1.0.5.tar.gz", hash = "sha256:c075544284eadc5cddc70f4757331d99dcbc16b2bbd4849d15f8aae4cf36d31c"},
@@ -2458 +2466 @@ name = "mypy"
-version = "1.0.1"
+version = "1.1.1"
@@ -2464,26 +2472,26 @@ files = [
- {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"},
- {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"},
- {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"},
- {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"},
- {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"},
- {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"},
- {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"},
- {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"},
- {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"},
- {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"},
- {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"},
- {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"},
- {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"},
- {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"},
- {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"},
- {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"},
- {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"},
- {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"},
- {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"},
- {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"},
- {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"},
- {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"},
- {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"},
- {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"},
- {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"},
- {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"},
+ {file = "mypy-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39c7119335be05630611ee798cc982623b9e8f0cff04a0b48dfc26100e0b97af"},
+ {file = "mypy-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:61bf08362e93b6b12fad3eab68c4ea903a077b87c90ac06c11e3d7a09b56b9c1"},
+ {file = "mypy-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dbb19c9f662e41e474e0cff502b7064a7edc6764f5262b6cd91d698163196799"},
+ {file = "mypy-1.1.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:315ac73cc1cce4771c27d426b7ea558fb4e2836f89cb0296cbe056894e3a1f78"},
+ {file = "mypy-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:5cb14ff9919b7df3538590fc4d4c49a0f84392237cbf5f7a816b4161c061829e"},
+ {file = "mypy-1.1.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:26cdd6a22b9b40b2fd71881a8a4f34b4d7914c679f154f43385ca878a8297389"},
+ {file = "mypy-1.1.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5b5f81b40d94c785f288948c16e1f2da37203c6006546c5d947aab6f90aefef2"},
+ {file = "mypy-1.1.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21b437be1c02712a605591e1ed1d858aba681757a1e55fe678a15c2244cd68a5"},
+ {file = "mypy-1.1.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d809f88734f44a0d44959d795b1e6f64b2bbe0ea4d9cc4776aa588bb4229fc1c"},
+ {file = "mypy-1.1.1-cp311-cp311-win_amd64.whl", hash = "sha256:a380c041db500e1410bb5b16b3c1c35e61e773a5c3517926b81dfdab7582be54"},
+ {file = "mypy-1.1.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:b7c7b708fe9a871a96626d61912e3f4ddd365bf7f39128362bc50cbd74a634d5"},
+ {file = "mypy-1.1.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c1c10fa12df1232c936830839e2e935d090fc9ee315744ac33b8a32216b93707"},
+ {file = "mypy-1.1.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:0a28a76785bf57655a8ea5eb0540a15b0e781c807b5aa798bd463779988fa1d5"},
+ {file = "mypy-1.1.1-cp37-cp37m-win_amd64.whl", hash = "sha256:ef6a01e563ec6a4940784c574d33f6ac1943864634517984471642908b30b6f7"},
+ {file = "mypy-1.1.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d64c28e03ce40d5303450f547e07418c64c241669ab20610f273c9e6290b4b0b"},
+ {file = "mypy-1.1.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:64cc3afb3e9e71a79d06e3ed24bb508a6d66f782aff7e56f628bf35ba2e0ba51"},
+ {file = "mypy-1.1.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce61663faf7a8e5ec6f456857bfbcec2901fbdb3ad958b778403f63b9e606a1b"},
+ {file = "mypy-1.1.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:2b0c373d071593deefbcdd87ec8db91ea13bd8f1328d44947e88beae21e8d5e9"},
+ {file = "mypy-1.1.1-cp38-cp38-win_amd64.whl", hash = "sha256:2888ce4fe5aae5a673386fa232473014056967f3904f5abfcf6367b5af1f612a"},
+ {file = "mypy-1.1.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:19ba15f9627a5723e522d007fe708007bae52b93faab00f95d72f03e1afa9598"},
+ {file = "mypy-1.1.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:59bbd71e5c58eed2e992ce6523180e03c221dcd92b52f0e792f291d67b15a71c"},
+ {file = "mypy-1.1.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9401e33814cec6aec8c03a9548e9385e0e228fc1b8b0a37b9ea21038e64cdd8a"},
+ {file = "mypy-1.1.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4b398d8b1f4fba0e3c6463e02f8ad3346f71956b92287af22c9b12c3ec965a9f"},
+ {file = "mypy-1.1.1-cp39-cp39-win_amd64.whl", hash = "sha256:69b35d1dcb5707382810765ed34da9db47e7f95b3528334a3c999b0c90fe523f"},
+ {file = "mypy-1.1.1-py3-none-any.whl", hash = "sha256:4e4e8b362cdf99ba00c2b218036002bdcdf1e0de085cdb296a49df03fb31dfc4"},
+ {file = "mypy-1.1.1.tar.gz", hash = "sha256:ae9ceae0f5b9059f33dbc62dea087e942c0ccab4b7a003719cb70f9b8abfa32f"},
@@ -2493 +2501 @@ files = [
-mypy-extensions = ">=0.4.3"
+mypy-extensions = ">=1.0.0"
@@ -2667 +2675 @@ name = "openpyxl"
-version = "3.1.1"
+version = "3.1.2"
@@ -2673,2 +2681,2 @@ files = [
- {file = "openpyxl-3.1.1-py2.py3-none-any.whl", hash = "sha256:a0266e033e65f33ee697254b66116a5793c15fc92daf64711080000df4cfe0a8"},
- {file = "openpyxl-3.1.1.tar.gz", hash = "sha256:f06d44e2c973781068bce5ecf860a09bcdb1c7f5ce1facd5e9aa82c92c93ae72"},
+ {file = "openpyxl-3.1.2-py2.py3-none-any.whl", hash = "sha256:f91456ead12ab3c6c2e9491cf33ba6d08357d802192379bb482f1033ade496f5"},
+ {file = "openpyxl-3.1.2.tar.gz", hash = "sha256:a6f5977418eff3b2d5500d54d9db50c8277a368436f4e4f8ddb1be3422870184"},
@@ -2701 +2709 @@ name = "orjson"
-version = "3.8.6"
+version = "3.8.7"
@@ -2707,44 +2715,44 @@ files = [
- {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"},
- {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"},
- {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"},
- {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"},
- {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"},
- {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"},
- {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"},
- {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"},
- {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"},
- {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"},
- {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"},
- {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"},
- {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"},
- {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"},
- {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"},
- {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"},
- {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"},
- {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"},
- {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"},
- {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"},
- {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"},
- {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"},
- {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"},
- {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"},
- {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"},
- {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"},
- {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"},
- {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"},
- {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"},
- {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"},
- {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"},
- {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"},
- {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"},
- {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"},
- {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"},
- {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"},
- {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"},
- {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"},
- {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"},
- {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"},
- {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"},
- {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"},
- {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"},
- {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"},
+ {file = "orjson-3.8.7-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:f98c82850b7b4b7e27785ca43706fa86c893cdb88d54576bbb9b0d9c1070e421"},
+ {file = "orjson-3.8.7-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:1dee503c6c1a0659c5b46f5f39d9ca9d3657b11ca8bb4af8506086df416887d9"},
+ {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cc4fa83831f42ce5c938f8cefc2e175fa1df6f661fdeaba3badf26d2b8cfcf73"},
+ {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9e432c6c9c8b97ad825276d5795286f7cc9689f377a97e3b7ecf14918413303f"},
+ {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee519964a5a0efb9633f38b1129fd242807c5c57162844efeeaab1c8de080051"},
+ {file = "orjson-3.8.7-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:109b539ce5bf60a121454d008fa67c3b67e5a3249e47d277012645922cf74bd0"},
+ {file = "orjson-3.8.7-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ad4d441fbde4133af6fee37f67dbf23181b9c537ecc317346ec8c3b4c8ec7705"},
+ {file = "orjson-3.8.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:89dc786419e1ce2588345f58dd6a434e6728bce66b94989644234bcdbe39b603"},
+ {file = "orjson-3.8.7-cp310-none-win_amd64.whl", hash = "sha256:697abde7350fb8076d44bcb6b4ab3ce415ae2b5a9bb91efc460e5ab0d96bb5d3"},
+ {file = "orjson-3.8.7-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:1c19f47b35b9966a3abadf341b18ee4a860431bf2b00fd8d58906d51cf78aa70"},
+ {file = "orjson-3.8.7-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:3ffaabb380cd0ee187b4fc362516df6bf739808130b1339445c7d8878fca36e7"},
+ {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d88837002c5a8af970745b8e0ca1b0fdb06aafbe7f1279e110d338ea19f3d23"},
+ {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ff60187d1b7e0bfab376b6002b08c560b7de06c87cf3a8ac639ecf58f84c5f3b"},
+ {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0110970aed35dec293f30ed1e09f8604afd5d15c5ef83de7f6c427619b3ba47b"},
+ {file = "orjson-3.8.7-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:51b275475d4e36118b65ad56f9764056a09d985c5d72e64579bf8816f1356a5e"},
+ {file = "orjson-3.8.7-cp311-none-win_amd64.whl", hash = "sha256:63144d27735f3b60f079f247ac9a289d80dfe49a7f03880dfa0c0ba64d6491d5"},
+ {file = "orjson-3.8.7-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:a16273d77db746bb1789a2bbfded81148a60743fd6f9d5185e02d92e3732fa18"},
+ {file = "orjson-3.8.7-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:5bb32259ea22cc9dd47a6fdc4b8f9f1e2f798fcf56c7c1122a7df0f4c5d33bf3"},
+ {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ad02e9102d4ba67db30a136e631e32aeebd1dce26c9f5942a457b02df131c5d0"},
+ {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:dbcfcec2b7ac52deb7be3685b551addc28ee8fa454ef41f8b714df6ba0e32a27"},
+ {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e1a0e5504a5fc86083cc210c6946e8d61e13fe9f1d7a7bf81b42f7050a49d4fb"},
+ {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:7bd4fd37adb03b1f2a1012d43c9f95973a02164e131dfe3ff804d7e180af5653"},
+ {file = "orjson-3.8.7-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:188ed9f9a781333ad802af54c55d5a48991e292239aef41bd663b6e314377eb8"},
+ {file = "orjson-3.8.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:cc52f58c688cb10afd810280e450f56fbcb27f52c053463e625c8335c95db0dc"},
+ {file = "orjson-3.8.7-cp37-none-win_amd64.whl", hash = "sha256:403c8c84ac8a02c40613b0493b74d5256379e65196d39399edbf2ed3169cbeb5"},
+ {file = "orjson-3.8.7-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:7d6ac5f8a2a17095cd927c4d52abbb38af45918e0d3abd60fb50cfd49d71ae24"},
+ {file = "orjson-3.8.7-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:0295a7bfd713fa89231fd0822c995c31fc2343c59a1d13aa1b8b6651335654f5"},
+ {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:feb32aaaa34cf2f891eb793ad320d4bb6731328496ae59b6c9eb1b620c42b529"},
+ {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7a3ab1a473894e609b6f1d763838c6689ba2b97620c256a32c4d9f10595ac179"},
+ {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e8c430d82b532c5ab95634e034bbf6ca7432ffe175a3e63eadd493e00b3a555"},
+ {file = "orjson-3.8.7-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:366cc75f7e09106f9dac95a675aef413367b284f25507d21e55bd7f45f445e80"},
+ {file = "orjson-3.8.7-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:84d154d07e8b17d97e990d5d710b719a031738eb1687d8a05b9089f0564ff3e0"},
+ {file = "orjson-3.8.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:06180014afcfdc167ca984b312218aa62ce20093965c437c5f9166764cb65ef7"},
+ {file = "orjson-3.8.7-cp38-none-win_amd64.whl", hash = "sha256:41244431ba13f2e6ef22b52c5cf0202d17954489f4a3c0505bd28d0e805c3546"},
+ {file = "orjson-3.8.7-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:b20f29fa8371b8023f1791df035a2c3ccbd98baa429ac3114fc104768f7db6f8"},
+ {file = "orjson-3.8.7-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:226bfc1da2f21ee74918cee2873ea9a0fec1a8830e533cb287d192d593e99d02"},
+ {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e75c11023ac29e29fd3e75038d0e8dd93f9ea24d7b9a5e871967a8921a88df24"},
+ {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:78604d3acfd7cd502f6381eea0c42281fe2b74755b334074ab3ebc0224100be1"},
+ {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e7129a6847f0494aa1427167486ef6aea2e835ba05f6c627df522692ee228f65"},
+ {file = "orjson-3.8.7-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:1a1a8f4980059f48483782c608145b0f74538c266e01c183d9bcd9f8b71dbada"},
+ {file = "orjson-3.8.7-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:d60304172a33705ce4bd25a6261ab84bed2dab0b3d3b79672ea16c7648af4832"},
+ {file = "orjson-3.8.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4f733062d84389c32c0492e5a4929056fac217034a94523debe0430bcc602cda"},
+ {file = "orjson-3.8.7-cp39-none-win_amd64.whl", hash = "sha256:010e2970ec9e826c332819e0da4b14b29b19641da0f1a6af4cec91629ef9b988"},
+ {file = "orjson-3.8.7.tar.gz", hash = "sha256:8460c8810652dba59c38c80d27c325b5092d189308d8d4f3e688dbd8d4f3b2dc"},
@@ -2828 +2836 @@ name = "pandas-stubs"
-version = "1.5.3.230227"
+version = "1.5.3.230304"
@@ -2834,2 +2842,2 @@ files = [
- {file = "pandas_stubs-1.5.3.230227-py3-none-any.whl", hash = "sha256:bc728e1c9a7fd69b4e9487846b035dcb0cfc194df1f5c03b4b3d71988d42cc2f"},
- {file = "pandas_stubs-1.5.3.230227.tar.gz", hash = "sha256:dbba9c2e08860f927fbb6b98722bdbe9b958a00afd4c1d1a28945f3e3333e840"},
+ {file = "pandas_stubs-1.5.3.230304-py3-none-any.whl", hash = "sha256:ebb545329109e98510cdf6357074497dfbdf62f173d8ec8b5430486716360639"},
+ {file = "pandas_stubs-1.5.3.230304.tar.gz", hash = "sha256:853174ea5aac1ee2001b3ae708dea9780c60d546ee490abb5eeb7c37ed57d255"},
@@ -2893,0 +2902,7 @@ files = [
+ {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
+ {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
+ {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
+ {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
+ {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
+ {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
+ {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -3041 +3056 @@ name = "platformdirs"
-version = "3.0.0"
+version = "3.1.1"
@@ -3047,2 +3062,2 @@ files = [
- {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"},
- {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"},
+ {file = "platformdirs-3.1.1-py3-none-any.whl", hash = "sha256:e5986afb596e4bb5bde29a79ac9061aa955b94fca2399b7aaac4090860920dd8"},
+ {file = "platformdirs-3.1.1.tar.gz", hash = "sha256:024996549ee88ec1a9aa99ff7f8fc819bb59e2c3477b410d90a16d32d6e707aa"},
@@ -3776 +3791 @@ name = "pytest"
-version = "7.2.1"
+version = "7.2.2"
@@ -3782,2 +3797,2 @@ files = [
- {file = "pytest-7.2.1-py3-none-any.whl", hash = "sha256:c7c6ca206e93355074ae32f7403e8ea12163b1163c976fee7d4d84027c162be5"},
- {file = "pytest-7.2.1.tar.gz", hash = "sha256:d45e0952f3727241918b8fd0f376f5ff6b301cc0777c6f9a556935c92d8a7d42"},
+ {file = "pytest-7.2.2-py3-none-any.whl", hash = "sha256:130328f552dcfac0b1cec75c12e3f005619dc5f874f0a06e8ff7263f0ee6225e"},
+ {file = "pytest-7.2.2.tar.gz", hash = "sha256:c99ab0c73aceb050f68929bc93af19ab6db0558791c6a0715723abe9d0ade9d4"},
@@ -3946,0 +3962,8 @@ files = [
+ {file = "pyzstd-0.15.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:df80324759cb74a10818f5ea4dfb0332b6064a31305246f66b043c4f73b37918"},
+ {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b4833842d5e6660c1b6cd316f102ffddb95566f12ec1f89c5630d4b511e7e084"},
+ {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9069da3bb94cac8af3fd2d7b622c7a78a248d3295a478587d880553287d56e6"},
+ {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:9db032965b8fabf0bac13b75706031444ebfa7214794afe80be41799d8a349fb"},
+ {file = "pyzstd-0.15.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:16eb2b2b206489bf8a7b690cf3b06b24352a16ee25717cb38c4381aefb3a94ce"},
+ {file = "pyzstd-0.15.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:821555ce9bc7f50130bf52dc7b7d342913e00435c68a5cac564c5f2e57f33cbf"},
+ {file = "pyzstd-0.15.4-cp36-cp36m-win32.whl", hash = "sha256:4f4a96bdffa773f8820e065ec840c5d3f252b0611951b73567121757388f171c"},
+ {file = "pyzstd-0.15.4-cp36-cp36m-win_amd64.whl", hash = "sha256:3b0d384a48592683edf812c1a3d75a0cf5af253b1629596567ce3a2991d6ccbc"},
@@ -4165 +4188 @@ name = "resolvelib"
-version = "0.9.0"
+version = "1.0.1"
@@ -4171,2 +4194,2 @@ files = [
- {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"},
- {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"},
+ {file = "resolvelib-1.0.1-py2.py3-none-any.whl", hash = "sha256:d2da45d1a8dfee81bdd591647783e340ef3bcb104b54c383f70d422ef5cc7dbf"},
+ {file = "resolvelib-1.0.1.tar.gz", hash = "sha256:04ce76cbd63fded2078ce224785da6ecd42b9564b1390793f64ddecbe997b309"},
@@ -4202 +4225 @@ name = "rich"
-version = "13.3.1"
+version = "13.3.2"
@@ -4208,2 +4231,2 @@ files = [
- {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"},
- {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"},
+ {file = "rich-13.3.2-py3-none-any.whl", hash = "sha256:a104f37270bf677148d8acb07d33be1569eeee87e2d1beb286a4e9113caf6f2f"},
+ {file = "rich-13.3.2.tar.gz", hash = "sha256:91954fe80cfb7985727a467ca98a7618e5dd15178cc2da10f553b36a93859001"},
@@ -4213,2 +4236,2 @@ files = [
-markdown-it-py = ">=2.1.0,<3.0.0"
-pygments = ">=2.14.0,<3.0.0"
+markdown-it-py = ">=2.2.0,<3.0.0"
+pygments = ">=2.13.0,<3.0.0"
@@ -4236 +4259 @@ name = "scikit-learn"
-version = "1.2.1"
+version = "1.2.2"
@@ -4242,21 +4265,21 @@ files = [
- {file = "scikit-learn-1.2.1.tar.gz", hash = "sha256:fbf8a5c893c9b4b99bcc7ed8fb3e8500957a113f4101860386d06635520f7cfb"},
- {file = "scikit_learn-1.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:bed9f75763bd392c094bf474c7ab75a01d68b15146ea7a20c0f9ff6fb3063dad"},
- {file = "scikit_learn-1.2.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:c9285275a435d1f8f47bbe3500346ab9ead2499e0e090518404d318ea90d1c1c"},
- {file = "scikit_learn-1.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:dc838b5a4057c55ba81b82316ea8bf443af445f96eb21500b0e40618017e0923"},
- {file = "scikit_learn-1.2.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d8bcd303dd982494842a3f482f844d539484c6043b4eed896b43ea8e5f609a21"},
- {file = "scikit_learn-1.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:a9abf17d177df54e529154f26acfd42930e19117d045e8a9a8e893ca82dd94ec"},
- {file = "scikit_learn-1.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:70fa30d146b7e9d0c256e73e271b3e17f23123b7c4adcbde1a385031adf59090"},
- {file = "scikit_learn-1.2.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:5a8111f3c7a314017ebf90d6feab861c11d1ca14f3dbafb39abcc31aa4c54ba6"},
- {file = "scikit_learn-1.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cba0c7c6bf1493f8ce670bab69f9317874826ee838988de377ae355abd4d74cf"},
- {file = "scikit_learn-1.2.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:479aedd0abedbda6b8b4529145fe4cd8622f69f726a72cef8f75548a93eeb1e1"},
- {file = "scikit_learn-1.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:5523e21ab2b4d52b2bd41bedd335dbe8f3c1b5f6dd7c9c001b2e17ec9818af8d"},
- {file = "scikit_learn-1.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:dcfab6a19b236194af88771d8e6e778a60c3339248ab0018696ebf2b7c8bed4b"},
- {file = "scikit_learn-1.2.1-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:559f66e12f93b34c8c85c0a5728c3b8af98f04eb12f2c9ee18ea3c82c3d2fad1"},
- {file = "scikit_learn-1.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:dbb7831b2308c67bb6dd83c5ea3cdaf8e8cafd2de4000b93d78bb689126bd2cf"},
- {file = "scikit_learn-1.2.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5b2c5d9930ced2b7821ad936b9940706ccb5471d89b8a516bb641cec87257d1c"},
- {file = "scikit_learn-1.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:54731e2c2fbff40da6d76cbb9022ace5f44a4020a10bd5cd92107e86882bad15"},
- {file = "scikit_learn-1.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:d00e46a2a7fce6e118ed0f4c6263785bf6c297a94ffd0cd7b32455043c508cc8"},
- {file = "scikit_learn-1.2.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:da0e2d50a8435ea8dc5cd21f1fc1a45d329bae03dcca92087ebed859d22d184e"},
- {file = "scikit_learn-1.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61bb9c654b5d2e6cdd4b1c7e6048fc66270c1682bda1b0f7d2726fdae09010f4"},
- {file = "scikit_learn-1.2.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e0ee4d4d32c94e082344308528f7b3c9294b60ab19c84eb37a2d9c88bdffd9d1"},
- {file = "scikit_learn-1.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:c722f3446ad8c4f1a93b2399fe1a188635b94709a3f25e6f4d61efbe75fe8eaa"},
+ {file = "scikit-learn-1.2.2.tar.gz", hash = "sha256:8429aea30ec24e7a8c7ed8a3fa6213adf3814a6efbea09e16e0a0c71e1a1a3d7"},
+ {file = "scikit_learn-1.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:99cc01184e347de485bf253d19fcb3b1a3fb0ee4cea5ee3c43ec0cc429b6d29f"},
+ {file = "scikit_learn-1.2.2-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:e6e574db9914afcb4e11ade84fab084536a895ca60aadea3041e85b8ac963edb"},
+ {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6fe83b676f407f00afa388dd1fdd49e5c6612e551ed84f3b1b182858f09e987d"},
+ {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e2642baa0ad1e8f8188917423dd73994bf25429f8893ddbe115be3ca3183584"},
+ {file = "scikit_learn-1.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:ad66c3848c0a1ec13464b2a95d0a484fd5b02ce74268eaa7e0c697b904f31d6c"},
+ {file = "scikit_learn-1.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:dfeaf8be72117eb61a164ea6fc8afb6dfe08c6f90365bde2dc16456e4bc8e45f"},
+ {file = "scikit_learn-1.2.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:fe0aa1a7029ed3e1dcbf4a5bc675aa3b1bc468d9012ecf6c6f081251ca47f590"},
+ {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:065e9673e24e0dc5113e2dd2b4ca30c9d8aa2fa90f4c0597241c93b63130d233"},
+ {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf036ea7ef66115e0d49655f16febfa547886deba20149555a41d28f56fd6d3c"},
+ {file = "scikit_learn-1.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:8b0670d4224a3c2d596fd572fb4fa673b2a0ccfb07152688ebd2ea0b8c61025c"},
+ {file = "scikit_learn-1.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9c710ff9f9936ba8a3b74a455ccf0dcf59b230caa1e9ba0223773c490cab1e51"},
+ {file = "scikit_learn-1.2.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:2dd3ffd3950e3d6c0c0ef9033a9b9b32d910c61bd06cb8206303fb4514b88a49"},
+ {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:44b47a305190c28dd8dd73fc9445f802b6ea716669cfc22ab1eb97b335d238b1"},
+ {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:953236889928d104c2ef14027539f5f2609a47ebf716b8cbe4437e85dce42744"},
+ {file = "scikit_learn-1.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:7f69313884e8eb311460cc2f28676d5e400bd929841a2c8eb8742ae78ebf7c20"},
+ {file = "scikit_learn-1.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8156db41e1c39c69aa2d8599ab7577af53e9e5e7a57b0504e116cc73c39138dd"},
+ {file = "scikit_learn-1.2.2-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:fe175ee1dab589d2e1033657c5b6bec92a8a3b69103e3dd361b58014729975c3"},
+ {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d5312d9674bed14f73773d2acf15a3272639b981e60b72c9b190a0cffed5bad"},
+ {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ea061bf0283bf9a9f36ea3c5d3231ba2176221bbd430abd2603b1c3b2ed85c89"},
+ {file = "scikit_learn-1.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:6477eed40dbce190f9f9e9d0d37e020815825b300121307942ec2110302b66a3"},
@@ -4318 +4341 @@ name = "setuptools"
-version = "67.4.0"
+version = "67.6.0"
@@ -4324,2 +4347,2 @@ files = [
- {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"},
- {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"},
+ {file = "setuptools-67.6.0-py3-none-any.whl", hash = "sha256:b78aaa36f6b90a074c1fa651168723acbf45d14cb1196b6f02c0fd07f17623b2"},
+ {file = "setuptools-67.6.0.tar.gz", hash = "sha256:2ee892cd5f29f3373097f5a814697e397cf3ce313616df0af11231e2ad118077"},
@@ -4406 +4429 @@ name = "soxr"
-version = "0.3.3"
+version = "0.3.4"
@@ -4412,28 +4435,28 @@ files = [
- {file = "soxr-0.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:086fb3e3837216cdac4673530af641bc9e7a65ff34b012d81d6b594690a4c5c6"},
- {file = "soxr-0.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:65431e8ab1b8cc86e8cab2fe0b0a284046814b4aad42ee6611d1ca1bc7e0bbf0"},
- {file = "soxr-0.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b7402e59294d3a39696990c63d45f24cbf8cb011a167da21f37b820d65a6f20"},
- {file = "soxr-0.3.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4637508a34b577ffb55207da7e8ec6e55a813b904fc56e87819a04cd16cf4bd4"},
- {file = "soxr-0.3.3-cp310-cp310-win32.whl", hash = "sha256:45e5c5bbfc605be646b944aa044bc3315bd81cac1935c220c4814f0659ceb8fb"},
- {file = "soxr-0.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:5b51eb310529bf19c57a02d3fbadf9d45896bb9981e94d013327ffb27a3b00b0"},
- {file = "soxr-0.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:52cf5b58abc82f0263b9c2d40da63073a57b4a70b14e8d4549b1aa09cdfef3c9"},
- {file = "soxr-0.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ccf76f0487d0db186f57a2a616feca0a11d37477776e64d260daf4c01013aa39"},
- {file = "soxr-0.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2202eb8807a834621b61f13230afaa3643eafb822e37c9cf19a50b20c6f6310"},
- {file = "soxr-0.3.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d7687274daf541125c508dd6509310927180a0de0aa0c268f5084d4dab70501b"},
- {file = "soxr-0.3.3-cp311-cp311-win32.whl", hash = "sha256:56fe880b97b8466166256b443ec45a1ec97f90cc1cd3167bc60922501f415b2c"},
- {file = "soxr-0.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:f6673054a609d862cc39cc90e1b1a05de6b7fd51fb18e45ebdd127d171656b02"},
- {file = "soxr-0.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:65c353979cb57f38ee3a24f68ea766706570075db216e079c05dd7d38c1df4ca"},
- {file = "soxr-0.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:da548494b2f5ae84d7c551165b442834db6c42807e959dbf742146cf9ee4d97f"},
- {file = "soxr-0.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a472c2380f624a45ea3d7c9f4c6db6a2a0dbeb610eafe1a57ec731ccb29cdd76"},
- {file = "soxr-0.3.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e0a4bbdc2246d05201ed1e080ddf3fcc1856ca604c9565fdb8f69477589f2c17"},
- {file = "soxr-0.3.3-cp38-cp38-win32.whl", hash = "sha256:a0996e682f6d8b796702823c21aca53707b2a8667c1154f8a12dc1ca97bfba1c"},
- {file = "soxr-0.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:a7ddd6db851530ca356d5db5f257a38a12e7484899c209c9bed10eceb8d8edeb"},
- {file = "soxr-0.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cb5d66b3b3a493f237ea80add2176a097a118488926f3f2fae369fe5caafe976"},
- {file = "soxr-0.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:54a9934b98b151562a08ff396d2c0ea65e7a40398319b622e31afc34a510a523"},
- {file = "soxr-0.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c41c5f1a75397e0255d62dc03e8759e011128d24385c1a7de54edd3f4bca2e36"},
- {file = "soxr-0.3.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2081e5534218d30d1fa88c1ef4a1b5399b439e2ee396cd210ef4fd5dbe3659b5"},
- {file = "soxr-0.3.3-cp39-cp39-win32.whl", hash = "sha256:54a6d7b17cad791ed91041abe28a645d36869d0410fac20cb97337a08b21a904"},
- {file = "soxr-0.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:9e179d8fe35cbf8584f025bee1516fdaa6a69056a1ca7ccc600022d0d3f8a6bc"},
- {file = "soxr-0.3.3-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:61bb9cd20f8aa52a39f7677b8edf587e017a1d97147a13a39bc836d989347760"},
- {file = "soxr-0.3.3-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d3b9e692be2adb959ef114ef632ffc78eb83f6f0a7fbc0be0c49e87fbb151d81"},
- {file = "soxr-0.3.3-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:e0b2334101798fa79e75218e94da47e0e5c056690559df84d5f321c9bb3893cf"},
- {file = "soxr-0.3.3.tar.gz", hash = "sha256:af874e9f14c4c261c22ce55972a1231c085a538c2d8fec78580b331370fdb551"},
+ {file = "soxr-0.3.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b7b84126643c063d5daa203f7f9137e21734dabbd7e68c097607b2ef457e2f2e"},
+ {file = "soxr-0.3.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:380d2d43871a68e8b1ef1702a0abe6f9e48ddb3933c7a303c45d67e121503e7c"},
+ {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4a1b4019c9972f57612482c4f85523d6e832e3d10935e2f070a9dcd334a4dcb"},
+ {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e613cee023b7c3f162b9da3f6b169cd7f58de345275be1fde9f19adc9cf144df"},
+ {file = "soxr-0.3.4-cp310-cp310-win32.whl", hash = "sha256:182c02a7ba45a159a0dbb0a297335df2381ead03a65377b19663ea0ff720ecb7"},
+ {file = "soxr-0.3.4-cp310-cp310-win_amd64.whl", hash = "sha256:1e95c96ce94524fae453b4331c9910d33f97506f99bae06d76a9c0649710619e"},
+ {file = "soxr-0.3.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2678d2719e7496803983584d661eb5fddc7017154a8dda4a774407c56ff07973"},
+ {file = "soxr-0.3.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:11bd1396052049e6d389225a0e96a9df15f706da501c619b35d3c72ac6bc7257"},
+ {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7e23de4dfe54ac30e583bbc9cc3feda1cd776fedce13206bc4b3115b75ecab82"},
+ {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3e7396498a5f5b7d8f23b656f65c24517a6ff5bdc3ee0623ccd491036a43ea08"},
+ {file = "soxr-0.3.4-cp311-cp311-win32.whl", hash = "sha256:e57e9703c2bff834cabc06800d3c11a259544891d2c24a78949f3cf2f5492cc5"},
+ {file = "soxr-0.3.4-cp311-cp311-win_amd64.whl", hash = "sha256:7c8350acd7150f74261a0569b47ccb1bb4aa39b2d575860bc97cfa69aab8aead"},
+ {file = "soxr-0.3.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:941f7355116fe77fe6a82938fa7799a0e466a494ebc093f676969ce32b2815b1"},
+ {file = "soxr-0.3.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:00fdbf24f64d8c3fb800425c383048cb24c32defac80901cde4a57fb6ce5d431"},
+ {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7bb6d4dc807d04c536674429e2b05ae08a1efac9815c4595e41ffd6b57c2c662"},
+ {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ff15853895b54f1b627799c6600be1ce5f7286724e7a93e4b7b9d79e5d4166f5"},
+ {file = "soxr-0.3.4-cp38-cp38-win32.whl", hash = "sha256:d858becbc1fcc7b38c3436d3276290fae09403cdcbdf1d5986a18dab7023a6c3"},
+ {file = "soxr-0.3.4-cp38-cp38-win_amd64.whl", hash = "sha256:068ab4df549df5783cc1eb4eb6c94f53823b164dc27134fc621fc9f5097f38cd"},
+ {file = "soxr-0.3.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:20130329985f9767c8417bbd125fe138790a71802b000481c386a800e2ad2bca"},
+ {file = "soxr-0.3.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:78090e97abfb326b7cf14ef37d08a17252b07d438388dcbbd82a6836a9d551b1"},
+ {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84e590e75b7e5dca12bf68bfb090276f34a88fbcd793781c62d47f5d7dbe525e"},
+ {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3479d265574b960e12bca0878baba0862c43278915e0319d84679bb4d4fcd33"},
+ {file = "soxr-0.3.4-cp39-cp39-win32.whl", hash = "sha256:83de825d6a713c7b2e76d9ec3f229a58a9ed290237e7adc05d80e8b39be995a6"},
+ {file = "soxr-0.3.4-cp39-cp39-win_amd64.whl", hash = "sha256:2082f88cae89de854c3e0d62f55d0cb31eb11764f5c2a28299121fb642a22472"},
+ {file = "soxr-0.3.4-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:fe8b5f92c802f1e7793c40344f5368dc6163718c9ffa82e79ee6ad779d318ac5"},
+ {file = "soxr-0.3.4-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d0063d5f9a4e1a367084f4705301e9da131cf4d2d32aa3fe0072a1245e18088f"},
+ {file = "soxr-0.3.4-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:a680bab57adae462cdc86abcc7330beb5daa3ba5101165583eedcda88b7ba551"},
+ {file = "soxr-0.3.4.tar.gz", hash = "sha256:fe68daf00e8f020977b187699903d219f9e39b9fb3d915f3f923eed8ba431449"},
@@ -4513,0 +4537,37 @@ files = [
+[[package]]
+name = "tensorflow-aarch64"
+version = "2.11.0"
+description = "TensorFlow is an open source machine learning framework for everyone."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "tensorflow_aarch64-2.11.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f0e562ea7dc9f9c02f3c0df27ff8e561d4dccd42fe9d42b1b9536bbc555db560"},
+ {file = "tensorflow_aarch64-2.11.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e0e915cfcba6d629ef9f2de78152836f94819041ff38b4ee6ac31f18e2137ce"},
+ {file = "tensorflow_aarch64-2.11.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b9f5139d57dee505c8f5b94bc17b99f75e51399ec4cf09c51ef6a71b1b099ea2"},
+ {file = "tensorflow_aarch64-2.11.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:97b6dbfb1307f76d54f8ddc4bddc1ad2470e00c6c457f962e12de8be4f59b7f6"},
+]
+
+[package.dependencies]
+absl-py = ">=1.0.0"
+astunparse = ">=1.6.0"
+flatbuffers = ">=2.0"
+gast = ">=0.2.1,<=0.4.0"
+google-pasta = ">=0.1.1"
+grpcio = ">=1.24.3,<2.0"
+h5py = ">=2.9.0"
+keras = ">=2.11.0,<2.12"
+libclang = ">=13.0.0"
+numpy = ">=1.20"
+opt-einsum = ">=2.3.2"
+packaging = "*"
+protobuf = ">=3.9.2,<3.20"
+setuptools = "*"
+six = ">=1.12.0"
+tensorboard = ">=2.11,<2.12"
+tensorflow-estimator = ">=2.11.0,<2.12"
+tensorflow-io-gcs-filesystem = {version = ">=0.23.1", markers = "platform_machine != \"arm64\" or platform_system != \"Darwin\""}
+termcolor = ">=1.1.0"
+typing-extensions = ">=3.6.6"
+wrapt = ">=1.11.0"
+
@@ -4715,0 +4776,2 @@ files = [
+ {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"},
+ {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"},
@@ -4720,0 +4783 @@ files = [
+ {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"},
@@ -4728,0 +4792 @@ files = [
+ {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"},
@@ -4816 +4880 @@ name = "tqdm"
-version = "4.64.1"
+version = "4.65.0"
@@ -4820 +4884 @@ optional = false
-python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,>=2.7"
+python-versions = ">=3.7"
@@ -4822,2 +4886,2 @@ files = [
- {file = "tqdm-4.64.1-py2.py3-none-any.whl", hash = "sha256:6fee160d6ffcd1b1c68c65f14c829c22832bc401726335ce92c52d395944a6a1"},
- {file = "tqdm-4.64.1.tar.gz", hash = "sha256:5f4f682a004951c1b450bc753c710e9280c5746ce6ffedee253ddbcbf54cf1e4"},
+ {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"},
+ {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"},
@@ -4943 +5007 @@ name = "types-psutil"
-version = "5.9.5.9"
+version = "5.9.5.10"
@@ -4949,2 +5013,2 @@ files = [
- {file = "types-psutil-5.9.5.9.tar.gz", hash = "sha256:c6c7e6b41b6f7ebb87afd5650df360751d0e8acf0d276b7ac64d00f419beb922"},
- {file = "types_psutil-5.9.5.9-py3-none-any.whl", hash = "sha256:f7830fb3a7bc3288113ff08199330fced7851f5219f0af1b2abc24aaa3112236"},
+ {file = "types-psutil-5.9.5.10.tar.gz", hash = "sha256:26206682efe011bc84f15f8d6539c714f3d598843e2b4c88ef371ecb1a34829b"},
+ {file = "types_psutil-5.9.5.10-py3-none-any.whl", hash = "sha256:04c54f95f95d7175649f21cec4528f0b0313f0c4d103cc4d675bb66a25ac0bca"},
@@ -5081 +5145 @@ name = "urllib3"
-version = "1.26.14"
+version = "1.26.15"
@@ -5087,2 +5151,2 @@ files = [
- {file = "urllib3-1.26.14-py2.py3-none-any.whl", hash = "sha256:75edcdc2f7d85b137124a6c3c9fc3933cdeaa12ecb9a6a959f22797a0feca7e1"},
- {file = "urllib3-1.26.14.tar.gz", hash = "sha256:076907bf8fd355cde77728471316625a4d2f7e713c125f51953bb5b3eecf4f72"},
+ {file = "urllib3-1.26.15-py2.py3-none-any.whl", hash = "sha256:aa751d169e23c7479ce47a0cb0da579e3ede798f994f5816a74e4f4500dcea42"},
+ {file = "urllib3-1.26.15.tar.gz", hash = "sha256:8a388717b9476f934a21484e8c8e61875ab60644d29b9b39e11e4b9dc1c6b305"},
@@ -5519 +5583 @@ python-versions = "3.9.15"
-content-hash = "2bf2596d9309ec8715c9bf1e89f590057e44c8cd7e143d9f4bff3d27b9a51c5b"
+content-hash = "f3944a99f45c8d6f063ec218eb246c3c6ef4015d6f633350ec026948efd54549"
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index 85fb9a5e..17e16b5e 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -34 +34,5 @@ scikit-learn = "^1.2.1"
-tensorflow-cpu = {version = "^2.11.0", markers = "sys_platform != 'darwin' or platform_machine != 'arm64'"}
+tensorflow-cpu = [
+ {version = "^2.11.0", markers = "sys_platform == 'linux' and platform_machine != 'aarch64'"},
+ {version = "^2.11.0", markers = "sys_platform == 'darwin' and platform_machine != 'arm64'"}
+]
+tensorflow-aarch64 = {version = "^2.11.0", markers = "sys_platform == 'linux' and platform_machine == 'aarch64'"}
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
index 617997bb..e3a60b41 100644
--- a/tools/docker-compose-dev-base.yml
+++ b/tools/docker-compose-dev-base.yml
@@ -23,0 +24,15 @@ services:
+ # volumes to local source directory for development
+ volumes:
+ - ../libs/libcommon/src:/src/libs/libcommon/src
+ admin:
+ extends:
+ service: common
+ # volumes to local source directory for development
+ volumes:
+ - ../services/admin/src:/src/services/admin/src
+ api:
+ extends:
+ service: common
+ # volumes to local source directory for development
+ volumes:
+ - ../services/api/src:/src/services/api/src
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 8d7ded1e..3797b83e 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -24 +24 @@ services:
- dockerfile: services/admin/Dockerfile
+ dockerfile: services/admin/dev.Dockerfile
@@ -27 +27 @@ services:
- service: common
+ service: admin
@@ -50 +50 @@ services:
- dockerfile: services/api/Dockerfile
+ dockerfile: services/api/dev.Dockerfile
@@ -53 +53 @@ services:
- service: common
+ service: api
|
|
ec38d5f2378c9c7e99aa95b9b79bf24e0b63cb6d
|
Quentin Lhoest
| 2023-03-13T09:52:30 |
Add partial cache entries for fan-in jobs (#909)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 18857e50..fa9fe261 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -22,0 +23,3 @@ from mongodb_migration.migrations._20230216141000_queue_split_names_from_streami
+from mongodb_migration.migrations._20230309123100_cache_add_progress import (
+ MigrationAddProgressToCacheResponse,
+)
@@ -50,0 +54,4 @@ class MigrationsCollector:
+ MigrationAddProgressToCacheResponse(
+ version="20230309123100",
+ description="add the 'progress' field with the default value (1.0) to the cached results",
+ ),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py
new file mode 100644
index 00000000..a56f1185
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py
@@ -0,0 +1,29 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.simple_cache import CachedResponse
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import Migration
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationAddProgressToCacheResponse(Migration):
+ def up(self) -> None:
+ # See https://docs.mongoengine.org/guide/migration.html#example-1-addition-of-a-field
+ logging.info("If missing, add the progress field with the default value (1.0) to the cached results")
+ db = get_db("cache")
+ db["cachedResponsesBlue"].update_many({"progress": {"$exists": False}}, {"$set": {"progress": 1.0}})
+
+ def down(self) -> None:
+ logging.info("Remove the progress field from all the cached results")
+ db = get_db("cache")
+ db["cachedResponsesBlue"].update_many({}, {"$unset": {"progress": ""}})
+
+ def validate(self) -> None:
+ logging.info("Ensure that a random selection of cached results have the 'progress' field")
+
+ check_documents(DocCls=CachedResponse, sample_size=10)
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index 37ce1098..aaa1c5fe 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -26,0 +27 @@ from mongoengine.fields import (
+ FloatField,
@@ -81,0 +83 @@ class CachedResponse(Document):
+ progress (`float`): Progress percentage (between 0. and 1.) if the result is not complete yet.
@@ -95,0 +98 @@ class CachedResponse(Document):
+ progress = FloatField(min_value=0.0, max_value=1.0)
@@ -133,0 +137 @@ def upsert_response(
+ progress: Optional[float] = None,
@@ -141,0 +146 @@ def upsert_response(
+ progress=progress,
@@ -160,0 +166 @@ class CacheEntryWithoutContent(TypedDict):
+ progress: Optional[float]
@@ -169 +175 @@ def get_response_without_content(
- .only("http_status", "error_code", "worker_version", "dataset_git_revision")
+ .only("http_status", "error_code", "worker_version", "dataset_git_revision", "progress")
@@ -176,0 +183 @@ def get_response_without_content(
+ "progress": response.progress,
@@ -182 +189 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L
- "http_status", "error_code", "worker_version", "dataset_git_revision"
+ "http_status", "error_code", "worker_version", "dataset_git_revision", "progress"
@@ -189,0 +197 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L
+ "progress": response.progress,
@@ -203 +211 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O
- .only("content", "http_status", "error_code", "worker_version", "dataset_git_revision")
+ .only("content", "http_status", "error_code", "worker_version", "dataset_git_revision", "progress")
@@ -211,0 +220 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O
+ "progress": response.progress,
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index 1b7f8d2d..e151c0f1 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -117,0 +118 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
+ "progress": None,
@@ -126,0 +128 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
+ "progress": None,
@@ -165,0 +168 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
+ "progress": None,
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index 25c42461..960b09dd 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -5,0 +6 @@ from abc import ABC, abstractmethod
+from dataclasses import dataclass, field
@@ -40,0 +42,16 @@ ERROR_CODES_TO_RETRY: list[str] = ["ClientConnectionError"]
+@dataclass
+class JobResult:
+ content: Mapping[str, Any]
+ progress: float
+
+ def __post_init__(self) -> None:
+ if self.progress < 0.0 or self.progress > 1.0:
+ raise ValueError(f"Progress should be between 0 and 1, but got {self.progress}")
+
+
+@dataclass
+class CompleteJobResult(JobResult):
+ content: Mapping[str, Any]
+ progress: float = field(init=False, default=1.0)
+
+
@@ -291,0 +309 @@ class JobRunner(ABC):
+ - and the cached entry is complete (has a progress of 1.)
@@ -316,0 +335,4 @@ class JobRunner(ABC):
+ if cached_response["progress"] is not None and cached_response["progress"] < 1.0:
+ # this job is still waiting for more inputs to be complete - we should not skip it.
+ # this can happen with fan-in jobs
+ return False
@@ -337 +359,2 @@ class JobRunner(ABC):
- content = self.compute()
+ job_result = self.compute()
+ content = job_result.content
@@ -356,0 +380 @@ class JobRunner(ABC):
+ progress=job_result.progress,
@@ -396 +420 @@ class JobRunner(ABC):
- def compute(self) -> Mapping[str, Any]:
+ def compute(self) -> JobResult:
diff --git a/services/worker/src/worker/job_runners/config_names.py b/services/worker/src/worker/job_runners/config_names.py
index 323dfe26..052f7c6a 100644
--- a/services/worker/src/worker/job_runners/config_names.py
+++ b/services/worker/src/worker/job_runners/config_names.py
@@ -12 +12 @@ from libcommon.simple_cache import SplitFullName
-from worker.job_runner import JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunnerError
@@ -107,2 +107,4 @@ class ConfigNamesJobRunner(DatasetsBasedJobRunner):
- def compute(self) -> Mapping[str, Any]:
- return compute_config_names_response(dataset=self.dataset, hf_token=self.common_config.hf_token)
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult(
+ compute_config_names_response(dataset=self.dataset, hf_token=self.common_config.hf_token)
+ )
diff --git a/services/worker/src/worker/job_runners/dataset_info.py b/services/worker/src/worker/job_runners/dataset_info.py
index 7093ca57..9f4019fa 100644
--- a/services/worker/src/worker/job_runners/dataset_info.py
+++ b/services/worker/src/worker/job_runners/dataset_info.py
@@ -11 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from worker.job_runner import JobRunner, JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError
@@ -99,2 +99,2 @@ class DatasetInfoJobRunner(JobRunner):
- def compute(self) -> Mapping[str, Any]:
- return compute_dataset_info_response(dataset=self.dataset)
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult(compute_dataset_info_response(dataset=self.dataset))
diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py
index 30888ddb..c594f5ba 100644
--- a/services/worker/src/worker/job_runners/first_rows.py
+++ b/services/worker/src/worker/job_runners/first_rows.py
@@ -43 +43,6 @@ from worker.features import get_cell_value
-from worker.job_runner import ConfigNotFoundError, JobRunnerError, SplitNotFoundError
+from worker.job_runner import (
+ CompleteJobResult,
+ ConfigNotFoundError,
+ JobRunnerError,
+ SplitNotFoundError,
+)
@@ -665 +670 @@ class FirstRowsJobRunner(DatasetsBasedJobRunner):
- def compute(self) -> Mapping[str, Any]:
+ def compute(self) -> CompleteJobResult:
@@ -668,12 +673,14 @@ class FirstRowsJobRunner(DatasetsBasedJobRunner):
- return compute_first_rows_response(
- dataset=self.dataset,
- config=self.config,
- split=self.split,
- assets_base_url=self.assets_base_url,
- assets_directory=self.assets_directory,
- hf_token=self.common_config.hf_token,
- min_cell_bytes=self.first_rows_config.min_cell_bytes,
- rows_max_bytes=self.first_rows_config.max_bytes,
- rows_max_number=self.first_rows_config.max_number,
- rows_min_number=self.first_rows_config.min_number,
- columns_max_number=self.first_rows_config.columns_max_number,
+ return CompleteJobResult(
+ compute_first_rows_response(
+ dataset=self.dataset,
+ config=self.config,
+ split=self.split,
+ assets_base_url=self.assets_base_url,
+ assets_directory=self.assets_directory,
+ hf_token=self.common_config.hf_token,
+ min_cell_bytes=self.first_rows_config.min_cell_bytes,
+ rows_max_bytes=self.first_rows_config.max_bytes,
+ rows_max_number=self.first_rows_config.max_number,
+ rows_min_number=self.first_rows_config.min_number,
+ columns_max_number=self.first_rows_config.columns_max_number,
+ )
diff --git a/services/worker/src/worker/job_runners/parquet.py b/services/worker/src/worker/job_runners/parquet.py
index e5807e2a..b76bb5d6 100644
--- a/services/worker/src/worker/job_runners/parquet.py
+++ b/services/worker/src/worker/job_runners/parquet.py
@@ -11 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from worker.job_runner import JobRunner, JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError
@@ -102,2 +102,2 @@ class ParquetJobRunner(JobRunner):
- def compute(self) -> Mapping[str, Any]:
- return compute_parquet_response(dataset=self.dataset)
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult(compute_parquet_response(dataset=self.dataset))
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index 02608877..bd2f1ed9 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -48 +48 @@ from worker.config import AppConfig, ParquetAndDatasetInfoConfig
-from worker.job_runner import JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunnerError
@@ -889,14 +889,16 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner):
- def compute(self) -> Mapping[str, Any]:
- return compute_parquet_and_dataset_info_response(
- dataset=self.dataset,
- hf_endpoint=self.common_config.hf_endpoint,
- hf_token=self.common_config.hf_token,
- committer_hf_token=self.parquet_and_dataset_info_config.committer_hf_token,
- source_revision=self.parquet_and_dataset_info_config.source_revision,
- target_revision=self.parquet_and_dataset_info_config.target_revision,
- commit_message=self.parquet_and_dataset_info_config.commit_message,
- url_template=self.parquet_and_dataset_info_config.url_template,
- supported_datasets=self.parquet_and_dataset_info_config.supported_datasets,
- blocked_datasets=self.parquet_and_dataset_info_config.blocked_datasets,
- max_dataset_size=self.parquet_and_dataset_info_config.max_dataset_size,
- max_external_data_files=self.parquet_and_dataset_info_config.max_external_data_files,
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult(
+ compute_parquet_and_dataset_info_response(
+ dataset=self.dataset,
+ hf_endpoint=self.common_config.hf_endpoint,
+ hf_token=self.common_config.hf_token,
+ committer_hf_token=self.parquet_and_dataset_info_config.committer_hf_token,
+ source_revision=self.parquet_and_dataset_info_config.source_revision,
+ target_revision=self.parquet_and_dataset_info_config.target_revision,
+ commit_message=self.parquet_and_dataset_info_config.commit_message,
+ url_template=self.parquet_and_dataset_info_config.url_template,
+ supported_datasets=self.parquet_and_dataset_info_config.supported_datasets,
+ blocked_datasets=self.parquet_and_dataset_info_config.blocked_datasets,
+ max_dataset_size=self.parquet_and_dataset_info_config.max_dataset_size,
+ max_external_data_files=self.parquet_and_dataset_info_config.max_external_data_files,
+ )
diff --git a/services/worker/src/worker/job_runners/sizes.py b/services/worker/src/worker/job_runners/sizes.py
index a18ec238..c8ae0a6a 100644
--- a/services/worker/src/worker/job_runners/sizes.py
+++ b/services/worker/src/worker/job_runners/sizes.py
@@ -11 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from worker.job_runner import JobRunner, JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError
@@ -183,2 +183,2 @@ class SizesJobRunner(JobRunner):
- def compute(self) -> Mapping[str, Any]:
- return compute_sizes_response(dataset=self.dataset)
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult(compute_sizes_response(dataset=self.dataset))
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
index 2eb8c100..a5fa8f46 100644
--- a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
+++ b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
@@ -11 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from worker.job_runner import JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunnerError
@@ -120 +120 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner):
- def compute(self) -> Mapping[str, Any]:
+ def compute(self) -> CompleteJobResult:
@@ -125 +125,3 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner):
- return compute_split_names_from_dataset_info_response(dataset=self.dataset, config=self.config)
+ return CompleteJobResult(
+ compute_split_names_from_dataset_info_response(dataset=self.dataset, config=self.config)
+ )
diff --git a/services/worker/src/worker/job_runners/split_names_from_streaming.py b/services/worker/src/worker/job_runners/split_names_from_streaming.py
index 6fc15d28..56a7477f 100644
--- a/services/worker/src/worker/job_runners/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split_names_from_streaming.py
@@ -12 +12 @@ from libcommon.simple_cache import SplitFullName
-from worker.job_runner import JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunnerError
@@ -124 +124 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner):
- def compute(self) -> Mapping[str, Any]:
+ def compute(self) -> CompleteJobResult:
@@ -127,2 +127,4 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner):
- return compute_split_names_from_streaming_response(
- dataset=self.dataset, config=self.config, hf_token=self.common_config.hf_token
+ return CompleteJobResult(
+ compute_split_names_from_streaming_response(
+ dataset=self.dataset, config=self.config, hf_token=self.common_config.hf_token
+ )
diff --git a/services/worker/src/worker/job_runners/splits.py b/services/worker/src/worker/job_runners/splits.py
index fa8a9487..d276cc2d 100644
--- a/services/worker/src/worker/job_runners/splits.py
+++ b/services/worker/src/worker/job_runners/splits.py
@@ -12 +12 @@ from libcommon.simple_cache import SplitFullName
-from worker.job_runner import JobRunnerError
+from worker.job_runner import CompleteJobResult, JobRunnerError
@@ -136,2 +136,2 @@ class SplitsJobRunner(DatasetsBasedJobRunner):
- def compute(self) -> Mapping[str, Any]:
- return compute_splits_response(dataset=self.dataset, hf_token=self.common_config.hf_token)
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult(compute_splits_response(dataset=self.dataset, hf_token=self.common_config.hf_token))
diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py
index 8f1a5996..44affd1e 100644
--- a/services/worker/tests/job_runners/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__datasets_based_worker.py
@@ -8 +8 @@ from pathlib import Path
-from typing import Any, Callable, Mapping, Optional
+from typing import Callable, Optional
@@ -17,0 +18 @@ from worker.config import AppConfig
+from worker.job_runner import CompleteJobResult
@@ -35 +36 @@ class DummyJobRunner(DatasetsBasedJobRunner):
- def compute(self) -> Mapping[str, Any]:
+ def compute(self) -> CompleteJobResult:
@@ -39 +40 @@ class DummyJobRunner(DatasetsBasedJobRunner):
- return {"col1": "a" * 200}
+ return CompleteJobResult({"col1": "a" * 200})
diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py
index b41fafc5..1a329ac0 100644
--- a/services/worker/tests/job_runners/test_config_names.py
+++ b/services/worker/tests/job_runners/test_config_names.py
@@ -126 +126 @@ def test_compute_splits_response_simple_csv(
- result = job_runner.compute()
+ result = job_runner.compute().content
diff --git a/services/worker/tests/job_runners/test_dataset_info.py b/services/worker/tests/job_runners/test_dataset_info.py
index 37fae457..739ca2b6 100644
--- a/services/worker/tests/job_runners/test_dataset_info.py
+++ b/services/worker/tests/job_runners/test_dataset_info.py
@@ -108 +108 @@ def test_compute(
- assert job_runner.compute() == expected_content
+ assert job_runner.compute().content == expected_content
diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py
index 23add41e..1d281884 100644
--- a/services/worker/tests/job_runners/test_first_rows.py
+++ b/services/worker/tests/job_runners/test_first_rows.py
@@ -159 +159 @@ def test_number_rows(
- result = job_runner.compute()
+ result = job_runner.compute().content
@@ -225 +225 @@ def test_truncation(
- response = job_runner.compute()
+ response = job_runner.compute().content
diff --git a/services/worker/tests/job_runners/test_parquet.py b/services/worker/tests/job_runners/test_parquet.py
index 6e0d71c2..ddf4adb2 100644
--- a/services/worker/tests/job_runners/test_parquet.py
+++ b/services/worker/tests/job_runners/test_parquet.py
@@ -108 +108 @@ def test_compute(
- assert job_runner.compute() == expected_content
+ assert job_runner.compute().content == expected_content
diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index 7192d30b..1b911c27 100644
--- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -409 +409 @@ def test_compute_splits_response_simple_csv_ok(
- result = job_runner.compute()
+ result = job_runner.compute().content
diff --git a/services/worker/tests/job_runners/test_sizes.py b/services/worker/tests/job_runners/test_sizes.py
index 6b731071..1895211e 100644
--- a/services/worker/tests/job_runners/test_sizes.py
+++ b/services/worker/tests/job_runners/test_sizes.py
@@ -263 +263 @@ def test_compute(
- assert job_runner.compute() == expected_content
+ assert job_runner.compute().content == expected_content
diff --git a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
index 1f5869a3..9c68913a 100644
--- a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
+++ b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
@@ -136 +136 @@ def test_compute(
- assert job_runner.compute() == content
+ assert job_runner.compute().content == content
diff --git a/services/worker/tests/job_runners/test_split_names_from_streaming.py b/services/worker/tests/job_runners/test_split_names_from_streaming.py
index 19418d68..19228d75 100644
--- a/services/worker/tests/job_runners/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/test_split_names_from_streaming.py
@@ -120 +120 @@ def test_compute_split_names_from_streaming_response(
- result = job_runner.compute()
+ result = job_runner.compute().content
diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py
index b69c7022..a987b48f 100644
--- a/services/worker/tests/job_runners/test_splits.py
+++ b/services/worker/tests/job_runners/test_splits.py
@@ -127 +127 @@ def test_compute_splits_response_simple_csv(
- result = job_runner.compute()
+ result = job_runner.compute().content
diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py
index 2b154439..36716405 100644
--- a/services/worker/tests/test_job_runner.py
+++ b/services/worker/tests/test_job_runner.py
@@ -13 +13 @@ from worker.config import WorkerConfig
-from worker.job_runner import ERROR_CODES_TO_RETRY, JobRunner
+from worker.job_runner import ERROR_CODES_TO_RETRY, CompleteJobResult, JobRunner
@@ -42,2 +42,2 @@ class DummyJobRunner(JobRunner):
- def compute(self) -> Mapping[str, Any]:
- return {"key": "value"}
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult({"key": "value"})
@@ -94,0 +95 @@ class CacheEntry:
+ progress: Optional[float] = None
@@ -177,0 +179,20 @@ class CacheEntry:
+ (
+ False,
+ CacheEntry(
+ error_code=None, # no error
+ worker_version=DummyJobRunner.get_version(),
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
+ progress=0.5, # incomplete result
+ ),
+ False, # process
+ ),
+ (
+ False,
+ CacheEntry(
+ error_code=None, # no error
+ worker_version=DummyJobRunner.get_version(),
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
+ progress=1.0, # complete result
+ ),
+ True, # skip
+ ),
@@ -212,0 +234 @@ def test_should_skip_job(
+ progress=cache_entry.progress,
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index d97341ff..11d755e7 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -1 +1 @@
-from typing import Any, Mapping, Optional
+from typing import Optional
@@ -9 +9 @@ from worker.config import AppConfig, WorkerConfig
-from worker.job_runner import JobRunner
+from worker.job_runner import CompleteJobResult, JobRunner
@@ -28,2 +28,2 @@ class DummyJobRunner(JobRunner):
- def compute(self) -> Mapping[str, Any]:
- return {"key": "value"}
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult({"key": "value"})
|
|
745a83b5d9c9f43465e0950da69396c6faa39b96
|
Quentin Lhoest
| 2023-03-10T18:40:47 |
Fix gated (gated=True -> gated="auto") (#920)
|
diff --git a/e2e/tests/fixtures/hub.py b/e2e/tests/fixtures/hub.py
index 5f6a3af8..4d960ac7 100644
--- a/e2e/tests/fixtures/hub.py
+++ b/e2e/tests/fixtures/hub.py
@@ -8 +8,11 @@ from contextlib import contextmanager, suppress
-from typing import Any, Callable, Iterator, Literal, Mapping, Optional, TypedDict
+from typing import (
+ Any,
+ Callable,
+ Dict,
+ Iterator,
+ Literal,
+ Mapping,
+ Optional,
+ TypedDict,
+ Union,
+)
@@ -32 +42 @@ def update_repo_settings(
- gated: Optional[bool] = None,
+ gated: Optional[str] = None,
@@ -48 +58 @@ def update_repo_settings(
- gated (`bool`, *optional*, defaults to `None`):
+ gated (`str`, *optional*, defaults to `None`):
@@ -49,0 +60 @@ def update_repo_settings(
+ Possible values are 'auto' and 'manual'
@@ -81 +92 @@ def update_repo_settings(
- json = {}
+ json: Dict[str, Union[bool, str]] = {}
@@ -132 +143,7 @@ def create_hf_dataset_repo_csv_data(
- hf_api: HfApi, hf_token: str, csv_path: str, *, private: bool = False, gated: bool = False, user: str = CI_HUB_USER
+ hf_api: HfApi,
+ hf_token: str,
+ csv_path: str,
+ *,
+ private: bool = False,
+ gated: Optional[str] = None,
+ user: str = CI_HUB_USER,
@@ -175 +192 @@ def hf_gated_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str)
- repo_id = create_hf_dataset_repo_csv_data(hf_api=hf_api, hf_token=hf_token, csv_path=csv_path, gated=True)
+ repo_id = create_hf_dataset_repo_csv_data(hf_api=hf_api, hf_token=hf_token, csv_path=csv_path, gated="auto")
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py
index 63ed4fb8..362bc3bd 100644
--- a/libs/libcommon/src/libcommon/dataset.py
+++ b/libs/libcommon/src/libcommon/dataset.py
@@ -147,0 +148,2 @@ def ask_access(
+ if r.headers and r.headers.get("X-Error-Code") == "RepoNotGated":
+ return # the dataset is not gated
diff --git a/services/admin/tests/fixtures/hub.py b/services/admin/tests/fixtures/hub.py
index 42aaba31..404d38ff 100644
--- a/services/admin/tests/fixtures/hub.py
+++ b/services/admin/tests/fixtures/hub.py
@@ -8 +8 @@ from contextlib import contextmanager, suppress
-from typing import Any, Callable, Iterator, Literal, Optional, TypedDict
+from typing import Any, Callable, Dict, Iterator, Literal, Optional, TypedDict, Union
@@ -28 +28 @@ def update_repo_settings(
- gated: Optional[bool] = None,
+ gated: Optional[str] = None,
@@ -44 +44 @@ def update_repo_settings(
- gated (`bool`, *optional*, defaults to `None`):
+ gated (`str`, *optional*, defaults to `None`):
@@ -45,0 +46 @@ def update_repo_settings(
+ Possible values are 'auto' and 'manual'
@@ -77 +78 @@ def update_repo_settings(
- json = {}
+ json: Dict[str, Union[bool, str]] = {}
@@ -133 +134,7 @@ def create_hf_dataset_repo(
- hf_api: HfApi, hf_token: str, prefix: str, *, private: bool = False, gated: bool = False, user: str = CI_HUB_USER
+ hf_api: HfApi,
+ hf_token: str,
+ prefix: str,
+ *,
+ private: bool = False,
+ gated: Optional[str] = None,
+ user: str = CI_HUB_USER,
@@ -153 +160 @@ def hf_gated_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterator[str]:
- repo_id = create_hf_dataset_repo(hf_api=hf_api, hf_token=hf_token, prefix="repo_empty", gated=True)
+ repo_id = create_hf_dataset_repo(hf_api=hf_api, hf_token=hf_token, prefix="repo_empty", gated="auto")
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 20fee544..1760fae2 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -11,0 +12 @@ from typing import (
+ Dict,
@@ -18,0 +20 @@ from typing import (
+ Union,
@@ -44 +46 @@ def update_repo_settings(
- gated: Optional[bool] = None,
+ gated: Optional[str] = None,
@@ -60 +62 @@ def update_repo_settings(
- gated (`bool`, *optional*, defaults to `None`):
+ gated (`str`, *optional*, defaults to `None`):
@@ -61,0 +64 @@ def update_repo_settings(
+ Possible values are 'auto' and 'manual'
@@ -93 +96 @@ def update_repo_settings(
- json = {}
+ json: Dict[str, Union[bool, str]] = {}
@@ -114 +117 @@ def create_hub_dataset_repo(
- gated: bool = False,
+ gated: Optional[str] = None,
@@ -183 +186 @@ def hub_gated_csv(csv_path: str) -> Iterator[str]:
- repo_id = create_hub_dataset_repo(prefix="csv_gated", file_paths=[csv_path], gated=True)
+ repo_id = create_hub_dataset_repo(prefix="csv_gated", file_paths=[csv_path], gated="auto")
@@ -198 +201 @@ def hub_gated_extra_fields_csv(csv_path: str, extra_fields_readme: str) -> Itera
- prefix="csv_extra_fields_gated", file_paths=[csv_path, extra_fields_readme], gated=True
+ prefix="csv_extra_fields_gated", file_paths=[csv_path, extra_fields_readme], gated="auto"
|
|
a80ba4ff6801ee38df3137c76798986ac605f103
|
Sylvain Lesage
| 2023-03-10T11:37:41 |
Revert "feat: 🎸 change log level to DEBUG (#917)" (#918)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 7323baca..54f1b731 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -71 +71 @@ common:
- logLevel: "DEBUG"
+ logLevel: "INFO"
diff --git a/libs/libcommon/src/libcommon/log.py b/libs/libcommon/src/libcommon/log.py
index e5b99911..c531c5dd 100644
--- a/libs/libcommon/src/libcommon/log.py
+++ b/libs/libcommon/src/libcommon/log.py
@@ -10,4 +9,0 @@ def init_logging(log_level: int = logging.INFO) -> None:
-
- requests_log = logging.getLogger("requests.packages.urllib3")
- requests_log.setLevel(log_level)
- requests_log.propagate = True
|
|
6643f943619a6b8c2b4449c12eafc883db1873fe
|
Sylvain Lesage
| 2023-03-10T10:13:15 |
feat: 🎸 change log level to DEBUG (#917)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 54f1b731..7323baca 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -71 +71 @@ common:
- logLevel: "INFO"
+ logLevel: "DEBUG"
diff --git a/libs/libcommon/src/libcommon/log.py b/libs/libcommon/src/libcommon/log.py
index c531c5dd..e5b99911 100644
--- a/libs/libcommon/src/libcommon/log.py
+++ b/libs/libcommon/src/libcommon/log.py
@@ -9,0 +10,4 @@ def init_logging(log_level: int = logging.INFO) -> None:
+
+ requests_log = logging.getLogger("requests.packages.urllib3")
+ requests_log.setLevel(log_level)
+ requests_log.propagate = True
|
|
1e5064e2ef711741bdf7ec579898ce63495a5c43
|
Sylvain Lesage
| 2023-03-09T18:59:54 |
Add some profiling (#915)
|
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index d7303b0c..e2b06d19 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -3,0 +4 @@
+import logging
@@ -10,0 +12 @@ from starlette.requests import Request
+from api.prometheus import StepProfiler
@@ -64,31 +66,44 @@ def auth_check(
- if external_auth_url is None:
- return True
- try:
- url = external_auth_url % dataset
- except TypeError as e:
- raise ValueError("external_auth_url must contain %s") from e
- try:
- response = requests.get(url, auth=RequestAuth(request), timeout=hf_timeout_seconds)
- except Exception as err:
- raise AuthCheckHubRequestError(
- (
- "Authentication check on the Hugging Face Hub failed or timed out. Please try again later, it's a"
- " temporary internal issue."
- ),
- err,
- ) from err
-
- if response.status_code == 200:
- return True
- elif response.status_code == 401:
- raise ExternalUnauthenticatedError(
- "The dataset does not exist, or is not accessible without authentication (private or gated). Please check"
- " the spelling of the dataset name or retry with authentication."
- )
- elif response.status_code in [403, 404]:
- raise ExternalAuthenticatedError(
- "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please"
- " check the spelling of the dataset name or retry with other authentication credentials."
- )
- else:
- raise ValueError(f"Unexpected status code {response.status_code}")
+ with StepProfiler(method="auth_check", step="all"):
+ with StepProfiler(method="auth_check", step="prepare parameters"):
+ if external_auth_url is None:
+ return True
+ try:
+ url = external_auth_url % dataset
+ except TypeError as e:
+ raise ValueError("external_auth_url must contain %s") from e
+ with StepProfiler(method="auth_check", step="create auth parameter"):
+ auth = RequestAuth(request)
+ with StepProfiler(
+ method="auth_check",
+ step="requests.get",
+ context=f"external_auth_url={external_auth_url} timeout={hf_timeout_seconds}",
+ ):
+ try:
+ logging.debug(
+ f"Checking authentication on the Hugging Face Hub for dataset {dataset}, url: {url}, timeout:"
+ f" {hf_timeout_seconds}"
+ )
+ response = requests.get(url, auth=auth, timeout=hf_timeout_seconds)
+ except Exception as err:
+ raise AuthCheckHubRequestError(
+ (
+ "Authentication check on the Hugging Face Hub failed or timed out. Please try again later,"
+ " it's a temporary internal issue."
+ ),
+ err,
+ ) from err
+ with StepProfiler(method="auth_check", step="return or raise"):
+ if response.status_code == 200:
+ return True
+ elif response.status_code == 401:
+ raise ExternalUnauthenticatedError(
+ "The dataset does not exist, or is not accessible without authentication (private or gated)."
+ " Please check the spelling of the dataset name or retry with authentication."
+ )
+ elif response.status_code in [403, 404]:
+ raise ExternalAuthenticatedError(
+ "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ " Please check the spelling of the dataset name or retry with other authentication credentials."
+ )
+ else:
+ raise ValueError(f"Unexpected status code {response.status_code}")
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 58a24209..e8ab75d6 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -226,2 +226,2 @@ def create_endpoint(
- f"endpoint={endpoint_name} dataset={dataset_parameter} ",
- f"config={config_parameter} split={split_parameter}",
+ f"endpoint={endpoint_name} dataset={dataset_parameter} config={config_parameter}"
+ + f" split={split_parameter}"
|
|
11ea7d077e9e34a7d709a4bfd8dbbd78bdcb2107
|
Sylvain Lesage
| 2023-03-09T16:41:02 |
Add "context" label to filter the metrics on the endpoint (#914)
|
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml
index cdc7ea4d..c6ab2b65 100644
--- a/.github/workflows/_e2e_tests.yml
+++ b/.github/workflows/_e2e_tests.yml
@@ -73,0 +74 @@ jobs:
+ API_HF_TIMEOUT_SECONDS: "10"
diff --git a/services/admin/Makefile b/services/admin/Makefile
index c177523c..1dc6ae35 100644
--- a/services/admin/Makefile
+++ b/services/admin/Makefile
@@ -27 +27,4 @@ test:
- PROMETHEUS_MULTIPROC_DIR=/tmp poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
+ rm -rf /tmp/admin.prometheus
+ mkdir /tmp/admin.prometheus
+ PROMETHEUS_MULTIPROC_DIR=/tmp/admin.prometheus poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
+ rm -rf /tmp/admin.prometheus
diff --git a/services/api/Makefile b/services/api/Makefile
index 004c8356..44aa2fbe 100644
--- a/services/api/Makefile
+++ b/services/api/Makefile
@@ -27 +27,4 @@ test:
- PROMETHEUS_MULTIPROC_DIR=/tmp poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
+ rm -rf /tmp/api.prometheus
+ mkdir /tmp/api.prometheus
+ PROMETHEUS_MULTIPROC_DIR=/tmp/api.prometheus poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
+ rm -rf /tmp/api.prometheus
diff --git a/services/api/src/api/prometheus.py b/services/api/src/api/prometheus.py
index 2f921a56..ec5d069c 100644
--- a/services/api/src/api/prometheus.py
+++ b/services/api/src/api/prometheus.py
@@ -26,2 +26,2 @@ METHOD_STEPS_PROCESSING_TIME = Histogram(
- "Histogram of the processing time of specific steps in methods (in seconds)",
- ["method", "step"],
+ "Histogram of the processing time of specific steps in methods for a given context (in seconds)",
+ ["method", "step", "context"],
@@ -38 +38 @@ class StepProfiler:
- >>> with StepProfiler("method", "step") as profiler:
+ >>> with StepProfiler("method", "step", "context") as profiler:
@@ -43,0 +44 @@ class StepProfiler:
+ context (str|None): An optional string that adds context. If None, the label "None" is used.
@@ -46 +47 @@ class StepProfiler:
- def __init__(self, method: str, step: str):
+ def __init__(self, method: str, step: str, context: Optional[str] = None):
@@ -48,0 +50 @@ class StepProfiler:
+ self.context = str(context)
@@ -61 +63,3 @@ class StepProfiler:
- METHOD_STEPS_PROCESSING_TIME.labels(method=self.method, step=self.step).observe(after_time - self.before_time)
+ METHOD_STEPS_PROCESSING_TIME.labels(method=self.method, step=self.step, context=self.context).observe(
+ after_time - self.before_time
+ )
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 52a62597..58a24209 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -211 +211,2 @@ def create_endpoint(
- with StepProfiler(method="processing_step_endpoint", step="all"):
+ context = f"endpoint: {endpoint_name}"
+ with StepProfiler(method="processing_step_endpoint", step="all", context=context):
@@ -214 +215,3 @@ def create_endpoint(
- method="processing_step_endpoint", step="validate parameters and get processing steps"
+ method="processing_step_endpoint",
+ step="validate parameters and get processing steps",
+ context=context,
@@ -240 +243 @@ def create_endpoint(
- with StepProfiler(method="processing_step_endpoint", step="check authentication"):
+ with StepProfiler(method="processing_step_endpoint", step="check authentication", context=context):
@@ -249 +252 @@ def create_endpoint(
- with StepProfiler(method="processing_step_endpoint", step="get cache entry"):
+ with StepProfiler(method="processing_step_endpoint", step="get cache entry", context=context):
@@ -258 +261 @@ def create_endpoint(
- with StepProfiler(method="processing_step_endpoint", step="generate OK response"):
+ with StepProfiler(method="processing_step_endpoint", step="generate OK response", context=context):
@@ -261 +264 @@ def create_endpoint(
- with StepProfiler(method="routes.endpoint", step="generate error response"):
+ with StepProfiler(method="processing_step_endpoint", step="generate error response", context=context):
@@ -267 +270,3 @@ def create_endpoint(
- with StepProfiler(method="processing_step_endpoint", step="generate API error response"):
+ with StepProfiler(
+ method="processing_step_endpoint", step="generate API error response", context=context
+ ):
diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py
index 2bb25276..78e4bccd 100644
--- a/services/api/tests/test_app.py
+++ b/services/api/tests/test_app.py
@@ -151 +151 @@ def test_metrics(client: TestClient) -> None:
- 'method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"}',
+ 'method_steps_processing_time_seconds_sum{context="None",method="healthcheck_endpoint",step="all"}',
diff --git a/services/api/tests/test_prometheus.py b/services/api/tests/test_prometheus.py
index 12ed41b5..4f74845f 100644
--- a/services/api/tests/test_prometheus.py
+++ b/services/api/tests/test_prometheus.py
@@ -1,0 +2,2 @@ import os
+import time
+from typing import Dict, Optional
@@ -3 +5,12 @@ import os
-from api.prometheus import Prometheus
+from api.prometheus import Prometheus, StepProfiler
+
+
+def parse_metrics(content: str) -> dict[str, float]:
+ # examples:
+ # starlette_requests_total{method="GET",path_template="/metrics"} 1.0
+ # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05
+ return {
+ parts[0]: float(parts[1])
+ for line in content.split("\n")
+ if line and line[0] != "#" and (parts := line.rsplit(" ", 1))
+ }
@@ -14,8 +27 @@ def test_prometheus() -> None:
- print("content:", content)
- lines = content.split("\n")
- # examples:
- # starlette_requests_total{method="GET",path_template="/metrics"} 1.0
- # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05
- metrics = {
- parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1))
- }
+ metrics = parse_metrics(content)
@@ -28,0 +35,66 @@ def test_prometheus() -> None:
+
+
+def create_key(suffix: str, labels: Dict[str, str], le: Optional[str] = None) -> str:
+ items = list(labels.items())
+ if le:
+ items.append(("le", le))
+ labels_string = ",".join([f'{key}="{value}"' for key, value in sorted(items)])
+ return f"method_steps_processing_time_seconds_{suffix}{{{labels_string}}}"
+
+
+def check_histogram_metric(
+ metrics: Dict[str, float], method: str, step: str, context: str, events: int, duration: float
+) -> None:
+ labels = {"context": context, "method": method, "step": step}
+ assert metrics[create_key("count", labels)] == events, metrics
+ assert metrics[create_key("bucket", labels, le="+Inf")] == events, metrics
+ assert metrics[create_key("bucket", labels, le="1.0")] == events, metrics
+ assert metrics[create_key("bucket", labels, le="0.05")] == 0, metrics
+ assert metrics[create_key("sum", labels)] >= duration, metrics
+ assert metrics[create_key("sum", labels)] <= duration * 1.1, metrics
+
+
+def test_step_profiler() -> None:
+ duration = 0.1
+ method = "test_step_profiler"
+ step_all = "all"
+ context = "None"
+ with StepProfiler(method=method, step=step_all):
+ time.sleep(duration)
+ metrics = parse_metrics(Prometheus().getLatestContent())
+ check_histogram_metric(metrics=metrics, method=method, step=step_all, context=context, events=1, duration=duration)
+
+
+def test_nested_step_profiler() -> None:
+ method = "test_nested_step_profiler"
+ step_all = "all"
+ context = "None"
+ step_1 = "step_1"
+ duration_1a = 0.1
+ duration_1b = 0.3
+ context_1 = "None"
+ step_2 = "step_2"
+ duration_2 = 0.5
+ context_2 = "endpoint: /splits"
+ with StepProfiler(method=method, step=step_all):
+ with StepProfiler(method, step_1):
+ time.sleep(duration_1a)
+ with StepProfiler(method, step_1, context_1):
+ time.sleep(duration_1b)
+ with StepProfiler(method, step_2, context_2):
+ time.sleep(duration_2)
+ metrics = parse_metrics(Prometheus().getLatestContent())
+ check_histogram_metric(
+ metrics=metrics,
+ method=method,
+ step=step_all,
+ context=context,
+ events=1,
+ duration=duration_1a + duration_1b + duration_2,
+ )
+ check_histogram_metric(
+ metrics=metrics, method=method, step=step_1, context=context_1, events=2, duration=duration_1a + duration_1b
+ )
+ check_histogram_metric(
+ metrics=metrics, method=method, step=step_2, context=context_2, events=1, duration=duration_2
+ )
|
|
077f369eb31f77a9ccd5519f03c6196b509e7db1
|
Sylvain Lesage
| 2023-03-09T10:10:19 |
Instrument api service (#911)
|
diff --git a/e2e/tests/test_21_api_metrics.py b/e2e/tests/test_21_api_metrics.py
index 67cadb1a..a3f88ab9 100644
--- a/e2e/tests/test_21_api_metrics.py
+++ b/e2e/tests/test_21_api_metrics.py
@@ -23 +23,6 @@ def test_metrics() -> None:
- metrics = {line.split(" ")[0]: float(line.split(" ")[1]) for line in lines if line and line[0] != "#"}
+ # examples:
+ # starlette_requests_total{method="GET",path_template="/metrics"} 1.0
+ # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05
+ metrics = {
+ parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1))
+ }
diff --git a/services/admin/pyproject.toml b/services/admin/pyproject.toml
index 7680b5f8..8efbf362 100644
--- a/services/admin/pyproject.toml
+++ b/services/admin/pyproject.toml
@@ -55,0 +56,5 @@ strict = true
+
+[[tool.mypy.overrides]]
+module = "prometheus_client.*"
+ignore_missing_imports = true
+# ^ prometheus_client is now typed, but starlette-prometheus requires an old version
diff --git a/services/admin/src/admin/prometheus.py b/services/admin/src/admin/prometheus.py
index f54bbe2c..a30b9647 100644
--- a/services/admin/src/admin/prometheus.py
+++ b/services/admin/src/admin/prometheus.py
@@ -12 +12 @@ from libcommon.storage import StrPath
-from prometheus_client import ( # type: ignore[import]
+from prometheus_client import (
@@ -19 +19 @@ from prometheus_client import ( # type: ignore[import]
-from prometheus_client.multiprocess import MultiProcessCollector # type: ignore[import]
+from prometheus_client.multiprocess import MultiProcessCollector
diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml
index fdd88d31..20a9ba25 100644
--- a/services/api/pyproject.toml
+++ b/services/api/pyproject.toml
@@ -56,0 +57,5 @@ strict = true
+
+[[tool.mypy.overrides]]
+module = "prometheus_client.*"
+ignore_missing_imports = true
+# ^ prometheus_client is now typed, but starlette-prometheus requires an old version
diff --git a/services/api/src/api/prometheus.py b/services/api/src/api/prometheus.py
index ba515612..2f921a56 100644
--- a/services/api/src/api/prometheus.py
+++ b/services/api/src/api/prometheus.py
@@ -5 +5,3 @@ import os
-from typing import Any
+import time
+from types import TracebackType
+from typing import Any, Optional, Type, TypeVar
@@ -7 +9 @@ from typing import Any
-from prometheus_client import ( # type: ignore
+from prometheus_client import (
@@ -10,0 +13 @@ from prometheus_client import ( # type: ignore
+ Histogram,
@@ -13 +16 @@ from prometheus_client import ( # type: ignore
-from prometheus_client.multiprocess import MultiProcessCollector # type: ignore
+from prometheus_client.multiprocess import MultiProcessCollector
@@ -19,0 +23,40 @@ from starlette.responses import Response
+# the metrics are global to the process
+METHOD_STEPS_PROCESSING_TIME = Histogram(
+ "method_steps_processing_time_seconds",
+ "Histogram of the processing time of specific steps in methods (in seconds)",
+ ["method", "step"],
+)
+
+T = TypeVar("T", bound="StepProfiler")
+
+
+class StepProfiler:
+ """
+ A context manager that measures the time spent in a step of a method and reports it to Prometheus.
+
+ Example:
+ >>> with StepProfiler("method", "step") as profiler:
+ ... pass
+
+ Args:
+ method (str): The name of the method.
+ step (str): The name of the step.
+ """
+
+ def __init__(self, method: str, step: str):
+ self.method = method
+ self.step = step
+ self.before_time = time.perf_counter()
+
+ def __enter__(self: T) -> T:
+ return self
+
+ def __exit__(
+ self,
+ exc_type: Optional[Type[BaseException]],
+ exc_value: Optional[BaseException],
+ traceback: Optional[TracebackType],
+ ) -> None:
+ after_time = time.perf_counter()
+ METHOD_STEPS_PROCESSING_TIME.labels(method=self.method, step=self.step).observe(after_time - self.before_time)
+
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index b4046f13..52a62597 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -17,0 +18 @@ from api.config import EndpointConfig
+from api.prometheus import StepProfiler
@@ -210,50 +211,58 @@ def create_endpoint(
- try:
- # validating request parameters
- dataset_parameter = request.query_params.get("dataset")
- config_parameter = request.query_params.get("config")
- split_parameter = request.query_params.get("split")
- validators = get_input_type_validators_by_priority(steps_by_input_type=steps_by_input_type)
-
- logging.debug(
- f"endpoint={endpoint_name} dataset={dataset_parameter} ",
- f"config={config_parameter} split={split_parameter}",
- )
-
- validator = get_input_type_validator_by_parameters(
- validators, dataset_parameter, config_parameter, split_parameter
- )
- processing_steps = steps_by_input_type[validator.input_type]
- dataset, config, split = validator.get_useful_parameters(
- dataset_parameter, config_parameter, split_parameter
- )
-
- # for now, dataset is always required in the endpoints.
- if not dataset:
- raise MissingRequiredParameterError("Parameter 'dataset' is required")
-
- # if auth_check fails, it will raise an exception that will be caught below
- auth_check(
- dataset,
- external_auth_url=external_auth_url,
- request=request,
- hf_timeout_seconds=hf_timeout_seconds,
- )
-
- # getting result based on processing steps
- result = get_cache_entry_from_steps(
- processing_steps, dataset, config, split, init_processing_steps, hf_endpoint, hf_token
- )
-
- content = result["content"]
- http_status = result["http_status"]
- error_code = result["error_code"]
- if http_status == HTTPStatus.OK:
- return get_json_ok_response(content=content, max_age=max_age_long)
-
- return get_json_error_response(
- content=content, status_code=http_status, max_age=max_age_short, error_code=error_code
- )
- except ApiCustomError as e:
- return get_json_api_error_response(error=e, max_age=max_age_short)
- except Exception as e:
- return get_json_api_error_response(error=UnexpectedError("Unexpected error.", e), max_age=max_age_short)
+ with StepProfiler(method="processing_step_endpoint", step="all"):
+ try:
+ with StepProfiler(
+ method="processing_step_endpoint", step="validate parameters and get processing steps"
+ ):
+ # validating request parameters
+ dataset_parameter = request.query_params.get("dataset")
+ config_parameter = request.query_params.get("config")
+ split_parameter = request.query_params.get("split")
+ validators = get_input_type_validators_by_priority(steps_by_input_type=steps_by_input_type)
+
+ logging.debug(
+ f"endpoint={endpoint_name} dataset={dataset_parameter} ",
+ f"config={config_parameter} split={split_parameter}",
+ )
+
+ validator = get_input_type_validator_by_parameters(
+ validators, dataset_parameter, config_parameter, split_parameter
+ )
+ processing_steps = steps_by_input_type[validator.input_type]
+ dataset, config, split = validator.get_useful_parameters(
+ dataset_parameter, config_parameter, split_parameter
+ )
+
+ # for now, dataset is always required in the endpoints.
+ if not dataset:
+ raise MissingRequiredParameterError("Parameter 'dataset' is required")
+
+ # if auth_check fails, it will raise an exception that will be caught below
+ with StepProfiler(method="processing_step_endpoint", step="check authentication"):
+ auth_check(
+ dataset,
+ external_auth_url=external_auth_url,
+ request=request,
+ hf_timeout_seconds=hf_timeout_seconds,
+ )
+
+ # getting result based on processing steps
+ with StepProfiler(method="processing_step_endpoint", step="get cache entry"):
+ result = get_cache_entry_from_steps(
+ processing_steps, dataset, config, split, init_processing_steps, hf_endpoint, hf_token
+ )
+
+ content = result["content"]
+ http_status = result["http_status"]
+ error_code = result["error_code"]
+ if http_status == HTTPStatus.OK:
+ with StepProfiler(method="processing_step_endpoint", step="generate OK response"):
+ return get_json_ok_response(content=content, max_age=max_age_long)
+
+ with StepProfiler(method="routes.endpoint", step="generate error response"):
+ return get_json_error_response(
+ content=content, status_code=http_status, max_age=max_age_short, error_code=error_code
+ )
+ except Exception as e:
+ error = e if isinstance(e, ApiCustomError) else UnexpectedError("Unexpected error.", e)
+ with StepProfiler(method="processing_step_endpoint", step="generate API error response"):
+ return get_json_api_error_response(error=error, max_age=max_age_short)
diff --git a/services/api/src/api/routes/healthcheck.py b/services/api/src/api/routes/healthcheck.py
index d5b66889..2b11dc59 100644
--- a/services/api/src/api/routes/healthcheck.py
+++ b/services/api/src/api/routes/healthcheck.py
@@ -8,0 +9,2 @@ from starlette.responses import PlainTextResponse, Response
+from api.prometheus import StepProfiler
+
@@ -12 +14,2 @@ async def healthcheck_endpoint(_: Request) -> Response:
- return PlainTextResponse("ok", headers={"Cache-Control": "no-store"})
+ with StepProfiler(method="healthcheck_endpoint", step="all"):
+ return PlainTextResponse("ok", headers={"Cache-Control": "no-store"})
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 2c8a976e..65f54177 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -12,0 +13 @@ from api.authentication import auth_check
+from api.prometheus import StepProfiler
@@ -55,6 +56,10 @@ def create_valid_endpoint(
- try:
- logging.info("/valid")
- content = {"valid": get_valid(processing_steps_for_valid=processing_steps_for_valid)}
- return get_json_ok_response(content, max_age=max_age_long)
- except Exception as e:
- return get_json_api_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age_short)
+ with StepProfiler(method="valid_endpoint", step="all"):
+ try:
+ logging.info("/valid")
+ with StepProfiler(method="valid_endpoint", step="prepare content"):
+ content = {"valid": get_valid(processing_steps_for_valid=processing_steps_for_valid)}
+ with StepProfiler(method="valid_endpoint", step="generate OK response"):
+ return get_json_ok_response(content, max_age=max_age_long)
+ except Exception as e:
+ with StepProfiler(method="valid_endpoint", step="generate API error response"):
+ return get_json_api_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age_short)
@@ -74,20 +79,25 @@ def create_is_valid_endpoint(
- try:
- dataset = request.query_params.get("dataset")
- logging.info(f"/is-valid, dataset={dataset}")
- if not are_valid_parameters([dataset]) or not dataset:
- raise MissingRequiredParameterError("Parameter 'dataset' is required")
- # if auth_check fails, it will raise an exception that will be caught below
- auth_check(
- dataset,
- external_auth_url=external_auth_url,
- request=request,
- hf_timeout_seconds=hf_timeout_seconds,
- )
- content = {
- "valid": is_valid(dataset=dataset, processing_steps_for_valid=processing_steps_for_valid),
- }
- return get_json_ok_response(content=content, max_age=max_age_long)
- except ApiCustomError as e:
- return get_json_api_error_response(error=e, max_age=max_age_short)
- except Exception as e:
- return get_json_api_error_response(error=UnexpectedError("Unexpected error.", e), max_age=max_age_short)
+ with StepProfiler(method="is_valid_endpoint", step="all"):
+ try:
+ with StepProfiler(method="is_valid_endpoint", step="validate parameters and get processing steps"):
+ dataset = request.query_params.get("dataset")
+ logging.info(f"/is-valid, dataset={dataset}")
+ if not are_valid_parameters([dataset]) or not dataset:
+ raise MissingRequiredParameterError("Parameter 'dataset' is required")
+ # if auth_check fails, it will raise an exception that will be caught below
+ with StepProfiler(method="is_valid_endpoint", step="check authentication"):
+ auth_check(
+ dataset,
+ external_auth_url=external_auth_url,
+ request=request,
+ hf_timeout_seconds=hf_timeout_seconds,
+ )
+ with StepProfiler(method="is_valid_endpoint", step="prepare content"):
+ content = {
+ "valid": is_valid(dataset=dataset, processing_steps_for_valid=processing_steps_for_valid),
+ }
+ with StepProfiler(method="is_valid_endpoint", step="generate OK response"):
+ return get_json_ok_response(content=content, max_age=max_age_long)
+ except Exception as e:
+ error = e if isinstance(e, ApiCustomError) else UnexpectedError("Unexpected error.", e)
+ with StepProfiler(method="is_valid_endpoint", step="generate API error response"):
+ return get_json_api_error_response(error=error, max_age=max_age_short)
diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py
index 131c3b30..caa9b17e 100644
--- a/services/api/src/api/routes/webhook.py
+++ b/services/api/src/api/routes/webhook.py
@@ -14,0 +15 @@ from starlette.responses import Response
+from api.prometheus import StepProfiler
@@ -105,29 +106,33 @@ def create_webhook_endpoint(
- try:
- json = await request.json()
- except Exception:
- content = {"status": "error", "error": "the body could not be parsed as a JSON"}
- return get_response(content, 400)
- logging.info(f"/webhook: {json}")
- try:
- payload = parse_payload(json)
- except ValidationError:
- content = {"status": "error", "error": "the JSON payload is invalid"}
- return get_response(content, 400)
- except Exception as e:
- logging.exception("Unexpected error", exc_info=e)
- content = {"status": "error", "error": "unexpected error"}
- return get_response(content, 500)
-
- try:
- process_payload(
- init_processing_steps=init_processing_steps,
- payload=payload,
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- hf_timeout_seconds=hf_timeout_seconds,
- )
- except DatasetError:
- content = {"status": "error", "error": "the dataset is not supported"}
- return get_response(content, 400)
- content = {"status": "ok"}
- return get_response(content, 200)
+ with StepProfiler(method="webhook_endpoint", step="all"):
+ with StepProfiler(method="webhook_endpoint", step="get JSON"):
+ try:
+ json = await request.json()
+ except Exception:
+ content = {"status": "error", "error": "the body could not be parsed as a JSON"}
+ return get_response(content, 400)
+ logging.info(f"/webhook: {json}")
+ with StepProfiler(method="webhook_endpoint", step="parse payload"):
+ try:
+ payload = parse_payload(json)
+ except ValidationError:
+ content = {"status": "error", "error": "the JSON payload is invalid"}
+ return get_response(content, 400)
+ except Exception as e:
+ logging.exception("Unexpected error", exc_info=e)
+ content = {"status": "error", "error": "unexpected error"}
+ return get_response(content, 500)
+
+ with StepProfiler(method="webhook_endpoint", step="process payload"):
+ try:
+ process_payload(
+ init_processing_steps=init_processing_steps,
+ payload=payload,
+ hf_endpoint=hf_endpoint,
+ hf_token=hf_token,
+ hf_timeout_seconds=hf_timeout_seconds,
+ )
+ except DatasetError:
+ content = {"status": "error", "error": "the dataset is not supported"}
+ return get_response(content, 400)
+ content = {"status": "ok"}
+ return get_response(content, 200)
diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py
index a3175613..2bb25276 100644
--- a/services/api/tests/test_app.py
+++ b/services/api/tests/test_app.py
@@ -135,0 +136 @@ def test_metrics(client: TestClient) -> None:
+ response = client.get("/healthcheck")
@@ -140,6 +141,14 @@ def test_metrics(client: TestClient) -> None:
- metrics = {line.split(" ")[0]: float(line.split(" ")[1]) for line in lines if line and line[0] != "#"}
-
- # the middleware should have recorded the request
- name = 'starlette_requests_total{method="GET",path_template="/metrics"}'
- assert name in metrics, metrics
- assert metrics[name] > 0, metrics
+ # examples:
+ # starlette_requests_total{method="GET",path_template="/metrics"} 1.0
+ # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05
+ metrics = {
+ parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1))
+ }
+
+ # the metrics should contain at least the following
+ for name in [
+ 'starlette_requests_total{method="GET",path_template="/metrics"}',
+ 'method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"}',
+ ]:
+ assert name in metrics, metrics
+ assert metrics[name] > 0, metrics
diff --git a/services/api/tests/test_prometheus.py b/services/api/tests/test_prometheus.py
index ab549de3..12ed41b5 100644
--- a/services/api/tests/test_prometheus.py
+++ b/services/api/tests/test_prometheus.py
@@ -16 +16,7 @@ def test_prometheus() -> None:
- metrics = {line.split(" ")[0]: float(line.split(" ")[1]) for line in lines if line and line[0] != "#"}
+ # examples:
+ # starlette_requests_total{method="GET",path_template="/metrics"} 1.0
+ # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05
+ metrics = {
+ parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1))
+ }
+
|
|
98182c29720554abb51ec13b6dcdf38926cf2c5a
|
Sylvain Lesage
| 2023-03-08T15:48:05 |
Add timeout to calls to hub (#910)
|
diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl
index c86609ac..40214e51 100644
--- a/chart/templates/services/api/_container.tpl
+++ b/chart/templates/services/api/_container.tpl
@@ -15,2 +15,2 @@
- - name: API_HF_AUTH_TIMEOUT_SECONDS
- value: {{ .Values.api.hfAuthTimeoutSeconds | quote }}
+ - name: API_HF_TIMEOUT_SECONDS
+ value: {{ .Values.api.hfTimeoutSeconds | quote }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 67feca4d..c621c6a2 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -256,2 +256,2 @@ api:
- # the timeout in seconds for the external authentication service.
- hfAuthTimeoutSeconds: "0.2"
+ # the timeout in seconds for the requests to the Hugging Face Hub.
+ hfTimeoutSeconds: "0.2"
diff --git a/e2e/Makefile b/e2e/Makefile
index a2be9a19..5322ef53 100644
--- a/e2e/Makefile
+++ b/e2e/Makefile
@@ -16 +16 @@ export COMMON_LOG_LEVEL := DEBUG
-export API_HF_AUTH_TIMEOUT_SECONDS := 10
+export API_HF_TIMEOUT_SECONDS := 10
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py
index d03d2e39..63ed4fb8 100644
--- a/libs/libcommon/src/libcommon/dataset.py
+++ b/libs/libcommon/src/libcommon/dataset.py
@@ -9 +9 @@ from huggingface_hub.hf_api import DatasetInfo, HfApi
-from huggingface_hub.utils._errors import RepositoryNotFoundError
+from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFoundError
@@ -14,0 +15,2 @@ DatasetErrorCode = Literal[
+ "AskAccessHubRequestError",
+ "DatasetInfoHubRequestError",
@@ -36,0 +39,26 @@ class DatasetError(CustomError):
+class AskAccessHubRequestError(DatasetError):
+ """Raised when the request to the Hub's ask-access endpoint times out."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(
+ message=message,
+ status_code=HTTPStatus.INTERNAL_SERVER_ERROR,
+ code="AskAccessHubRequestError",
+ cause=cause,
+ disclose_cause=False,
+ )
+
+
+class DatasetInfoHubRequestError(DatasetError):
+ """Raised when the request to the Hub's dataset-info endpoint times out."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(
+ message=message,
+ status_code=HTTPStatus.INTERNAL_SERVER_ERROR,
+ code="DatasetInfoHubRequestError",
+ cause=cause,
+ disclose_cause=False,
+ )
+
+
@@ -76 +104,3 @@ class GatedExtraFieldsError(DatasetError):
-def ask_access(dataset: str, hf_endpoint: str, hf_token: Optional[str]) -> None:
+def ask_access(
+ dataset: str, hf_endpoint: str, hf_token: Optional[str], hf_timeout_seconds: Optional[float] = None
+) -> None:
@@ -86,0 +117,2 @@ def ask_access(dataset: str, hf_endpoint: str, hf_token: Optional[str]) -> None:
+ hf_timeout_seconds (`float`, *optional*, defaults to None):
+ The timeout in seconds for the request to the Hub.
@@ -89,0 +122,2 @@ def ask_access(dataset: str, hf_endpoint: str, hf_token: Optional[str]) -> None:
+ - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the
+ dataset failed or timed out.
@@ -100 +134,10 @@ def ask_access(dataset: str, hf_endpoint: str, hf_token: Optional[str]) -> None:
- r = requests.post(path, headers=build_hf_headers(token=hf_token))
+ try:
+ r = requests.post(path, headers=build_hf_headers(token=hf_token), timeout=hf_timeout_seconds)
+ except Exception as err:
+ raise AskAccessHubRequestError(
+ (
+ "Request to the Hub to get access to the dataset failed or timed out. Please try again later, it's a"
+ " temporary internal issue."
+ ),
+ err,
+ ) from err
@@ -120,0 +164 @@ def get_dataset_info_for_supported_datasets(
+ hf_timeout_seconds: Optional[float] = None,
@@ -131,0 +176,2 @@ def get_dataset_info_for_supported_datasets(
+ hf_timeout_seconds (`float`, *optional*, defaults to None):
+ The timeout in seconds for the request to the Hub.
@@ -135,0 +182,4 @@ def get_dataset_info_for_supported_datasets(
+ - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the
+ dataset failed or timed out.
+ - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset
+ info failed or timed out.
@@ -142 +192,2 @@ def get_dataset_info_for_supported_datasets(
- (private datasets are not supported by the datasets server)
+ (private datasets are not supported by the datasets server), or if the default branch cannot
+ be found in the dataset.
@@ -147 +198,3 @@ def get_dataset_info_for_supported_datasets(
- dataset_info = HfApi(endpoint=hf_endpoint).dataset_info(repo_id=dataset, token=hf_token)
+ dataset_info = HfApi(endpoint=hf_endpoint).dataset_info(
+ repo_id=dataset, token=hf_token, timeout=hf_timeout_seconds
+ )
@@ -149 +202,13 @@ def get_dataset_info_for_supported_datasets(
- ask_access(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token)
+ ask_access(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds)
+ except RevisionNotFoundError as err:
+ raise DatasetNotFoundError(
+ f"The default branch cannot be found in dataset {dataset} on the Hub.", cause=err
+ ) from err
+ except Exception as err:
+ raise DatasetInfoHubRequestError(
+ (
+ "Request to the Hub to get the dataset info failed or timed out. Please try again later, it's a"
+ " temporary internal issue."
+ ),
+ err,
+ ) from err
@@ -158,0 +224 @@ def get_dataset_git_revision(
+ hf_timeout_seconds: Optional[float] = None,
@@ -169,0 +236,2 @@ def get_dataset_git_revision(
+ hf_timeout_seconds (`float`, *optional*, defaults to None):
+ The timeout in seconds for the request to the Hub.
@@ -173,0 +242,4 @@ def get_dataset_git_revision(
+ - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the
+ dataset failed or timed out.
+ - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset
+ info failed or timed out.
@@ -184 +256,3 @@ def get_dataset_git_revision(
- return get_dataset_info_for_supported_datasets(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token).sha
+ return get_dataset_info_for_supported_datasets(
+ dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds
+ ).sha
@@ -190,0 +265 @@ def check_support(
+ hf_timeout_seconds: Optional[float] = None,
@@ -201,0 +277,2 @@ def check_support(
+ hf_timeout_seconds (`float`, *optional*, defaults to None):
+ The timeout in seconds for the request to the Hub.
@@ -204,0 +282,4 @@ def check_support(
+ - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the
+ dataset failed or timed out.
+ - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset
+ info failed or timed out.
@@ -214 +295,3 @@ def check_support(
- get_dataset_info_for_supported_datasets(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token)
+ get_dataset_info_for_supported_datasets(
+ dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds
+ )
@@ -218,0 +302 @@ def get_supported_datasets(hf_endpoint: str, hf_token: Optional[str] = None) ->
+ # no timeout on this function. It's used only in the admin service
diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py
index db2f97a4..fd59a273 100644
--- a/libs/libcommon/src/libcommon/operations.py
+++ b/libs/libcommon/src/libcommon/operations.py
@@ -29,0 +30 @@ def update_dataset(
+ hf_timeout_seconds: Optional[float] = None,
@@ -41,0 +43,2 @@ def update_dataset(
+ hf_timeout_seconds (Optional[float], optional): The timeout for requests to the hub. None means no timeout.
+ Defaults to None.
@@ -45,0 +49,4 @@ def update_dataset(
+ - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the
+ dataset failed or timed out.
+ - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset
+ info failed or timed out.
@@ -49 +56,3 @@ def update_dataset(
- check_support(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token)
+ check_support(
+ dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds
+ )
@@ -77,0 +87 @@ def move_dataset(
+ hf_timeout_seconds: Optional[float] = None,
@@ -92,0 +103,2 @@ def move_dataset(
+ hf_timeout_seconds (Optional[float], optional): The timeout for requests to the hub. None means no timeout.
+ Defaults to None.
@@ -96,0 +109,4 @@ def move_dataset(
+ - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the
+ dataset failed or timed out.
+ - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset
+ info failed or timed out.
@@ -106,0 +123 @@ def move_dataset(
+ hf_timeout_seconds=hf_timeout_seconds,
@@ -119,0 +137 @@ def check_in_process(
+ hf_timeout_seconds: Optional[float] = None,
@@ -130,0 +149,3 @@ def check_in_process(
+ hf_timeout_seconds (Optional[float], optional): The timeout for requests to the hub. None means no timeout.
+ Defaults to None.
+
@@ -134,0 +156,4 @@ def check_in_process(
+ - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the
+ dataset failed or timed out.
+ - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset
+ info failed or timed out.
@@ -157,0 +183 @@ def check_in_process(
+ hf_timeout_seconds=hf_timeout_seconds,
@@ -170,0 +197 @@ def check_in_process(
+ hf_timeout_seconds=hf_timeout_seconds,
diff --git a/libs/libcommon/tests/test_dataset.py b/libs/libcommon/tests/test_dataset.py
index 2db05f67..d5900604 100644
--- a/libs/libcommon/tests/test_dataset.py
+++ b/libs/libcommon/tests/test_dataset.py
@@ -6 +6 @@ import pytest
-from libcommon.dataset import check_support
+from libcommon.dataset import DatasetInfoHubRequestError, check_support
@@ -14,0 +15,9 @@ def test_check_support() -> None:
+
+
[email protected]_dataset
+def test_check_support_timeout() -> None:
+ dataset = "glue"
+ hf_endpoint = "https://huggingface.co"
+ hf_token = None
+ with pytest.raises(DatasetInfoHubRequestError):
+ check_support(dataset, hf_endpoint, hf_token, hf_timeout_seconds=0.01)
diff --git a/services/api/README.md b/services/api/README.md
index e5078604..49320e7f 100644
--- a/services/api/README.md
+++ b/services/api/README.md
@@ -14 +14 @@ Set environment variables to configure the application (`API_` prefix):
-- `API_HF_AUTH_TIMEOUT_SECONDS`: the timeout in seconds for the external authentication service. Defaults to `0.2` (200 ms).
+- `API_HF_TIMEOUT_SECONDS`: the timeout in seconds for the requests to the Hugging Face Hub. Defaults to `0.2` (200 ms).
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 20b542a8..1514a2e3 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -66 +66 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- external_auth_timeout_seconds=app_config.api.hf_auth_timeout_seconds,
+ hf_timeout_seconds=app_config.api.hf_timeout_seconds,
@@ -85 +85 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- external_auth_timeout_seconds=app_config.api.hf_auth_timeout_seconds,
+ hf_timeout_seconds=app_config.api.hf_timeout_seconds,
@@ -100,0 +101 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ hf_timeout_seconds=app_config.api.hf_timeout_seconds,
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index bbfd456f..d7303b0c 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -7 +7 @@ import requests
-from requests import PreparedRequest, Timeout
+from requests import PreparedRequest
@@ -11,0 +12 @@ from api.utils import (
+ AuthCheckHubRequestError,
@@ -13 +13,0 @@ from api.utils import (
- ExternalTimeoutError,
@@ -42 +42 @@ def auth_check(
- external_auth_timeout_seconds: Optional[float] = None,
+ hf_timeout_seconds: Optional[float] = None,
@@ -58 +58 @@ def auth_check(
- external_auth_timeout_seconds (float|None): the timeout in seconds for the external authentication service. It
+ hf_timeout_seconds (float|None): the timeout in seconds for the external authentication service. It
@@ -71,5 +71 @@ def auth_check(
- response = requests.get(url, auth=RequestAuth(request), timeout=external_auth_timeout_seconds)
- except Timeout as err:
- raise ExternalTimeoutError(
- "Authentication check timed out. Please try again later, it's a temporary internal issue.", err
- ) from err
+ response = requests.get(url, auth=RequestAuth(request), timeout=hf_timeout_seconds)
@@ -77 +73,7 @@ def auth_check(
- raise RuntimeError("External authentication check failed", err) from err
+ raise AuthCheckHubRequestError(
+ (
+ "Authentication check on the Hugging Face Hub failed or timed out. Please try again later, it's a"
+ " temporary internal issue."
+ ),
+ err,
+ ) from err
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 49ea94fa..8265621e 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -40 +40 @@ API_HF_AUTH_PATH = "/api/datasets/%s/auth-check"
-API_HF_AUTH_TIMEOUT_SECONDS = 0.2
+API_HF_TIMEOUT_SECONDS = 0.2
@@ -49 +49 @@ class ApiConfig:
- hf_auth_timeout_seconds: Optional[float] = API_HF_AUTH_TIMEOUT_SECONDS
+ hf_timeout_seconds: Optional[float] = API_HF_TIMEOUT_SECONDS
@@ -62 +62 @@ class ApiConfig:
- hf_auth_timeout_seconds=env.float(name="HF_AUTH_TIMEOUT_SECONDS", default=API_HF_AUTH_TIMEOUT_SECONDS),
+ hf_timeout_seconds=env.float(name="HF_TIMEOUT_SECONDS", default=API_HF_TIMEOUT_SECONDS),
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 3e993bad..b4046f13 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -205 +205 @@ def create_endpoint(
- external_auth_timeout_seconds: Optional[float] = None,
+ hf_timeout_seconds: Optional[float] = None,
@@ -239 +239 @@ def create_endpoint(
- external_auth_timeout_seconds=external_auth_timeout_seconds,
+ hf_timeout_seconds=hf_timeout_seconds,
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 0faa99d8..2c8a976e 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -68 +68 @@ def create_is_valid_endpoint(
- external_auth_timeout_seconds: Optional[float] = None,
+ hf_timeout_seconds: Optional[float] = None,
@@ -84 +84 @@ def create_is_valid_endpoint(
- external_auth_timeout_seconds=external_auth_timeout_seconds,
+ hf_timeout_seconds=hf_timeout_seconds,
diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py
index c1931b2c..131c3b30 100644
--- a/services/api/src/api/routes/webhook.py
+++ b/services/api/src/api/routes/webhook.py
@@ -61,0 +62 @@ def process_payload(
+ hf_timeout_seconds: Optional[float] = None,
@@ -76,0 +78 @@ def process_payload(
+ hf_timeout_seconds=hf_timeout_seconds,
@@ -91,0 +94 @@ def process_payload(
+ hf_timeout_seconds=hf_timeout_seconds,
@@ -96 +99,4 @@ def create_webhook_endpoint(
- init_processing_steps: List[ProcessingStep], hf_endpoint: str, hf_token: Optional[str] = None
+ init_processing_steps: List[ProcessingStep],
+ hf_endpoint: str,
+ hf_token: Optional[str] = None,
+ hf_timeout_seconds: Optional[float] = None,
@@ -120,0 +127 @@ def create_webhook_endpoint(
+ hf_timeout_seconds=hf_timeout_seconds,
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index 43e43f6c..52a04024 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -20 +20 @@ ApiErrorCode = Literal[
- "ExternalTimeoutError",
+ "AuthCheckHubRequestError",
@@ -85,2 +85,2 @@ class ExternalAuthenticatedError(ApiCustomError):
-class ExternalTimeoutError(ApiCustomError):
- """Raised when the external authentication check failed due to timeout."""
+class AuthCheckHubRequestError(ApiCustomError):
+ """Raised when the external authentication check failed or timed out."""
@@ -90 +90 @@ class ExternalTimeoutError(ApiCustomError):
- message, HTTPStatus.INTERNAL_SERVER_ERROR, "ExternalTimeoutError", cause=cause, disclose_cause=False
+ message, HTTPStatus.INTERNAL_SERVER_ERROR, "AuthCheckHubRequestError", cause=cause, disclose_cause=False
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index 4022bc1a..0784eda0 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -24 +24 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]:
- monkeypatch_session.setenv("API_HF_AUTH_TIMEOUT_SECONDS", "10")
+ monkeypatch_session.setenv("API_HF_TIMEOUT_SECONDS", "10")
diff --git a/services/api/tests/test_authentication.py b/services/api/tests/test_authentication.py
index fcd0d88b..4ffabefc 100644
--- a/services/api/tests/test_authentication.py
+++ b/services/api/tests/test_authentication.py
@@ -15,0 +16 @@ from api.utils import (
+ AuthCheckHubRequestError,
@@ -17 +17,0 @@ from api.utils import (
- ExternalTimeoutError,
@@ -34 +34 @@ def test_unreachable_external_auth_check_service() -> None:
- with pytest.raises(RuntimeError):
+ with pytest.raises(AuthCheckHubRequestError):
@@ -71 +71 @@ def sleeping(_: werkzeug.wrappers.Request) -> werkzeug.wrappers.Response:
- "external_auth_timeout_seconds,expectation",
+ "hf_timeout_seconds,expectation",
@@ -75 +75 @@ def sleeping(_: werkzeug.wrappers.Request) -> werkzeug.wrappers.Response:
- (TIMEOUT_TIME / 2, pytest.raises(ExternalTimeoutError)),
+ (TIMEOUT_TIME / 2, pytest.raises(AuthCheckHubRequestError)),
@@ -78 +78 @@ def sleeping(_: werkzeug.wrappers.Request) -> werkzeug.wrappers.Response:
-def test_external_auth_timeout(
+def test_hf_timeout_seconds(
@@ -82 +82 @@ def test_external_auth_timeout(
- external_auth_timeout_seconds: Optional[float],
+ hf_timeout_seconds: Optional[float],
@@ -89,3 +89 @@ def test_external_auth_timeout(
- auth_check(
- dataset, external_auth_url=external_auth_url, external_auth_timeout_seconds=external_auth_timeout_seconds
- )
+ auth_check(dataset, external_auth_url=external_auth_url, hf_timeout_seconds=hf_timeout_seconds)
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index d73d01a5..455a7021 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -57 +57 @@ services:
- API_HF_AUTH_TIMEOUT_SECONDS: ${API_HF_AUTH_TIMEOUT_SECONDS-0.2}
+ API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-0.2}
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 914c2c27..8d7ded1e 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -57 +57 @@ services:
- API_HF_AUTH_TIMEOUT_SECONDS: ${API_HF_AUTH_TIMEOUT_SECONDS-0.2}
+ API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-0.2}
|
|
d95b9716587ad862a17b1c40937ecea2e1c4a24f
|
Sylvain Lesage
| 2023-03-08T09:14:52 |
feat: 🎸 add timeout when requesting Hub authentication check (#896)
|
diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl
index 196695fc..c86609ac 100644
--- a/chart/templates/services/api/_container.tpl
+++ b/chart/templates/services/api/_container.tpl
@@ -14,0 +15,2 @@
+ - name: API_HF_AUTH_TIMEOUT_SECONDS
+ value: {{ .Values.api.hfAuthTimeoutSeconds | quote }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 32d4a6db..67feca4d 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -255,0 +256,2 @@ api:
+ # the timeout in seconds for the external authentication service.
+ hfAuthTimeoutSeconds: "0.2"
diff --git a/e2e/Makefile b/e2e/Makefile
index 97ec2bee..a2be9a19 100644
--- a/e2e/Makefile
+++ b/e2e/Makefile
@@ -15,0 +16,2 @@ export COMMON_LOG_LEVEL := DEBUG
+export API_HF_AUTH_TIMEOUT_SECONDS := 10
+# ^ allow hub-ci to be slow to answer
diff --git a/services/api/README.md b/services/api/README.md
index 1fe75f77..e5078604 100644
--- a/services/api/README.md
+++ b/services/api/README.md
@@ -13,0 +14 @@ Set environment variables to configure the application (`API_` prefix):
+- `API_HF_AUTH_TIMEOUT_SECONDS`: the timeout in seconds for the external authentication service. Defaults to `0.2` (200 ms).
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index ab7c76d1..20b542a8 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -65,0 +66 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ external_auth_timeout_seconds=app_config.api.hf_auth_timeout_seconds,
@@ -83,0 +85 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ external_auth_timeout_seconds=app_config.api.hf_auth_timeout_seconds,
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index 3c4d81b8..bbfd456f 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -7 +7 @@ import requests
-from requests import PreparedRequest
+from requests import PreparedRequest, Timeout
@@ -11 +11,5 @@ from starlette.requests import Request
-from api.utils import ExternalAuthenticatedError, ExternalUnauthenticatedError
+from api.utils import (
+ ExternalAuthenticatedError,
+ ExternalTimeoutError,
+ ExternalUnauthenticatedError,
+)
@@ -35 +39,4 @@ def auth_check(
- dataset: str, external_auth_url: Optional[str] = None, request: Optional[Request] = None
+ dataset: str,
+ external_auth_url: Optional[str] = None,
+ request: Optional[Request] = None,
+ external_auth_timeout_seconds: Optional[float] = None,
@@ -38,0 +46,4 @@ def auth_check(
+ It sends a request to the Hugging Face API to check if the dataset is authorized for the input request. The request
+ to the Hugging Face API is authenticated with the same authentication headers as the input request. It timeouts
+ after 200ms.
+
@@ -46,0 +58,2 @@ def auth_check(
+ external_auth_timeout_seconds (float|None): the timeout in seconds for the external authentication service. It
+ is used both for the connection timeout and the read timeout. If None, the request never timeouts.
@@ -58 +71,5 @@ def auth_check(
- response = requests.get(url, auth=RequestAuth(request))
+ response = requests.get(url, auth=RequestAuth(request), timeout=external_auth_timeout_seconds)
+ except Timeout as err:
+ raise ExternalTimeoutError(
+ "Authentication check timed out. Please try again later, it's a temporary internal issue.", err
+ ) from err
@@ -60,0 +78 @@ def auth_check(
+
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 9398cde4..49ea94fa 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -39,0 +40 @@ API_HF_AUTH_PATH = "/api/datasets/%s/auth-check"
+API_HF_AUTH_TIMEOUT_SECONDS = 0.2
@@ -47,0 +49 @@ class ApiConfig:
+ hf_auth_timeout_seconds: Optional[float] = API_HF_AUTH_TIMEOUT_SECONDS
@@ -59,0 +62 @@ class ApiConfig:
+ hf_auth_timeout_seconds=env.float(name="HF_AUTH_TIMEOUT_SECONDS", default=API_HF_AUTH_TIMEOUT_SECONDS),
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 84e3e891..3e993bad 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -204,0 +205 @@ def create_endpoint(
+ external_auth_timeout_seconds: Optional[float] = None,
@@ -234 +235,6 @@ def create_endpoint(
- auth_check(dataset, external_auth_url=external_auth_url, request=request)
+ auth_check(
+ dataset,
+ external_auth_url=external_auth_url,
+ request=request,
+ external_auth_timeout_seconds=external_auth_timeout_seconds,
+ )
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index cd3450fe..0faa99d8 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -67,0 +68 @@ def create_is_valid_endpoint(
+ external_auth_timeout_seconds: Optional[float] = None,
@@ -79 +80,6 @@ def create_is_valid_endpoint(
- auth_check(dataset, external_auth_url=external_auth_url, request=request)
+ auth_check(
+ dataset,
+ external_auth_url=external_auth_url,
+ request=request,
+ external_auth_timeout_seconds=external_auth_timeout_seconds,
+ )
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index 4bd7aac8..43e43f6c 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -19,0 +20 @@ ApiErrorCode = Literal[
+ "ExternalTimeoutError",
@@ -83,0 +85,9 @@ class ExternalAuthenticatedError(ApiCustomError):
+class ExternalTimeoutError(ApiCustomError):
+ """Raised when the external authentication check failed due to timeout."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(
+ message, HTTPStatus.INTERNAL_SERVER_ERROR, "ExternalTimeoutError", cause=cause, disclose_cause=False
+ )
+
+
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index df390bbe..4022bc1a 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -23,0 +24 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]:
+ monkeypatch_session.setenv("API_HF_AUTH_TIMEOUT_SECONDS", "10")
diff --git a/services/api/tests/test_authentication.py b/services/api/tests/test_authentication.py
index 64c5c554..fcd0d88b 100644
--- a/services/api/tests/test_authentication.py
+++ b/services/api/tests/test_authentication.py
@@ -3,0 +4 @@
+import time
@@ -5 +6 @@ from contextlib import nullcontext as does_not_raise
-from typing import Any, Mapping
+from typing import Any, Mapping, Optional
@@ -7,0 +9 @@ import pytest
+import werkzeug.wrappers
@@ -13 +15,5 @@ from api.authentication import auth_check
-from api.utils import ExternalAuthenticatedError, ExternalUnauthenticatedError
+from api.utils import (
+ ExternalAuthenticatedError,
+ ExternalTimeoutError,
+ ExternalUnauthenticatedError,
+)
@@ -55,0 +62,32 @@ def test_external_auth_responses_without_request(
+TIMEOUT_TIME = 0.2
+
+
+def sleeping(_: werkzeug.wrappers.Request) -> werkzeug.wrappers.Response:
+ time.sleep(TIMEOUT_TIME)
+ return werkzeug.wrappers.Response(status=200)
+
+
[email protected](
+ "external_auth_timeout_seconds,expectation",
+ [
+ (TIMEOUT_TIME * 2, does_not_raise()),
+ (None, does_not_raise()),
+ (TIMEOUT_TIME / 2, pytest.raises(ExternalTimeoutError)),
+ ],
+)
+def test_external_auth_timeout(
+ httpserver: HTTPServer,
+ hf_endpoint: str,
+ hf_auth_path: str,
+ external_auth_timeout_seconds: Optional[float],
+ expectation: Any,
+) -> None:
+ dataset = "dataset"
+ external_auth_url = hf_endpoint + hf_auth_path
+ httpserver.expect_request(hf_auth_path % dataset).respond_with_handler(func=sleeping)
+ with expectation:
+ auth_check(
+ dataset, external_auth_url=external_auth_url, external_auth_timeout_seconds=external_auth_timeout_seconds
+ )
+
+
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index fec255aa..d73d01a5 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -56,0 +57 @@ services:
+ API_HF_AUTH_TIMEOUT_SECONDS: ${API_HF_AUTH_TIMEOUT_SECONDS-0.2}
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index e65b329f..914c2c27 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -56,0 +57 @@ services:
+ API_HF_AUTH_TIMEOUT_SECONDS: ${API_HF_AUTH_TIMEOUT_SECONDS-0.2}
|
|
1b3c02007caa6ad285de5419a9132af9bb3eba58
|
Andrea Francis Soria Jimenez
| 2023-03-07T15:52:13 |
Renaming duplicated name (#906)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 2f1b18e6..54f1b731 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -306 +306 @@ workers:
- deployName: "split-names-from-streaming"
+ deployName: "split-names-from-dataset-info"
@@ -308 +308 @@ workers:
- workerOnlyJobTypes: "/split-names-from-streaming"
+ workerOnlyJobTypes: "/split-names-from-dataset-info"
|
|
d148931e8462b6e48f0475c3f015d8e7b72df72a
|
Andrea Francis Soria Jimenez
| 2023-03-07T15:37:46 |
Adding deploy for missing JobRunners (#904)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index b253edaf..2f1b18e6 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -216 +216 @@ workers:
- deployName: "split-names"
+ deployName: "split-names-from-streaming"
@@ -218 +218 @@ workers:
- workerOnlyJobTypes: "/split-names"
+ workerOnlyJobTypes: "/split-names-from-streaming"
@@ -304,0 +305,15 @@ workers:
+ -
+ deployName: "split-names-from-streaming"
+ maxJobsPerNamespace: 20
+ workerOnlyJobTypes: "/split-names-from-streaming"
+ nodeSelector:
+ role-datasets-server-worker: "true"
+ replicas: 2
+ resources:
+ requests:
+ cpu: 200m
+ memory: "100Mi"
+ limits:
+ cpu: 2
+ memory: "1Gi"
+ tolerations: []
diff --git a/chart/values.yaml b/chart/values.yaml
index c5da2bdd..32d4a6db 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -315 +315 @@ workers:
- deployName: "split-names"
+ deployName: "split-names-from-streaming"
@@ -319 +319 @@ workers:
- workerOnlyJobTypes: "/split-names"
+ workerOnlyJobTypes: "/split-names-from-streaming"
@@ -402,0 +403,15 @@ workers:
+ -
+ # name of the deployment
+ deployName: "split-names-from-dataset-info"
+ # Maximum number of jobs running at the same time for the same namespace
+ maxJobsPerNamespace: 1
+ # job types that this worker can process
+ workerOnlyJobTypes: "/split-names-from-dataset-info"
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
|
|
727d11d817e08a7ee10a5fd3a3ddaab1495f7b3b
|
Rémy
| 2023-03-07T14:51:03 |
fix: chart dependencies (#905)
|
diff --git a/chart/.gitignore b/chart/.gitignore
deleted file mode 100644
index 1326c841..00000000
--- a/chart/.gitignore
+++ /dev/null
@@ -1,2 +0,0 @@
-# we don't version the Helm dependencies charts
-charts/
diff --git a/chart/Chart.lock b/chart/Chart.lock
index ea38003d..dd641516 100644
--- a/chart/Chart.lock
+++ b/chart/Chart.lock
@@ -4 +4 @@ dependencies:
- version: 0.0.0
+ version: 0.2.1
@@ -8,2 +8,2 @@ dependencies:
-digest: sha256:3be4b001a1d350839f37b7f7b14f39ab5dd221b7bfd1030250d317acc28fb5d6
-generated: "2023-03-06T20:52:19.58506+01:00"
+digest: sha256:d9066105851469168acee5780e48ebde2b116c20b0146798f2921f17a80ba3dc
+generated: "2023-03-07T15:32:05.802355+01:00"
diff --git a/chart/charts/common-0.2.1.tgz b/chart/charts/common-0.2.1.tgz
new file mode 100644
index 00000000..e19a9738
Binary files /dev/null and b/chart/charts/common-0.2.1.tgz differ
diff --git a/chart/charts/mongodb-13.6.4.tgz b/chart/charts/mongodb-13.6.4.tgz
new file mode 100644
index 00000000..386a954c
Binary files /dev/null and b/chart/charts/mongodb-13.6.4.tgz differ
|
|
4291a0688bf36ce2a8ebb960cbb010ca7a5fa911
|
Rémy
| 2023-03-06T20:56:06 |
ci: update dependencies before lint (#895)
|
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml
index e8821b61..d2c0f629 100644
--- a/.github/workflows/cd.yml
+++ b/.github/workflows/cd.yml
@@ -70,0 +71,3 @@ jobs:
+ - name: Update dependencies
+ run: helm dependencies update
+ working-directory: chart
diff --git a/.github/workflows/chart-pr.yml b/.github/workflows/chart-pr.yml
index 08c206fe..6e33e252 100644
--- a/.github/workflows/chart-pr.yml
+++ b/.github/workflows/chart-pr.yml
@@ -16,4 +15,0 @@ jobs:
- - name: Tailscale
- uses: tailscale/github-action@v1
- with:
- authkey: ${{ secrets.TAILSCALE_AUTHKEY }}
@@ -37,0 +34,3 @@ jobs:
+ - name: Update dependencies
+ run: helm dependencies update
+ working-directory: chart
diff --git a/chart/Chart.lock b/chart/Chart.lock
index b3602fa0..ea38003d 100644
--- a/chart/Chart.lock
+++ b/chart/Chart.lock
@@ -3,2 +3,2 @@ dependencies:
- repository: https://registry.internal.huggingface.tech/chartrepo/charts
- version: 0.1.2
+ repository: oci://ghcr.io/huggingface/helm-common
+ version: 0.0.0
@@ -8,2 +8,2 @@ dependencies:
-digest: sha256:7b30ce342604b5ef63973ab9446d5425bbd41720c94f3589a7c912dfdde2fec5
-generated: "2023-03-03T14:45:23.93166+01:00"
+digest: sha256:3be4b001a1d350839f37b7f7b14f39ab5dd221b7bfd1030250d317acc28fb5d6
+generated: "2023-03-06T20:52:19.58506+01:00"
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 8faeacff..06d69c0e 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -34,2 +34,2 @@ dependencies:
- version: 0.1.2
- repository: https://registry.internal.huggingface.tech/chartrepo/charts
+ version: 0.2.1
+ repository: oci://ghcr.io/huggingface/helm-common
|
|
bccbe848454d20e9479cb360e24d6f3f75e92459
|
Sylvain Lesage
| 2023-03-06T13:06:51 |
fix: 🐛 ensure the dataset parameter is provided in endpoints (#893)
|
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 29044b51..84e3e891 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -228,0 +229,4 @@ def create_endpoint(
+ # for now, dataset is always required in the endpoints.
+ if not dataset:
+ raise MissingRequiredParameterError("Parameter 'dataset' is required")
+
@@ -230 +234 @@ def create_endpoint(
- auth_check(dataset, external_auth_url=external_auth_url, request=request) # type: ignore
+ auth_check(dataset, external_auth_url=external_auth_url, request=request)
@@ -234 +238 @@ def create_endpoint(
- processing_steps, dataset, config, split, init_processing_steps, hf_endpoint, hf_token # type: ignore
+ processing_steps, dataset, config, split, init_processing_steps, hf_endpoint, hf_token
|
|
bb2eb334e3c4c4a8985c201b02a2c35e3cb1c2a2
|
Rémy
| 2023-03-06T10:28:02 |
feat: allow to use service nodeport & use common chart (#890)
|
diff --git a/.github/workflows/chart-pr.yml b/.github/workflows/chart-pr.yml
index 606db1a5..08c206fe 100644
--- a/.github/workflows/chart-pr.yml
+++ b/.github/workflows/chart-pr.yml
@@ -15,0 +16,7 @@ jobs:
+ - name: Tailscale
+ uses: tailscale/github-action@v1
+ with:
+ authkey: ${{ secrets.TAILSCALE_AUTHKEY }}
+ - name: Update dependencies
+ run: helm dependencies update
+ working-directory: chart
diff --git a/chart/Chart.lock b/chart/Chart.lock
index 1273ec41..b3602fa0 100644
--- a/chart/Chart.lock
+++ b/chart/Chart.lock
@@ -1,0 +2,3 @@ dependencies:
+- name: common
+ repository: https://registry.internal.huggingface.tech/chartrepo/charts
+ version: 0.1.2
@@ -5,2 +8,2 @@ dependencies:
-digest: sha256:d9d99bcc06040ce2432a014455faf0e7b3b27c20d3a80bbe50b4e9259351e1ac
-generated: "2023-01-20T11:34:50.070944+01:00"
+digest: sha256:7b30ce342604b5ef63973ab9446d5425bbd41720c94f3589a7c912dfdde2fec5
+generated: "2023-03-03T14:45:23.93166+01:00"
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 2064a137..8faeacff 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.4.0
+version: 1.5.0
@@ -30,0 +31 @@ appVersion: "0.21.0"
+icon: https://huggingface.co/front/assets/huggingface_logo-noborder.svg
@@ -31,0 +33,3 @@ dependencies:
+ - name: common
+ version: 0.1.2
+ repository: https://registry.internal.huggingface.tech/chartrepo/charts
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 2163ba4b..b253edaf 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -5,0 +6,9 @@
+global:
+ huggingface:
+ service:
+ ports:
+ datasetsServer:
+ proxy: 30931
+ admin: 32702
+ api: 31370
+
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index 15e997f2..56f838af 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -29,40 +28,0 @@ Docker image management
-{{- define "datasetsServer.images.image" -}}
-{{- $registryName := .imageRoot.registry -}}
-{{- $repositoryName := .imageRoot.repository -}}
-{{- $separator := ":" -}}
-{{- $termination := .imageRoot.tag | toString -}}
-{{- if .global }}
- {{- if and .global.imageRegistry .imageRoot.useGlobalRegistry }}
- {{- $registryName = .global.imageRegistry -}}
- {{- end -}}
-{{- end -}}
-{{- if .imageRoot.digest }}
- {{- $separator = "@" -}}
- {{- $termination = .imageRoot.digest | toString -}}
-{{- end -}}
-{{- printf "%s/%s%s%s" $registryName $repositoryName $separator $termination -}}
-{{- end -}}
-
-{{- define "common.images.pullSecrets" -}}
- {{- $pullSecrets := list }}
-
- {{- if .global }}
- {{- range .global.imagePullSecrets -}}
- {{- $pullSecrets = append $pullSecrets . -}}
- {{- end -}}
- {{- end -}}
-
- {{- range .images -}}
- {{- range .pullSecrets -}}
- {{- $pullSecrets = append $pullSecrets . -}}
- {{- end -}}
- {{- end -}}
-
- {{- if (not (empty $pullSecrets)) }}
-imagePullSecrets:
- {{- range $pullSecrets }}
- - name: {{ . }}
- {{- end }}
- {{- end }}
-{{- end -}}
-
@@ -70 +30 @@ imagePullSecrets:
-{{ include "datasetsServer.images.image" (dict "imageRoot" .Values.images.reverseProxy "global" .Values.global.huggingface) }}
+{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.reverseProxy "global" .Values.global.huggingface) }}
@@ -74 +34 @@ imagePullSecrets:
-{{ include "datasetsServer.images.image" (dict "imageRoot" .Values.images.jobs.mongodbMigration "global" .Values.global.huggingface) }}
+{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.jobs.mongodbMigration "global" .Values.global.huggingface) }}
@@ -78 +38 @@ imagePullSecrets:
-{{ include "datasetsServer.images.image" (dict "imageRoot" .Values.images.services.admin "global" .Values.global.huggingface) }}
+{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.services.admin "global" .Values.global.huggingface) }}
@@ -82 +42 @@ imagePullSecrets:
-{{ include "datasetsServer.images.image" (dict "imageRoot" .Values.images.services.api "global" .Values.global.huggingface) }}
+{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.services.api "global" .Values.global.huggingface) }}
@@ -86 +46 @@ imagePullSecrets:
-{{ include "datasetsServer.images.image" (dict "imageRoot" .Values.images.services.worker "global" .Values.global.huggingface) }}
+{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.services.worker "global" .Values.global.huggingface) }}
@@ -90 +50 @@ imagePullSecrets:
-{{ include "common.images.pullSecrets" (dict "images" (list .Values.images) "global" .Values.global.huggingface) }}
+{{- include "hf.common.images.renderPullSecrets" (dict "images" (list .Values.images) "context" $) -}}
@@ -97,7 +56,0 @@ Common labels
-{{- define "datasetServer.labels" -}}
-app.kubernetes.io/name: {{ include "name" . }}
-helm.sh/chart: {{ .Chart.Name }}
-app.kubernetes.io/instance: {{ .Release.Name }}
-app.kubernetes.io/managed-by: {{ .Release.Service }}
-{{- end }}
-
@@ -105 +58 @@ app.kubernetes.io/managed-by: {{ .Release.Service }}
-{{ include "datasetServer.labels" . }}
+{{ include "hf.labels.commons" . }}
@@ -110 +63 @@ app.kubernetes.io/component: "{{ include "name" . }}-reverse-proxy"
-{{ include "datasetServer.labels" . }}
+{{ include "hf.labels.commons" . }}
@@ -115 +68 @@ app.kubernetes.io/component: "{{ include "name" . }}-storage-admin"
-{{ include "datasetServer.labels" . }}
+{{ include "hf.labels.commons" . }}
@@ -120 +73 @@ app.kubernetes.io/component: "{{ include "name" . }}-mongodb-migration"
-{{ include "datasetServer.labels" . }}
+{{ include "hf.labels.commons" . }}
@@ -125 +78 @@ app.kubernetes.io/component: "{{ include "name" . }}-admin"
-{{ include "datasetServer.labels" . }}
+{{ include "hf.labels.commons" . }}
@@ -130 +83 @@ app.kubernetes.io/component: "{{ include "name" . }}-api"
-{{ include "datasetServer.labels" . }}
+{{ include "hf.labels.commons" . }}
@@ -208,13 +160,0 @@ See https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#a-a
-{{/*
-Return true if cert-manager required annotations for TLS signed
-certificates are set in the Ingress annotations
-Ref: https://cert-manager.io/docs/usage/ingress/#supported-annotations
-Usage:
-{{ include "common.ingress.certManagerRequest" ( dict "annotations" .Values.path.to.the.ingress.annotations ) }}
-*/}}
-{{- define "common.ingress.certManagerRequest" -}}
-{{ if or (hasKey .annotations "cert-manager.io/cluster-issuer") (hasKey .annotations "cert-manager.io/issuer") (hasKey .annotations "kubernetes.io/tls-acme") }}
- {{- true -}}
-{{- end -}}
-{{- end -}}
-
diff --git a/chart/templates/ingress.yaml b/chart/templates/ingress.yaml
index c484c005..69b4c2ad 100644
--- a/chart/templates/ingress.yaml
+++ b/chart/templates/ingress.yaml
@@ -0,0 +1 @@
+{{- if and .Values.global.huggingface.ingress.enabled .Values.ingress.enabled -}}
@@ -21 +22 @@ spec:
- {{- if include "common.ingress.certManagerRequest" ( dict "annotations" $annotations ) }}
+ {{- if include "hf.common.ingress.certManagerRequest" ( dict "annotations" $annotations ) }}
@@ -31,0 +33 @@ spec:
+{{- end }}
diff --git a/chart/templates/jobs/mongodb-migration/job.yaml b/chart/templates/jobs/mongodb-migration/job.yaml
index a5dc5343..84a18ef5 100644
--- a/chart/templates/jobs/mongodb-migration/job.yaml
+++ b/chart/templates/jobs/mongodb-migration/job.yaml
@@ -26 +25,0 @@ spec:
- volumes: {{ include "volumeData" . | nindent 8 }}
diff --git a/chart/templates/pvc.yaml b/chart/templates/pvc.yaml
index f4370320..a4302245 100644
--- a/chart/templates/pvc.yaml
+++ b/chart/templates/pvc.yaml
@@ -5 +5 @@ metadata:
- labels: {{ include "datasetServer.labels" . | nindent 4 }}
+ labels: {{ include "hf.labels.commons" . | nindent 4 }}
diff --git a/chart/templates/reverse-proxy/service.yaml b/chart/templates/reverse-proxy/service.yaml
index d8449c67..66c5a63b 100644
--- a/chart/templates/reverse-proxy/service.yaml
+++ b/chart/templates/reverse-proxy/service.yaml
@@ -3,0 +4 @@
+{{ $serviceType := .Values.reverseProxy.service.type | default .Values.global.huggingface.service.type }}
@@ -15,0 +17,3 @@ spec:
+ {{- if eq "NodePort" $serviceType }}
+ nodePort: {{ .Values.global.huggingface.service.ports.datasetsServer.proxy }}
+ {{- end }}
@@ -18 +22 @@ spec:
- type: {{ .Values.reverseProxy.service.type }}
+ type: {{ $serviceType }}
diff --git a/chart/templates/services/admin/service.yaml b/chart/templates/services/admin/service.yaml
index 8541ed4d..8b31482e 100644
--- a/chart/templates/services/admin/service.yaml
+++ b/chart/templates/services/admin/service.yaml
@@ -3,0 +4 @@
+{{ $serviceType := .Values.reverseProxy.service.type | default .Values.global.huggingface.service.type }}
@@ -15,0 +17,3 @@ spec:
+ {{- if eq "NodePort" $serviceType }}
+ nodePort: {{ .Values.global.huggingface.service.ports.datasetsServer.admin }}
+ {{- end }}
@@ -18 +22 @@ spec:
- type: {{ .Values.admin.service.type }}
+ type: {{ $serviceType }}
diff --git a/chart/templates/services/api/service.yaml b/chart/templates/services/api/service.yaml
index fde05c46..98ed995b 100644
--- a/chart/templates/services/api/service.yaml
+++ b/chart/templates/services/api/service.yaml
@@ -3,0 +4 @@
+{{ $serviceType := .Values.reverseProxy.service.type | default .Values.global.huggingface.service.type }}
@@ -15,0 +17,3 @@ spec:
+ {{- if eq "NodePort" $serviceType }}
+ nodePort: {{ .Values.global.huggingface.service.ports.datasetsServer.api }}
+ {{- end }}
@@ -18 +22 @@ spec:
- type: {{ .Values.api.service.type }}
+ type: {{ $serviceType }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 34441a69..c5da2bdd 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -7,0 +8 @@ global:
+ enabled: true
@@ -8,0 +10 @@ global:
+ ssl: true
@@ -11 +13,7 @@ global:
- ssl: true
+ service:
+ type: ClusterIP
+ ports:
+ datasetsServer:
+ proxy: 30020
+ admin: 30021
+ api: 30022
@@ -178 +186 @@ storageAdmin:
- type: ClusterIP
+ type: ""
@@ -200 +208 @@ reverseProxy:
- type: ClusterIP
+ type: ""
@@ -204,0 +213 @@ ingress:
+ enabled: true
@@ -239 +248 @@ admin:
- type: ClusterIP
+ type: ""
@@ -269 +278 @@ api:
- type: ClusterIP
+ type: ""
|
|
0539fbd1950bfbe89605ddd40db1aaebfc757438
|
Sylvain Lesage
| 2023-03-03T09:42:40 |
feat: 🎸 add an index to the jobs collection (#888)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 79a57eed..b50bd11b 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -137,0 +138 @@ class Job(Document):
+ ("priority", "status", "created_at", "namespace", "unicity_id"),
|
|
3f74aeb3c8df7e7dfee6db798318c79c666b172a
|
Sylvain Lesage
| 2023-03-02T14:24:32 |
feat: 🎸 add an index (#883)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 0243d36c..79a57eed 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -137,0 +138 @@ class Job(Document):
+ ("priority", "status", "created_at", "type", "namespace"),
|
|
f13ec97bae4ac886c21402fc6135c7377dd6c56d
|
Sylvain Lesage
| 2023-03-02T14:04:15 |
feat: 🎸 reduce the requests (and limits) in k8s resources (#886)
|
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index a3cd9aef..57ff300e 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -92,2 +92,2 @@ storageAdmin:
- cpu: 100m
- memory: "256Mi"
+ cpu: 50m
+ memory: "64Mi"
@@ -105 +105 @@ reverseProxy:
- memory: "256Mi"
+ memory: "64Mi"
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 36455b12..2163ba4b 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -91,2 +91,2 @@ storageAdmin:
- cpu: 1
- memory: "256Mi"
+ cpu: 50m
+ memory: "64Mi"
@@ -106 +106 @@ reverseProxy:
- memory: "256Mi"
+ memory: "64Mi"
@@ -150 +150 @@ admin:
- cpu: 4
+ cpu: 1
@@ -169 +169 @@ api:
- cpu: 4
+ cpu: 1
@@ -186 +186 @@ workers:
- memory: "1Gi"
+ memory: "2Gi"
@@ -189 +189 @@ workers:
- memory: "30Gi"
+ memory: "8Gi"
@@ -201 +201 @@ workers:
- memory: "1Gi"
+ memory: "2Gi"
@@ -204 +204 @@ workers:
- memory: "30Gi"
+ memory: "8Gi"
@@ -216 +216 @@ workers:
- memory: "8Gi"
+ memory: "2Gi"
@@ -219 +219 @@ workers:
- memory: "30Gi"
+ memory: "8Gi"
@@ -231 +231 @@ workers:
- memory: "8Gi"
+ memory: "2Gi"
@@ -234 +234 @@ workers:
- memory: "30Gi"
+ memory: "8Gi"
@@ -246 +246 @@ workers:
- memory: "8Gi"
+ memory: "2Gi"
@@ -249 +249 @@ workers:
- memory: "30Gi"
+ memory: "8Gi"
@@ -261 +261 @@ workers:
- memory: "8Gi"
+ memory: "2Gi"
@@ -264 +264 @@ workers:
- memory: "30Gi"
+ memory: "8Gi"
@@ -275 +275 @@ workers:
- cpu: 1
+ cpu: 200m
@@ -290 +290 @@ workers:
- cpu: 1
+ cpu: 200m
@@ -305 +305 @@ workers:
- cpu: 1
+ cpu: 200m
|
|
f44ee34c07ba9009b847bfb9e16919d46b94a15c
|
Sylvain Lesage
| 2023-03-02T12:44:54 |
Publish /healthcheck again (#885)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 21fe769a..2064a137 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.3.1
+version: 1.4.0
diff --git a/chart/nginx-templates/default.conf.template b/chart/nginx-templates/default.conf.template
index 04858d50..3c8d9d89 100644
--- a/chart/nginx-templates/default.conf.template
+++ b/chart/nginx-templates/default.conf.template
@@ -61 +61 @@ server {
- location ~* ^(/admin)?(/healthcheck|/metrics)$ {
+ location ~* ^(/admin)?/metrics$ {
diff --git a/e2e/tests/test_10_healthcheck.py b/e2e/tests/test_10_healthcheck.py
index 42d59cb4..39a52a0a 100644
--- a/e2e/tests/test_10_healthcheck.py
+++ b/e2e/tests/test_10_healthcheck.py
@@ -9 +9 @@ from .utils import poll
[email protected]("endpoint", ["/", "/healthcheck", "/metrics"])
[email protected]("endpoint", ["/healthcheck", "/admin/healthcheck"])
@@ -11 +11,9 @@ def test_healthcheck(endpoint: str) -> None:
- # this tests ensures the /healthcheck and the /metrics endpoints are hidden
+ # this tests ensures the /healthcheck are accessible
+ response = poll(endpoint, expected_code=200)
+ assert response.status_code == 200, f"{response.status_code} - {response.text}"
+ assert "ok" in response.text, response.text
+
+
[email protected]("endpoint", ["/", "/metrics", "/admin/metrics"])
+def test_hidden(endpoint: str) -> None:
+ # this tests ensures the root / and the /metrics endpoints are hidden
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 5edb2f6b..ab7c76d1 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -4,2 +3,0 @@
-from typing import List
-
@@ -14 +12 @@ from starlette.middleware.gzip import GZipMiddleware
-from starlette.routing import BaseRoute, Route
+from starlette.routing import Route
@@ -58 +56,16 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- valid: List[BaseRoute] = [
+ routes = [
+ Route(
+ endpoint_name,
+ endpoint=create_endpoint(
+ endpoint_name=endpoint_name,
+ steps_by_input_type=steps_by_input_type,
+ init_processing_steps=init_processing_steps,
+ hf_endpoint=app_config.common.hf_endpoint,
+ hf_token=app_config.common.hf_token,
+ external_auth_url=app_config.api.external_auth_url,
+ max_age_long=app_config.api.max_age_long,
+ max_age_short=app_config.api.max_age_short,
+ ),
+ )
+ for endpoint_name, steps_by_input_type in endpoints_definition.steps_by_input_type_and_endpoint.items()
+ ] + [
@@ -75 +88 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- )
+ ),
@@ -77,19 +90,3 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- ]
- endpoints: List[BaseRoute] = [
- Route(
- endpoint_name,
- endpoint=create_endpoint(
- endpoint_name=endpoint_name,
- steps_by_input_type=steps_by_input_type,
- init_processing_steps=init_processing_steps,
- hf_endpoint=app_config.common.hf_endpoint,
- hf_token=app_config.common.hf_token,
- external_auth_url=app_config.api.external_auth_url,
- max_age_long=app_config.api.max_age_long,
- max_age_short=app_config.api.max_age_short,
- ),
- )
- for endpoint_name, steps_by_input_type in endpoints_definition.steps_by_input_type_and_endpoint.items()
- ]
- to_protect: List[BaseRoute] = [
- # called by the Hub webhooks
+ Route("/healthcheck", endpoint=healthcheck_endpoint),
+ Route("/metrics", endpoint=prometheus.endpoint),
+ # ^ called by Prometheus
@@ -104,0 +102 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ # ^ called by the Hub webhooks
@@ -106,6 +103,0 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- protected: List[BaseRoute] = [
- Route("/healthcheck", endpoint=healthcheck_endpoint),
- # called by Prometheus
- Route("/metrics", endpoint=prometheus.endpoint),
- ]
- routes: List[BaseRoute] = valid + endpoints + to_protect + protected
|
|
fb48ed36cd41b13fba88b6b18072faf1bbd1fb62
|
Sylvain Lesage
| 2023-03-02T11:15:17 |
chore: 🤖 use other nodes for the workers (#884)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 3efdfed3..36455b12 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -181 +181 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -196 +196 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -211 +211 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -226 +226 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -241 +241 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -256 +256 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -271 +271 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -286 +286 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
@@ -301 +301 @@ workers:
- role-datasets-server: "true"
+ role-datasets-server-worker: "true"
|
|
405f64c12460df3dbb5bd5918aa6ca1218263a02
|
Andrea Francis Soria Jimenez
| 2023-03-01T17:47:31 |
Increase version of workers (#879)
|
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
index 1705c64e..2eb8c100 100644
--- a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
+++ b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
@@ -118 +118 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner):
- return "1.0.0"
+ return "2.0.0"
diff --git a/services/worker/src/worker/job_runners/split_names_from_streaming.py b/services/worker/src/worker/job_runners/split_names_from_streaming.py
index 879db6b7..6fc15d28 100644
--- a/services/worker/src/worker/job_runners/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split_names_from_streaming.py
@@ -122 +122 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner):
- return "1.0.0"
+ return "2.0.0"
|
|
23945ca8189026f17a8cf49164f1b6de0f38c5cf
|
Andrea Francis Soria Jimenez
| 2023-03-01T16:28:41 |
Renaming split-names response to splits (#878)
|
diff --git a/e2e/tests/test_12_splits.py b/e2e/tests/test_12_splits.py
index 20d8715d..26017ae7 100644
--- a/e2e/tests/test_12_splits.py
+++ b/e2e/tests/test_12_splits.py
@@ -78 +78 @@ def test_splits_with_config_using_openapi(status: int, dataset: str, config: str
- assert all(split["config"] == config for split in r_splits.json()["split_names"])
+ assert all(split["config"] == config for split in r_splits.json()["splits"])
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
index 2147358a..1705c64e 100644
--- a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
+++ b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
@@ -57 +57 @@ class SplitNamesFromDatasetInfoResponseContent(TypedDict):
- split_names: List[SplitNameItem]
+ splits: List[SplitNameItem]
@@ -108 +108 @@ def compute_split_names_from_dataset_info_response(
- return {"split_names": split_name_items}
+ return {"splits": split_name_items}
@@ -129,3 +129 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner):
- return {
- SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["split_names"]
- }
+ return {SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["splits"]}
diff --git a/services/worker/src/worker/job_runners/split_names_from_streaming.py b/services/worker/src/worker/job_runners/split_names_from_streaming.py
index 309c2422..879db6b7 100644
--- a/services/worker/src/worker/job_runners/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split_names_from_streaming.py
@@ -58 +58 @@ class SplitNamesFromStreamingResponseContent(TypedDict):
- split_names: List[SplitNameItem]
+ splits: List[SplitNameItem]
@@ -112 +112 @@ def compute_split_names_from_streaming_response(
- return {"split_names": split_name_items}
+ return {"splits": split_name_items}
@@ -133,3 +133 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner):
- return {
- SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["split_names"]
- }
+ return {SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["splits"]}
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index afef7d98..20fee544 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -247 +246,0 @@ class HubDatasetTest(TypedDict):
- split_names_response: Any
@@ -268,13 +266,0 @@ def create_config_names_response(dataset: str) -> Any:
-def create_split_names_response(dataset: str) -> Any:
- dataset, config, split = get_default_config_split(dataset)
- return {
- "split_names": [
- {
- "dataset": dataset,
- "config": config,
- "split": split,
- }
- ]
- }
-
-
@@ -526 +511,0 @@ def hub_datasets(
- "split_names_response": None,
@@ -534 +518,0 @@ def hub_datasets(
- "split_names_response": None,
@@ -542 +525,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_public_csv),
@@ -552 +534,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_private_csv),
@@ -562 +543,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_gated_csv),
@@ -572 +552,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_public_jsonl),
@@ -580 +559,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_gated_extra_fields_csv),
@@ -590 +568,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_public_audio),
@@ -602 +579,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_public_image),
@@ -612 +588,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_public_images_list),
@@ -622 +597,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_public_big),
@@ -630 +604,0 @@ def hub_datasets(
- "split_names_response": create_split_names_response(hub_public_external_files),
diff --git a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
index 2870218c..1f5869a3 100644
--- a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
+++ b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
@@ -83 +83 @@ def get_job_runner(
- "split_names": [
+ "splits": [
diff --git a/services/worker/tests/job_runners/test_split_names_from_streaming.py b/services/worker/tests/job_runners/test_split_names_from_streaming.py
index 04493ba3..19418d68 100644
--- a/services/worker/tests/job_runners/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/test_split_names_from_streaming.py
@@ -75 +75 @@ def test_process(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public
- assert len(content["split_names"]) == 1
+ assert len(content["splits"]) == 1
@@ -102 +102 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No
-def test_compute_split_names_response(
+def test_compute_split_names_from_streaming_response(
@@ -112 +112 @@ def test_compute_split_names_response(
- expected_configs_response = hub_datasets[name]["split_names_response"]
+ expected_configs_response = hub_datasets[name]["splits_response"]
|
|
b61a1ce1426b665cea4604abcce3fe3daa3b128a
|
Andrea Francis Soria Jimenez
| 2023-03-01T14:37:43 |
Endpoint respond by input type (#834)
|
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json
index 356e24a8..e21190e1 100644
--- a/chart/static-files/openapi.json
+++ b/chart/static-files/openapi.json
@@ -1874 +1874 @@
- "error": "Parameter 'dataset' is required"
+ "error": "Parameters 'split', 'config' and 'dataset' are required"
@@ -1880 +1880 @@
- "error": "Parameters 'config' and 'split' are required"
+ "error": "Parameters 'split', 'config' and 'dataset' are required"
@@ -1886 +1886 @@
- "error": "Parameters 'config' and 'split' are required"
+ "error": "Parameters 'split', 'config' and 'dataset' are required"
@@ -1892 +1892 @@
- "error": "Parameter 'dataset' is required"
+ "error": "Parameters 'split', 'config' and 'dataset' are required"
@@ -1898 +1898 @@
- "error": "Parameters 'config' and 'split' are required"
+ "error": "Parameters 'split', 'config' and 'dataset' are required"
@@ -1904 +1904 @@
- "error": "Parameters 'config' and 'split' are required"
+ "error": "Parameters 'split', 'config' and 'dataset' are required"
diff --git a/e2e/tests/test_11_auth.py b/e2e/tests/test_11_auth.py
index f4e0df3b..8d38fab2 100644
--- a/e2e/tests/test_11_auth.py
+++ b/e2e/tests/test_11_auth.py
@@ -43,2 +42,0 @@ def test_auth_e2e(
- f"/split-names-from-streaming?dataset={dataset}&config={config}",
- # TODO: ^ this will be changed to /splits in an uncoming PR
diff --git a/e2e/tests/test_12_splits.py b/e2e/tests/test_12_splits.py
index 97d7c476..20d8715d 100644
--- a/e2e/tests/test_12_splits.py
+++ b/e2e/tests/test_12_splits.py
@@ -57,0 +58,26 @@ def test_splits_using_openapi(status: int, name: str, dataset: str, error_code:
+
+
[email protected](
+ "status,dataset,config,error_code",
+ [
+ # (200, "duorc", "SelfRC", None),
+ (401, "missing-parameter", None, "ExternalUnauthenticatedError")
+ # missing config will result in asking dataset but it does not exist
+ ],
+)
+def test_splits_with_config_using_openapi(status: int, dataset: str, config: str, error_code: str) -> None:
+ r_splits = (
+ poll(f"/splits?dataset={dataset}&config=", error_field="error")
+ if error_code
+ else poll(f"/splits?dataset={dataset}&config={config}")
+ )
+
+ assert r_splits.status_code == status, f"{r_splits.status_code} - {r_splits.text}"
+
+ if error_code is None:
+ assert all(split["config"] == config for split in r_splits.json()["split_names"])
+ # all splits must belong to the provided config
+
+ assert "X-Error-Code" not in r_splits.headers, r_splits.headers["X-Error-Code"]
+ else:
+ assert r_splits.headers["X-Error-Code"] == error_code, r_splits.headers["X-Error-Code"]
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index bb67a1ca..5edb2f6b 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -26,0 +27,2 @@ def create_app() -> Starlette:
+ endpoint_config = EndpointConfig.from_env()
+ return create_app_with_config(app_config=app_config, endpoint_config=endpoint_config)
@@ -27,0 +30,2 @@ def create_app() -> Starlette:
+
+def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfig) -> Starlette:
@@ -34 +38 @@ def create_app() -> Starlette:
- endpoints_definition = EndpointsDefinition(processing_graph, EndpointConfig.from_env())
+ endpoints_definition = EndpointsDefinition(processing_graph, endpoint_config)
@@ -78 +82,2 @@ def create_app() -> Starlette:
- processing_steps=processing_steps,
+ endpoint_name=endpoint_name,
+ steps_by_input_type=steps_by_input_type,
@@ -87 +92 @@ def create_app() -> Starlette:
- for endpoint_name, processing_steps in endpoints_definition.definition.items()
+ for endpoint_name, steps_by_input_type in endpoints_definition.steps_by_input_type_and_endpoint.items()
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 39af9f83..9398cde4 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -13,0 +14 @@ from libcommon.config import (
+from libcommon.processing_graph import InputType
@@ -84 +85,3 @@ class AppConfig:
-EndpointProcessingStepNamesMapping = Mapping[str, List[str]]
+StepNamesByInputType = Mapping[InputType, List[str]]
+
+StepNamesByInputTypeAndEndpoint = Mapping[str, StepNamesByInputType]
@@ -89 +92,9 @@ class EndpointConfig:
- specification: EndpointProcessingStepNamesMapping = field(
+ """Contains the endpoint config specification to relate with step names.
+ The list of processing steps corresponds to the priority in which the response
+ has to be reached. The cache from the first step in the list will be used first
+ then, if it's an error or missing, the second one, etc.
+ The related steps depend on the query parameters passed in the request
+ (dataset, config, split)
+ """
+
+ step_names_by_input_type_and_endpoint: StepNamesByInputTypeAndEndpoint = field(
@@ -91,8 +102,10 @@ class EndpointConfig:
- "/config-names": ["/config-names"],
- "/split-names-from-streaming": ["/split-names-from-streaming"],
- "/splits": ["/splits"],
- "/first-rows": ["/first-rows"],
- "/parquet-and-dataset-info": ["/parquet-and-dataset-info"],
- "/parquet": ["/parquet"],
- "/dataset-info": ["/dataset-info"],
- "/sizes": ["/sizes"],
+ "/config-names": {"dataset": ["/config-names"]},
+ "/splits": {
+ "dataset": ["/splits"],
+ "config": ["/split-names-from-streaming", "/split-names-from-dataset-info"],
+ },
+ "/first-rows": {"split": ["/first-rows"]},
+ "/parquet-and-dataset-info": {"dataset": ["/parquet-and-dataset-info"]},
+ "/parquet": {"dataset": ["/parquet"]},
+ "/dataset-info": {"dataset": ["/dataset-info"]},
+ "/sizes": {"dataset": ["/sizes"]},
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 9936dc62..29044b51 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -4,0 +5 @@ import logging
+from abc import ABC, abstractmethod
@@ -6 +7 @@ from http import HTTPStatus
-from typing import List, Mapping, Optional
+from typing import List, Mapping, Optional, Tuple
@@ -10,2 +11,2 @@ from libcommon.operations import PreviousStepError, check_in_process
-from libcommon.processing_graph import ProcessingGraph, ProcessingStep
-from libcommon.simple_cache import DoesNotExist, get_response
+from libcommon.processing_graph import InputType, ProcessingGraph, ProcessingStep
+from libcommon.simple_cache import CacheEntry, DoesNotExist, get_response
@@ -20 +20,0 @@ from api.utils import (
- MissingProcessingStepsError,
@@ -30,0 +31,4 @@ from api.utils import (
+StepsByInputType = Mapping[InputType, List[ProcessingStep]]
+
+StepsByInputTypeAndEndpoint = Mapping[str, StepsByInputType]
+
@@ -35 +39 @@ class EndpointsDefinition:
- processing_steps_by_endpoint: Mapping[str, List[ProcessingStep]]
+ steps_by_input_type_and_endpoint: StepsByInputTypeAndEndpoint
@@ -38,3 +42,6 @@ class EndpointsDefinition:
- self.definition = {
- endpoint: [graph.get_step(step) for step in processing_steps]
- for endpoint, processing_steps in endpoint_config.specification.items()
+ self.steps_by_input_type_and_endpoint = {
+ endpoint: {
+ input_type: [graph.get_step(step_name) for step_name in step_names]
+ for input_type, step_names in step_names_by_input_type.items()
+ }
+ for endpoint, step_names_by_input_type in endpoint_config.step_names_by_input_type_and_endpoint.items()
@@ -44 +51 @@ class EndpointsDefinition:
-def create_endpoint(
+def get_cache_entry_from_steps(
@@ -45,0 +53,148 @@ def create_endpoint(
+ dataset: str,
+ config: Optional[str],
+ split: Optional[str],
+ init_processing_steps: List[ProcessingStep],
+ hf_endpoint: str,
+ hf_token: Optional[str] = None,
+) -> CacheEntry:
+ """Gets the cache from the first successful step in the processing steps list.
+ If no successful result is found, it will return the last one even if it's an error,
+ Checks if job is still in progress by each processing step in case of no entry found.
+ Raises:
+ ResponseNotReadyError: if no result is found.
+
+ Returns: the cached record
+ """
+ last_result = None
+ for processing_step in processing_steps:
+ try:
+ last_result = get_response(
+ kind=processing_step.cache_kind,
+ dataset=dataset,
+ config=config,
+ split=split,
+ )
+
+ if last_result["http_status"] == HTTPStatus.OK:
+ return last_result
+ except DoesNotExist:
+ logging.debug(
+ f"processing_step={processing_step.name} dataset={dataset} "
+ f"config={config} split={split} no entry found"
+ )
+ try:
+ check_in_process(
+ processing_step=processing_step,
+ init_processing_steps=init_processing_steps,
+ dataset=dataset,
+ config=config,
+ split=split,
+ hf_endpoint=hf_endpoint,
+ hf_token=hf_token,
+ )
+ except (PreviousStepError, DatasetError):
+ raise ResponseNotFoundError("Not found.")
+ if last_result:
+ return last_result
+
+ raise ResponseNotReadyError(
+ "The server is busier than usual and the response is not ready yet. Please retry later."
+ )
+
+
+class InputTypeValidator(ABC):
+ input_type: InputType = NotImplemented
+
+ @abstractmethod
+ def are_parameters_sufficient(self, dataset: Optional[str], config: Optional[str], split: Optional[str]) -> bool:
+ pass
+
+ @abstractmethod
+ def get_error_message(self) -> str:
+ pass
+
+ @abstractmethod
+ def get_useful_parameters(
+ self, dataset: Optional[str], config: Optional[str], split: Optional[str]
+ ) -> Tuple[Optional[str], Optional[str], Optional[str]]:
+ pass
+
+ @staticmethod
+ def from_input_type(input_type: InputType) -> "InputTypeValidator":
+ return (
+ DatasetInputTypeValidator()
+ if input_type == "dataset"
+ else ConfigInputTypeValidator()
+ if input_type == "config"
+ else SplitInputTypeValidator()
+ )
+
+
+class DatasetInputTypeValidator(InputTypeValidator):
+ input_type: InputType = "dataset"
+
+ def are_parameters_sufficient(self, dataset: Optional[str], config: Optional[str], split: Optional[str]) -> bool:
+ return are_valid_parameters([dataset])
+
+ def get_error_message(self) -> str:
+ return "Parameter 'dataset' is required"
+
+ def get_useful_parameters(
+ self, dataset: Optional[str], config: Optional[str], split: Optional[str]
+ ) -> Tuple[Optional[str], Optional[str], Optional[str]]:
+ return (dataset, None, None)
+
+
+class ConfigInputTypeValidator(InputTypeValidator):
+ input_type: InputType = "config"
+
+ def are_parameters_sufficient(self, dataset: Optional[str], config: Optional[str], split: Optional[str]) -> bool:
+ return are_valid_parameters([dataset, config])
+
+ def get_error_message(self) -> str:
+ return "Parameters 'config' and 'dataset' are required"
+
+ def get_useful_parameters(
+ self, dataset: Optional[str], config: Optional[str], split: Optional[str]
+ ) -> Tuple[Optional[str], Optional[str], Optional[str]]:
+ return (dataset, config, None)
+
+
+class SplitInputTypeValidator(InputTypeValidator):
+ input_type: InputType = "split"
+
+ def are_parameters_sufficient(self, dataset: Optional[str], config: Optional[str], split: Optional[str]) -> bool:
+ return are_valid_parameters([dataset, config, split])
+
+ def get_error_message(self) -> str:
+ return "Parameters 'split', 'config' and 'dataset' are required"
+
+ def get_useful_parameters(
+ self, dataset: Optional[str], config: Optional[str], split: Optional[str]
+ ) -> Tuple[Optional[str], Optional[str], Optional[str]]:
+ return (dataset, config, split)
+
+
+def get_input_type_validators_by_priority(steps_by_input_type: StepsByInputType) -> List[InputTypeValidator]:
+ input_type_order: List[InputType] = ["split", "config", "dataset"]
+ return [
+ InputTypeValidator.from_input_type(input_type)
+ for input_type in input_type_order
+ if input_type in steps_by_input_type
+ ]
+
+
+def get_input_type_validator_by_parameters(
+ validators: List[InputTypeValidator], dataset: Optional[str], config: Optional[str], split: Optional[str]
+) -> InputTypeValidator:
+ error_message = "No processing steps supported for parameters"
+ for validator in validators:
+ error_message = validator.get_error_message()
+ if validator.are_parameters_sufficient(dataset=dataset, config=config, split=split):
+ return validator
+ raise MissingRequiredParameterError(error_message)
+
+
+def create_endpoint(
+ endpoint_name: str,
+ steps_by_input_type: StepsByInputType,
@@ -54,5 +208,0 @@ def create_endpoint(
- if not processing_steps:
- raise MissingProcessingStepsError("Missing processing steps or empty list")
-
- processing_step = processing_steps[0]
- # TODO: Add support to iterate over the processing steps list in case of failure
@@ -60,17 +210,18 @@ def create_endpoint(
- dataset = request.query_params.get("dataset")
- if not are_valid_parameters([dataset]) or not dataset:
- raise MissingRequiredParameterError("Parameter 'dataset' is required")
- if processing_step.input_type == "dataset":
- config = None
- split = None
- elif processing_step.input_type == "config":
- config = request.query_params.get("config")
- split = None
- if not are_valid_parameters([config]):
- raise MissingRequiredParameterError("Parameter 'config' is required")
- else:
- config = request.query_params.get("config")
- split = request.query_params.get("split")
- if not are_valid_parameters([config, split]):
- raise MissingRequiredParameterError("Parameters 'config' and 'split' are required")
- logging.info(f"{processing_step.job_type}, dataset={dataset}, config={config}, split={split}")
+ # validating request parameters
+ dataset_parameter = request.query_params.get("dataset")
+ config_parameter = request.query_params.get("config")
+ split_parameter = request.query_params.get("split")
+ validators = get_input_type_validators_by_priority(steps_by_input_type=steps_by_input_type)
+
+ logging.debug(
+ f"endpoint={endpoint_name} dataset={dataset_parameter} ",
+ f"config={config_parameter} split={split_parameter}",
+ )
+
+ validator = get_input_type_validator_by_parameters(
+ validators, dataset_parameter, config_parameter, split_parameter
+ )
+ processing_steps = steps_by_input_type[validator.input_type]
+ dataset, config, split = validator.get_useful_parameters(
+ dataset_parameter, config_parameter, split_parameter
+ )
@@ -79,29 +230,16 @@ def create_endpoint(
- auth_check(dataset, external_auth_url=external_auth_url, request=request)
- try:
- result = get_response(kind=processing_step.cache_kind, dataset=dataset, config=config, split=split)
- content = result["content"]
- http_status = result["http_status"]
- error_code = result["error_code"]
- if http_status == HTTPStatus.OK:
- return get_json_ok_response(content=content, max_age=max_age_long)
- else:
- return get_json_error_response(
- content=content, status_code=http_status, max_age=max_age_short, error_code=error_code
- )
- except DoesNotExist as e:
- # maybe the response is in process
- try:
- check_in_process(
- processing_step=processing_step,
- init_processing_steps=init_processing_steps,
- dataset=dataset,
- config=config,
- split=split,
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- )
- except (PreviousStepError, DatasetError):
- raise ResponseNotFoundError("Not found.") from e
- raise ResponseNotReadyError(
- "The server is busier than usual and the response is not ready yet. Please retry later."
- ) from e
+ auth_check(dataset, external_auth_url=external_auth_url, request=request) # type: ignore
+
+ # getting result based on processing steps
+ result = get_cache_entry_from_steps(
+ processing_steps, dataset, config, split, init_processing_steps, hf_endpoint, hf_token # type: ignore
+ )
+
+ content = result["content"]
+ http_status = result["http_status"]
+ error_code = result["error_code"]
+ if http_status == HTTPStatus.OK:
+ return get_json_ok_response(content=content, max_age=max_age_long)
+
+ return get_json_error_response(
+ content=content, status_code=http_status, max_age=max_age_short, error_code=error_code
+ )
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index dfbc1d05..4bd7aac8 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -84,7 +83,0 @@ class ExternalAuthenticatedError(ApiCustomError):
-class MissingProcessingStepsError(ApiCustomError):
- """Raised when an endpoint does not have related processing steps."""
-
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "MissingProcessingStepsError")
-
-
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index 7f22af4c..df390bbe 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -4 +4 @@
-from typing import Iterator, List, Mapping
+from typing import Iterator
@@ -6 +6 @@ from typing import Iterator, List, Mapping
-from libcommon.processing_graph import ProcessingGraph, ProcessingStep
+from libcommon.processing_graph import ProcessingGraph
@@ -13 +13 @@ from api.config import AppConfig, EndpointConfig, UvicornConfig
-from api.routes.endpoint import EndpointsDefinition
+from api.routes.endpoint import EndpointsDefinition, StepsByInputTypeAndEndpoint
@@ -40 +40,14 @@ def app_config(monkeypatch_session: MonkeyPatch) -> AppConfig:
-def endpoint_definition(app_config: AppConfig) -> Mapping[str, List[ProcessingStep]]:
+def endpoint_config(monkeypatch_session: MonkeyPatch) -> EndpointConfig:
+ return EndpointConfig(
+ step_names_by_input_type_and_endpoint={
+ "/config-names": {"dataset": ["/config-names"]},
+ "/splits": {
+ "config": ["/split-names-from-streaming"],
+ },
+ "/first-rows": {"split": ["/first-rows"]},
+ }
+ )
+
+
+@fixture(scope="session")
+def endpoint_definition(endpoint_config: EndpointConfig, app_config: AppConfig) -> StepsByInputTypeAndEndpoint:
@@ -42,2 +55 @@ def endpoint_definition(app_config: AppConfig) -> Mapping[str, List[ProcessingSt
- endpoint_specification = EndpointsDefinition(processing_graph, EndpointConfig.from_env())
- return endpoint_specification.definition
+ return EndpointsDefinition(processing_graph, endpoint_config=endpoint_config).steps_by_input_type_and_endpoint
@@ -47 +59 @@ def endpoint_definition(app_config: AppConfig) -> Mapping[str, List[ProcessingSt
-def first_dataset_endpoint(endpoint_definition: Mapping[str, List[ProcessingStep]]) -> str:
+def first_dataset_endpoint(endpoint_definition: StepsByInputTypeAndEndpoint) -> str:
@@ -50,2 +62,2 @@ def first_dataset_endpoint(endpoint_definition: Mapping[str, List[ProcessingStep
- for endpoint, processing_steps in endpoint_definition.items()
- if next((step for step in processing_steps if step.input_type == "dataset"), None)
+ for endpoint, input_types in endpoint_definition.items()
+ if next((endpoint for input_type, _ in input_types.items() if input_type == "dataset"), None)
@@ -56 +68 @@ def first_dataset_endpoint(endpoint_definition: Mapping[str, List[ProcessingStep
-def first_config_endoint(endpoint_definition: Mapping[str, List[ProcessingStep]]) -> str:
+def first_config_endoint(endpoint_definition: StepsByInputTypeAndEndpoint) -> str:
@@ -59,2 +71,2 @@ def first_config_endoint(endpoint_definition: Mapping[str, List[ProcessingStep]]
- for endpoint, processing_steps in endpoint_definition.items()
- if next((step for step in processing_steps if step.input_type == "config"), None)
+ for endpoint, input_types in endpoint_definition.items()
+ if next((endpoint for input_type, _ in input_types.items() if input_type == "config"), None)
@@ -65 +77 @@ def first_config_endoint(endpoint_definition: Mapping[str, List[ProcessingStep]]
-def first_split_endpoint(endpoint_definition: Mapping[str, List[ProcessingStep]]) -> str:
+def first_split_endpoint(endpoint_definition: StepsByInputTypeAndEndpoint) -> str:
@@ -68,2 +80,2 @@ def first_split_endpoint(endpoint_definition: Mapping[str, List[ProcessingStep]]
- for endpoint, processing_steps in endpoint_definition.items()
- if next((step for step in processing_steps if step.input_type == "split"), None)
+ for endpoint, input_types in endpoint_definition.items()
+ if next((endpoint for input_type, _ in input_types.items() if input_type == "split"), None)
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
new file mode 100644
index 00000000..aefa84d7
--- /dev/null
+++ b/services/api/tests/routes/test_endpoint.py
@@ -0,0 +1,134 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from http import HTTPStatus
+
+from libcommon.config import ProcessingGraphConfig
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.queue import Queue
+from libcommon.simple_cache import upsert_response
+from pytest import raises
+
+from api.config import AppConfig, EndpointConfig
+from api.routes.endpoint import EndpointsDefinition, get_cache_entry_from_steps
+from api.utils import ResponseNotReadyError
+
+
+def test_endpoints_definition() -> None:
+ config = ProcessingGraphConfig()
+ graph = ProcessingGraph(config.specification)
+ endpoint_config = EndpointConfig.from_env()
+
+ endpoints_definition = EndpointsDefinition(graph, endpoint_config)
+ assert endpoints_definition
+
+ definition = endpoints_definition.steps_by_input_type_and_endpoint
+ assert definition
+
+ config_names = definition["/config-names"]
+ assert config_names is not None
+ assert config_names["dataset"] is not None
+ assert len(config_names["dataset"]) == 1 # Only has one processing step
+
+ splits = definition["/splits"]
+ assert splits is not None
+ assert splits["dataset"] is not None
+ assert len(splits["dataset"]) == 1 # Only has one processing step
+ assert len(splits["config"]) == 2 # Has two processing step
+
+ first_rows = definition["/first-rows"]
+ assert first_rows is not None
+ assert first_rows["split"] is not None
+ assert len(first_rows["split"]) == 1 # Only has one processing step
+
+ parquet_and_dataset_info = definition["/parquet-and-dataset-info"]
+ assert parquet_and_dataset_info is not None
+ assert parquet_and_dataset_info["dataset"] is not None
+ assert len(parquet_and_dataset_info["dataset"]) == 1 # Only has one processing step
+
+ parquet = definition["/parquet"]
+ assert parquet is not None
+ assert parquet["dataset"] is not None
+ assert len(parquet["dataset"]) == 1 # Only has one processing step
+
+ dataset_info = definition["/dataset-info"]
+ assert dataset_info is not None
+ assert dataset_info["dataset"] is not None
+ assert len(dataset_info["dataset"]) == 1 # Only has one processing step
+
+ sizes = definition["/sizes"]
+ assert sizes is not None
+ assert sizes["dataset"] is not None
+ assert len(sizes["dataset"]) == 1 # Only has one processing step
+
+
+def test_get_cache_entry_from_steps() -> None:
+ dataset = "dataset"
+ config = "config"
+
+ app_config = AppConfig.from_env()
+ graph_config = ProcessingGraphConfig()
+ graph = ProcessingGraph(graph_config.specification)
+ init_processing_steps = graph.get_first_steps()
+
+ cache_with_error = "/split-names-from-streaming"
+ cache_without_error = "/split-names-from-dataset-info"
+
+ step_with_error = graph.get_step(cache_with_error)
+ step_whitout_error = graph.get_step(cache_without_error)
+
+ upsert_response(
+ kind=cache_without_error,
+ dataset=dataset,
+ config=config,
+ content={},
+ http_status=HTTPStatus.OK,
+ )
+
+ upsert_response(
+ kind=cache_with_error,
+ dataset=dataset,
+ config=config,
+ content={},
+ http_status=HTTPStatus.INTERNAL_SERVER_ERROR,
+ )
+
+ # succeeded result is returned
+ result = get_cache_entry_from_steps(
+ [step_whitout_error, step_with_error],
+ dataset,
+ config,
+ None,
+ init_processing_steps,
+ app_config.common.hf_endpoint,
+ )
+ assert result
+ assert result["http_status"] == HTTPStatus.OK
+
+ # succeeded result is returned even if first step failed
+ result = get_cache_entry_from_steps(
+ [step_with_error, step_whitout_error],
+ dataset,
+ config,
+ None,
+ init_processing_steps,
+ app_config.common.hf_endpoint,
+ )
+ assert result
+ assert result["http_status"] == HTTPStatus.OK
+
+ # error result is returned if all steps failed
+ result = get_cache_entry_from_steps(
+ [step_with_error, step_with_error], dataset, config, None, init_processing_steps, app_config.common.hf_endpoint
+ )
+ assert result
+ assert result["http_status"] == HTTPStatus.INTERNAL_SERVER_ERROR
+
+ # peding job thows exception
+ queue = Queue()
+ queue.upsert_job(job_type="/splits", dataset=dataset, config=config, force=True)
+ non_existent_step = graph.get_step("/splits")
+ with raises(ResponseNotReadyError):
+ get_cache_entry_from_steps(
+ [non_existent_step], dataset, config, None, init_processing_steps, app_config.common.hf_endpoint
+ )
diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py
index 8cacbda0..a3175613 100644
--- a/services/api/tests/test_app.py
+++ b/services/api/tests/test_app.py
@@ -10 +10,2 @@ from starlette.testclient import TestClient
-from api.app import create_app
+from api.app import create_app_with_config
+from api.config import AppConfig, EndpointConfig
@@ -16,2 +17,4 @@ from .utils import auth_callback
-def client(monkeypatch_session: pytest.MonkeyPatch) -> TestClient:
- return TestClient(create_app())
+def client(
+ monkeypatch_session: pytest.MonkeyPatch, app_config: AppConfig, endpoint_config: EndpointConfig
+) -> TestClient:
+ return TestClient(create_app_with_config(app_config=app_config, endpoint_config=endpoint_config))
|
|
dae20a72b8c8d923ffefd3cb3130bb948f30c5c9
|
Sylvain Lesage
| 2023-03-01T09:48:03 |
feat: 🎸 reduce a lot the resources used in production (#874)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 39cde838..3efdfed3 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -178 +178 @@ workers:
- maxJobsPerNamespace: 4
+ maxJobsPerNamespace: 20
@@ -182 +182 @@ workers:
- replicas: 20
+ replicas: 12
@@ -186 +186 @@ workers:
- memory: "8Gi"
+ memory: "1Gi"
@@ -193 +193 @@ workers:
- maxJobsPerNamespace: 4
+ maxJobsPerNamespace: 20
@@ -197 +197 @@ workers:
- replicas: 8
+ replicas: 2
@@ -208 +208 @@ workers:
- maxJobsPerNamespace: 5
+ maxJobsPerNamespace: 20
@@ -212 +212 @@ workers:
- replicas: 12
+ replicas: 2
@@ -223 +223 @@ workers:
- maxJobsPerNamespace: 5
+ maxJobsPerNamespace: 20
@@ -227 +227 @@ workers:
- replicas: 8
+ replicas: 2
@@ -238 +238 @@ workers:
- maxJobsPerNamespace: 4
+ maxJobsPerNamespace: 20
@@ -242 +242 @@ workers:
- replicas: 80
+ replicas: 2
@@ -253 +253 @@ workers:
- maxJobsPerNamespace: 4
+ maxJobsPerNamespace: 20
@@ -257 +257 @@ workers:
- replicas: 20
+ replicas: 2
@@ -268 +268 @@ workers:
- maxJobsPerNamespace: 2
+ maxJobsPerNamespace: 20
@@ -283 +283 @@ workers:
- maxJobsPerNamespace: 2
+ maxJobsPerNamespace: 20
@@ -298 +298 @@ workers:
- maxJobsPerNamespace: 2
+ maxJobsPerNamespace: 20
|
|
ae19dfb828415927e331109962018d983f7b3e6e
|
Sylvain Lesage
| 2023-03-01T09:19:01 |
fix: 🐛 manage duplicates in cache entries migration (#873)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
index b1e7774d..e10637ff 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
@@ -13,0 +14 @@ split_names_from_streaming = "/split-names-from-streaming"
+split_names_tmp = "/split-names-TMP"
@@ -21,0 +23,5 @@ class MigrationCacheUpdateSplitNames(Migration):
+ # update existing documents with the new kind (if any) to avoid duplicates (will be deleted later)
+ db["cachedResponsesBlue"].update_many(
+ {"kind": split_names_from_streaming}, {"$set": {"kind": split_names_tmp}}
+ )
+ # update existing documents with the old kind
@@ -22,0 +29,2 @@ class MigrationCacheUpdateSplitNames(Migration):
+ # delete the duplicates
+ db["cachedResponsesBlue"].delete_many({"kind": split_names_tmp})
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
index 301bfa53..3684b9a0 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
@@ -35,2 +35,2 @@ class MigrationQueueUpdateSplitNames(Migration):
- "find": f"Job[{split_names}",
- "replacement": f"Job[{split_names_from_streaming}",
+ "find": f"Job[{split_names}]",
+ "replacement": f"Job[{split_names_from_streaming}]",
@@ -61,2 +61,2 @@ class MigrationQueueUpdateSplitNames(Migration):
- "find": f"Job[{split_names_from_streaming}",
- "replacement": f"Job[{split_names}",
+ "find": f"Job[{split_names_from_streaming}]",
+ "replacement": f"Job[{split_names}]",
|
|
ae635fd0cdd6f773d587384b4e2b2f401c12d6dc
|
Sylvain Lesage
| 2023-02-28T20:09:21 |
fix: 🐛 set the list of supported job types as a csv string (#871)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 14a66117..21fe769a 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.3.0
+version: 1.3.1
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index e7ef62d0..a3cd9aef 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -164 +164 @@ workers:
- workerOnlyJobTypes: []
+ workerOnlyJobTypes: ""
@@ -178 +178 @@ workers:
- workerOnlyJobTypes: ["/first-rows"]
+ workerOnlyJobTypes: "/first-rows"
@@ -192 +192 @@ workers:
- workerOnlyJobTypes: []
+ workerOnlyJobTypes: "/parquet-and-dataset-info"
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 75d1382a..39cde838 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -179 +179 @@ workers:
- workerOnlyJobTypes: []
+ workerOnlyJobTypes: ""
@@ -194 +194 @@ workers:
- workerOnlyJobTypes: ["/config-names"]
+ workerOnlyJobTypes: "/config-names"
@@ -209 +209 @@ workers:
- workerOnlyJobTypes: ["/split-names"]
+ workerOnlyJobTypes: "/split-names"
@@ -224 +224 @@ workers:
- workerOnlyJobTypes: ["/splits"]
+ workerOnlyJobTypes: "/splits"
@@ -239 +239 @@ workers:
- workerOnlyJobTypes: ["/first-rows"]
+ workerOnlyJobTypes: "/first-rows"
@@ -254 +254 @@ workers:
- workerOnlyJobTypes: ["/parquet-and-dataset-info"]
+ workerOnlyJobTypes: "/parquet-and-dataset-info"
@@ -269 +269 @@ workers:
- workerOnlyJobTypes: ["/parquet"]
+ workerOnlyJobTypes: "/parquet"
@@ -284 +284 @@ workers:
- workerOnlyJobTypes: ["/dataset-info"]
+ workerOnlyJobTypes: "/dataset-info"
@@ -299 +299 @@ workers:
- workerOnlyJobTypes: ["/sizes"]
+ workerOnlyJobTypes: "/sizes"
diff --git a/chart/values.yaml b/chart/values.yaml
index 34ac87aa..34441a69 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -280 +280 @@ workers:
- workerOnlyJobTypes: []
+ workerOnlyJobTypes: ""
@@ -295 +295 @@ workers:
- workerOnlyJobTypes: ["/config-names"]
+ workerOnlyJobTypes: "/config-names"
@@ -310 +310 @@ workers:
- workerOnlyJobTypes: ["/split-names"]
+ workerOnlyJobTypes: "/split-names"
@@ -325 +325 @@ workers:
- workerOnlyJobTypes: ["/splits"]
+ workerOnlyJobTypes: "/splits"
@@ -340 +340 @@ workers:
- workerOnlyJobTypes: ["/first-rows"]
+ workerOnlyJobTypes: "/first-rows"
@@ -355 +355 @@ workers:
- workerOnlyJobTypes: ["/parquet-and-dataset-info"]
+ workerOnlyJobTypes: "/parquet-and-dataset-info"
@@ -370 +370 @@ workers:
- workerOnlyJobTypes: ["/parquet"]
+ workerOnlyJobTypes: "/parquet"
@@ -385 +385 @@ workers:
- workerOnlyJobTypes: ["/dataset-info"]
+ workerOnlyJobTypes: "/dataset-info"
@@ -400 +400 @@ workers:
- workerOnlyJobTypes: ["/sizes"]
+ workerOnlyJobTypes: "/sizes"
|
|
d034d83cd8b08e29f8f77e1930d793ed350a55e7
|
Sylvain Lesage
| 2023-02-28T16:15:25 |
ci: 🎡 add missing step to get the source code (#869)
|
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml
index e3bfb380..e8821b61 100644
--- a/.github/workflows/cd.yml
+++ b/.github/workflows/cd.yml
@@ -86,0 +87,2 @@ jobs:
+ - name: Checkout sources
+ uses: actions/checkout@v3
|
|
7e9aa1dd89ff19c0d866987cf9e222a5fbb15f0d
|
Sylvain Lesage
| 2023-02-28T14:48:17 |
Add doc, simplify docker, update chart (#868)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 218b9243..14a66117 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.2.0
+version: 1.3.0
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
index 22a93428..11b01b55 100644
--- a/chart/templates/_envWorker.tpl
+++ b/chart/templates/_envWorker.tpl
@@ -6,0 +7,4 @@
+- name: WORKER_HEARTBEAT_INTERVAL_SECONDS
+ value: {{ .Values.worker.heartbeatIntervalSeconds | quote}}
+- name: WORKER_KILL_ZOMBIES_INTERVAL_SECONDS
+ value: {{ .Values.worker.killZombiesIntervalSeconds | quote}}
@@ -12,0 +17,2 @@
+- name: WORKER_MAX_MISSING_HEARTBEATS
+ value: {{ .Values.worker.maxMissingHeartbeats | quote }}
@@ -14,0 +21,3 @@
+- name: TMPDIR
+ value: "/tmp"
+ # ^ensure the temporary files are created in /tmp, which is writable
diff --git a/chart/values.yaml b/chart/values.yaml
index 4de55d6d..34ac87aa 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -99,0 +100,4 @@ worker:
+ # the time interval between two heartbeats. Each heartbeat updates the job "last_heartbeat" field in the queue.
+ heartbeatIntervalSeconds: 60
+ # the time interval at which the worker looks for zombie jobs to kill them
+ killZombiesIntervalSeconds: 600
@@ -105,0 +110,2 @@ worker:
+ # the number of hearbeats a job must have missed to be considered a zombie job.
+ maxMissingHeartbeats: 5
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index ab45a3bb..7989d2a3 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -253 +253 @@ name = "coverage"
-version = "7.2.0"
+version = "7.2.1"
@@ -259,51 +259,51 @@ files = [
- {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
- {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
- {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
- {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
- {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
- {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
- {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
- {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
- {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
- {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
- {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
- {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
- {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
- {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
- {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"},
+ {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"},
+ {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"},
+ {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"},
+ {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"},
+ {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"},
+ {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"},
+ {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"},
+ {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"},
+ {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"},
+ {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"},
+ {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"},
+ {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"},
+ {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"},
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index 522ec9a9..ae883283 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -253 +253 @@ name = "coverage"
-version = "7.2.0"
+version = "7.2.1"
@@ -259,51 +259,51 @@ files = [
- {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
- {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
- {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
- {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
- {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
- {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
- {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
- {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
- {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
- {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
- {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
- {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
- {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
- {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
- {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"},
+ {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"},
+ {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"},
+ {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"},
+ {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"},
+ {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"},
+ {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"},
+ {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"},
+ {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"},
+ {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"},
+ {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"},
+ {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"},
+ {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"},
+ {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"},
@@ -1496 +1496 @@ name = "types-psutil"
-version = "5.9.5.8"
+version = "5.9.5.9"
@@ -1502,2 +1502,14 @@ files = [
- {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
- {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
+ {file = "types-psutil-5.9.5.9.tar.gz", hash = "sha256:c6c7e6b41b6f7ebb87afd5650df360751d0e8acf0d276b7ac64d00f419beb922"},
+ {file = "types_psutil-5.9.5.9-py3-none-any.whl", hash = "sha256:f7830fb3a7bc3288113ff08199330fced7851f5219f0af1b2abc24aaa3112236"},
+]
+
+[[package]]
+name = "types-pytz"
+version = "2022.7.1.2"
+description = "Typing stubs for pytz"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-pytz-2022.7.1.2.tar.gz", hash = "sha256:487d3e8e9f4071eec8081746d53fa982bbc05812e719dcbf2ebf3d55a1a4cd28"},
+ {file = "types_pytz-2022.7.1.2-py3-none-any.whl", hash = "sha256:40ca448a928d566f7d44ddfde0066e384f7ffbd4da2778e42a4570eaca572446"},
@@ -1520 +1532 @@ name = "types-requests"
-version = "2.28.11.14"
+version = "2.28.11.15"
@@ -1526,2 +1538,2 @@ files = [
- {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
- {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
+ {file = "types-requests-2.28.11.15.tar.gz", hash = "sha256:fc8eaa09cc014699c6b63c60c2e3add0c8b09a410c818b5ac6e65f92a26dde09"},
+ {file = "types_requests-2.28.11.15-py3-none-any.whl", hash = "sha256:a05e4c7bc967518fba5789c341ea8b0c942776ee474c7873129a61161978e586"},
@@ -1535 +1547 @@ name = "types-urllib3"
-version = "1.26.25.7"
+version = "1.26.25.8"
@@ -1541,2 +1553,2 @@ files = [
- {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
- {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
+ {file = "types-urllib3-1.26.25.8.tar.gz", hash = "sha256:ecf43c42d8ee439d732a1110b4901e9017a79a38daca26f08e42c8460069392c"},
+ {file = "types_urllib3-1.26.25.8-py3-none-any.whl", hash = "sha256:95ea847fbf0bf675f50c8ae19a665baedcf07e6b4641662c4c3c72e7b2edf1a9"},
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index 76b6602d..ffb5ed5d 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -274 +274 @@ name = "coverage"
-version = "7.2.0"
+version = "7.2.1"
@@ -280,51 +280,51 @@ files = [
- {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
- {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
- {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
- {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
- {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
- {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
- {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
- {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
- {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
- {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
- {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
- {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
- {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
- {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
- {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"},
+ {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"},
+ {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"},
+ {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"},
+ {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"},
+ {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"},
+ {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"},
+ {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"},
+ {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"},
+ {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"},
+ {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"},
+ {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"},
+ {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"},
+ {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"},
@@ -1702 +1702 @@ name = "types-psutil"
-version = "5.9.5.8"
+version = "5.9.5.9"
@@ -1708,2 +1708,2 @@ files = [
- {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
- {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
+ {file = "types-psutil-5.9.5.9.tar.gz", hash = "sha256:c6c7e6b41b6f7ebb87afd5650df360751d0e8acf0d276b7ac64d00f419beb922"},
+ {file = "types_psutil-5.9.5.9-py3-none-any.whl", hash = "sha256:f7830fb3a7bc3288113ff08199330fced7851f5219f0af1b2abc24aaa3112236"},
@@ -1714 +1714 @@ name = "types-requests"
-version = "2.28.11.14"
+version = "2.28.11.15"
@@ -1720,2 +1720,2 @@ files = [
- {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
- {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
+ {file = "types-requests-2.28.11.15.tar.gz", hash = "sha256:fc8eaa09cc014699c6b63c60c2e3add0c8b09a410c818b5ac6e65f92a26dde09"},
+ {file = "types_requests-2.28.11.15-py3-none-any.whl", hash = "sha256:a05e4c7bc967518fba5789c341ea8b0c942776ee474c7873129a61161978e586"},
@@ -1741 +1741 @@ name = "types-urllib3"
-version = "1.26.25.7"
+version = "1.26.25.8"
@@ -1747,2 +1747,2 @@ files = [
- {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
- {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
+ {file = "types-urllib3-1.26.25.8.tar.gz", hash = "sha256:ecf43c42d8ee439d732a1110b4901e9017a79a38daca26f08e42c8460069392c"},
+ {file = "types_urllib3-1.26.25.8-py3-none-any.whl", hash = "sha256:95ea847fbf0bf675f50c8ae19a665baedcf07e6b4641662c4c3c72e7b2edf1a9"},
@@ -1801 +1801 @@ name = "watchdog"
-version = "2.3.0"
+version = "2.3.1"
@@ -1807,28 +1807,28 @@ files = [
- {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:c1b3962e5463a848ba2a342cb66c80251dca27a102933b8f38d231d2a9e5a543"},
- {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9e651b4874477c1bf239417d43818bbfd047aaf641b029fa60d6f5109ede0db0"},
- {file = "watchdog-2.3.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d04662017efd00a014cff9068708e085d67f2fac43f48bbbb95a7f97490487f3"},
- {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0f7d759299ce21a3d2a77e18d430c24811369c3432453701790acc6ff45a7101"},
- {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:a4b9bece40d46bf6fb8621817ea7d903eae2b9b3ebac55a51ed50354a79061a8"},
- {file = "watchdog-2.3.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:242e57253e84a736e6777ba756c48cf6a68d3d90cb9e01bd6bfd371a949ace3a"},
- {file = "watchdog-2.3.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:3fa74b0ef4825f9112932675a002296cb2d3d3e400d7a44c32fafd1ecc83ada0"},
- {file = "watchdog-2.3.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:15bf5b165d7a6b48265411dad74fb0d33053f8270eb6575faad0e016035cf9f7"},
- {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:139262f678b4e6a7013261c772059bca358441de04fb0e0087489a34db9e3db0"},
- {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8a214955769d2ef0f7aaa82f31863e3bdf6b083ce1b5f1c2e85cab0f66fba024"},
- {file = "watchdog-2.3.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4e648df44a4c6ea6da4d9eb6722745c986b9d70268f25ae60f140082d7c8908e"},
- {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:473164a2de473f708ca194a992466eeefff73b58273bbb88e089c5a5a98fcda1"},
- {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebe756f788cb130fdc5c150ea8a4fda39cb4ee3a5873a345607c8b84fecf018b"},
- {file = "watchdog-2.3.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a623de186477e9e05f8461087f856412eae5cd005cc4bcb232ed5c6f9a8709f5"},
- {file = "watchdog-2.3.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:43d76d7888b26850b908208bb82383a193e8b0f25d0abaa84452f191b4acdea4"},
- {file = "watchdog-2.3.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:5ddbbe87f9ed726940d174076da030cd01ec45433ef2b1b2e6094c84f2af17f1"},
- {file = "watchdog-2.3.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3fa1572f5a2f6d17d4d860edbc04488fef31b007c25c2f3b11203fb8179b7c67"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_aarch64.whl", hash = "sha256:1d9c656495172873bf1ddc7e39e80055fcdd21c4608cf68f23a28116dcba0b43"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_armv7l.whl", hash = "sha256:00f93782c67042d9525ec51628330b5faf5fb84bcb7ebaac05ea8528cfb20bba"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_i686.whl", hash = "sha256:f1a655f4a49f9232311b9967f42cc2eaf43fd4903f3bed850dd4570fda5d5eff"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64.whl", hash = "sha256:aa4773160b9cb21ba369cb42d59a947087330b3a02480173033a6a6cc137a510"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:982f5416a2817003172994d865285dd6a2b3836f033cd3fa87d1a62096a162cc"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_s390x.whl", hash = "sha256:45c13e7e6eea1013da419bf9aa9a8f5df7bbf3e5edce40bc6df84130febf39d5"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_x86_64.whl", hash = "sha256:7767a3da3307d9cf597832f692702441a97c259e5d0d560f2e57c43ad0d191d2"},
- {file = "watchdog-2.3.0-py3-none-win32.whl", hash = "sha256:8863913ea2c3f256d18c33d84546518636e391cd8f50d209b9a31221e0f7d3fd"},
- {file = "watchdog-2.3.0-py3-none-win_amd64.whl", hash = "sha256:6d79b5954db8f41d6a7f5763042b988f7a4afd40b7d141456061fa7c5b7f2159"},
- {file = "watchdog-2.3.0-py3-none-win_ia64.whl", hash = "sha256:a3559ee82a10976de1ec544b6ebe3b4aa398d491860a283d80ec0f550076d068"},
- {file = "watchdog-2.3.0.tar.gz", hash = "sha256:9d39effe6909be898ba3e7286a9e9b17a6a9f734fb1ef9dde3e9bb68715fca39"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:d1f1200d4ec53b88bf04ab636f9133cb703eb19768a39351cee649de21a33697"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:564e7739abd4bd348aeafbf71cc006b6c0ccda3160c7053c4a53b67d14091d42"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:95ad708a9454050a46f741ba5e2f3468655ea22da1114e4c40b8cbdaca572565"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:a073c91a6ef0dda488087669586768195c3080c66866144880f03445ca23ef16"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aa8b028750b43e80eea9946d01925168eeadb488dfdef1d82be4b1e28067f375"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:964fd236cd443933268ae49b59706569c8b741073dbfd7ca705492bae9d39aab"},
+ {file = "watchdog-2.3.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:91fd146d723392b3e6eb1ac21f122fcce149a194a2ba0a82c5e4d0ee29cd954c"},
+ {file = "watchdog-2.3.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:efe3252137392a471a2174d721e1037a0e6a5da7beb72a021e662b7000a9903f"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:85bf2263290591b7c5fa01140601b64c831be88084de41efbcba6ea289874f44"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8f2df370cd8e4e18499dd0bfdef476431bcc396108b97195d9448d90924e3131"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ea5d86d1bcf4a9d24610aa2f6f25492f441960cf04aed2bd9a97db439b643a7b"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:6f5d0f7eac86807275eba40b577c671b306f6f335ba63a5c5a348da151aba0fc"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5b848c71ef2b15d0ef02f69da8cc120d335cec0ed82a3fa7779e27a5a8527225"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0d9878be36d2b9271e3abaa6f4f051b363ff54dbbe7e7df1af3c920e4311ee43"},
+ {file = "watchdog-2.3.1-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4cd61f98cb37143206818cb1786d2438626aa78d682a8f2ecee239055a9771d5"},
+ {file = "watchdog-2.3.1-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3d2dbcf1acd96e7a9c9aefed201c47c8e311075105d94ce5e899f118155709fd"},
+ {file = "watchdog-2.3.1-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:03f342a9432fe08107defbe8e405a2cb922c5d00c4c6c168c68b633c64ce6190"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_aarch64.whl", hash = "sha256:7a596f9415a378d0339681efc08d2249e48975daae391d58f2e22a3673b977cf"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_armv7l.whl", hash = "sha256:0e1dd6d449267cc7d6935d7fe27ee0426af6ee16578eed93bacb1be9ff824d2d"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_i686.whl", hash = "sha256:7a1876f660e32027a1a46f8a0fa5747ad4fcf86cb451860eae61a26e102c8c79"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64.whl", hash = "sha256:2caf77ae137935c1466f8cefd4a3aec7017b6969f425d086e6a528241cba7256"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:53f3e95081280898d9e4fc51c5c69017715929e4eea1ab45801d5e903dd518ad"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_s390x.whl", hash = "sha256:9da7acb9af7e4a272089bd2af0171d23e0d6271385c51d4d9bde91fe918c53ed"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_x86_64.whl", hash = "sha256:8a4d484e846dcd75e96b96d80d80445302621be40e293bfdf34a631cab3b33dc"},
+ {file = "watchdog-2.3.1-py3-none-win32.whl", hash = "sha256:a74155398434937ac2780fd257c045954de5b11b5c52fc844e2199ce3eecf4cf"},
+ {file = "watchdog-2.3.1-py3-none-win_amd64.whl", hash = "sha256:5defe4f0918a2a1a4afbe4dbb967f743ac3a93d546ea4674567806375b024adb"},
+ {file = "watchdog-2.3.1-py3-none-win_ia64.whl", hash = "sha256:4109cccf214b7e3462e8403ab1e5b17b302ecce6c103eb2fc3afa534a7f27b96"},
+ {file = "watchdog-2.3.1.tar.gz", hash = "sha256:d9f9ed26ed22a9d331820a8432c3680707ea8b54121ddcc9dc7d9f2ceeb36906"},
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index 4fae0e53..264f58cc 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -274 +274 @@ name = "coverage"
-version = "7.2.0"
+version = "7.2.1"
@@ -280,51 +280,51 @@ files = [
- {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
- {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
- {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
- {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
- {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
- {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
- {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
- {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
- {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
- {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
- {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
- {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
- {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
- {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
- {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"},
+ {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"},
+ {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"},
+ {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"},
+ {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"},
+ {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"},
+ {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"},
+ {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"},
+ {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"},
+ {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"},
+ {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"},
+ {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"},
+ {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"},
+ {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"},
@@ -1825 +1825 @@ name = "types-psutil"
-version = "5.9.5.8"
+version = "5.9.5.9"
@@ -1831,2 +1831,2 @@ files = [
- {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
- {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
+ {file = "types-psutil-5.9.5.9.tar.gz", hash = "sha256:c6c7e6b41b6f7ebb87afd5650df360751d0e8acf0d276b7ac64d00f419beb922"},
+ {file = "types_psutil-5.9.5.9-py3-none-any.whl", hash = "sha256:f7830fb3a7bc3288113ff08199330fced7851f5219f0af1b2abc24aaa3112236"},
@@ -1837 +1837 @@ name = "types-requests"
-version = "2.28.11.14"
+version = "2.28.11.15"
@@ -1843,2 +1843,2 @@ files = [
- {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
- {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
+ {file = "types-requests-2.28.11.15.tar.gz", hash = "sha256:fc8eaa09cc014699c6b63c60c2e3add0c8b09a410c818b5ac6e65f92a26dde09"},
+ {file = "types_requests-2.28.11.15-py3-none-any.whl", hash = "sha256:a05e4c7bc967518fba5789c341ea8b0c942776ee474c7873129a61161978e586"},
@@ -1852 +1852 @@ name = "types-urllib3"
-version = "1.26.25.7"
+version = "1.26.25.8"
@@ -1858,2 +1858,2 @@ files = [
- {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
- {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
+ {file = "types-urllib3-1.26.25.8.tar.gz", hash = "sha256:ecf43c42d8ee439d732a1110b4901e9017a79a38daca26f08e42c8460069392c"},
+ {file = "types_urllib3-1.26.25.8-py3-none-any.whl", hash = "sha256:95ea847fbf0bf675f50c8ae19a665baedcf07e6b4641662c4c3c72e7b2edf1a9"},
@@ -1912 +1912 @@ name = "watchdog"
-version = "2.3.0"
+version = "2.3.1"
@@ -1918,28 +1918,28 @@ files = [
- {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:c1b3962e5463a848ba2a342cb66c80251dca27a102933b8f38d231d2a9e5a543"},
- {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9e651b4874477c1bf239417d43818bbfd047aaf641b029fa60d6f5109ede0db0"},
- {file = "watchdog-2.3.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d04662017efd00a014cff9068708e085d67f2fac43f48bbbb95a7f97490487f3"},
- {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0f7d759299ce21a3d2a77e18d430c24811369c3432453701790acc6ff45a7101"},
- {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:a4b9bece40d46bf6fb8621817ea7d903eae2b9b3ebac55a51ed50354a79061a8"},
- {file = "watchdog-2.3.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:242e57253e84a736e6777ba756c48cf6a68d3d90cb9e01bd6bfd371a949ace3a"},
- {file = "watchdog-2.3.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:3fa74b0ef4825f9112932675a002296cb2d3d3e400d7a44c32fafd1ecc83ada0"},
- {file = "watchdog-2.3.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:15bf5b165d7a6b48265411dad74fb0d33053f8270eb6575faad0e016035cf9f7"},
- {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:139262f678b4e6a7013261c772059bca358441de04fb0e0087489a34db9e3db0"},
- {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8a214955769d2ef0f7aaa82f31863e3bdf6b083ce1b5f1c2e85cab0f66fba024"},
- {file = "watchdog-2.3.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4e648df44a4c6ea6da4d9eb6722745c986b9d70268f25ae60f140082d7c8908e"},
- {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:473164a2de473f708ca194a992466eeefff73b58273bbb88e089c5a5a98fcda1"},
- {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebe756f788cb130fdc5c150ea8a4fda39cb4ee3a5873a345607c8b84fecf018b"},
- {file = "watchdog-2.3.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a623de186477e9e05f8461087f856412eae5cd005cc4bcb232ed5c6f9a8709f5"},
- {file = "watchdog-2.3.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:43d76d7888b26850b908208bb82383a193e8b0f25d0abaa84452f191b4acdea4"},
- {file = "watchdog-2.3.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:5ddbbe87f9ed726940d174076da030cd01ec45433ef2b1b2e6094c84f2af17f1"},
- {file = "watchdog-2.3.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3fa1572f5a2f6d17d4d860edbc04488fef31b007c25c2f3b11203fb8179b7c67"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_aarch64.whl", hash = "sha256:1d9c656495172873bf1ddc7e39e80055fcdd21c4608cf68f23a28116dcba0b43"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_armv7l.whl", hash = "sha256:00f93782c67042d9525ec51628330b5faf5fb84bcb7ebaac05ea8528cfb20bba"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_i686.whl", hash = "sha256:f1a655f4a49f9232311b9967f42cc2eaf43fd4903f3bed850dd4570fda5d5eff"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64.whl", hash = "sha256:aa4773160b9cb21ba369cb42d59a947087330b3a02480173033a6a6cc137a510"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:982f5416a2817003172994d865285dd6a2b3836f033cd3fa87d1a62096a162cc"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_s390x.whl", hash = "sha256:45c13e7e6eea1013da419bf9aa9a8f5df7bbf3e5edce40bc6df84130febf39d5"},
- {file = "watchdog-2.3.0-py3-none-manylinux2014_x86_64.whl", hash = "sha256:7767a3da3307d9cf597832f692702441a97c259e5d0d560f2e57c43ad0d191d2"},
- {file = "watchdog-2.3.0-py3-none-win32.whl", hash = "sha256:8863913ea2c3f256d18c33d84546518636e391cd8f50d209b9a31221e0f7d3fd"},
- {file = "watchdog-2.3.0-py3-none-win_amd64.whl", hash = "sha256:6d79b5954db8f41d6a7f5763042b988f7a4afd40b7d141456061fa7c5b7f2159"},
- {file = "watchdog-2.3.0-py3-none-win_ia64.whl", hash = "sha256:a3559ee82a10976de1ec544b6ebe3b4aa398d491860a283d80ec0f550076d068"},
- {file = "watchdog-2.3.0.tar.gz", hash = "sha256:9d39effe6909be898ba3e7286a9e9b17a6a9f734fb1ef9dde3e9bb68715fca39"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:d1f1200d4ec53b88bf04ab636f9133cb703eb19768a39351cee649de21a33697"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:564e7739abd4bd348aeafbf71cc006b6c0ccda3160c7053c4a53b67d14091d42"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:95ad708a9454050a46f741ba5e2f3468655ea22da1114e4c40b8cbdaca572565"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:a073c91a6ef0dda488087669586768195c3080c66866144880f03445ca23ef16"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aa8b028750b43e80eea9946d01925168eeadb488dfdef1d82be4b1e28067f375"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:964fd236cd443933268ae49b59706569c8b741073dbfd7ca705492bae9d39aab"},
+ {file = "watchdog-2.3.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:91fd146d723392b3e6eb1ac21f122fcce149a194a2ba0a82c5e4d0ee29cd954c"},
+ {file = "watchdog-2.3.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:efe3252137392a471a2174d721e1037a0e6a5da7beb72a021e662b7000a9903f"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:85bf2263290591b7c5fa01140601b64c831be88084de41efbcba6ea289874f44"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8f2df370cd8e4e18499dd0bfdef476431bcc396108b97195d9448d90924e3131"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ea5d86d1bcf4a9d24610aa2f6f25492f441960cf04aed2bd9a97db439b643a7b"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:6f5d0f7eac86807275eba40b577c671b306f6f335ba63a5c5a348da151aba0fc"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5b848c71ef2b15d0ef02f69da8cc120d335cec0ed82a3fa7779e27a5a8527225"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0d9878be36d2b9271e3abaa6f4f051b363ff54dbbe7e7df1af3c920e4311ee43"},
+ {file = "watchdog-2.3.1-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4cd61f98cb37143206818cb1786d2438626aa78d682a8f2ecee239055a9771d5"},
+ {file = "watchdog-2.3.1-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3d2dbcf1acd96e7a9c9aefed201c47c8e311075105d94ce5e899f118155709fd"},
+ {file = "watchdog-2.3.1-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:03f342a9432fe08107defbe8e405a2cb922c5d00c4c6c168c68b633c64ce6190"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_aarch64.whl", hash = "sha256:7a596f9415a378d0339681efc08d2249e48975daae391d58f2e22a3673b977cf"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_armv7l.whl", hash = "sha256:0e1dd6d449267cc7d6935d7fe27ee0426af6ee16578eed93bacb1be9ff824d2d"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_i686.whl", hash = "sha256:7a1876f660e32027a1a46f8a0fa5747ad4fcf86cb451860eae61a26e102c8c79"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64.whl", hash = "sha256:2caf77ae137935c1466f8cefd4a3aec7017b6969f425d086e6a528241cba7256"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:53f3e95081280898d9e4fc51c5c69017715929e4eea1ab45801d5e903dd518ad"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_s390x.whl", hash = "sha256:9da7acb9af7e4a272089bd2af0171d23e0d6271385c51d4d9bde91fe918c53ed"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_x86_64.whl", hash = "sha256:8a4d484e846dcd75e96b96d80d80445302621be40e293bfdf34a631cab3b33dc"},
+ {file = "watchdog-2.3.1-py3-none-win32.whl", hash = "sha256:a74155398434937ac2780fd257c045954de5b11b5c52fc844e2199ce3eecf4cf"},
+ {file = "watchdog-2.3.1-py3-none-win_amd64.whl", hash = "sha256:5defe4f0918a2a1a4afbe4dbb967f743ac3a93d546ea4674567806375b024adb"},
+ {file = "watchdog-2.3.1-py3-none-win_ia64.whl", hash = "sha256:4109cccf214b7e3462e8403ab1e5b17b302ecce6c103eb2fc3afa534a7f27b96"},
+ {file = "watchdog-2.3.1.tar.gz", hash = "sha256:d9f9ed26ed22a9d331820a8432c3680707ea8b54121ddcc9dc7d9f2ceeb36906"},
diff --git a/services/worker/README.md b/services/worker/README.md
index b71a0609..dd27e6b2 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -17 +16,0 @@ Set environment variables to configure the worker.
-- `WORKER_MAX_MISSING_HEARTBEATS`: the number of hearbeats a job must have missed to be considered a zombie job. Defaults to `5`.
@@ -19,0 +19 @@ Set environment variables to configure the worker.
+- `WORKER_MAX_MISSING_HEARTBEATS`: the number of hearbeats a job must have missed to be considered a zombie job. Defaults to `5`.
@@ -23,0 +24,2 @@ Set environment variables to configure the worker.
+Also, it's possible to force the parent directory in which the temporary files (as the current job state file and its associated lock file) will be created by setting `TMPDIR` to a writable directory. If not set, the worker will use the default temporary directory of the system, as described in https://docs.python.org/3/library/tempfile.html#tempfile.gettempdir.
+
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 5e2aebcd..23b97f42 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -2828 +2828 @@ name = "pandas-stubs"
-version = "1.5.3.230214"
+version = "1.5.3.230227"
@@ -2834,2 +2834,2 @@ files = [
- {file = "pandas_stubs-1.5.3.230214-py3-none-any.whl", hash = "sha256:be0f3056fef543c37248951a9868cba77e17ace3b815b0a23c72ed2ec13fbd9e"},
- {file = "pandas_stubs-1.5.3.230214.tar.gz", hash = "sha256:5fc1fc9c470d5332f9025a53ecab4c36cb80b2c52a596d597d6c5e33b4078153"},
+ {file = "pandas_stubs-1.5.3.230227-py3-none-any.whl", hash = "sha256:bc728e1c9a7fd69b4e9487846b035dcb0cfc194df1f5c03b4b3d71988d42cc2f"},
+ {file = "pandas_stubs-1.5.3.230227.tar.gz", hash = "sha256:dbba9c2e08860f927fbb6b98722bdbe9b958a00afd4c1d1a28945f3e3333e840"},
@@ -2894,7 +2893,0 @@ files = [
- {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
- {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
- {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
- {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
- {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
- {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
- {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -3050 +3043 @@ description = "A small Python package for determining appropriate platform-speci
-category = "dev"
+category = "main"
@@ -3080 +3073 @@ name = "pooch"
-version = "1.6.0"
+version = "1.7.0"
@@ -3084 +3077 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -3086,2 +3079,2 @@ files = [
- {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"},
- {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"},
+ {file = "pooch-1.7.0-py3-none-any.whl", hash = "sha256:74258224fc33d58f53113cf955e8d51bf01386b91492927d0d1b6b341a765ad7"},
+ {file = "pooch-1.7.0.tar.gz", hash = "sha256:f174a1041b6447f0eef8860f76d17f60ed2f857dc0efa387a7f08228af05d998"},
@@ -3091 +3083,0 @@ files = [
-appdirs = ">=1.3.0"
@@ -3092,0 +3085 @@ packaging = ">=20.0"
+platformdirs = ">=2.5.0"
@@ -4723,2 +4715,0 @@ files = [
- {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"},
- {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"},
@@ -4730 +4720,0 @@ files = [
- {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"},
@@ -4739 +4728,0 @@ files = [
- {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"},
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index c6a10e8e..85fb9a5e 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -85,5 +84,0 @@ ignore_missing_imports = true
-
-[[tool.mypy.overrides]]
-ignore_missing_imports = true
-module = "mirakuru.*"
-# ^ until https://github.com/ClearcodeHQ/mirakuru/pull/633 is merged and released
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index 2d6733c4..3b6a320d 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -24,0 +25 @@ WORKER_SLEEP_SECONDS = 15
+WORKER_STATE_FILE_PATH = None
@@ -40 +41 @@ class WorkerConfig:
- state_path: Optional[str] = None
+ state_file_path: Optional[str] = WORKER_STATE_FILE_PATH
@@ -57 +58,3 @@ class WorkerConfig:
- state_path=env.str(name="STATE_PATH", default=None),
+ state_file_path=env.str(
+ name="STATE_FILE_PATH", default=WORKER_STATE_FILE_PATH
+ ), # this environment variable is not expected to be set explicitly, it's set by the worker executor
diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py
index 96e888ff..bba453b6 100644
--- a/services/worker/src/worker/executor.py
+++ b/services/worker/src/worker/executor.py
@@ -39 +39 @@ class WorkerExecutor:
- def __init__(self, app_config: AppConfig, job_runner_factory: JobRunnerFactory) -> None:
+ def __init__(self, app_config: AppConfig, job_runner_factory: JobRunnerFactory, state_file_path: str) -> None:
@@ -41,0 +42 @@ class WorkerExecutor:
+ self.state_file_path = state_file_path
@@ -48,3 +49 @@ class WorkerExecutor:
- banner = self.app_config.worker.state_path
- if not banner:
- raise ValueError("Failed to create the executor because WORKER_STATE_PATH is missing.")
+ banner = self.state_file_path
@@ -71,12 +70,2 @@ class WorkerExecutor:
- worker_state_path = self.app_config.worker.state_path
- if not worker_state_path:
- raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
- if os.path.exists(worker_state_path):
- with FileLock(worker_state_path + ".lock"):
- try:
- with open(worker_state_path, "r") as worker_state_f:
- worker_state = json.load(worker_state_f)
- return WorkerState(current_job_info=worker_state.get("current_job_info"))
- except json.JSONDecodeError as err:
- raise BadWorkerState(f"Failed to read worker state at {worker_state_path}") from err
- else:
+ worker_state_file_path = self.state_file_path
+ if not os.path.exists(worker_state_file_path):
@@ -83,0 +73,7 @@ class WorkerExecutor:
+ with FileLock(f"{worker_state_file_path}.lock"):
+ try:
+ with open(worker_state_file_path, "r") as worker_state_f:
+ worker_state = json.load(worker_state_f)
+ return WorkerState(current_job_info=worker_state.get("current_job_info"))
+ except json.JSONDecodeError as err:
+ raise BadWorkerState(f"Failed to read worker state at {worker_state_file_path}") from err
diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py
index 24b013f1..86ab375f 100644
--- a/services/worker/src/worker/loop.py
+++ b/services/worker/src/worker/loop.py
@@ -45,0 +46,2 @@ class Loop:
+ state_file_path (`str`):
+ The path of the file where the state of the loop will be saved.
@@ -51,0 +54 @@ class Loop:
+ state_file_path: str
@@ -145,4 +148,3 @@ class Loop:
- if self.worker_config.state_path:
- with FileLock(self.worker_config.state_path + ".lock"):
- with open(self.worker_config.state_path, "w") as worker_state_f:
- json.dump(worker_state, worker_state_f)
+ with FileLock(f"{self.state_file_path}.lock"):
+ with open(self.state_file_path, "w") as worker_state_f:
+ json.dump(worker_state, worker_state_f)
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index 71e39036..7237718d 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -21,2 +21,2 @@ if __name__ == "__main__":
- if "WORKER_STATE_PATH" not in os.environ:
- os.environ["WORKER_STATE_PATH"] = os.path.join(tmp_dir, WORKER_STATE_FILE_NAME)
+ state_file_path = os.path.join(tmp_dir, WORKER_STATE_FILE_NAME)
+ os.environ["WORKER_STATE_FILE_PATH"] = state_file_path
@@ -56 +56,5 @@ if __name__ == "__main__":
- worker_executor = WorkerExecutor(app_config, job_runner_factory)
+ worker_executor = WorkerExecutor(
+ app_config=app_config,
+ job_runner_factory=job_runner_factory,
+ state_file_path=state_file_path,
+ )
diff --git a/services/worker/src/worker/start_worker_loop.py b/services/worker/src/worker/start_worker_loop.py
index 311971ee..e5ea2a91 100644
--- a/services/worker/src/worker/start_worker_loop.py
+++ b/services/worker/src/worker/start_worker_loop.py
@@ -17,0 +18,2 @@ if __name__ == "__main__":
+
+ state_file_path = app_config.worker.state_file_path
@@ -19 +21,3 @@ if __name__ == "__main__":
- print(app_config.worker.state_path, flush=True)
+ print(state_file_path, flush=True)
+ if not state_file_path:
+ raise RuntimeError("The worker state file path is not set. Exiting.")
@@ -54,0 +59 @@ if __name__ == "__main__":
+ state_file_path=state_file_path,
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index 4456d4a9..49c949a1 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -32,2 +32,2 @@ def modules_cache_directory(tmp_path: Path) -> Path:
-def worker_state_path(tmp_path: Path) -> Path:
- return tmp_path / WORKER_STATE_FILE_NAME
+def worker_state_file_path(tmp_path: Path) -> str:
+ return str(tmp_path / WORKER_STATE_FILE_NAME)
@@ -51 +51 @@ def set_env_vars(
- datasets_cache_directory: Path, modules_cache_directory: Path, worker_state_path: Path
+ datasets_cache_directory: Path, modules_cache_directory: Path, worker_state_file_path: str
@@ -66 +66 @@ def set_env_vars(
- mp.setenv("WORKER_STATE_PATH", str(worker_state_path))
+ mp.setenv("WORKER_STATE_FILE_PATH", worker_state_file_path)
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index b3612a22..a6c06a6c 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -45,3 +45,3 @@ def get_job_info(prefix: str = "base") -> JobInfo:
-def write_worker_state(worker_state: WorkerState, worker_state_path: str) -> None:
- with FileLock(worker_state_path + ".lock"):
- with open(worker_state_path, "w") as worker_state_f:
+def write_worker_state(worker_state: WorkerState, worker_state_file_path: str) -> None:
+ with FileLock(worker_state_file_path + ".lock"):
+ with open(worker_state_file_path, "w") as worker_state_f:
@@ -53,2 +53,2 @@ def start_worker_loop() -> None:
- if not app_config.worker.state_path:
- raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
+ if not app_config.worker.state_file_path:
+ raise ValueError("Failed to get worker state because 'state_file_path' is missing.")
@@ -56 +56 @@ def start_worker_loop() -> None:
- print(app_config.worker.state_path, flush=True)
+ print(app_config.worker.state_file_path, flush=True)
@@ -59 +59 @@ def start_worker_loop() -> None:
- write_worker_state(worker_state, app_config.worker.state_path)
+ write_worker_state(worker_state, app_config.worker.state_file_path)
@@ -64,2 +64,2 @@ def start_worker_loop_that_crashes() -> None:
- if not app_config.worker.state_path:
- raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
+ if not app_config.worker.state_file_path:
+ raise ValueError("Failed to get worker state because 'state_file_path' is missing.")
@@ -67 +67 @@ def start_worker_loop_that_crashes() -> None:
- print(app_config.worker.state_path, flush=True)
+ print(app_config.worker.state_file_path, flush=True)
@@ -76 +76 @@ def start_worker_loop_that_times_out() -> None:
-def set_worker_state(worker_state_path: Path) -> Iterator[WorkerState]:
+def set_worker_state(worker_state_file_path: str) -> Iterator[WorkerState]:
@@ -79 +79 @@ def set_worker_state(worker_state_path: Path) -> Iterator[WorkerState]:
- write_worker_state(worker_state, str(worker_state_path))
+ write_worker_state(worker_state, worker_state_file_path)
@@ -81 +81 @@ def set_worker_state(worker_state_path: Path) -> Iterator[WorkerState]:
- os.remove(worker_state_path)
+ os.remove(worker_state_file_path)
@@ -185,2 +185,4 @@ def job_runner_factory(
-def executor(app_config: AppConfig, job_runner_factory: JobRunnerFactory) -> WorkerExecutor:
- return WorkerExecutor(app_config, job_runner_factory)
+def executor(
+ app_config: AppConfig, job_runner_factory: JobRunnerFactory, worker_state_file_path: str
+) -> WorkerExecutor:
+ return WorkerExecutor(app_config, job_runner_factory, state_file_path=worker_state_file_path)
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 21ed88ad..d97341ff 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -52,0 +53 @@ def test_process_next_job(
+ worker_state_file_path: str,
@@ -59,0 +61 @@ def test_process_next_job(
+ state_file_path=worker_state_file_path,
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index 4a3e818b..fec255aa 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -25 +24,0 @@ services:
- # image: ${IMAGE_SERVICE_ADMIN?IMAGE_SERVICE_ADMIN env var must be provided}
@@ -52 +50,0 @@ services:
- # image: ${IMAGE_SERVICE_API?IMAGE_SERVICE_API env var must be provided}
@@ -73 +71 @@ services:
- worker-config-names:
+ worker:
@@ -79,145 +76,0 @@ services:
- - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/config-names" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-split-names:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/split-names" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-splits:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/splits" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-first-rows:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - first-rows-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - first-rows-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - first-rows-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- ASSETS_BASE_URL: "http://localhost:${PORT_REVERSE_PROXY-8000}/assets" # hard-coded to work with the reverse-proxy
- ASSETS_STORAGE_DIRECTORY: ${ASSETS_STORAGE_DIRECTORY-/assets}
- WORKER_ONLY_JOB_TYPES: "/first-rows" # hard-coded
- FIRST_ROWS_MAX_BYTES: ${FIRST_ROWS_MAX_BYTES-1_000_000}
- FIRST_ROWS_MAX_NUMBER: ${FIRST_ROWS_MAX_NUMBER-100}
- FIRST_ROWS_MIN_CELL_BYTES: ${FIRST_ROWS_MIN_CELL_BYTES-100}
- FIRST_ROWS_MIN_NUMBER: ${FIRST_ROWS_MIN_NUMBER-10}
- FIRST_ROWS_COLUMNS_MAX_NUMBER: ${FIRST_ROWS_COLUMNS_MAX_NUMBER-1_000}
- WORKER_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
- # ^ note: the datasets cache is automatically added, so no need to add it here
- depends_on:
- - mongodb
- restart: always
- worker-parquet-and-dataset-info:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - parquet-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - parquet-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - parquet-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/parquet-and-dataset-info" # hard-coded
- PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS: ${PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS-}
- PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE: ${PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE-Update parquet files}
- PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-}
- PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE-100_000_000}
- PARQUET_AND_DATASET_INFO_SOURCE_REVISION: ${PARQUET_AND_DATASET_INFO_SOURCE_REVISION-main}
- PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS-}
- PARQUET_AND_DATASET_INFO_TARGET_REVISION: ${PARQUET_AND_DATASET_INFO_TARGET_REVISION-refs/convert/parquet}
- PARQUET_AND_DATASET_INFO_URL_TEMPLATE: ${PARQUET_AND_DATASET_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
- depends_on:
- - mongodb
- restart: always
- worker-parquet:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/parquet" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-dataset-info:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/dataset-info" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-sizes:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- extends:
- file: docker-compose-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/sizes" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-anything:
- build:
- context: ..
- dockerfile: services/worker/Dockerfile
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index e713b72f..e65b329f 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -25 +24,0 @@ services:
- # image: ${IMAGE_SERVICE_ADMIN?IMAGE_SERVICE_ADMIN env var must be provided}
@@ -52 +50,0 @@ services:
- # image: ${IMAGE_SERVICE_API?IMAGE_SERVICE_API env var must be provided}
@@ -73 +71 @@ services:
- worker-config-names:
+ worker:
@@ -79,54 +76,0 @@ services:
- - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/config-names" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-split-names:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/split-names" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-splits:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/splits" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-first-rows:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - first-rows-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - first-rows-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - first-rows-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
@@ -136,88 +79,0 @@ services:
- environment:
- ASSETS_BASE_URL: "http://localhost:${PORT_REVERSE_PROXY-8000}/assets" # hard-coded to work with the reverse-proxy
- ASSETS_STORAGE_DIRECTORY: ${ASSETS_STORAGE_DIRECTORY-/assets}
- WORKER_ONLY_JOB_TYPES: "/first-rows" # hard-coded
- FIRST_ROWS_MAX_BYTES: ${FIRST_ROWS_MAX_BYTES-1_000_000}
- FIRST_ROWS_MAX_NUMBER: ${FIRST_ROWS_MAX_NUMBER-100}
- FIRST_ROWS_MIN_CELL_BYTES: ${FIRST_ROWS_MIN_CELL_BYTES-100}
- FIRST_ROWS_MIN_NUMBER: ${FIRST_ROWS_MIN_NUMBER-10}
- FIRST_ROWS_COLUMNS_MAX_NUMBER: ${FIRST_ROWS_COLUMNS_MAX_NUMBER-1_000}
- WORKER_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
- # ^ note: the datasets cache is automatically added, so no need to add it here
- depends_on:
- - mongodb
- restart: always
- worker-parquet-and-dataset-info:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- - parquet-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
- - parquet-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
- - parquet-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/parquet-and-dataset-info" # hard-coded
- PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS: ${PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS-}
- PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE: ${PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE-Update parquet files}
- PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-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}
- PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE-100_000_000}
- PARQUET_AND_DATASET_INFO_SOURCE_REVISION: ${PARQUET_AND_DATASET_INFO_SOURCE_REVISION-main}
- PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS-}
- PARQUET_AND_DATASET_INFO_TARGET_REVISION: ${PARQUET_AND_DATASET_INFO_TARGET_REVISION-refs/convert/parquet}
- PARQUET_AND_DATASET_INFO_URL_TEMPLATE: ${PARQUET_AND_DATASET_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
- depends_on:
- - mongodb
- restart: always
- worker-parquet:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/parquet" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-dataset-info:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/dataset-info" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-sizes:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
- environment:
- WORKER_ONLY_JOB_TYPES: "/sizes" # hard-coded
- depends_on:
- - mongodb
- restart: always
- worker-anything:
- build:
- context: ..
- dockerfile: services/worker/dev.Dockerfile
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
- volumes:
- - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -242,3 +97,0 @@ services:
- extends:
- file: docker-compose-dev-base.yml
- service: datasets-worker
|
|
d55cbcc6f92564f042a4dafdbb3bb5e919099a15
|
Quentin Lhoest
| 2023-02-28T09:49:12 |
Reorganize executor (#850)
|
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index d5902509..ab45a3bb 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -562,0 +563 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+pytz = "^2020.1"
@@ -1284,0 +1286,12 @@ cli = ["click (>=5.0)"]
+[[package]]
+name = "pytz"
+version = "2020.5"
+description = "World timezone definitions, modern and historical"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"},
+ {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"},
+]
+
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index 72864289..522ec9a9 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -1259,0 +1260,12 @@ cli = ["click (>=5.0)"]
+[[package]]
+name = "pytz"
+version = "2020.5"
+description = "World timezone definitions, modern and historical"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"},
+ {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"},
+]
+
@@ -1493,0 +1506,12 @@ files = [
+[[package]]
+name = "types-pytz"
+version = "2022.7.1.2"
+description = "Typing stubs for pytz"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-pytz-2022.7.1.2.tar.gz", hash = "sha256:487d3e8e9f4071eec8081746d53fa982bbc05812e719dcbf2ebf3d55a1a4cd28"},
+ {file = "types_pytz-2022.7.1.2-py3-none-any.whl", hash = "sha256:40ca448a928d566f7d44ddfde0066e384f7ffbd4da2778e42a4570eaca572446"},
+]
+
@@ -1565 +1589 @@ python-versions = "3.9.15"
-content-hash = "f56e77566d857fade5ef9b4a8a6ee7bf99c5bafbc6ec625894646eeb1221e655"
+content-hash = "cfcb2f1c0c312d601ff81a829435ff17e651b1be8959b8302f336b6d6ef9d2aa"
diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml
index 141d7a4d..b238db41 100644
--- a/libs/libcommon/pyproject.toml
+++ b/libs/libcommon/pyproject.toml
@@ -17,0 +18 @@ python = "3.9.15"
+pytz = "^2020.1"
@@ -29,0 +31 @@ types-psutil = "^5.9.5"
+types-pytz = "^2022.1.1"
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 71f929de..0243d36c 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -13,0 +14 @@ from typing import Dict, Generic, List, Literal, Optional, Type, TypedDict, Type
+import pytz
@@ -174,0 +176,13 @@ class Job(Document):
+ def info(self) -> JobInfo:
+ return JobInfo(
+ {
+ "job_id": str(self.pk), # job.pk is the id. job.id is not recognized by mypy
+ "type": self.type,
+ "dataset": self.dataset,
+ "config": self.config,
+ "split": self.split,
+ "force": self.force,
+ "priority": self.priority,
+ }
+ )
+
@@ -199 +213,4 @@ class Queue:
- def __init__(self, max_jobs_per_namespace: Optional[int] = None):
+ def __init__(
+ self,
+ max_jobs_per_namespace: Optional[int] = None,
+ ):
@@ -432,9 +449,14 @@ class Queue:
- return {
- "job_id": str(next_waiting_job.pk), # job.pk is the id. job.id is not recognized by mypy
- "type": next_waiting_job.type,
- "dataset": next_waiting_job.dataset,
- "config": next_waiting_job.config,
- "split": next_waiting_job.split,
- "force": next_waiting_job.force,
- "priority": next_waiting_job.priority,
- }
+ return next_waiting_job.info()
+
+ def get_job_with_id(self, job_id: str) -> Job:
+ """Get the job for a given job id.
+
+ Args:
+ job_id (`str`, required): id of the job
+
+ Returns: the requested job
+
+ Raises:
+ DoesNotExist: if the job does not exist
+ """
+ return Job.objects(pk=job_id).get()
@@ -453 +475 @@ class Queue:
- job = Job.objects(pk=job_id).get()
+ job = self.get_job_with_id(job_id=job_id)
@@ -603,0 +626,56 @@ class Queue:
+ def heartbeat(self, job_id: str) -> None:
+ """Update the job `last_heartbeat` field with the current date.
+ This is used to keep track of running jobs.
+ If a job doesn't have recent heartbeats, it means it crashed at one point and is considered a zombie.
+ """
+ try:
+ job = self.get_job_with_id(job_id)
+ except DoesNotExist:
+ logging.warning(f"Heartbeat skipped because job {job_id} doesn't exist in the queue.")
+ return
+ job.update(last_heartbeat=get_datetime())
+
+ def get_zombies(self, max_seconds_without_heartbeat: int) -> List[JobInfo]:
+ """Get the zombie jobs.
+ It returns jobs without recent heartbeats, which means they crashed at one point and became zombies.
+ Usually `max_seconds_without_heartbeat` is a factor of the time between two heartbeats.
+
+ Returns: an array of the zombie job infos.
+ """
+ started_jobs = Job.objects(status=Status.STARTED)
+ if max_seconds_without_heartbeat <= 0:
+ return []
+ zombies = [
+ job
+ for job in started_jobs
+ if (
+ job.last_heartbeat is not None
+ and get_datetime()
+ >= pytz.UTC.localize(job.last_heartbeat) + timedelta(seconds=max_seconds_without_heartbeat)
+ )
+ or (
+ job.last_heartbeat is None
+ and job.started_at is not None
+ and get_datetime()
+ >= pytz.UTC.localize(job.started_at) + timedelta(seconds=max_seconds_without_heartbeat)
+ )
+ ]
+ return [zombie.info() for zombie in zombies]
+
+ def kill_zombies(self, zombies: List[JobInfo]) -> int:
+ """Kill the zombie jobs in the queue, setting their status to ERROR.
+ It does nothing if the input list of zombies contain jobs that have already been updated and
+ are not in the STARTED status anymore.
+
+ Returns: number of killed zombies.
+ """
+ if not zombies:
+ return 0
+ zombie_job_ids = [zombie["job_id"] for zombie in zombies]
+ zombies_examples = zombie_job_ids[:10]
+ zombies_examples_str = ", ".join(zombies_examples) + ("..." if len(zombies_examples) != len(zombies) else "")
+ logging.info(f"Killing {len(zombies)} zombies. Job ids = " + zombies_examples_str)
+ return Job.objects(pk__in=zombie_job_ids, status=Status.STARTED).update(
+ status=Status.ERROR, finished_at=get_datetime()
+ )
+
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 291373e4..3bb6a1fc 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -4,0 +5 @@ import time
+from datetime import datetime, timedelta
@@ -5,0 +7 @@ from typing import Iterator, Optional
+from unittest.mock import patch
@@ -7,0 +10 @@ import pytest
+import pytz
@@ -14,0 +18 @@ from libcommon.queue import (
+ get_datetime,
@@ -18,0 +23,4 @@ from libcommon.resources import QueueMongoResource
+def get_old_datetime() -> datetime:
+ return get_datetime() - timedelta(days=1)
+
+
@@ -286,0 +295,45 @@ def test_get_total_duration_per_dataset() -> None:
+
+
+def test_queue_heartbeat() -> None:
+ job_type = "test_type"
+ queue = Queue()
+ job = queue.upsert_job(job_type=job_type, dataset="dataset1", config="config", split="split1")
+ queue.start_job([job_type])
+ assert job.last_heartbeat is None
+ queue.heartbeat(job.pk)
+ job.reload()
+ assert job.last_heartbeat is not None
+ last_heartbeat_datetime = pytz.UTC.localize(job.last_heartbeat)
+ assert last_heartbeat_datetime >= get_datetime() - timedelta(seconds=1)
+
+
+def test_queue_get_zombies() -> None:
+ job_type = "test_type"
+ queue = Queue()
+ with patch("libcommon.queue.get_datetime", get_old_datetime):
+ zombie = queue.upsert_job(job_type=job_type, dataset="dataset1", config="config", split="split1")
+ queue.start_job([job_type])
+ queue.upsert_job(job_type=job_type, dataset="dataset1", config="config", split="split2")
+ queue.start_job([job_type])
+ assert queue.get_zombies(max_seconds_without_heartbeat=10) == [zombie.info()]
+ assert queue.get_zombies(max_seconds_without_heartbeat=-1) == []
+ assert queue.get_zombies(max_seconds_without_heartbeat=0) == []
+ assert queue.get_zombies(max_seconds_without_heartbeat=9999999) == []
+
+
+def test_queue_kill_zombies() -> None:
+ job_type = "test_type"
+ queue = Queue()
+ with patch("libcommon.queue.get_datetime", get_old_datetime):
+ zombie = queue.upsert_job(job_type=job_type, dataset="dataset1", config="config", split="split1")
+ queue.start_job([job_type])
+ another_job = queue.upsert_job(job_type=job_type, dataset="dataset1", config="config", split="split2")
+ queue.start_job([job_type])
+
+ assert queue.get_zombies(max_seconds_without_heartbeat=10) == [zombie.info()]
+ queue.kill_zombies([zombie.info()])
+ assert queue.get_zombies(max_seconds_without_heartbeat=10) == []
+ zombie.reload()
+ another_job.reload()
+ assert zombie.status == Status.ERROR
+ assert another_job.status == Status.STARTED
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index 2a0769e7..76b6602d 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -641,0 +642 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+pytz = "^2020.1"
@@ -1378,0 +1380,12 @@ cli = ["click (>=5.0)"]
+[[package]]
+name = "pytz"
+version = "2020.5"
+description = "World timezone definitions, modern and historical"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"},
+ {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"},
+]
+
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index 900cf425..4fae0e53 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -661,0 +662 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+pytz = "^2020.1"
@@ -1511,0 +1513,12 @@ cli = ["click (>=5.0)"]
+name = "pytz"
+version = "2020.5"
+description = "World timezone definitions, modern and historical"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"},
+ {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"},
+]
+
+[[package]]
diff --git a/services/worker/README.md b/services/worker/README.md
index 38e6d783..b71a0609 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -13,0 +14,2 @@ Set environment variables to configure the worker.
+- `WORKER_HEARTBEAT_INTERVAL_SECONDS`: the time interval between two heartbeats. Each heartbeat updates the job "last_heartbeat" field in the queue. Defaults to `60` (1 minute).
+- `WORKER_KILL_ZOMBIES_INTERVAL_SECONDS`: the time interval at which the worker looks for zombie jobs to kill them. Defaults to `600` (10 minutes).
@@ -14,0 +17 @@ Set environment variables to configure the worker.
+- `WORKER_MAX_MISSING_HEARTBEATS`: the number of hearbeats a job must have missed to be considered a zombie job. Defaults to `5`.
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index e113ba65..5e2aebcd 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -785 +785 @@ name = "coverage"
-version = "7.2.0"
+version = "7.2.1"
@@ -791,51 +791,51 @@ files = [
- {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
- {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
- {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
- {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
- {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
- {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
- {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
- {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
- {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
- {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
- {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
- {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
- {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
- {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
- {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
- {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
- {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
- {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
- {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
- {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
- {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
- {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
- {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
- {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
- {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
- {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
- {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
- {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"},
+ {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"},
+ {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"},
+ {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"},
+ {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"},
+ {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"},
+ {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"},
+ {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"},
+ {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"},
+ {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"},
+ {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"},
+ {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"},
+ {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"},
+ {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"},
+ {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"},
+ {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"},
+ {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"},
+ {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"},
+ {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"},
+ {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"},
+ {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"},
+ {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"},
+ {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"},
+ {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"},
+ {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"},
+ {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"},
+ {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"},
+ {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"},
@@ -1887,0 +1888 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+pytz = "^2020.1"
@@ -2236 +2237 @@ name = "mirakuru"
-version = "2.5.0"
+version = "2.5.1"
@@ -2242,2 +2243,2 @@ files = [
- {file = "mirakuru-2.5.0-py3-none-any.whl", hash = "sha256:20c440f07e11589735e2634bcf5ab9df0e9f901a1a18e323e15e0200112499fc"},
- {file = "mirakuru-2.5.0.tar.gz", hash = "sha256:4484f9405610886de3eb1bb1405b3ccd9b0d88b3ea4de68ee7873a873c11956b"},
+ {file = "mirakuru-2.5.1-py3-none-any.whl", hash = "sha256:0a16f897841741f8cd784f790e54d74e61456ba36be9cb9de731b49e2e7a45dc"},
+ {file = "mirakuru-2.5.1.tar.gz", hash = "sha256:5a60d641fa92c8bfcd383f6e52f7a0bf3f081da0467fc6e3e6a3f6b3e3e47a7b"},
@@ -2866 +2867 @@ name = "pdf2image"
-version = "1.16.2"
+version = "1.16.3"
@@ -2872,2 +2873 @@ files = [
- {file = "pdf2image-1.16.2-py3-none-any.whl", hash = "sha256:1469335050a17657f94c2f1ef3a23e57807d631ad5bcbaec997c2c42a8186f4a"},
- {file = "pdf2image-1.16.2.tar.gz", hash = "sha256:86761091eee35f4641ea98dfddb254254361d018be698a199aff7c1d37331803"},
+ {file = "pdf2image-1.16.3-py3-none-any.whl", hash = "sha256:b6154164af3677211c22cbb38b2bd778b43aca02758e962fe1e231f6d3b0e380"},
@@ -2893,0 +2894,7 @@ files = [
+ {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
+ {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
+ {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
+ {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
+ {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
+ {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
+ {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -3850 +3857 @@ name = "pytz"
-version = "2022.7.1"
+version = "2020.5"
@@ -3856,2 +3863,2 @@ files = [
- {file = "pytz-2022.7.1-py2.py3-none-any.whl", hash = "sha256:78f4f37d8198e0627c5f1143240bb0206b8691d8d7ac6d78fee88b78733f8c4a"},
- {file = "pytz-2022.7.1.tar.gz", hash = "sha256:01a0681c4b9684a28304615eba55d1ab31ae00bf68ec157ec3708a8182dbbcd0"},
+ {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"},
+ {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"},
@@ -3912,2 +3919,2 @@ name = "pyzstd"
-version = "0.15.3"
-description = "Python bindings to Zstandard (zstd) compression library, the API is similar to Python's bz2/lzma/zlib modules."
+version = "0.15.4"
+description = "Python bindings to Zstandard (zstd) compression library, the API style is similar to Python's bz2/lzma/zlib modules."
@@ -3918,105 +3925,91 @@ files = [
- {file = "pyzstd-0.15.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:da9907e447d41650c00b8023be6789f7c2133eca3c6a0f72200ff25df29c0bf5"},
- {file = "pyzstd-0.15.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:fe24509154d3d7c16195fa2c0b546e160c947b09fd49ca08702abdcc5bc0c933"},
- {file = "pyzstd-0.15.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:db93acdde600fd91ef329300de3d03d36d5dcb99533d69c85a0b58c27a0f4e53"},
- {file = "pyzstd-0.15.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:946eefc64d3a4d756920986dcc5043b7fdef179fa2f9dcb1a77dac2821abd934"},
- {file = "pyzstd-0.15.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8a316c14ba55191a2eb32b04b7e1468fcac73e5fcd287b189650e1238dec9183"},
- {file = "pyzstd-0.15.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:082b08c1c0e8a9441e5f48c0f44fe4cee201c682f730c838ef86ed8619cda760"},
- {file = "pyzstd-0.15.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3ef2ef67b2e54168d303f95f074f21e3e98e491f9d700ec2fa3266f6c0c71b5e"},
- {file = "pyzstd-0.15.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:5bb9e6581db70e99973bb4267014e2dd48ed271d6ae3581fc92f73ad82f36dc1"},
- {file = "pyzstd-0.15.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:92894764d892d50a076607848fdbac4d011fcd35fc73eaedb93bab4442e502d7"},
- {file = "pyzstd-0.15.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:08d1b45aed0df32b07ba1d1c49fe892d6fef4d4c97d1133d8836350a78a93244"},
- {file = "pyzstd-0.15.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:22e38237ace69bb6810734da1a3c2dd267fb5d7f68a35e175565bdc7ce6bb04a"},
- {file = "pyzstd-0.15.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:fd98100fce2d6eb5ad3f7292c3498986e37397d50c6e3b04adc3c91c3d26bd9d"},
- {file = "pyzstd-0.15.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:49cd5bb0220f16bf41b80a0301f63fd4f5713cdc1795919891f8904f82f31768"},
- {file = "pyzstd-0.15.3-cp310-cp310-win32.whl", hash = "sha256:8218f953895e6d43a789b9a53a4945531d1ad4e76b64ac2d88e0a8980cc1d9e4"},
- {file = "pyzstd-0.15.3-cp310-cp310-win_amd64.whl", hash = "sha256:4af74993d8eb032105b7640c4e4af161f8093447a62de3f2c9f14493576a95e7"},
- {file = "pyzstd-0.15.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b7841e3bcfdf18818bb3408b151351a51440895525d62a55f157586a55032e40"},
- {file = "pyzstd-0.15.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:a381a0d1f3e3fdb93460ae93c5c1057d894d436c5d97b2c2bf8f015d28632afe"},
- {file = "pyzstd-0.15.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b087a31541d81d0fc5769f95dda8c5a45e50d9c67098c643d5cd55d565b6441b"},
- {file = "pyzstd-0.15.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a91a8d60a3a42e920d3ed02b327010c670ce83c6c79206657af3f6a61dde418c"},
- {file = "pyzstd-0.15.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e11bdfe4478449244eb2d677c8881a80ae94ae08cded8051e82c73d6725fde17"},
- {file = "pyzstd-0.15.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:23cad4e6b6760de73f952312c0770069876358122a7e8e296183d833cbf465c5"},
- {file = "pyzstd-0.15.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d1631cc546c30da82b4bfb07bfc53aa46ce765800c4c839aabdd9df0f49c6bf6"},
- {file = "pyzstd-0.15.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:35f1aa7c87b613a09fba9b6cfc1b6fbeddeee8fd1b3ba25facabdb53fa1b17fe"},
- {file = "pyzstd-0.15.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:49b27434e7d247a8326713f4a30d8d2447120e5f8b523400df1b5274b6a721e6"},
- {file = "pyzstd-0.15.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:9253059f2ae2721405e9f45b34f907ab29f6e671e2bfda1593c3114a46673bed"},
- {file = "pyzstd-0.15.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5367743837ca7f46fbbdbb0faafc3e99b22bb6132fe78cf40892b8ba10367b5b"},
- {file = "pyzstd-0.15.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8d7af343043d27330caa915a942fc6579b7360d5fce7a1065476c846f9988a0f"},
- {file = "pyzstd-0.15.3-cp311-cp311-win32.whl", hash = "sha256:df7c22c019249031da18ca350e087c8357576cfaf2970be6cc6e5b9604a4255f"},
- {file = "pyzstd-0.15.3-cp311-cp311-win_amd64.whl", hash = "sha256:b543ae7a2449caa96fe4427fb83e0b004a9f4ca9fd943edf8296a73dfa7c0a69"},
- {file = "pyzstd-0.15.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:b68c2f51a8c6dd9bc66e7bba8e59e99c2a91112ec75c18e53a880b2dbc6e8e68"},
- {file = "pyzstd-0.15.3-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3b0576ce2165a3a95b222e6514013105218d56b81857a1b694514eb63fbbdc5a"},
- {file = "pyzstd-0.15.3-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a6fe2fad80743e60f969b02f9ab056f8a005974d5c34f1a9b3eca1df8f56b756"},
- {file = "pyzstd-0.15.3-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5ae0e5532bb41e830c257dec2abfe94bf8ab09afebe7ecf710d6d3cfa35d6aea"},
- {file = "pyzstd-0.15.3-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bd5eee632a69c8c0ab78215cae44f9944d906771622564f2a90fd7374739eeb5"},
- {file = "pyzstd-0.15.3-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:dadd0cbeda15cf89abcd814b5478f1f17e22444113d35428cd62d0b651a35a19"},
- {file = "pyzstd-0.15.3-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:36754d70db264943ad9cb21a5130d3bce9d62ac98a645a2d90adfb9cd548eb21"},
- {file = "pyzstd-0.15.3-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:761bcbf2aa04fcf33fee3e114d832b0858dfefb6824ce585757e7b793a0b2deb"},
- {file = "pyzstd-0.15.3-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:3e7fb44fd42abf6db9850f27662763c52f2f282851c7a5af790655cf593016ce"},
- {file = "pyzstd-0.15.3-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:b81786383dcc62e1bc40055e14f268a6bea5818b63efbfb4514083e91f3ba111"},
- {file = "pyzstd-0.15.3-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3e073dbbb16c2c815299037134c697f044bae142ca02142a10cb72de874580ea"},
- {file = "pyzstd-0.15.3-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:04cbf37ac09ca272143dff90d54f4856869bdd3b40eb262f625e4cc785efdd3b"},
- {file = "pyzstd-0.15.3-cp36-cp36m-win32.whl", hash = "sha256:4d7e876fea4ded82233c2fc2df4c56b00433db351bb21f401507e7dea7c16819"},
- {file = "pyzstd-0.15.3-cp36-cp36m-win_amd64.whl", hash = "sha256:4fb7d0267a025509b22c1eaa4110563aa60ca27ca4cab24e3aac7a8770ce944b"},
- {file = "pyzstd-0.15.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:287b3f15d1f473674b3689fda5c7143b396d5dd53360b450560823485dbfdd8e"},
- {file = "pyzstd-0.15.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b840c7056759da13f8634856945f2b6855335b7e759ee27003f4c42c57676d8"},
- {file = "pyzstd-0.15.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d1310deeb8af68a0c1b32b069776b4352c7e5f2d8ac60f79955606c49a9852bc"},
- {file = "pyzstd-0.15.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30e58666f97cbfea43e12d93b482a86d1e79771609dbb8f095d30a0cbb69d0d6"},
- {file = "pyzstd-0.15.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:66f9de313d1ba7746d61698c14573f597b5a9d562041828139a3eecd62efa240"},
- {file = "pyzstd-0.15.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0c84b349b4dbabbc1e004b80b1cfcb8dc78442c10a81636dfa9ee94c028ed9b"},
- {file = "pyzstd-0.15.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:3e8e11d7503f48a6d46d5962c953f17f12b7e001af9c64d58d3ab195981066cc"},
- {file = "pyzstd-0.15.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:d7204054567bd41e57f10f9134c4210edbb9eab9cea55e9081dd388461b2c794"},
- {file = "pyzstd-0.15.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:eb27499fab77ce5838d42067a964b454b5784913e6fa0e1e6841e3b183c11154"},
- {file = "pyzstd-0.15.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:23903d2da8b650358ce67c4a2125e8d1d9a7c9ebf959011832dcb2779f7fb51c"},
- {file = "pyzstd-0.15.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:d41a41e81f66002eed7e0df49ee2893b41068c1866612f59fe2751823a1c650c"},
- {file = "pyzstd-0.15.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:86f56db8a082da130d1ca67e9181bcf42deab75527b3f2a35e5e6144f3f0691a"},
- {file = "pyzstd-0.15.3-cp37-cp37m-win32.whl", hash = "sha256:a10ef9ab262f117a379158cd2ff262caf48ec4e35f54554a971bfa698a33a530"},
- {file = "pyzstd-0.15.3-cp37-cp37m-win_amd64.whl", hash = "sha256:48e81e5e4f315164790163ff503b0dce7b4ad519cc954215033c683d0fd0f9cd"},
- {file = "pyzstd-0.15.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7fc780a067b3754b913268481aa0bd9d80cac1d2a9c1e1c7abb7102ab4726903"},
- {file = "pyzstd-0.15.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:db69b7bf80935d3c3da0dff4000e8a94f0224f98c312914190af79932ae421d5"},
- {file = "pyzstd-0.15.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ed0826eab6ab133b8f8bb0369d76e546dad70a94b372b6d6351ba8320ec33615"},
- {file = "pyzstd-0.15.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0faf892a3f20258da72fd83ad0b394e8ebcbe3a637735870528529f3aef3c676"},
- {file = "pyzstd-0.15.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4e4694db47c8816c499d7d4240abcec5154a227f454a30041de5632faef11a41"},
- {file = "pyzstd-0.15.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:539f0157f2283e395a503022aab915a9f1577fd97d92ed27b85adceeaea3d24c"},
- {file = "pyzstd-0.15.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e887757277409fd02535d24dc0bfc48aa3b8c1990b0451dcb5157776c64cf0d3"},
- {file = "pyzstd-0.15.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:650fce9212410fdc82f1cb32d562e89f6dd1480d1cdbc0769e09235e236317c2"},
- {file = "pyzstd-0.15.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:17de64690976caf4a355da8a9b06d6ad55b424899e7cf305c6b08b96c8b764f4"},
- {file = "pyzstd-0.15.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8cd3792aa54a6c4933d2b63e90252220d8f8347e424f39c5eaec10f3bc415f10"},
- {file = "pyzstd-0.15.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4e92706f6f579d78768f942cde4359195fc2750e58c4bf3c1c91929693e10fd0"},
- {file = "pyzstd-0.15.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4fa699945cbd1316657550c00e1fa998c1ab6df5e0aff60254b0eb768be38003"},
- {file = "pyzstd-0.15.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:c3604d118493f2c09d387f9e569c2ebc71f07be148f57397bd485773945f192f"},
- {file = "pyzstd-0.15.3-cp38-cp38-win32.whl", hash = "sha256:11bcf59b869abc10cf7cd872bd3d113642c94e92a5b68fe990154945096f8c4e"},
- {file = "pyzstd-0.15.3-cp38-cp38-win_amd64.whl", hash = "sha256:3ad35fd4de3591d8c538fe1fc4192a5cfc8715727dd9d7bedf6aceae67ff3408"},
- {file = "pyzstd-0.15.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:041bc30b7c47e52ee418786fa806fbe42094f990353d3e685a9c96ed6a4d2212"},
- {file = "pyzstd-0.15.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a08887c9ea43f5b08f2c33dd92ffc8a26afb9d9e23e8cebc962bbba134719f3c"},
- {file = "pyzstd-0.15.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e33e9367f32c2422bbf1a33a4e965e5e2d076eb4042f97971b6acd19c0a16ae6"},
- {file = "pyzstd-0.15.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0134944e00345f89657716aca9a1d2280bef69aca7a0cd326dd10d33f3caae78"},
- {file = "pyzstd-0.15.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:67a9f383ee8054a72e7389ba51b131cd5acf26c3c8137e45a460d30d350da3ac"},
- {file = "pyzstd-0.15.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:66ffba0844b84b742764455244e582f24a172390d8e1f479900fa549b2acc96a"},
- {file = "pyzstd-0.15.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0bdedc13c0d67aaaa706310eb41248fb78e7bd3fdf335d2de8fdeb2d71574645"},
- {file = "pyzstd-0.15.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:9ddc5a9ef8b09241fff58bbcb780bffaf85437d29ce516f2ac522c3c6d9f5fee"},
- {file = "pyzstd-0.15.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:2b368e6f601824401d3f5e5f78319bb09b0d6d1c0d23175f71b82739de9d2218"},
- {file = "pyzstd-0.15.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:1eb94243194db49c8d1d7ffdc51982d88459cb74b4ac5a6ecd64313a93927cf3"},
- {file = "pyzstd-0.15.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:9d1688fc5cd6c32cf648e6e86162b5f2a9bddfc317deb19893c0d53fa15145f4"},
- {file = "pyzstd-0.15.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0a4e098058d8262f33ab550eed3824bb9f044a62120c17f0bf886529b32bf1cc"},
- {file = "pyzstd-0.15.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:089e70d146d3d37cd1c8adbdb8700311752a2d3ad42323037c5fb4032a00f7f7"},
- {file = "pyzstd-0.15.3-cp39-cp39-win32.whl", hash = "sha256:6e486d38fd247fdecde5bafe4af47a6e583c46c0a0c34c098e4d8a291603a2f8"},
- {file = "pyzstd-0.15.3-cp39-cp39-win_amd64.whl", hash = "sha256:728be92bc42bdccfecef88fc93a56e6ea561919fe9e00a8ddccde644dc1ecc53"},
- {file = "pyzstd-0.15.3-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3cfe3961fef371f616255d36c5629b421ea1adf6eed341cc64223e84d544429f"},
- {file = "pyzstd-0.15.3-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:106928b5926ead3cee7a121d50568ffaac89966e31a061f3faa2ec2c9dad8904"},
- {file = "pyzstd-0.15.3-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f8378d0f0726d12bf01f52e2448725236b98b2e629e4b1183433274213eb576"},
- {file = "pyzstd-0.15.3-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4695713ced2d1b3e34ffbe644c9bd855e5eceb85d6bff6b113302a2878951e2b"},
- {file = "pyzstd-0.15.3-pp37-pypy37_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:6fb1866aead975ff17c8094819f361b015704a0fb01468b65b5a82d2686b75d1"},
- {file = "pyzstd-0.15.3-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:992823978523ee3107cc75ea5ed49907212e04dd4beb0f2e5b22587c8ed9e395"},
- {file = "pyzstd-0.15.3-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:10edcb025eade0b92ecc3d801094b0511b5484c78cf43ac9b68be7d27710ba77"},
- {file = "pyzstd-0.15.3-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7bd4c60832749e34b23bd2e652d233c0283cff71a68f54f015f12bd682b780c4"},
- {file = "pyzstd-0.15.3-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1bb599d2a4275f5708216fd701756b956233f4cccd576bc3e10f7114e69779c2"},
- {file = "pyzstd-0.15.3-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:71ad945b118ef231b0e7475ed998c9b4b62af8964e73510b66a2a71fbd977109"},
- {file = "pyzstd-0.15.3-pp38-pypy38_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:d4f972628947451154285a460aad40626301b269b949f205467a1947003583f6"},
- {file = "pyzstd-0.15.3-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:ba2775501b126a3edec424a29d2afabdd6e65b36991c404ec29cbde713b1cfb0"},
- {file = "pyzstd-0.15.3-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:d265e6c92fda25059452c604fa407c35d3a6ae51416b874c37f7c7bbccc4c1c7"},
- {file = "pyzstd-0.15.3-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0ef8e9886c96d59d9357a30b06862fd29887c9de8652454de4cc5d021d706ff9"},
- {file = "pyzstd-0.15.3-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:73dfdd5b9ceea88b53ae2896054bc6b1e6e7e5d4c04b9a4a8c800d85a6b62056"},
- {file = "pyzstd-0.15.3-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:33e94fbccea132044ffbd3523a376c1de5afb521ecfd54f44de4ac8d3681dddb"},
- {file = "pyzstd-0.15.3-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:c0fbef0dbe12a34b8c1eb980c18ea39c432565a77922bc692eeb666fa77fe97b"},
- {file = "pyzstd-0.15.3.tar.gz", hash = "sha256:ac4edab5d3955343e8f7f287e62cd2882907d46bcba4b406a1e9f84aa2887472"},
+ {file = "pyzstd-0.15.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:7aea5a1474cacdd0285f16bc5392271645147806986e17be8b100ef750520548"},
+ {file = "pyzstd-0.15.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1e53d12e026605b254569a384191850b6e9f2fcd0fba8b3c80e09b9683fc86af"},
+ {file = "pyzstd-0.15.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a6250aa075408753cde58d79a80923c29c1791f32c4b3e58f25701ae5dbf1678"},
+ {file = "pyzstd-0.15.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ca8f520618a3a1c66ee0f6bc3a6b4fc6f9d55b7b1ce80a8f8a422c82a2f79418"},
+ {file = "pyzstd-0.15.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:bca4d7285363b73966e8c6d8af8da515d23bcaf8eb91293e866d5d4615ea5227"},
+ {file = "pyzstd-0.15.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9aed2fda29875d42a7f96b741582e6e9adb48f3903760439f60b0f8fb8f822f1"},
+ {file = "pyzstd-0.15.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:448d4412462d00bbb7c4857f9a4bedc879f9834f01e2723c23f9de4edde8bde3"},
+ {file = "pyzstd-0.15.4-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:8a7f8be2493356bf55a1261b7a9bb476324738ced6210c3f876f452df8288785"},
+ {file = "pyzstd-0.15.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:15b84a04edb257cc7d75c208be43dc4c66e8bd2a44ad9e2af523a829386f0e4c"},
+ {file = "pyzstd-0.15.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:e1e7e43714b0646fb1d61b6b9bbf6e6ede6c77c89667f2addb086f75196ce389"},
+ {file = "pyzstd-0.15.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:29f692a74f3d9778c71ad6899a1c7578fd44dfb5b6187fe37abdfc7cc034e9d3"},
+ {file = "pyzstd-0.15.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:c2e35d5a169ef9107760bd276974331e0be8fafe2e26ec824020c4999013dc40"},
+ {file = "pyzstd-0.15.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:ab88c8123e2c8ad298d1ac79f5b5aeab68ef27a0afe968e5c5be1a2f24577e2e"},
+ {file = "pyzstd-0.15.4-cp310-cp310-win32.whl", hash = "sha256:2e25838c7410245201a455ff000372c65654c781a7a168b2249432dae4ce260f"},
+ {file = "pyzstd-0.15.4-cp310-cp310-win_amd64.whl", hash = "sha256:b28e6f095fd56ac373844ee76ab74b011dbd128e4113a033911675421742ce91"},
+ {file = "pyzstd-0.15.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2988e5282c807634911d9336ff4797de84e308d94a1f57bfab4223d927e992c5"},
+ {file = "pyzstd-0.15.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:00c9e9aa800b7ded58bdddd90930f949205475c17d1e5923ab936321b805ede0"},
+ {file = "pyzstd-0.15.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9c681938be80e4ceedc2f43b5d7c08e8e3719492a8250cc742382633c7176919"},
+ {file = "pyzstd-0.15.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:65dc2ac46389823214dd21141a05009cc2fc8d139d8a0aae5f7eb6830ffe5ded"},
+ {file = "pyzstd-0.15.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:91cbd6d8a2fa0b18f9b50076a46e837b90e9f1974578483fcc4b094b43cbd89b"},
+ {file = "pyzstd-0.15.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:46e95ec6b499571eca1683c078d2f6cfe3815a5103bf4f488a4edbb1607be652"},
+ {file = "pyzstd-0.15.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d1da1bcbac16a075675609865e2705d9ec90a2312450d4d99b65a003fdef84f7"},
+ {file = "pyzstd-0.15.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:55dd0cf038e9053147e58fd4d0509fc1b735f63ce276b88151560bae271a25af"},
+ {file = "pyzstd-0.15.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:7740d39f74f9154a7f42d15c1b16f87ba062ac16439d62525f2c85165ea9ebc2"},
+ {file = "pyzstd-0.15.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:c03e1bfc6c391a05ea04dfff485f4884b07805a16f198c1d5b4e24f4df1e1723"},
+ {file = "pyzstd-0.15.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:e7361f4e5a354715d00c43ad3472188cb9eab83d2e75a2ab7ca75734779579bf"},
+ {file = "pyzstd-0.15.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:56cb56e24e9d33ef571c23b0157593c44c052ffe5810599a76bc6c60ec77f67a"},
+ {file = "pyzstd-0.15.4-cp311-cp311-win32.whl", hash = "sha256:ab55575f1c6c12ca65918dba8196bfb281df86038a9f9e376fb6ac5a501ac724"},
+ {file = "pyzstd-0.15.4-cp311-cp311-win_amd64.whl", hash = "sha256:b73b18d8ebf5caba863bfccb6ad422b9322c7a806b50a555bb44f51f16b7af5e"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e3b4ee242e00029c68fc5d8cddc3e01f6cdf7ac466d844a1440ebde4ae1f9e79"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d62f3e5ca97150ddd32e83275170bcb7ea462f271df03d39de434e8e02889b09"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:01e723cafb6e681dccba134524f9f012c5f04078d6cb05e03b2f5d871716e0db"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:7da76175bbbec45ecd35e611eaa50e40c0f29a05e86a8cd50b1b20b5a52cd218"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d6f2bbd863e674875882de8dd6f54ba98916cfeb89d9224393e9846c9e9aea1d"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:23c07257fa5df97ebde0a01a54e7c4ae295dc684908f8b1c4ac2f407b881d91b"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:b5f921cf4ba10cf2f1dbaaba3b9e7afe876a0962cf43f16455c4e5052c1dc0bd"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:c2c6f976e45e63a4f6cbaf20240bb0842ed2ed65703718d54a6fad8642d3c230"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:60079134dd0967fc4c81ea50a7deedd99c52bb263a46750326cba9f53b03f93c"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:55a368a2195bf9cabf8f20ea2841c6a7463c21d8345c8c7d50b809635fec262e"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:78a5e30f90f499f6961e56263e8f029e228b92550aa4e8a5649ed220f8f19230"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab21cc20922de335feef32245a9329bc3cb443346dcda243a711aab15bc130cc"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-win32.whl", hash = "sha256:0d742edd7340e4b8ba4cbe240e873bbc88c197c0b3838212eb5c4885094f191b"},
+ {file = "pyzstd-0.15.4-cp37-cp37m-win_amd64.whl", hash = "sha256:39fa99cac5abe0cdb34afda54f54735593646d8c57de1b734d1fe3c9761e0575"},
+ {file = "pyzstd-0.15.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:2146d6ab27f3b5dd02b093d748151897e3c929ec009768f20b7a5f3627de7d9c"},
+ {file = "pyzstd-0.15.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7422e119a9dd10695dee636707881fa6e1af872df5a4f5f1ae6feec34e3cdab6"},
+ {file = "pyzstd-0.15.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e9074eff37261effb5c7312911e11f0a4c26475ab55bae803d3b095fa291e631"},
+ {file = "pyzstd-0.15.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:aa9ab4ca7e32c219ad0d420786897cdb3c09f1ffc3fd1a2785e7652ace5684e2"},
+ {file = "pyzstd-0.15.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:452e1e774fd7e693f2f5065de61f8cf42eb449ea5614e4cd15ae458e70977d98"},
+ {file = "pyzstd-0.15.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27e1b4a3a6fdab664f1d415c3bbc63700a85f8ec46ad2bf7287e6a86bbea0581"},
+ {file = "pyzstd-0.15.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21b7f72717d58b39487c903854556fc363be3329b94aa696e9111291a13d714c"},
+ {file = "pyzstd-0.15.4-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:eb300366e9bf62776ef8c5c5fca57d5b18fdf6a62d055244e9f9f71ea263597e"},
+ {file = "pyzstd-0.15.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:1e138c585fc2d07fb20e61f6cbac77fdc77497299199a610a1fca28933128948"},
+ {file = "pyzstd-0.15.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:242467cc26e2e4b344028887c934fce3e48354e5eddbc4bb4e695f2c011d3755"},
+ {file = "pyzstd-0.15.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:929f894d0b1439826f45a248d7569f8c9ceb43ff3d5739e7ac9ff264acdcb070"},
+ {file = "pyzstd-0.15.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:e96731ecfff34aeefb5e0f93222c9831880e8a94ea700f4432d4ee45995badc1"},
+ {file = "pyzstd-0.15.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:f437e316cd24f21ea13ad635d7e24550458217af75182554963f461bae0bc189"},
+ {file = "pyzstd-0.15.4-cp38-cp38-win32.whl", hash = "sha256:912345726aa45dba8fb4930e172149f558caf606bf22c219cde9deb984a1523a"},
+ {file = "pyzstd-0.15.4-cp38-cp38-win_amd64.whl", hash = "sha256:2589c484c59a0ef2d52d1fd7ab3efd648ef38a396ef22b892c79c2cb878487e3"},
+ {file = "pyzstd-0.15.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:33d42c56fde20a2a98bd200d91d09e44957ce782ffeef5041fc4c544df7640d1"},
+ {file = "pyzstd-0.15.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:72bf7435dad00fe8a24815f67fcd9fddf44265aa8155a507ac79472cd13c97c1"},
+ {file = "pyzstd-0.15.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d1bcfb77bf00ab5e134321ddf9123845c80b1f49500d499998a17881aa644e28"},
+ {file = "pyzstd-0.15.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4e18f986325eedf0b5d86feb87fad70a0430e3da0bff1a1c33ddce8c70fc371e"},
+ {file = "pyzstd-0.15.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9860aeb47f88e61f8bd615f4f3e8372990c8e680c52d8a2f893157ebdb15e8a2"},
+ {file = "pyzstd-0.15.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6be764774531007ce83c32d24379c32474329655bcc177f1b154e2d2e7e4dba1"},
+ {file = "pyzstd-0.15.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4b41ca76a847286cd54ebabbfd37fa732a68dc2daf5178aede34ee377598cfa5"},
+ {file = "pyzstd-0.15.4-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:1659da1d388a85c8297083772fdd16543a43305ec0213f1d6f5a6b65702c8402"},
+ {file = "pyzstd-0.15.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:94302b3a36c562a31f791a5e0112bcc129e0245c6f462d86627a4cefda0f5764"},
+ {file = "pyzstd-0.15.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:0820d84c272865e996696fc84c8c63cf6898d66bfda8ac4109880e24272ac13a"},
+ {file = "pyzstd-0.15.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e111d203153837fca01a3c6ed48167af4b590f148998a2d8edefd98338615006"},
+ {file = "pyzstd-0.15.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:67ee34825be593151ba0da78d6914914afce9a7ddc77683319d9406bf422c578"},
+ {file = "pyzstd-0.15.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:7048a34197205cb4459492a1d397059ec35a758e29cee226ffa8e593148300af"},
+ {file = "pyzstd-0.15.4-cp39-cp39-win32.whl", hash = "sha256:cb03d3fca1fab2b33b4d0de9b80ab0efc9bb872eb28575eeb73af0d9aac08322"},
+ {file = "pyzstd-0.15.4-cp39-cp39-win_amd64.whl", hash = "sha256:b06eabacbefaf440785ee1f499582c79083822fb3d2640fb74dd5d3391bbc2ae"},
+ {file = "pyzstd-0.15.4-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:649f23ab5c203dade1594f593c6b65247e848f99ba5b11ef26b74541e1f0b605"},
+ {file = "pyzstd-0.15.4-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c173226b2fab75bfcf9c0315d551e9b484e50f4d84869676bcf33d080f5b43f8"},
+ {file = "pyzstd-0.15.4-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d774cf6406c33c9af6e1fde10a1b0aab7e1f573b2510c491ed24c6f4dbad677"},
+ {file = "pyzstd-0.15.4-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f4d47ea4a9a3a43a556da63ee59bf941aecfb529bb85957960e91339c6d6cf3d"},
+ {file = "pyzstd-0.15.4-pp37-pypy37_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:be71183e8e7a03b298d4e547f58f68ee1664889bef4e7460464fdf1e70349b66"},
+ {file = "pyzstd-0.15.4-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:74750bee4fd2b584bf502cdc03861c9decb0875a7febcb2d629cf72a0470188d"},
+ {file = "pyzstd-0.15.4-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4bf88010b15470f00810e8fc3cb46608af7dceb0d8ca696552d635be652bdfd7"},
+ {file = "pyzstd-0.15.4-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5ec07186d5150a6129327e906c0c1f15fd505d4f5538957b2481913a959121ee"},
+ {file = "pyzstd-0.15.4-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3e11b0534f958668dbcac6d3d943af93d1f5c4dfa7857147589398b405a9ee6"},
+ {file = "pyzstd-0.15.4-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bf63ad569ebc08fee48bc1080c1eb98c9620af4d52ecc56b2a54d6f01797d7be"},
+ {file = "pyzstd-0.15.4-pp38-pypy38_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:88e9f435a68ab9d04a4a90d5b11c587080295ab4057a403730990944b98d0b4e"},
+ {file = "pyzstd-0.15.4-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:9b433e0495b0231c20733b4649d33c5ab51f3670db4b47e0964b0b4dd240e03f"},
+ {file = "pyzstd-0.15.4-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:e884faae04b2f52cc6aa5aa18de2631ab75786440fd8a708f6b886d75aade937"},
+ {file = "pyzstd-0.15.4-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:229626e1ee592fd8a922935b997e9b91b8b8b874a72565fe1839041cf3ae1a54"},
+ {file = "pyzstd-0.15.4-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8576440b3212d648fbff49d1127e2bffe0f74aa8be40c3ec8b783f276135cb30"},
+ {file = "pyzstd-0.15.4-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ddfbaad47edf70932a6e0fba37348ca240a34408e674c056ff16631fa07ea5fd"},
+ {file = "pyzstd-0.15.4-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:be4e4bc65de7093e4781d98016b15224311f52eacb220f67b1cbffcae3ac5497"},
+ {file = "pyzstd-0.15.4.tar.gz", hash = "sha256:de07ac54f57642f186732075cdce2be3d4a30228c3b17a6d8c6053765dc6eec8"},
@@ -4586 +4579 @@ name = "tensorflow-io-gcs-filesystem"
-version = "0.30.0"
+version = "0.31.0"
@@ -4592,19 +4585,19 @@ files = [
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:f5aa1aa0ec88b7d309ebf520c950ed12894ec1908c3d7335f080de9e16e88360"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:5093c9afd1e7e5a8d23f878b8074ee50d25d2fb66269a350542d6d92d643b608"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:95812ed532f7b9087eb0cb4597dcc8443708b2698ba1c07367333233e20074ea"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp310-cp310-win_amd64.whl", hash = "sha256:ab82d9a39dcdad2f525840c42387e2f064666e8d3e65c46d64873ad8eaa92742"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e4cea42fe5ab47cc6af7e146935489620ce2c4606a9483090bc9cac8f32ef111"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp311-cp311-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:8d7d99a61d68d3ad750404b8f402fafceeae81bd6b8f14cb81a1313182411571"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bd5ef0e38ed87696c2fbfbc9534a056484b6ee4ddd68967d644cc17e7d111018"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp311-cp311-win_amd64.whl", hash = "sha256:51f2ca3f4376b26d0e511c91a1468f089654a8736c76433404a8c4405c767d76"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:65f9cefcf52ef04caea1481faa0c3553b3cfd8ee65a01bcf7d9baf617361aaca"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:cf2940b843c6f4d91d4abb0df181af80b4cb8c680f34ebed61082c1e388157f3"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp37-cp37m-win_amd64.whl", hash = "sha256:c16dcee548a0aaff31793ac410a880a45a61401f1a1a8faee24f5ef506900796"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:e3bc68c76402b38a2486a0e2e710095c3eb6d6e84c131ad349f7ca034dfc345b"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp38-cp38-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f3cfcdd94de2cd4adffcb5659b95b2e5714e280c617b922c134b9d61b7f20429"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23eb36218b939e6a814cc4e4f4d9d4a8e2574d8589a5979e882f5f056a4264be"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp38-cp38-win_amd64.whl", hash = "sha256:86169a799752cf61c07d1a5a818e00d6233e3cb3ebe6bb144af5f0fed1dc2b89"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:649166ef120fa3af43e7550cb9f1c26ff54e41b0dcfc106ab13f92435fb9d21f"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:b2eb2a48f0d31359603f49b813453e4532958db3ef686e2738396cba54b7dc28"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:49264b8a7b04e18d516ad07c2f75e660d4d607a6320d3f4a16e00c2bbecf4db4"},
- {file = "tensorflow_io_gcs_filesystem-0.30.0-cp39-cp39-win_amd64.whl", hash = "sha256:ed4244b6a2963972ca86bb2e1855b8b7dced99d12a60641221db4f0b8cd83e32"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:a71421f8d75a093b6aac65b4c8c8d2f768c3ca6215307cf8c16192e62d992bcf"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:359134ecbd3bf938bb0cf65be4526106c30da461b2e2ce05446a229ed35f6832"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b658b33567552f155af2ed848130f787bfda29381fa78cd905d5ee8254364f3c"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp310-cp310-win_amd64.whl", hash = "sha256:961353b38c76471fa296bb7d883322c66b91415e7d47087236a6706db3ab2758"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:8909c4344b0e96aa356230ab460ffafe5900c33c1aaced65fafae71d177a1966"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp311-cp311-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:e417faf8755aafe52d8f8c6b5ae5bae6e4fae8326ee3acd5e9181b83bbfbae87"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:37c40e3c4ee1f8dda3b545deea6b8839192c82037d8021db9f589908034ad975"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp311-cp311-win_amd64.whl", hash = "sha256:4bb37d23f21c434687b11059cb7ffd094d52a7813368915ba1b7057e3c16e414"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:a7e8d4bd0a25de7637e562997c011294d7ea595a76f315427a5dd522d56e9d49"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:fbcfb4aa2eaa9a3038d2487e570ff93feb1dbe51c3a4663d7d9ab9f9a9f9a9d8"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp37-cp37m-win_amd64.whl", hash = "sha256:e3933059b1c53e062075de2e355ec136b655da5883c3c26736c45dfeb1901945"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f0adfbcd264262797d429311843733da2d5c1ffb119fbfa6339269b6c0414113"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp38-cp38-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:20e3ee5df01f2bd81d37fc715816c329b7533ccca967c47946eb458a5b7a7280"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bd628609b77aee0e385eadf1628222486f19b8f1d81b5f0a344f2470204df116"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp38-cp38-win_amd64.whl", hash = "sha256:b4ebb30ad7ce5f3769e3d959ea99bd95d80a44099bcf94da6042f9755ac6e850"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:68b89ef9f63f297de1cd9d545bc45dddc7d8fe12bcda4266279b244e8cf3b7c0"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:e6d8cc7b14ade870168b9704ee44f9c55b468b9a00ed40e12d20fffd321193b5"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:97ebb9a8001a38f615aa1f90d2e998b7bd6eddae7aafc92897833610b039401b"},
+ {file = "tensorflow_io_gcs_filesystem-0.31.0-cp39-cp39-win_amd64.whl", hash = "sha256:cb7459c15608fe42973a78e4d3ad7ac79cfc7adae1ccb1b1846db3165fbc081a"},
@@ -4729,0 +4723,2 @@ files = [
+ {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"},
+ {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"},
@@ -4734,0 +4730 @@ files = [
+ {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"},
@@ -4742,0 +4739 @@ files = [
+ {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"},
@@ -4808,0 +4806 @@ url = "https://download.pytorch.org/whl/cpu/torch-1.13.1%2Bcpu-cp39-cp39-linux_x
+
@@ -4825,0 +4824 @@ url = "https://download.pytorch.org/whl/cpu/torchaudio-0.13.1%2Bcpu-cp39-cp39-li
+
@@ -4955 +4954 @@ name = "types-psutil"
-version = "5.9.5.8"
+version = "5.9.5.9"
@@ -4961,2 +4960,2 @@ files = [
- {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
- {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
+ {file = "types-psutil-5.9.5.9.tar.gz", hash = "sha256:c6c7e6b41b6f7ebb87afd5650df360751d0e8acf0d276b7ac64d00f419beb922"},
+ {file = "types_psutil-5.9.5.9-py3-none-any.whl", hash = "sha256:f7830fb3a7bc3288113ff08199330fced7851f5219f0af1b2abc24aaa3112236"},
@@ -4979 +4978 @@ name = "types-requests"
-version = "2.28.11.14"
+version = "2.28.11.15"
@@ -4985,2 +4984,2 @@ files = [
- {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
- {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
+ {file = "types-requests-2.28.11.15.tar.gz", hash = "sha256:fc8eaa09cc014699c6b63c60c2e3add0c8b09a410c818b5ac6e65f92a26dde09"},
+ {file = "types_requests-2.28.11.15-py3-none-any.whl", hash = "sha256:a05e4c7bc967518fba5789c341ea8b0c942776ee474c7873129a61161978e586"},
@@ -4994 +4993 @@ name = "types-urllib3"
-version = "1.26.25.7"
+version = "1.26.25.8"
@@ -5000,2 +4999,2 @@ files = [
- {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
- {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
+ {file = "types-urllib3-1.26.25.8.tar.gz", hash = "sha256:ecf43c42d8ee439d732a1110b4901e9017a79a38daca26f08e42c8460069392c"},
+ {file = "types_urllib3-1.26.25.8-py3-none-any.whl", hash = "sha256:95ea847fbf0bf675f50c8ae19a665baedcf07e6b4641662c4c3c72e7b2edf1a9"},
@@ -5166 +5165 @@ name = "wrapt"
-version = "1.14.1"
+version = "1.15.0"
@@ -5172,64 +5171,75 @@ files = [
- {file = "wrapt-1.14.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:1b376b3f4896e7930f1f772ac4b064ac12598d1c38d04907e696cc4d794b43d3"},
- {file = "wrapt-1.14.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:903500616422a40a98a5a3c4ff4ed9d0066f3b4c951fa286018ecdf0750194ef"},
- {file = "wrapt-1.14.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:5a9a0d155deafd9448baff28c08e150d9b24ff010e899311ddd63c45c2445e28"},
- {file = "wrapt-1.14.1-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:ddaea91abf8b0d13443f6dac52e89051a5063c7d014710dcb4d4abb2ff811a59"},
- {file = "wrapt-1.14.1-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:36f582d0c6bc99d5f39cd3ac2a9062e57f3cf606ade29a0a0d6b323462f4dd87"},
- {file = "wrapt-1.14.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:7ef58fb89674095bfc57c4069e95d7a31cfdc0939e2a579882ac7d55aadfd2a1"},
- {file = "wrapt-1.14.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:e2f83e18fe2f4c9e7db597e988f72712c0c3676d337d8b101f6758107c42425b"},
- {file = "wrapt-1.14.1-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:ee2b1b1769f6707a8a445162ea16dddf74285c3964f605877a20e38545c3c462"},
- {file = "wrapt-1.14.1-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:833b58d5d0b7e5b9832869f039203389ac7cbf01765639c7309fd50ef619e0b1"},
- {file = "wrapt-1.14.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:80bb5c256f1415f747011dc3604b59bc1f91c6e7150bd7db03b19170ee06b320"},
- {file = "wrapt-1.14.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:07f7a7d0f388028b2df1d916e94bbb40624c59b48ecc6cbc232546706fac74c2"},
- {file = "wrapt-1.14.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02b41b633c6261feff8ddd8d11c711df6842aba629fdd3da10249a53211a72c4"},
- {file = "wrapt-1.14.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2fe803deacd09a233e4762a1adcea5db5d31e6be577a43352936179d14d90069"},
- {file = "wrapt-1.14.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:257fd78c513e0fb5cdbe058c27a0624c9884e735bbd131935fd49e9fe719d310"},
- {file = "wrapt-1.14.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4fcc4649dc762cddacd193e6b55bc02edca674067f5f98166d7713b193932b7f"},
- {file = "wrapt-1.14.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:11871514607b15cfeb87c547a49bca19fde402f32e2b1c24a632506c0a756656"},
- {file = "wrapt-1.14.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:8ad85f7f4e20964db4daadcab70b47ab05c7c1cf2a7c1e51087bfaa83831854c"},
- {file = "wrapt-1.14.1-cp310-cp310-win32.whl", hash = "sha256:a9a52172be0b5aae932bef82a79ec0a0ce87288c7d132946d645eba03f0ad8a8"},
- {file = "wrapt-1.14.1-cp310-cp310-win_amd64.whl", hash = "sha256:6d323e1554b3d22cfc03cd3243b5bb815a51f5249fdcbb86fda4bf62bab9e164"},
- {file = "wrapt-1.14.1-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:43ca3bbbe97af00f49efb06e352eae40434ca9d915906f77def219b88e85d907"},
- {file = "wrapt-1.14.1-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:6b1a564e6cb69922c7fe3a678b9f9a3c54e72b469875aa8018f18b4d1dd1adf3"},
- {file = "wrapt-1.14.1-cp35-cp35m-manylinux2010_i686.whl", hash = "sha256:00b6d4ea20a906c0ca56d84f93065b398ab74b927a7a3dbd470f6fc503f95dc3"},
- {file = "wrapt-1.14.1-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:a85d2b46be66a71bedde836d9e41859879cc54a2a04fad1191eb50c2066f6e9d"},
- {file = "wrapt-1.14.1-cp35-cp35m-win32.whl", hash = "sha256:dbcda74c67263139358f4d188ae5faae95c30929281bc6866d00573783c422b7"},
- {file = "wrapt-1.14.1-cp35-cp35m-win_amd64.whl", hash = "sha256:b21bb4c09ffabfa0e85e3a6b623e19b80e7acd709b9f91452b8297ace2a8ab00"},
- {file = "wrapt-1.14.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:9e0fd32e0148dd5dea6af5fee42beb949098564cc23211a88d799e434255a1f4"},
- {file = "wrapt-1.14.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9736af4641846491aedb3c3f56b9bc5568d92b0692303b5a305301a95dfd38b1"},
- {file = "wrapt-1.14.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5b02d65b9ccf0ef6c34cba6cf5bf2aab1bb2f49c6090bafeecc9cd81ad4ea1c1"},
- {file = "wrapt-1.14.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21ac0156c4b089b330b7666db40feee30a5d52634cc4560e1905d6529a3897ff"},
- {file = "wrapt-1.14.1-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:9f3e6f9e05148ff90002b884fbc2a86bd303ae847e472f44ecc06c2cd2fcdb2d"},
- {file = "wrapt-1.14.1-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:6e743de5e9c3d1b7185870f480587b75b1cb604832e380d64f9504a0535912d1"},
- {file = "wrapt-1.14.1-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d79d7d5dc8a32b7093e81e97dad755127ff77bcc899e845f41bf71747af0c569"},
- {file = "wrapt-1.14.1-cp36-cp36m-win32.whl", hash = "sha256:81b19725065dcb43df02b37e03278c011a09e49757287dca60c5aecdd5a0b8ed"},
- {file = "wrapt-1.14.1-cp36-cp36m-win_amd64.whl", hash = "sha256:b014c23646a467558be7da3d6b9fa409b2c567d2110599b7cf9a0c5992b3b471"},
- {file = "wrapt-1.14.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:88bd7b6bd70a5b6803c1abf6bca012f7ed963e58c68d76ee20b9d751c74a3248"},
- {file = "wrapt-1.14.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b5901a312f4d14c59918c221323068fad0540e34324925c8475263841dbdfe68"},
- {file = "wrapt-1.14.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d77c85fedff92cf788face9bfa3ebaa364448ebb1d765302e9af11bf449ca36d"},
- {file = "wrapt-1.14.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8d649d616e5c6a678b26d15ece345354f7c2286acd6db868e65fcc5ff7c24a77"},
- {file = "wrapt-1.14.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:7d2872609603cb35ca513d7404a94d6d608fc13211563571117046c9d2bcc3d7"},
- {file = "wrapt-1.14.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:ee6acae74a2b91865910eef5e7de37dc6895ad96fa23603d1d27ea69df545015"},
- {file = "wrapt-1.14.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:2b39d38039a1fdad98c87279b48bc5dce2c0ca0d73483b12cb72aa9609278e8a"},
- {file = "wrapt-1.14.1-cp37-cp37m-win32.whl", hash = "sha256:60db23fa423575eeb65ea430cee741acb7c26a1365d103f7b0f6ec412b893853"},
- {file = "wrapt-1.14.1-cp37-cp37m-win_amd64.whl", hash = "sha256:709fe01086a55cf79d20f741f39325018f4df051ef39fe921b1ebe780a66184c"},
- {file = "wrapt-1.14.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8c0ce1e99116d5ab21355d8ebe53d9460366704ea38ae4d9f6933188f327b456"},
- {file = "wrapt-1.14.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:e3fb1677c720409d5f671e39bac6c9e0e422584e5f518bfd50aa4cbbea02433f"},
- {file = "wrapt-1.14.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:642c2e7a804fcf18c222e1060df25fc210b9c58db7c91416fb055897fc27e8cc"},
- {file = "wrapt-1.14.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7b7c050ae976e286906dd3f26009e117eb000fb2cf3533398c5ad9ccc86867b1"},
- {file = "wrapt-1.14.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ef3f72c9666bba2bab70d2a8b79f2c6d2c1a42a7f7e2b0ec83bb2f9e383950af"},
- {file = "wrapt-1.14.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:01c205616a89d09827986bc4e859bcabd64f5a0662a7fe95e0d359424e0e071b"},
- {file = "wrapt-1.14.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:5a0f54ce2c092aaf439813735584b9537cad479575a09892b8352fea5e988dc0"},
- {file = "wrapt-1.14.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:2cf71233a0ed05ccdabe209c606fe0bac7379fdcf687f39b944420d2a09fdb57"},
- {file = "wrapt-1.14.1-cp38-cp38-win32.whl", hash = "sha256:aa31fdcc33fef9eb2552cbcbfee7773d5a6792c137b359e82879c101e98584c5"},
- {file = "wrapt-1.14.1-cp38-cp38-win_amd64.whl", hash = "sha256:d1967f46ea8f2db647c786e78d8cc7e4313dbd1b0aca360592d8027b8508e24d"},
- {file = "wrapt-1.14.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3232822c7d98d23895ccc443bbdf57c7412c5a65996c30442ebe6ed3df335383"},
- {file = "wrapt-1.14.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:988635d122aaf2bdcef9e795435662bcd65b02f4f4c1ae37fbee7401c440b3a7"},
- {file = "wrapt-1.14.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9cca3c2cdadb362116235fdbd411735de4328c61425b0aa9f872fd76d02c4e86"},
- {file = "wrapt-1.14.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d52a25136894c63de15a35bc0bdc5adb4b0e173b9c0d07a2be9d3ca64a332735"},
- {file = "wrapt-1.14.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:40e7bc81c9e2b2734ea4bc1aceb8a8f0ceaac7c5299bc5d69e37c44d9081d43b"},
- {file = "wrapt-1.14.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b9b7a708dd92306328117d8c4b62e2194d00c365f18eff11a9b53c6f923b01e3"},
- {file = "wrapt-1.14.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:6a9a25751acb379b466ff6be78a315e2b439d4c94c1e99cb7266d40a537995d3"},
- {file = "wrapt-1.14.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:34aa51c45f28ba7f12accd624225e2b1e5a3a45206aa191f6f9aac931d9d56fe"},
- {file = "wrapt-1.14.1-cp39-cp39-win32.whl", hash = "sha256:dee0ce50c6a2dd9056c20db781e9c1cfd33e77d2d569f5d1d9321c641bb903d5"},
- {file = "wrapt-1.14.1-cp39-cp39-win_amd64.whl", hash = "sha256:dee60e1de1898bde3b238f18340eec6148986da0455d8ba7848d50470a7a32fb"},
- {file = "wrapt-1.14.1.tar.gz", hash = "sha256:380a85cf89e0e69b7cfbe2ea9f765f004ff419f34194018a6827ac0e3edfed4d"},
+ {file = "wrapt-1.15.0-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:ca1cccf838cd28d5a0883b342474c630ac48cac5df0ee6eacc9c7290f76b11c1"},
+ {file = "wrapt-1.15.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:e826aadda3cae59295b95343db8f3d965fb31059da7de01ee8d1c40a60398b29"},
+ {file = "wrapt-1.15.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:5fc8e02f5984a55d2c653f5fea93531e9836abbd84342c1d1e17abc4a15084c2"},
+ {file = "wrapt-1.15.0-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:96e25c8603a155559231c19c0349245eeb4ac0096fe3c1d0be5c47e075bd4f46"},
+ {file = "wrapt-1.15.0-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:40737a081d7497efea35ab9304b829b857f21558acfc7b3272f908d33b0d9d4c"},
+ {file = "wrapt-1.15.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:f87ec75864c37c4c6cb908d282e1969e79763e0d9becdfe9fe5473b7bb1e5f09"},
+ {file = "wrapt-1.15.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:1286eb30261894e4c70d124d44b7fd07825340869945c79d05bda53a40caa079"},
+ {file = "wrapt-1.15.0-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:493d389a2b63c88ad56cdc35d0fa5752daac56ca755805b1b0c530f785767d5e"},
+ {file = "wrapt-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:58d7a75d731e8c63614222bcb21dd992b4ab01a399f1f09dd82af17bbfc2368a"},
+ {file = "wrapt-1.15.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:21f6d9a0d5b3a207cdf7acf8e58d7d13d463e639f0c7e01d82cdb671e6cb7923"},
+ {file = "wrapt-1.15.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ce42618f67741d4697684e501ef02f29e758a123aa2d669e2d964ff734ee00ee"},
+ {file = "wrapt-1.15.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:41d07d029dd4157ae27beab04d22b8e261eddfc6ecd64ff7000b10dc8b3a5727"},
+ {file = "wrapt-1.15.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:54accd4b8bc202966bafafd16e69da9d5640ff92389d33d28555c5fd4f25ccb7"},
+ {file = "wrapt-1.15.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2fbfbca668dd15b744418265a9607baa970c347eefd0db6a518aaf0cfbd153c0"},
+ {file = "wrapt-1.15.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:76e9c727a874b4856d11a32fb0b389afc61ce8aaf281ada613713ddeadd1cfec"},
+ {file = "wrapt-1.15.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:e20076a211cd6f9b44a6be58f7eeafa7ab5720eb796975d0c03f05b47d89eb90"},
+ {file = "wrapt-1.15.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a74d56552ddbde46c246b5b89199cb3fd182f9c346c784e1a93e4dc3f5ec9975"},
+ {file = "wrapt-1.15.0-cp310-cp310-win32.whl", hash = "sha256:26458da5653aa5b3d8dc8b24192f574a58984c749401f98fff994d41d3f08da1"},
+ {file = "wrapt-1.15.0-cp310-cp310-win_amd64.whl", hash = "sha256:75760a47c06b5974aa5e01949bf7e66d2af4d08cb8c1d6516af5e39595397f5e"},
+ {file = "wrapt-1.15.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ba1711cda2d30634a7e452fc79eabcadaffedf241ff206db2ee93dd2c89a60e7"},
+ {file = "wrapt-1.15.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56374914b132c702aa9aa9959c550004b8847148f95e1b824772d453ac204a72"},
+ {file = "wrapt-1.15.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a89ce3fd220ff144bd9d54da333ec0de0399b52c9ac3d2ce34b569cf1a5748fb"},
+ {file = "wrapt-1.15.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3bbe623731d03b186b3d6b0d6f51865bf598587c38d6f7b0be2e27414f7f214e"},
+ {file = "wrapt-1.15.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3abbe948c3cbde2689370a262a8d04e32ec2dd4f27103669a45c6929bcdbfe7c"},
+ {file = "wrapt-1.15.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b67b819628e3b748fd3c2192c15fb951f549d0f47c0449af0764d7647302fda3"},
+ {file = "wrapt-1.15.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:7eebcdbe3677e58dd4c0e03b4f2cfa346ed4049687d839adad68cc38bb559c92"},
+ {file = "wrapt-1.15.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:74934ebd71950e3db69960a7da29204f89624dde411afbfb3b4858c1409b1e98"},
+ {file = "wrapt-1.15.0-cp311-cp311-win32.whl", hash = "sha256:bd84395aab8e4d36263cd1b9308cd504f6cf713b7d6d3ce25ea55670baec5416"},
+ {file = "wrapt-1.15.0-cp311-cp311-win_amd64.whl", hash = "sha256:a487f72a25904e2b4bbc0817ce7a8de94363bd7e79890510174da9d901c38705"},
+ {file = "wrapt-1.15.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:4ff0d20f2e670800d3ed2b220d40984162089a6e2c9646fdb09b85e6f9a8fc29"},
+ {file = "wrapt-1.15.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:9ed6aa0726b9b60911f4aed8ec5b8dd7bf3491476015819f56473ffaef8959bd"},
+ {file = "wrapt-1.15.0-cp35-cp35m-manylinux2010_i686.whl", hash = "sha256:896689fddba4f23ef7c718279e42f8834041a21342d95e56922e1c10c0cc7afb"},
+ {file = "wrapt-1.15.0-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:75669d77bb2c071333417617a235324a1618dba66f82a750362eccbe5b61d248"},
+ {file = "wrapt-1.15.0-cp35-cp35m-win32.whl", hash = "sha256:fbec11614dba0424ca72f4e8ba3c420dba07b4a7c206c8c8e4e73f2e98f4c559"},
+ {file = "wrapt-1.15.0-cp35-cp35m-win_amd64.whl", hash = "sha256:fd69666217b62fa5d7c6aa88e507493a34dec4fa20c5bd925e4bc12fce586639"},
+ {file = "wrapt-1.15.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:b0724f05c396b0a4c36a3226c31648385deb6a65d8992644c12a4963c70326ba"},
+ {file = "wrapt-1.15.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bbeccb1aa40ab88cd29e6c7d8585582c99548f55f9b2581dfc5ba68c59a85752"},
+ {file = "wrapt-1.15.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:38adf7198f8f154502883242f9fe7333ab05a5b02de7d83aa2d88ea621f13364"},
+ {file = "wrapt-1.15.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:578383d740457fa790fdf85e6d346fda1416a40549fe8db08e5e9bd281c6a475"},
+ {file = "wrapt-1.15.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:a4cbb9ff5795cd66f0066bdf5947f170f5d63a9274f99bdbca02fd973adcf2a8"},
+ {file = "wrapt-1.15.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:af5bd9ccb188f6a5fdda9f1f09d9f4c86cc8a539bd48a0bfdc97723970348418"},
+ {file = "wrapt-1.15.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:b56d5519e470d3f2fe4aa7585f0632b060d532d0696c5bdfb5e8319e1d0f69a2"},
+ {file = "wrapt-1.15.0-cp36-cp36m-win32.whl", hash = "sha256:77d4c1b881076c3ba173484dfa53d3582c1c8ff1f914c6461ab70c8428b796c1"},
+ {file = "wrapt-1.15.0-cp36-cp36m-win_amd64.whl", hash = "sha256:077ff0d1f9d9e4ce6476c1a924a3332452c1406e59d90a2cf24aeb29eeac9420"},
+ {file = "wrapt-1.15.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5c5aa28df055697d7c37d2099a7bc09f559d5053c3349b1ad0c39000e611d317"},
+ {file = "wrapt-1.15.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3a8564f283394634a7a7054b7983e47dbf39c07712d7b177b37e03f2467a024e"},
+ {file = "wrapt-1.15.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:780c82a41dc493b62fc5884fb1d3a3b81106642c5c5c78d6a0d4cbe96d62ba7e"},
+ {file = "wrapt-1.15.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e169e957c33576f47e21864cf3fc9ff47c223a4ebca8960079b8bd36cb014fd0"},
+ {file = "wrapt-1.15.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b02f21c1e2074943312d03d243ac4388319f2456576b2c6023041c4d57cd7019"},
+ {file = "wrapt-1.15.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:f2e69b3ed24544b0d3dbe2c5c0ba5153ce50dcebb576fdc4696d52aa22db6034"},
+ {file = "wrapt-1.15.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d787272ed958a05b2c86311d3a4135d3c2aeea4fc655705f074130aa57d71653"},
+ {file = "wrapt-1.15.0-cp37-cp37m-win32.whl", hash = "sha256:02fce1852f755f44f95af51f69d22e45080102e9d00258053b79367d07af39c0"},
+ {file = "wrapt-1.15.0-cp37-cp37m-win_amd64.whl", hash = "sha256:abd52a09d03adf9c763d706df707c343293d5d106aea53483e0ec8d9e310ad5e"},
+ {file = "wrapt-1.15.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:cdb4f085756c96a3af04e6eca7f08b1345e94b53af8921b25c72f096e704e145"},
+ {file = "wrapt-1.15.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:230ae493696a371f1dbffaad3dafbb742a4d27a0afd2b1aecebe52b740167e7f"},
+ {file = "wrapt-1.15.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:63424c681923b9f3bfbc5e3205aafe790904053d42ddcc08542181a30a7a51bd"},
+ {file = "wrapt-1.15.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6bcbfc99f55655c3d93feb7ef3800bd5bbe963a755687cbf1f490a71fb7794b"},
+ {file = "wrapt-1.15.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c99f4309f5145b93eca6e35ac1a988f0dc0a7ccf9ccdcd78d3c0adf57224e62f"},
+ {file = "wrapt-1.15.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:b130fe77361d6771ecf5a219d8e0817d61b236b7d8b37cc045172e574ed219e6"},
+ {file = "wrapt-1.15.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:96177eb5645b1c6985f5c11d03fc2dbda9ad24ec0f3a46dcce91445747e15094"},
+ {file = "wrapt-1.15.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:d5fe3e099cf07d0fb5a1e23d399e5d4d1ca3e6dfcbe5c8570ccff3e9208274f7"},
+ {file = "wrapt-1.15.0-cp38-cp38-win32.whl", hash = "sha256:abd8f36c99512755b8456047b7be10372fca271bf1467a1caa88db991e7c421b"},
+ {file = "wrapt-1.15.0-cp38-cp38-win_amd64.whl", hash = "sha256:b06fa97478a5f478fb05e1980980a7cdf2712015493b44d0c87606c1513ed5b1"},
+ {file = "wrapt-1.15.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2e51de54d4fb8fb50d6ee8327f9828306a959ae394d3e01a1ba8b2f937747d86"},
+ {file = "wrapt-1.15.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0970ddb69bba00670e58955f8019bec4a42d1785db3faa043c33d81de2bf843c"},
+ {file = "wrapt-1.15.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:76407ab327158c510f44ded207e2f76b657303e17cb7a572ffe2f5a8a48aa04d"},
+ {file = "wrapt-1.15.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cd525e0e52a5ff16653a3fc9e3dd827981917d34996600bbc34c05d048ca35cc"},
+ {file = "wrapt-1.15.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9d37ac69edc5614b90516807de32d08cb8e7b12260a285ee330955604ed9dd29"},
+ {file = "wrapt-1.15.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:078e2a1a86544e644a68422f881c48b84fef6d18f8c7a957ffd3f2e0a74a0d4a"},
+ {file = "wrapt-1.15.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:2cf56d0e237280baed46f0b5316661da892565ff58309d4d2ed7dba763d984b8"},
+ {file = "wrapt-1.15.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:7dc0713bf81287a00516ef43137273b23ee414fe41a3c14be10dd95ed98a2df9"},
+ {file = "wrapt-1.15.0-cp39-cp39-win32.whl", hash = "sha256:46ed616d5fb42f98630ed70c3529541408166c22cdfd4540b88d5f21006b0eff"},
+ {file = "wrapt-1.15.0-cp39-cp39-win_amd64.whl", hash = "sha256:eef4d64c650f33347c1f9266fa5ae001440b232ad9b98f1f43dfe7a79435c0a6"},
+ {file = "wrapt-1.15.0-py3-none-any.whl", hash = "sha256:64b1df0f83706b4ef4cfb4fb0e4c2669100fd7ecacfb59e091fad300d4e04640"},
+ {file = "wrapt-1.15.0.tar.gz", hash = "sha256:d06730c6aed78cee4126234cf2d071e01b44b915e725a6cb439a879ec9754a3a"},
@@ -5436 +5446 @@ name = "zipp"
-version = "3.14.0"
+version = "3.15.0"
@@ -5442,2 +5452,2 @@ files = [
- {file = "zipp-3.14.0-py3-none-any.whl", hash = "sha256:188834565033387710d046e3fe96acfc9b5e86cbca7f39ff69cf21a4128198b7"},
- {file = "zipp-3.14.0.tar.gz", hash = "sha256:9e5421e176ef5ab4c0ad896624e87a7b2f07aca746c9b2aa305952800cb8eecb"},
+ {file = "zipp-3.15.0-py3-none-any.whl", hash = "sha256:48904fc76a60e542af151aded95726c1a5c34ed43ab4134b597665c86d7ad556"},
+ {file = "zipp-3.15.0.tar.gz", hash = "sha256:112929ad649da941c23de50f356a2b5570c954b65150642bccdd66bf194d224b"},
@@ -5448 +5458 @@ docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker
-testing = ["flake8 (<5)", "func-timeout", "jaraco.functools", "jaraco.itertools", "more-itertools", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)"]
+testing = ["big-O", "flake8 (<5)", "jaraco.functools", "jaraco.itertools", "more-itertools", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)"]
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index 53a36969..2d6733c4 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -17,0 +18,2 @@ WORKER_ENDPOINT = "/config-names"
+WORKER_HEARTBEAT_INTERVAL_SECONDS = 60
+WORKER_KILL_ZOMBIES_INTERVAL_SECONDS = 10 * 60
@@ -21,2 +22,0 @@ WORKER_MAX_MEMORY_PCT = 80
-WORKER_SLEEP_SECONDS = 15
-WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS = 60
@@ -24 +24 @@ WORKER_MAX_MISSING_HEARTBEATS = 5
-WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS = 10 * 60
+WORKER_SLEEP_SECONDS = 15
@@ -41 +41 @@ class WorkerConfig:
- heartbeat_time_interval_seconds: int = WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS
+ heartbeat_interval_seconds: int = WORKER_HEARTBEAT_INTERVAL_SECONDS
@@ -43 +43 @@ class WorkerConfig:
- kill_zombies_time_interval_seconds: int = WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS
+ kill_zombies_interval_seconds: int = WORKER_KILL_ZOMBIES_INTERVAL_SECONDS
@@ -58,2 +58,2 @@ class WorkerConfig:
- heartbeat_time_interval_seconds=env.int(
- name="HEARTBEAT_TIME_INTERVAL_SECONDS", default=WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS
+ heartbeat_interval_seconds=env.int(
+ name="HEARTBEAT_INTERVAL_SECONDS", default=WORKER_HEARTBEAT_INTERVAL_SECONDS
@@ -62,2 +62,2 @@ class WorkerConfig:
- kill_zombies_time_interval_seconds=env.int(
- name="KILL_ZOMBIES_TIME_INTERVAL_SECONDS", default=WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS
+ kill_zombies_interval_seconds=env.int(
+ name="KILL_ZOMBIES_INTERVAL_SECONDS", default=WORKER_KILL_ZOMBIES_INTERVAL_SECONDS
diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py
new file mode 100644
index 00000000..96e888ff
--- /dev/null
+++ b/services/worker/src/worker/executor.py
@@ -0,0 +1,104 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+import asyncio
+import json
+import logging
+import os
+import sys
+from typing import Any, Callable, Optional
+
+from filelock import FileLock
+from libcommon.queue import Queue
+from mirakuru import OutputExecutor
+
+from worker import start_worker_loop
+from worker.config import AppConfig
+from worker.job_runner_factory import JobRunnerFactory
+from worker.loop import WorkerState
+
+START_WORKER_LOOP_PATH = start_worker_loop.__file__
+
+
+async def every(
+ func: Callable[..., Optional[Any]], *args: Any, seconds: int, stop_on: Optional[Any] = None, **kwargs: Any
+) -> None:
+ while True:
+ out = func(*args, **kwargs)
+ if stop_on is not None and out == stop_on:
+ break
+ await asyncio.sleep(seconds)
+
+
+class BadWorkerState(RuntimeError):
+ """Raised when the worker state from the worker read by the executor is not valid."""
+
+ pass
+
+
+class WorkerExecutor:
+ def __init__(self, app_config: AppConfig, job_runner_factory: JobRunnerFactory) -> None:
+ self.app_config = app_config
+ self.job_runner_factory = job_runner_factory
+
+ max_missing_heartbeats = self.app_config.worker.max_missing_heartbeats
+ heartbeat_interval_seconds = self.app_config.worker.heartbeat_interval_seconds
+ self.max_seconds_without_heartbeat_for_zombies = heartbeat_interval_seconds * max_missing_heartbeats
+
+ def _create_worker_loop_executor(self) -> OutputExecutor:
+ banner = self.app_config.worker.state_path
+ if not banner:
+ raise ValueError("Failed to create the executor because WORKER_STATE_PATH is missing.")
+ start_worker_loop_command = [
+ sys.executable,
+ START_WORKER_LOOP_PATH,
+ "--print-worker-state-path",
+ ]
+ return OutputExecutor(start_worker_loop_command, banner, timeout=10)
+
+ def start(self) -> None:
+ worker_loop_executor = self._create_worker_loop_executor()
+ worker_loop_executor.start() # blocking until the banner is printed
+
+ loop = asyncio.get_event_loop()
+ logging.info("Starting heartbeat.")
+ loop.create_task(every(self.heartbeat, seconds=self.app_config.worker.heartbeat_interval_seconds))
+ loop.create_task(every(self.kill_zombies, seconds=self.app_config.worker.kill_zombies_interval_seconds))
+ loop.run_until_complete(
+ every(self.is_worker_alive, worker_loop_executor=worker_loop_executor, seconds=1, stop_on=False)
+ )
+
+ def get_state(self) -> WorkerState:
+ worker_state_path = self.app_config.worker.state_path
+ if not worker_state_path:
+ raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
+ if os.path.exists(worker_state_path):
+ with FileLock(worker_state_path + ".lock"):
+ try:
+ with open(worker_state_path, "r") as worker_state_f:
+ worker_state = json.load(worker_state_f)
+ return WorkerState(current_job_info=worker_state.get("current_job_info"))
+ except json.JSONDecodeError as err:
+ raise BadWorkerState(f"Failed to read worker state at {worker_state_path}") from err
+ else:
+ return WorkerState(current_job_info=None)
+
+ def heartbeat(self) -> None:
+ worker_state = self.get_state()
+ if worker_state["current_job_info"]:
+ Queue().heartbeat(job_id=worker_state["current_job_info"]["job_id"])
+
+ def kill_zombies(self) -> None:
+ queue = Queue()
+ zombies = queue.get_zombies(max_seconds_without_heartbeat=self.max_seconds_without_heartbeat_for_zombies)
+ queue.kill_zombies(zombies)
+ message = "Job runner crashed while running this job (missing heartbeats)."
+ for zombie in zombies:
+ job_runner = self.job_runner_factory.create_job_runner(zombie)
+ job_runner.set_crashed(message=message)
+
+ def is_worker_alive(self, worker_loop_executor: OutputExecutor) -> bool:
+ if not worker_loop_executor.running():
+ worker_loop_executor.stop() # raises an error if the worker returned exit code 1
+ return False
+ else:
+ return True
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index df0d5f2c..25c42461 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -33,0 +34 @@ GeneralJobRunnerErrorCode = Literal[
+ "JobRunnerCrashedError",
@@ -137,0 +139,13 @@ class UnexpectedError(GeneralJobRunnerError):
+class JobRunnerCrashedError(GeneralJobRunnerError):
+ """Raised when the job runner crashed and the job became a zombie."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(
+ message=message,
+ status_code=HTTPStatus.NOT_IMPLEMENTED,
+ code="JobRunnerCrashedError",
+ cause=cause,
+ disclose_cause=False,
+ )
+
+
@@ -462,0 +477,20 @@ class JobRunner(ABC):
+
+ def set_crashed(self, message: str, cause: Optional[BaseException] = None) -> None:
+ error = JobRunnerCrashedError(message=message, cause=cause)
+ upsert_response(
+ kind=self.processing_step.cache_kind,
+ dataset=self.dataset,
+ config=self.config,
+ split=self.split,
+ content=dict(error.as_response()),
+ http_status=error.status_code,
+ error_code=error.code,
+ details=dict(error.as_response_with_cause()),
+ worker_version=self.get_version(),
+ dataset_git_revision=self.get_dataset_git_revision(),
+ )
+ logging.debug(
+ "response for"
+ f" dataset={self.dataset} config={self.config} split={self.split} had an error (crashed),"
+ " cache updated"
+ )
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index 7996438f..71e39036 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -1,3 +1,2 @@
-import asyncio
-import json
-import logging
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
@@ -5 +3,0 @@ import os
-import sys
@@ -7,3 +4,0 @@ import tempfile
-from datetime import timedelta
-from http import HTTPStatus
-from typing import Any, Callable, List, Optional
@@ -11,3 +5,0 @@ from typing import Any, Callable, List, Optional
-import pytz
-from filelock import FileLock
-from libcommon.exceptions import CustomError
@@ -16 +7,0 @@ from libcommon.processing_graph import ProcessingGraph
-from libcommon.queue import Job, JobInfo, Status, get_datetime
@@ -18 +8,0 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.simple_cache import upsert_response
@@ -20 +9,0 @@ from libcommon.storage import init_assets_dir
-from mirakuru import OutputExecutor
@@ -22 +10,0 @@ from mirakuru import OutputExecutor
-from worker import start_worker_loop
@@ -23,0 +12 @@ from worker.config import AppConfig
+from worker.executor import WorkerExecutor
@@ -25 +13,0 @@ from worker.job_runner_factory import JobRunnerFactory
-from worker.loop import WorkerState
@@ -29,146 +16,0 @@ WORKER_STATE_FILE_NAME = "worker_state.json"
-START_WORKER_LOOP_PATH = start_worker_loop.__file__
-
-
-async def every(
- func: Callable[..., Optional[Any]], *args: Any, seconds: int, stop_on: Optional[Any] = None, **kwargs: Any
-) -> None:
- while True:
- out = func(*args, **kwargs)
- if stop_on is not None and out == stop_on:
- break
- await asyncio.sleep(seconds)
-
-
-class WorkerCrashedError(CustomError):
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "WorkerCrashedError", cause, False)
-
-
-class WorkerExecutor:
- def __init__(self, app_config: AppConfig, job_runner_factory: JobRunnerFactory) -> None:
- self.app_config = app_config
- self.job_runner_factory = job_runner_factory
-
- def _create_worker_loop_executor(self) -> OutputExecutor:
- banner = self.app_config.worker.state_path
- if not banner:
- raise ValueError("Failed to create the executor because WORKER_STATE_PATH is missing.")
- start_worker_loop_command = [
- sys.executable,
- START_WORKER_LOOP_PATH,
- "--print-worker-state-path",
- ]
- return OutputExecutor(start_worker_loop_command, banner, timeout=10)
-
- def start(self) -> None:
- worker_loop_executor = self._create_worker_loop_executor()
- worker_loop_executor.start() # blocking until the banner is printed
-
- loop = asyncio.get_event_loop()
- logging.info("Starting heartbeat.")
- loop.create_task(every(self.heartbeat, seconds=self.app_config.worker.heartbeat_time_interval_seconds))
- loop.create_task(every(self.kill_zombies, seconds=self.app_config.worker.kill_zombies_time_interval_seconds))
- loop.run_until_complete(
- every(self.is_worker_alive, worker_loop_executor=worker_loop_executor, seconds=1, stop_on=False)
- )
-
- def get_state(self) -> WorkerState:
- worker_state_path = self.app_config.worker.state_path
- if not worker_state_path:
- raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
- if os.path.exists(worker_state_path):
- with FileLock(worker_state_path + ".lock"):
- try:
- with open(worker_state_path, "r") as worker_state_f:
- worker_state = json.load(worker_state_f)
- return WorkerState(current_job_info=worker_state.get("current_job_info"))
- except json.JSONDecodeError:
- return WorkerState(current_job_info=None)
- else:
- return WorkerState(current_job_info=None)
-
- def get_current_job(self) -> Optional[Job]:
- worker_state = self.get_state()
- if worker_state["current_job_info"]:
- job = Job.objects.with_id(worker_state["current_job_info"]["job_id"]) # type: ignore
- if job and isinstance(job, Job) and job.status == Status.STARTED:
- return job
- return None
-
- def heartbeat(self) -> None:
- current_job = self.get_current_job()
- if current_job:
- current_job.update(last_heartbeat=get_datetime())
-
- def get_zombies(self) -> List[Job]:
- started_jobs = Job.objects(status=Status.STARTED)
- max_missing_heartbeats = self.app_config.worker.max_missing_heartbeats
- heartbeat_time_interval_seconds = self.app_config.worker.heartbeat_time_interval_seconds
- if max_missing_heartbeats >= 1:
- return [
- job
- for job in started_jobs
- if (
- job.last_heartbeat is not None
- and get_datetime()
- >= pytz.UTC.localize(job.last_heartbeat)
- + timedelta(seconds=max_missing_heartbeats * heartbeat_time_interval_seconds)
- )
- or (
- job.last_heartbeat is None
- and job.started_at is not None
- and get_datetime()
- >= pytz.UTC.localize(job.started_at)
- + timedelta(seconds=max_missing_heartbeats * heartbeat_time_interval_seconds)
- )
- ]
- else:
- return []
-
- def kill_zombies(self) -> None:
- zombies = self.get_zombies()
- if zombies:
- zombies_examples = [str(zombie.pk) for zombie in zombies[:10]]
- zombies_examples_str = ", ".join(zombies_examples) + (
- "..." if len(zombies_examples) != len(zombies) else ""
- )
- logging.info(f"Killing {len(zombies)} zombies. Job ids = " + zombies_examples_str)
- Job.objects(pk__in=[zombie.pk for zombie in zombies]).update(
- status=Status.ERROR, finished_at=get_datetime()
- )
- for zombie in zombies:
- job_info = JobInfo(
- job_id=str(zombie.pk),
- type=zombie.type,
- dataset=zombie.dataset,
- config=zombie.config,
- split=zombie.split,
- force=zombie.force,
- priority=zombie.priority,
- )
- job_runner = self.job_runner_factory.create_job_runner(job_info)
- error = WorkerCrashedError("Worker crashed while running this job.")
- upsert_response(
- kind=job_runner.processing_step.cache_kind,
- dataset=job_runner.dataset,
- config=job_runner.config,
- split=job_runner.split,
- content=dict(error.as_response()),
- http_status=error.status_code,
- error_code=error.code,
- details=dict(error.as_response_with_cause()),
- worker_version=job_runner.get_version(),
- dataset_git_revision=job_runner.get_dataset_git_revision(),
- )
- logging.debug(
- "response for"
- f" dataset={job_runner.dataset} config={job_runner.config} split={job_runner.split} had an error,"
- " cache updated"
- )
-
- def is_worker_alive(self, worker_loop_executor: OutputExecutor) -> bool:
- if not worker_loop_executor.running():
- worker_loop_executor.stop() # raises an error if the worker returned exit code 1
- return False
- else:
- return True
@@ -183 +24,0 @@ if __name__ == "__main__":
- init_logging(log_level=app_config.common.log_level)
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index 77325b05..4456d4a9 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -67,2 +67,2 @@ def set_env_vars(
- mp.setenv("WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS", "1")
- mp.setenv("WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS", "1")
+ mp.setenv("WORKER_HEARTBEAT_INTERVAL_SECONDS", "1")
+ mp.setenv("WORKER_KILL_ZOMBIES_INTERVAL_SECONDS", "1")
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index 938856ae..b3612a22 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -15 +15 @@ from libcommon.processing_graph import ProcessingGraph
-from libcommon.queue import Job, JobInfo, Priority, Status, get_datetime
+from libcommon.queue import DoesNotExist, Job, JobInfo, Priority, Status, get_datetime
@@ -22,0 +23 @@ from worker.config import AppConfig
+from worker.executor import WorkerExecutor
@@ -25 +25,0 @@ from worker.loop import WorkerState
-from worker.main import WorkerExecutor
@@ -89,2 +89,4 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I
- if Job.objects.with_id(job_info["job_id"]): # type: ignore
- Job.objects.with_id(job_info["job_id"]).delete() # type: ignore
+ try:
+ Job.objects(pk=job_info["job_id"]).get().delete()
+ except DoesNotExist:
+ pass
@@ -115,2 +117,4 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q
- if Job.objects.with_id(job_info["job_id"]): # type: ignore
- Job.objects.with_id(job_info["job_id"]).delete() # type: ignore
+ try:
+ Job.objects(pk=job_info["job_id"]).get().delete()
+ except DoesNotExist:
+ pass
@@ -118 +122 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q
- last_heartbeat = get_datetime() - timedelta(seconds=app_config.worker.heartbeat_time_interval_seconds)
+ last_heartbeat = get_datetime() - timedelta(seconds=app_config.worker.heartbeat_interval_seconds)
@@ -143,2 +147,4 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato
- if Job.objects.with_id(job_info["job_id"]): # type: ignore
- Job.objects.with_id(job_info["job_id"]).delete() # type: ignore
+ try:
+ Job.objects(pk=job_info["job_id"]).get().delete()
+ except DoesNotExist:
+ pass
@@ -193,20 +198,0 @@ def test_executor_get_empty_state(
-def test_executor_get_current_job(
- executor: WorkerExecutor, set_just_started_job_in_queue: Job, set_worker_state: WorkerState
-) -> None:
- assert executor.get_current_job() == set_just_started_job_in_queue
-
-
-def test_executor_get_nonexisting_current_job(executor: WorkerExecutor) -> None:
- assert executor.get_current_job() is None
-
-
-def test_executor_get_zombies(
- executor: WorkerExecutor,
- set_just_started_job_in_queue: Job,
- set_long_running_job_in_queue: Job,
- set_zombie_job_in_queue: Job,
-) -> None:
- zombie = set_zombie_job_in_queue
- assert executor.get_zombies() == [zombie]
-
-
@@ -218,2 +204 @@ def test_executor_heartbeat(
- current_job = executor.get_current_job()
- assert current_job is not None
+ current_job = set_just_started_job_in_queue
@@ -222,2 +207 @@ def test_executor_heartbeat(
- current_job = executor.get_current_job()
- assert current_job is not None
+ current_job.reload()
@@ -238,0 +223 @@ def test_executor_kill_zombies(
+ normal_job = set_just_started_job_in_queue
@@ -240 +224,0 @@ def test_executor_kill_zombies(
- assert executor.get_zombies() == [zombie]
@@ -243,2 +227,2 @@ def test_executor_kill_zombies(
- assert executor.get_zombies() == []
- assert Job.objects.with_id(zombie.pk).status == Status.ERROR # type: ignore
+ assert Job.objects(pk=zombie.pk).get().status == Status.ERROR
+ assert Job.objects(pk=normal_job.pk).get().status == Status.STARTED
@@ -247 +231 @@ def test_executor_kill_zombies(
- "error": "Worker crashed while running this job.",
+ "error": "Job runner crashed while running this job (missing heartbeats).",
@@ -250 +234 @@ def test_executor_kill_zombies(
- assert response.error_code == "WorkerCrashedError"
+ assert response.error_code == "JobRunnerCrashedError"
@@ -270 +254 @@ def test_executor_start(
- with patch("worker.main.START_WORKER_LOOP_PATH", __file__):
+ with patch("worker.executor.START_WORKER_LOOP_PATH", __file__):
@@ -272 +256 @@ def test_executor_start(
- current_job = executor.get_current_job()
+ current_job = set_just_started_job_in_queue
@@ -276 +260 @@ def test_executor_start(
- assert Job.objects.with_id(set_just_started_job_in_queue.pk).last_heartbeat is not None # type: ignore
+ assert Job.objects(pk=set_just_started_job_in_queue.pk).get().last_heartbeat is not None
@@ -278 +262 @@ def test_executor_start(
- assert Job.objects.with_id(set_zombie_job_in_queue.pk).status == Status.ERROR # type: ignore
+ assert Job.objects(pk=set_zombie_job_in_queue.pk).get().status == Status.ERROR
@@ -293 +277 @@ def test_executor_raises_on_bad_worker(
- with patch("worker.main.START_WORKER_LOOP_PATH", __file__):
+ with patch("worker.executor.START_WORKER_LOOP_PATH", __file__):
diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py
index cd4cef03..2b154439 100644
--- a/services/worker/tests/test_job_runner.py
+++ b/services/worker/tests/test_job_runner.py
@@ -10 +10 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.simple_cache import SplitFullName, upsert_response
+from libcommon.simple_cache import CachedResponse, SplitFullName, upsert_response
@@ -318,0 +319,36 @@ def test_create_children_jobs() -> None:
+
+
+def test_job_runner_set_crashed(
+ test_processing_step: ProcessingStep,
+) -> None:
+ job_id = "job_id"
+ dataset = "dataset"
+ config = "config"
+ split = "split"
+ force = False
+ message = "I'm crashed :("
+ job_runner = DummyJobRunner(
+ job_info={
+ "job_id": job_id,
+ "type": test_processing_step.job_type,
+ "dataset": dataset,
+ "config": config,
+ "split": split,
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ processing_step=test_processing_step,
+ common_config=CommonConfig(),
+ worker_config=WorkerConfig(),
+ )
+ job_runner.set_crashed(message=message)
+ response = CachedResponse.objects()[0]
+ expected_error = {"error": message}
+ assert response.http_status == HTTPStatus.NOT_IMPLEMENTED
+ assert response.error_code == "JobRunnerCrashedError"
+ assert response.dataset == dataset
+ assert response.config == config
+ assert response.split == split
+ assert response.content == expected_error
+ assert response.details == expected_error
+ # TODO: check if it stores the correct dataset git sha and job version when it's implemented
|
|
912199fe63265dd67d6f34bc02e20a087086b92d
|
Andrea Francis Soria Jimenez
| 2023-02-27T18:51:08 |
POC: Adding mongo TTL index to Jobs (#851)
|
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
index 7b707218..a9ea4c95 100644
--- a/libs/libcommon/src/libcommon/constants.py
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -6,0 +7 @@ QUEUE_MONGOENGINE_ALIAS = "queue"
+QUEUE_TTL_SECONDS = 604_800 # 7 days
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 60ae169a..71f929de 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -18 +18 @@ from mongoengine.queryset.queryset import QuerySet
-from libcommon.constants import QUEUE_MONGOENGINE_ALIAS
+from libcommon.constants import QUEUE_MONGOENGINE_ALIAS, QUEUE_TTL_SECONDS
@@ -138,0 +139 @@ class Job(Document):
+ {"fields": ["finished_at"], "expireAfterSeconds": QUEUE_TTL_SECONDS},
|
|
d5b5ebf52cd46b9df53eb0be6c34d79b40861be6
|
Andrea Francis Soria Jimenez
| 2023-02-27T16:19:24 |
Fix typos in split-names-from-streaming (#837)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 7ba24295..18857e50 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -43 +43 @@ class MigrationsCollector:
- description="update 'kind' field in cache from /split-names to /split-names-streaming",
+ description="update 'kind' field in cache from /split-names to /split-names-from-streaming",
@@ -47 +47,3 @@ class MigrationsCollector:
- description="update 'type' and 'unicity_id' fields in job from /split-names to /split-names-streaming",
+ description=(
+ "update 'type' and 'unicity_id' fields in job from /split-names to /split-names-from-streaming"
+ ),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
index 0083817d..b1e7774d 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
@@ -11,0 +12,2 @@ from mongodb_migration.migration import Migration
+split_names = "/split-names"
+split_names_from_streaming = "/split-names-from-streaming"
@@ -18 +20 @@ class MigrationCacheUpdateSplitNames(Migration):
- logging.info("Rename cache_kind field from /split-name to /split-names-streaming")
+ logging.info(f"Rename cache_kind field from {split_names} to {split_names_from_streaming}")
@@ -20 +22 @@ class MigrationCacheUpdateSplitNames(Migration):
- db["cachedResponsesBlue"].update_many({"kind": "/split-names"}, {"$set": {"kind": "/split-names-streaming"}})
+ db["cachedResponsesBlue"].update_many({"kind": split_names}, {"$set": {"kind": split_names_from_streaming}})
@@ -23 +25 @@ class MigrationCacheUpdateSplitNames(Migration):
- logging.info("Rollback cache_kind field from /split-name-streaming to /split-names")
+ logging.info(f"Rollback cache_kind field from {split_names_from_streaming} to {split_names}")
@@ -25 +27 @@ class MigrationCacheUpdateSplitNames(Migration):
- db["cachedResponsesBlue"].update_many({"kind": "/split-names-streaming"}, {"$set": {"kind": "/split-names"}})
+ db["cachedResponsesBlue"].update_many({"kind": split_names_from_streaming}, {"$set": {"kind": split_names}})
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
index b9c2cc4d..301bfa53 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
@@ -21,2 +21,3 @@ class MigrationQueueUpdateSplitNames(Migration):
- f"Rename unicity_id field from Job[{split_names}][<dataset>][<config>][<split>] to"
- f" Job[{split_names_from_streaming}][<dataset>][<config>][<split>]"
+ f"Rename unicity_id field from Job[{split_names}][<dataset>][<config>][None] to"
+ f" Job[{split_names_from_streaming}][<dataset>][<config>][None] and change type from {split_names} to"
+ f" {split_names_from_streaming}"
@@ -25,4 +25,0 @@ class MigrationQueueUpdateSplitNames(Migration):
- for job in Job.objects(type=split_names):
- job.update(unicity_id=f"Job[{split_names_from_streaming}][{job.dataset}][{job.config}][{job.split}]")
-
- logging.info(f"Rename type field from {split_names} to {split_names_from_streaming}")
@@ -30 +27,17 @@ class MigrationQueueUpdateSplitNames(Migration):
- db["jobsBlue"].update_many({"type": split_names}, {"$set": {"type": split_names_from_streaming}})
+ db["jobsBlue"].update_many(
+ {"type": split_names},
+ [
+ {
+ "$set": {
+ "unicity_id": {
+ "$replaceOne": {
+ "input": "$unicity_id",
+ "find": f"Job[{split_names}",
+ "replacement": f"Job[{split_names_from_streaming}",
+ }
+ },
+ "type": split_names_from_streaming,
+ }
+ },
+ ], # type: ignore
+ )
@@ -34,2 +47,3 @@ class MigrationQueueUpdateSplitNames(Migration):
- f"Rename unicity_id field from Job[{split_names_from_streaming}][<dataset>][<config>][<split>] to"
- f" Job[{split_names}][<dataset>][<config>][<split>]"
+ f"Rename unicity_id field from Job[{split_names_from_streaming}][<dataset>][<config>][None] to"
+ f" Job[{split_names}][<dataset>][<config>][None] and change type from {split_names_from_streaming} to"
+ f" {split_names}"
@@ -38,4 +51,0 @@ class MigrationQueueUpdateSplitNames(Migration):
- for job in Job.objects(type=split_names_from_streaming):
- job.update(unicity_id=f"Job[{split_names}][{job.dataset}][{job.config}][{job.split}]")
-
- logging.info(f"Rename type field from {split_names_from_streaming} to {split_names}")
@@ -43 +53,17 @@ class MigrationQueueUpdateSplitNames(Migration):
- db["jobsBlue"].update_many({"type": split_names_from_streaming}, {"$set": {"type": split_names}})
+ db["jobsBlue"].update_many(
+ {"type": split_names_from_streaming},
+ [
+ {
+ "$set": {
+ "unicity_id": {
+ "$replaceOne": {
+ "input": "$unicity_id",
+ "find": f"Job[{split_names_from_streaming}",
+ "replacement": f"Job[{split_names}",
+ }
+ },
+ "type": split_names_from_streaming,
+ }
+ },
+ ], # type: ignore
+ )
|
|
391cf55e831dab8aee7989e69d99973955244062
|
Andrea Francis Soria Jimenez
| 2023-02-24T20:30:38 |
Adding missing function on split-names-from-dataset-info worker (#839)
|
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
index be24011f..2147358a 100644
--- a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
+++ b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
@@ -9 +9 @@ from libcommon.dataset import DatasetNotFoundError
-from libcommon.simple_cache import DoesNotExist, get_response
+from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
@@ -125,0 +126,6 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner):
+
+ def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]:
+ """Get the set of new splits, from the content created by the compute."""
+ return {
+ SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["split_names"]
+ }
|
|
fcde0bb3f1669f80355dc0fa3768e5b39bc309e4
|
Sylvain Lesage
| 2023-02-24T15:32:32 |
Setup argocd action (#848)
|
diff --git a/.github/workflows/build_push_docker_hub.yml b/.github/workflows/cd.yml
similarity index 50%
rename from .github/workflows/build_push_docker_hub.yml
rename to .github/workflows/cd.yml
index dd1ac06e..e3bfb380 100644
--- a/.github/workflows/build_push_docker_hub.yml
+++ b/.github/workflows/cd.yml
@@ -4 +4 @@
-name: Build and push docker images to public Docker Hub
+name: Continuous deployment
@@ -9,0 +10,5 @@ on:
+
+concurrency:
+ group: ${{ github.workflow }}-${{ github.ref }}
+ cancel-in-progress: true
+
@@ -11,0 +17 @@ env:
+
@@ -13 +19,2 @@ jobs:
- build-and-push-image-to-docker-hub:
+ build-and-push-images:
+ name: Build and push docker images to public Docker Hub
@@ -57,0 +65,51 @@ jobs:
+
+ code-quality-helm:
+ runs-on: ubuntu-latest
+ steps:
+ - name: Checkout sources
+ uses: actions/checkout@v3
+ - name: Lint chart with default values
+ run: helm lint
+ working-directory: chart
+ - name: Lint chart with dev values
+ run: helm lint --values env/dev.yaml
+ working-directory: chart
+ - name: Lint chart with prod values
+ run: helm lint --values env/prod.yaml
+ working-directory: chart
+
+ deploy-dev-and-prod:
+ if: ${{ endsWith(github.ref, '/main') }}
+ runs-on: ubuntu-latest
+ needs: [ build-and-push-images ]
+ environment: production
+ steps:
+ - name: Set outputs
+ id: vars
+ run: echo "sha_short=$(git rev-parse --short HEAD)" >> $GITHUB_OUTPUT
+ - name: Gen values
+ run: |
+ VALUES=$(cat <<-END
+ images:
+ jobs:
+ mongodbMigration:
+ tag: sha-${{ steps.vars.outputs.sha_short }}
+ services:
+ admin:
+ tag: sha-${{ steps.vars.outputs.sha_short }}
+ api:
+ tag: sha-${{ steps.vars.outputs.sha_short }}
+ worker:
+ tag: sha-${{ steps.vars.outputs.sha_short }}
+ END
+ )
+ echo "VALUES=$(echo "$VALUES" | yq -o=json | jq tostring)" >> $GITHUB_ENV
+ - name: Deploy on infra-deployments
+ uses: aurelien-baudet/workflow-dispatch@v2
+ with:
+ workflow: Update application values
+ repo: huggingface/infra-deployments
+ wait-for-completion: true
+ ref: refs/heads/main
+ token: ${{ secrets.ARGO_CD_TOKEN }}
+ inputs: '{"path": "datasets-server/*.yaml", "values": ${{ env.VALUES }}, "url": "${{ github.event.head_commit.url }}"}'
diff --git a/.github/workflows/chart-pr.yml b/.github/workflows/chart-pr.yml
new file mode 100644
index 00000000..606db1a5
--- /dev/null
+++ b/.github/workflows/chart-pr.yml
@@ -0,0 +1,42 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+name: chart PR
+on:
+ pull_request:
+ paths:
+ - 'chart/**'
+ - '.github/workflows/chart-pr.yml'
+jobs:
+ code-quality-helm:
+ runs-on: ubuntu-latest
+ steps:
+ - name: Checkout sources
+ uses: actions/checkout@v3
+ - name: Lint chart with default values
+ run: helm lint
+ working-directory: chart
+ - name: Lint chart with dev values
+ run: helm lint --values env/dev.yaml
+ working-directory: chart
+ - name: Lint chart with prod values
+ run: helm lint --values env/prod.yaml
+ working-directory: chart
+ helm-diff:
+ name: Diff
+ runs-on: ubuntu-latest
+ steps:
+ - name: Checkout repository
+ uses: actions/checkout@v3
+ - name: Tailscale
+ uses: tailscale/github-action@v1
+ with:
+ authkey: ${{ secrets.TAILSCALE_AUTHKEY }}
+ - uses: XciD/argocd-diff-action@master
+ name: ArgoCD Diff
+ with:
+ argocd-server-url: ${{ secrets.ARGO_CD_URL }}
+ argocd-token: ${{ secrets.ARGO_CD_CI_TOKEN }}
+ github-token: ${{ secrets.GITHUB_TOKEN }}
+ argocd-version: v2.6.1
+ argocd-extra-cli-args: --grpc-web --loglevel error
diff --git a/.github/workflows/chart.yml b/.github/workflows/chart.yml
deleted file mode 100644
index 0a2ff9d0..00000000
--- a/.github/workflows/chart.yml
+++ /dev/null
@@ -1,31 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-name: chart
-on:
- workflow_dispatch:
- push:
- branches:
- - main
- paths:
- - 'chart/**'
- - '.github/workflows/chart.yml'
- pull_request:
- paths:
- - 'chart/**'
- - '.github/workflows/chart.yml'
-jobs:
- code-quality-helm:
- runs-on: ubuntu-20.04
- steps:
- - name: Checkout sources
- uses: actions/checkout@v3
- - name: Lint chart with default values
- run: helm lint
- working-directory: chart
- - name: Lint chart with dev values
- run: helm lint --values env/dev.yaml
- working-directory: chart
- - name: Lint chart with prod values
- run: helm lint --values env/prod.yaml
- working-directory: chart
|
|
6f35d61f2225dcc487d06dde5c54d0d26e691ea7
|
Albert Villanova del Moral
| 2023-02-24T14:40:50 |
Update datasets dependency to 2.10.0 version (#829)
|
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 56a2060f..e113ba65 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -966 +966 @@ name = "datasets"
-version = "2.9.0"
+version = "2.10.0"
@@ -972,2 +972,2 @@ files = [
- {file = "datasets-2.9.0-py3-none-any.whl", hash = "sha256:f1aa5b98959cddb30f5077448204c8ce4235a4f1c8ec2473920660ebd6fc304f"},
- {file = "datasets-2.9.0.tar.gz", hash = "sha256:c82458d635539b5a5dbed0fba8837006dfc3c213a5bcc00e18a67789f0f0f16f"},
+ {file = "datasets-2.10.0-py3-none-any.whl", hash = "sha256:42968d0cc88608ebc4a92a2fc26b4cd65e38f3ce3f5bb90a468cae6a6f0de989"},
+ {file = "datasets-2.10.0.tar.gz", hash = "sha256:c2ddab0899e17590fe39bc5783828ac9a7331c74508ab7a345ea0dca3112ead0"},
@@ -978 +978 @@ aiohttp = "*"
-dill = "<0.3.7"
+dill = ">=0.3.0,<0.3.7"
@@ -997,2 +997,2 @@ audio = ["librosa"]
-benchmarks = ["numpy (==1.18.5)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"]
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=22.0,<23.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "flake8 (>=3.8.3)", "isort (>=5.0.0)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "s3fs", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"]
+benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"]
@@ -1000,2 +1000,3 @@ docs = ["s3fs"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "sqlalchemy", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
-quality = ["black (>=22.0,<23.0)", "flake8 (>=3.8.3)", "isort (>=5.0.0)", "pyyaml (>=5.3.1)"]
+jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "sqlalchemy (<2.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"]
@@ -1005 +1006 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"]
@@ -5519 +5520 @@ python-versions = "3.9.15"
-content-hash = "c20f6820064117295d9db223142b9ccdafd534b12c3304e3fa51ccf3dbcd622f"
+content-hash = "2bf2596d9309ec8715c9bf1e89f590057e44c8cd7e143d9f4bff3d27b9a51c5b"
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index 4edc2342..c6a10e8e 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -15 +15 @@ conllu = "^4.5.2"
-datasets = { extras = ["audio", "vision"], version = "~2.9.0" }
+datasets = { extras = ["audio", "vision"], version = "~2.10.0" }
@@ -83,3 +83,2 @@ strict = true
-# could be solved in datasets by adding __add__ to the __init__.py file
-implicit_reexport = true
-module = "datasets"
+module = "datasets.*"
+ignore_missing_imports = true
diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py
index 020d0feb..30888ddb 100644
--- a/services/worker/src/worker/job_runners/first_rows.py
+++ b/services/worker/src/worker/job_runners/first_rows.py
@@ -381 +381 @@ def to_features_list(features: Features) -> List[FeatureItem]:
- features_dict = features.to_dict() # type: ignore
+ features_dict = features.to_dict()
@@ -524 +524 @@ def compute_first_rows_response(
- iterable_dataset = iterable_dataset._resolve_features() # type: ignore
+ iterable_dataset = iterable_dataset._resolve_features()
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index 8f9b688d..02608877 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -507 +507 @@ def _request_size(url: str, hf_token: Optional[str] = None) -> Optional[int]:
-class _MockStreamingDownloadManager(StreamingDownloadManager):
+class _MockStreamingDownloadManager(StreamingDownloadManager): # type: ignore
@@ -517 +517 @@ class _MockStreamingDownloadManager(StreamingDownloadManager):
- parallel_min_length=np.inf, # type: ignore
+ parallel_min_length=np.inf,
@@ -547 +547 @@ def raise_if_too_big_from_external_data_files(
- builder._split_generators(mock_dl_manager) # type: ignore
+ builder._split_generators(mock_dl_manager)
@@ -793 +793 @@ def compute_parquet_and_dataset_info_response(
- dataset_info[config] = asdict(builder.info) # type: ignore
+ dataset_info[config] = asdict(builder.info)
diff --git a/services/worker/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py
index b2b11a0e..ea07ce7f 100644
--- a/services/worker/tests/fixtures/datasets.py
+++ b/services/worker/tests/fixtures/datasets.py
@@ -35 +35 @@ def other(content: Any, feature_type: Optional[FeatureType] = None) -> Dataset:
- features = Features({"col": feature_type}) # type: ignore
+ features = Features({"col": feature_type})
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 33039aec..afef7d98 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -333 +333 @@ def create_dataset_info_response_for_csv(dataset: str, config: str) -> Any:
- "checksum": "441b6927a5442803821415bdcb0f418731b0d2a525a7f2e68ce0df0e95d444de",
+ "checksum": None,
@@ -352,2 +352,2 @@ def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any:
- "num_bytes": 1124,
- "checksum": "3b630ef6ede66c5ced336df78fd99d98f835b459baadbe88a2cdf180709e9543",
+ "num_bytes": 1383,
+ "checksum": None,
@@ -356 +356 @@ def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any:
- "download_size": 1124,
+ "download_size": 1383,
@@ -358 +358 @@ def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any:
- "size_in_bytes": 1183,
+ "size_in_bytes": 1442,
|
|
af31a82b6b34cdf56e8ad18bbd317230620d7029
|
Sylvain Lesage
| 2023-02-24T13:55:19 |
chore: 🤖 upgrade dependencies to fix vulnerability (#845)
|
diff --git a/e2e/poetry.lock b/e2e/poetry.lock
index a611e9bd..210871b4 100644
--- a/e2e/poetry.lock
+++ b/e2e/poetry.lock
@@ -241 +241 @@ name = "coverage"
-version = "7.1.0"
+version = "7.2.0"
@@ -247,51 +247,51 @@ files = [
- {file = "coverage-7.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3b946bbcd5a8231383450b195cfb58cb01cbe7f8949f5758566b881df4b33baf"},
- {file = "coverage-7.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ec8e767f13be637d056f7e07e61d089e555f719b387a7070154ad80a0ff31801"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4a5a5879a939cb84959d86869132b00176197ca561c664fc21478c1eee60d75"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b643cb30821e7570c0aaf54feaf0bfb630b79059f85741843e9dc23f33aaca2c"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:32df215215f3af2c1617a55dbdfb403b772d463d54d219985ac7cd3bf124cada"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:33d1ae9d4079e05ac4cc1ef9e20c648f5afabf1a92adfaf2ccf509c50b85717f"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:29571503c37f2ef2138a306d23e7270687c0efb9cab4bd8038d609b5c2393a3a"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:63ffd21aa133ff48c4dff7adcc46b7ec8b565491bfc371212122dd999812ea1c"},
- {file = "coverage-7.1.0-cp310-cp310-win32.whl", hash = "sha256:4b14d5e09c656de5038a3f9bfe5228f53439282abcab87317c9f7f1acb280352"},
- {file = "coverage-7.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:8361be1c2c073919500b6601220a6f2f98ea0b6d2fec5014c1d9cfa23dd07038"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:da9b41d4539eefd408c46725fb76ecba3a50a3367cafb7dea5f250d0653c1040"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c5b15ed7644ae4bee0ecf74fee95808dcc34ba6ace87e8dfbf5cb0dc20eab45a"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d12d076582507ea460ea2a89a8c85cb558f83406c8a41dd641d7be9a32e1274f"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2617759031dae1bf183c16cef8fcfb3de7617f394c813fa5e8e46e9b82d4222"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4e4881fa9e9667afcc742f0c244d9364d197490fbc91d12ac3b5de0bf2df146"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:9d58885215094ab4a86a6aef044e42994a2bd76a446dc59b352622655ba6621b"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:ffeeb38ee4a80a30a6877c5c4c359e5498eec095878f1581453202bfacc8fbc2"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3baf5f126f30781b5e93dbefcc8271cb2491647f8283f20ac54d12161dff080e"},
- {file = "coverage-7.1.0-cp311-cp311-win32.whl", hash = "sha256:ded59300d6330be27bc6cf0b74b89ada58069ced87c48eaf9344e5e84b0072f7"},
- {file = "coverage-7.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:6a43c7823cd7427b4ed763aa7fb63901ca8288591323b58c9cd6ec31ad910f3c"},
- {file = "coverage-7.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:7a726d742816cb3a8973c8c9a97539c734b3a309345236cd533c4883dda05b8d"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc7c85a150501286f8b56bd8ed3aa4093f4b88fb68c0843d21ff9656f0009d6a"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f5b4198d85a3755d27e64c52f8c95d6333119e49fd001ae5798dac872c95e0f8"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ddb726cb861c3117a553f940372a495fe1078249ff5f8a5478c0576c7be12050"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:51b236e764840a6df0661b67e50697aaa0e7d4124ca95e5058fa3d7cbc240b7c"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:7ee5c9bb51695f80878faaa5598040dd6c9e172ddcf490382e8aedb8ec3fec8d"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c31b75ae466c053a98bf26843563b3b3517b8f37da4d47b1c582fdc703112bc3"},
- {file = "coverage-7.1.0-cp37-cp37m-win32.whl", hash = "sha256:3b155caf3760408d1cb903b21e6a97ad4e2bdad43cbc265e3ce0afb8e0057e73"},
- {file = "coverage-7.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2a60d6513781e87047c3e630b33b4d1e89f39836dac6e069ffee28c4786715f5"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f2cba5c6db29ce991029b5e4ac51eb36774458f0a3b8d3137241b32d1bb91f06"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:beeb129cacea34490ffd4d6153af70509aa3cda20fdda2ea1a2be870dfec8d52"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c45948f613d5d18c9ec5eaa203ce06a653334cf1bd47c783a12d0dd4fd9c851"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef382417db92ba23dfb5864a3fc9be27ea4894e86620d342a116b243ade5d35d"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7c7c0d0827e853315c9bbd43c1162c006dd808dbbe297db7ae66cd17b07830f0"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e5cdbb5cafcedea04924568d990e20ce7f1945a1dd54b560f879ee2d57226912"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:9817733f0d3ea91bea80de0f79ef971ae94f81ca52f9b66500c6a2fea8e4b4f8"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:218fe982371ac7387304153ecd51205f14e9d731b34fb0568181abaf7b443ba0"},
- {file = "coverage-7.1.0-cp38-cp38-win32.whl", hash = "sha256:04481245ef966fbd24ae9b9e537ce899ae584d521dfbe78f89cad003c38ca2ab"},
- {file = "coverage-7.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ae125d1134bf236acba8b83e74c603d1b30e207266121e76484562bc816344c"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2bf1d5f2084c3932b56b962a683074a3692bce7cabd3aa023c987a2a8e7612f6"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:98b85dd86514d889a2e3dd22ab3c18c9d0019e696478391d86708b805f4ea0fa"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38da2db80cc505a611938d8624801158e409928b136c8916cd2e203970dde4dc"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3164d31078fa9efe406e198aecd2a02d32a62fecbdef74f76dad6a46c7e48311"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db61a79c07331e88b9a9974815c075fbd812bc9dbc4dc44b366b5368a2936063"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9ccb092c9ede70b2517a57382a601619d20981f56f440eae7e4d7eaafd1d1d09"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:33ff26d0f6cc3ca8de13d14fde1ff8efe1456b53e3f0273e63cc8b3c84a063d8"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d47dd659a4ee952e90dc56c97d78132573dc5c7b09d61b416a9deef4ebe01a0c"},
- {file = "coverage-7.1.0-cp39-cp39-win32.whl", hash = "sha256:d248cd4a92065a4d4543b8331660121b31c4148dd00a691bfb7a5cdc7483cfa4"},
- {file = "coverage-7.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:7ed681b0f8e8bcbbffa58ba26fcf5dbc8f79e7997595bf071ed5430d8c08d6f3"},
- {file = "coverage-7.1.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:755e89e32376c850f826c425ece2c35a4fc266c081490eb0a841e7c1cb0d3bda"},
- {file = "coverage-7.1.0.tar.gz", hash = "sha256:10188fe543560ec4874f974b5305cd1a8bdcfa885ee00ea3a03733464c4ca265"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
+ {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
+ {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
+ {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
+ {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
+ {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
+ {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
+ {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
+ {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
+ {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
+ {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
+ {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
+ {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
+ {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
@@ -386,2 +386,2 @@ name = "gitpython"
-version = "3.1.30"
-description = "GitPython is a python library used to interact with Git repositories"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
@@ -392,2 +392,2 @@ files = [
- {file = "GitPython-3.1.30-py3-none-any.whl", hash = "sha256:cd455b0000615c60e286208ba540271af9fe531fa6a87cc590a7298785ab2882"},
- {file = "GitPython-3.1.30.tar.gz", hash = "sha256:769c2d83e13f5d938b7688479da374c4e3d49f71549aaf462b646db9602ea6f8"},
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
@@ -423 +423 @@ name = "huggingface-hub"
-version = "0.12.0"
+version = "0.12.1"
@@ -429,2 +429,2 @@ files = [
- {file = "huggingface_hub-0.12.0-py3-none-any.whl", hash = "sha256:93809eabbfb2058a808bddf8b2a70f645de3f9df73ce87ddf5163d4c74b71c0c"},
- {file = "huggingface_hub-0.12.0.tar.gz", hash = "sha256:da82c9ec8f9d8f976ffd3fd8249d20bb35c2dd3145a9f7ca1106f0ebefd9afa0"},
+ {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"},
+ {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"},
@@ -592 +592 @@ name = "markdown-it-py"
-version = "2.1.0"
+version = "2.2.0"
@@ -598,2 +598,2 @@ files = [
- {file = "markdown-it-py-2.1.0.tar.gz", hash = "sha256:cf7e59fed14b5ae17c0006eff14a2d9a00ed5f3a846148153899a0224e2c07da"},
- {file = "markdown_it_py-2.1.0-py3-none-any.whl", hash = "sha256:93de681e5c021a432c63147656fe21790bc01231e0cd2da73626f1aa3ac0fe27"},
+ {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"},
+ {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"},
@@ -606,4 +606,4 @@ mdurl = ">=0.1,<1.0"
-benchmarking = ["psutil", "pytest", "pytest-benchmark (>=3.2,<4.0)"]
-code-style = ["pre-commit (==2.6)"]
-compare = ["commonmark (>=0.9.1,<0.10.0)", "markdown (>=3.3.6,<3.4.0)", "mistletoe (>=0.8.1,<0.9.0)", "mistune (>=2.0.2,<2.1.0)", "panflute (>=2.1.3,<2.2.0)"]
-linkify = ["linkify-it-py (>=1.0,<2.0)"]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
@@ -703 +703 @@ name = "mypy"
-version = "1.0.0"
+version = "1.0.1"
@@ -709,26 +709,26 @@ files = [
- {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
- {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
- {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
- {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
- {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
- {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
- {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
- {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
- {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
- {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
- {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
- {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
- {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
- {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
- {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
- {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
- {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
- {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
- {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
- {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"},
+ {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"},
+ {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"},
+ {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"},
+ {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"},
+ {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"},
+ {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"},
+ {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"},
+ {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"},
+ {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"},
+ {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"},
+ {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"},
+ {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"},
+ {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"},
+ {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"},
+ {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"},
+ {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"},
+ {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"},
+ {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"},
@@ -863 +863 @@ name = "pip"
-version = "23.0"
+version = "23.0.1"
@@ -869,2 +869,2 @@ files = [
- {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
- {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
+ {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"},
+ {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"},
@@ -1230 +1230 @@ name = "setuptools"
-version = "67.3.1"
+version = "67.4.0"
@@ -1236,2 +1236,2 @@ files = [
- {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
- {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
+ {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"},
+ {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"},
@@ -1343 +1343 @@ name = "types-requests"
-version = "2.28.11.12"
+version = "2.28.11.14"
@@ -1349,2 +1349,2 @@ files = [
- {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
- {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+ {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
+ {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
@@ -1358 +1358 @@ name = "types-urllib3"
-version = "1.26.25.5"
+version = "1.26.25.7"
@@ -1364,2 +1364,2 @@ files = [
- {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
- {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
+ {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
+ {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index c5d13c02..3aba1994 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -802 +802 @@ name = "huggingface-hub"
-version = "0.12.0"
+version = "0.12.1"
@@ -808,2 +808,2 @@ files = [
- {file = "huggingface_hub-0.12.0-py3-none-any.whl", hash = "sha256:93809eabbfb2058a808bddf8b2a70f645de3f9df73ce87ddf5163d4c74b71c0c"},
- {file = "huggingface_hub-0.12.0.tar.gz", hash = "sha256:da82c9ec8f9d8f976ffd3fd8249d20bb35c2dd3145a9f7ca1106f0ebefd9afa0"},
+ {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"},
+ {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"},
@@ -845 +845 @@ name = "importlib-resources"
-version = "5.10.2"
+version = "5.12.0"
@@ -851,2 +851,2 @@ files = [
- {file = "importlib_resources-5.10.2-py3-none-any.whl", hash = "sha256:7d543798b0beca10b6a01ac7cafda9f822c54db9e8376a6bf57e0cbd74d486b6"},
- {file = "importlib_resources-5.10.2.tar.gz", hash = "sha256:e4a96c8cc0339647ff9a5e0550d9f276fc5a01ffa276012b58ec108cfd7b8484"},
+ {file = "importlib_resources-5.12.0-py3-none-any.whl", hash = "sha256:7b1deeebbf351c7578e09bf2f63fa2ce8b5ffec296e0d349139d43cca061a81a"},
+ {file = "importlib_resources-5.12.0.tar.gz", hash = "sha256:4be82589bf5c1d7999aedf2a45159d10cb3ca4f19b2271f8792bc8e6da7b22f6"},
@@ -980 +980 @@ name = "linkify-it-py"
-version = "1.0.3"
+version = "2.0.0"
@@ -986,2 +986,2 @@ files = [
- {file = "linkify-it-py-1.0.3.tar.gz", hash = "sha256:2b3f168d5ce75e3a425e34b341a6b73e116b5d9ed8dbbbf5dc7456843b7ce2ee"},
- {file = "linkify_it_py-1.0.3-py3-none-any.whl", hash = "sha256:11e29f00150cddaa8f434153f103c14716e7e097a8fd372d9eb1ed06ed91524d"},
+ {file = "linkify-it-py-2.0.0.tar.gz", hash = "sha256:476464480906bed8b2fa3813bf55566282e55214ad7e41b7d1c2b564666caf2f"},
+ {file = "linkify_it_py-2.0.0-py3-none-any.whl", hash = "sha256:1bff43823e24e507a099e328fc54696124423dd6320c75a9da45b4b754b748ad"},
@@ -1001 +1001 @@ name = "markdown-it-py"
-version = "2.1.0"
+version = "2.2.0"
@@ -1007,2 +1007,2 @@ files = [
- {file = "markdown-it-py-2.1.0.tar.gz", hash = "sha256:cf7e59fed14b5ae17c0006eff14a2d9a00ed5f3a846148153899a0224e2c07da"},
- {file = "markdown_it_py-2.1.0-py3-none-any.whl", hash = "sha256:93de681e5c021a432c63147656fe21790bc01231e0cd2da73626f1aa3ac0fe27"},
+ {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"},
+ {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"},
@@ -1012 +1012 @@ files = [
-linkify-it-py = {version = ">=1.0,<2.0", optional = true, markers = "extra == \"linkify\""}
+linkify-it-py = {version = ">=1,<3", optional = true, markers = "extra == \"linkify\""}
@@ -1017,4 +1017,4 @@ mdurl = ">=0.1,<1.0"
-benchmarking = ["psutil", "pytest", "pytest-benchmark (>=3.2,<4.0)"]
-code-style = ["pre-commit (==2.6)"]
-compare = ["commonmark (>=0.9.1,<0.10.0)", "markdown (>=3.3.6,<3.4.0)", "mistletoe (>=0.8.1,<0.9.0)", "mistune (>=2.0.2,<2.1.0)", "panflute (>=2.1.3,<2.2.0)"]
-linkify = ["linkify-it-py (>=1.0,<2.0)"]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
@@ -1151 +1151 @@ name = "mdit-py-plugins"
-version = "0.3.3"
+version = "0.3.4"
@@ -1157,2 +1157,2 @@ files = [
- {file = "mdit-py-plugins-0.3.3.tar.gz", hash = "sha256:5cfd7e7ac582a594e23ba6546a2f406e94e42eb33ae596d0734781261c251260"},
- {file = "mdit_py_plugins-0.3.3-py3-none-any.whl", hash = "sha256:36d08a29def19ec43acdcd8ba471d3ebab132e7879d442760d963f19913e04b9"},
+ {file = "mdit-py-plugins-0.3.4.tar.gz", hash = "sha256:3278aab2e2b692539082f05e1243f24742194ffd92481f48844f057b51971283"},
+ {file = "mdit_py_plugins-0.3.4-py3-none-any.whl", hash = "sha256:4f1441264ac5cb39fa40a5901921c2acf314ea098d75629750c138f80d552cdf"},
@@ -1543 +1543 @@ name = "pydantic"
-version = "1.10.4"
+version = "1.10.5"
@@ -1549,36 +1549,36 @@ files = [
- {file = "pydantic-1.10.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b5635de53e6686fe7a44b5cf25fcc419a0d5e5c1a1efe73d49d48fe7586db854"},
- {file = "pydantic-1.10.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:6dc1cc241440ed7ca9ab59d9929075445da6b7c94ced281b3dd4cfe6c8cff817"},
- {file = "pydantic-1.10.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:51bdeb10d2db0f288e71d49c9cefa609bca271720ecd0c58009bd7504a0c464c"},
- {file = "pydantic-1.10.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78cec42b95dbb500a1f7120bdf95c401f6abb616bbe8785ef09887306792e66e"},
- {file = "pydantic-1.10.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:8775d4ef5e7299a2f4699501077a0defdaac5b6c4321173bcb0f3c496fbadf85"},
- {file = "pydantic-1.10.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:572066051eeac73d23f95ba9a71349c42a3e05999d0ee1572b7860235b850cc6"},
- {file = "pydantic-1.10.4-cp310-cp310-win_amd64.whl", hash = "sha256:7feb6a2d401f4d6863050f58325b8d99c1e56f4512d98b11ac64ad1751dc647d"},
- {file = "pydantic-1.10.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:39f4a73e5342b25c2959529f07f026ef58147249f9b7431e1ba8414a36761f53"},
- {file = "pydantic-1.10.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:983e720704431a6573d626b00662eb78a07148c9115129f9b4351091ec95ecc3"},
- {file = "pydantic-1.10.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75d52162fe6b2b55964fbb0af2ee58e99791a3138588c482572bb6087953113a"},
- {file = "pydantic-1.10.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fdf8d759ef326962b4678d89e275ffc55b7ce59d917d9f72233762061fd04a2d"},
- {file = "pydantic-1.10.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:05a81b006be15655b2a1bae5faa4280cf7c81d0e09fcb49b342ebf826abe5a72"},
- {file = "pydantic-1.10.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d88c4c0e5c5dfd05092a4b271282ef0588e5f4aaf345778056fc5259ba098857"},
- {file = "pydantic-1.10.4-cp311-cp311-win_amd64.whl", hash = "sha256:6a05a9db1ef5be0fe63e988f9617ca2551013f55000289c671f71ec16f4985e3"},
- {file = "pydantic-1.10.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:887ca463c3bc47103c123bc06919c86720e80e1214aab79e9b779cda0ff92a00"},
- {file = "pydantic-1.10.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fdf88ab63c3ee282c76d652fc86518aacb737ff35796023fae56a65ced1a5978"},
- {file = "pydantic-1.10.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a48f1953c4a1d9bd0b5167ac50da9a79f6072c63c4cef4cf2a3736994903583e"},
- {file = "pydantic-1.10.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:a9f2de23bec87ff306aef658384b02aa7c32389766af3c5dee9ce33e80222dfa"},
- {file = "pydantic-1.10.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:cd8702c5142afda03dc2b1ee6bc358b62b3735b2cce53fc77b31ca9f728e4bc8"},
- {file = "pydantic-1.10.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6e7124d6855b2780611d9f5e1e145e86667eaa3bd9459192c8dc1a097f5e9903"},
- {file = "pydantic-1.10.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b53e1d41e97063d51a02821b80538053ee4608b9a181c1005441f1673c55423"},
- {file = "pydantic-1.10.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:55b1625899acd33229c4352ce0ae54038529b412bd51c4915349b49ca575258f"},
- {file = "pydantic-1.10.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:301d626a59edbe5dfb48fcae245896379a450d04baeed50ef40d8199f2733b06"},
- {file = "pydantic-1.10.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b6f9d649892a6f54a39ed56b8dfd5e08b5f3be5f893da430bed76975f3735d15"},
- {file = "pydantic-1.10.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:d7b5a3821225f5c43496c324b0d6875fde910a1c2933d726a743ce328fbb2a8c"},
- {file = "pydantic-1.10.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:f2f7eb6273dd12472d7f218e1fef6f7c7c2f00ac2e1ecde4db8824c457300416"},
- {file = "pydantic-1.10.4-cp38-cp38-win_amd64.whl", hash = "sha256:4b05697738e7d2040696b0a66d9f0a10bec0efa1883ca75ee9e55baf511909d6"},
- {file = "pydantic-1.10.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:a9a6747cac06c2beb466064dda999a13176b23535e4c496c9d48e6406f92d42d"},
- {file = "pydantic-1.10.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:eb992a1ef739cc7b543576337bebfc62c0e6567434e522e97291b251a41dad7f"},
- {file = "pydantic-1.10.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:990406d226dea0e8f25f643b370224771878142155b879784ce89f633541a024"},
- {file = "pydantic-1.10.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2e82a6d37a95e0b1b42b82ab340ada3963aea1317fd7f888bb6b9dfbf4fff57c"},
- {file = "pydantic-1.10.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:9193d4f4ee8feca58bc56c8306bcb820f5c7905fd919e0750acdeeeef0615b28"},
- {file = "pydantic-1.10.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:2b3ce5f16deb45c472dde1a0ee05619298c864a20cded09c4edd820e1454129f"},
- {file = "pydantic-1.10.4-cp39-cp39-win_amd64.whl", hash = "sha256:9cbdc268a62d9a98c56e2452d6c41c0263d64a2009aac69246486f01b4f594c4"},
- {file = "pydantic-1.10.4-py3-none-any.whl", hash = "sha256:4948f264678c703f3877d1c8877c4e3b2e12e549c57795107f08cf70c6ec7774"},
- {file = "pydantic-1.10.4.tar.gz", hash = "sha256:b9a3859f24eb4e097502a3be1fb4b2abb79b6103dd9e2e0edb70613a4459a648"},
+ {file = "pydantic-1.10.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:5920824fe1e21cbb3e38cf0f3dd24857c8959801d1031ce1fac1d50857a03bfb"},
+ {file = "pydantic-1.10.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3bb99cf9655b377db1a9e47fa4479e3330ea96f4123c6c8200e482704bf1eda2"},
+ {file = "pydantic-1.10.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2185a3b3d98ab4506a3f6707569802d2d92c3a7ba3a9a35683a7709ea6c2aaa2"},
+ {file = "pydantic-1.10.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f582cac9d11c227c652d3ce8ee223d94eb06f4228b52a8adaafa9fa62e73d5c9"},
+ {file = "pydantic-1.10.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c9e5b778b6842f135902e2d82624008c6a79710207e28e86966cd136c621bfee"},
+ {file = "pydantic-1.10.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:72ef3783be8cbdef6bca034606a5de3862be6b72415dc5cb1fb8ddbac110049a"},
+ {file = "pydantic-1.10.5-cp310-cp310-win_amd64.whl", hash = "sha256:45edea10b75d3da43cfda12f3792833a3fa70b6eee4db1ed6aed528cef17c74e"},
+ {file = "pydantic-1.10.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:63200cd8af1af2c07964546b7bc8f217e8bda9d0a2ef0ee0c797b36353914984"},
+ {file = "pydantic-1.10.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:305d0376c516b0dfa1dbefeae8c21042b57b496892d721905a6ec6b79494a66d"},
+ {file = "pydantic-1.10.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1fd326aff5d6c36f05735c7c9b3d5b0e933b4ca52ad0b6e4b38038d82703d35b"},
+ {file = "pydantic-1.10.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6bb0452d7b8516178c969d305d9630a3c9b8cf16fcf4713261c9ebd465af0d73"},
+ {file = "pydantic-1.10.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:9a9d9155e2a9f38b2eb9374c88f02fd4d6851ae17b65ee786a87d032f87008f8"},
+ {file = "pydantic-1.10.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f836444b4c5ece128b23ec36a446c9ab7f9b0f7981d0d27e13a7c366ee163f8a"},
+ {file = "pydantic-1.10.5-cp311-cp311-win_amd64.whl", hash = "sha256:8481dca324e1c7b715ce091a698b181054d22072e848b6fc7895cd86f79b4449"},
+ {file = "pydantic-1.10.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:87f831e81ea0589cd18257f84386bf30154c5f4bed373b7b75e5cb0b5d53ea87"},
+ {file = "pydantic-1.10.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ce1612e98c6326f10888df951a26ec1a577d8df49ddcaea87773bfbe23ba5cc"},
+ {file = "pydantic-1.10.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:58e41dd1e977531ac6073b11baac8c013f3cd8706a01d3dc74e86955be8b2c0c"},
+ {file = "pydantic-1.10.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:6a4b0aab29061262065bbdede617ef99cc5914d1bf0ddc8bcd8e3d7928d85bd6"},
+ {file = "pydantic-1.10.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:36e44a4de37b8aecffa81c081dbfe42c4d2bf9f6dff34d03dce157ec65eb0f15"},
+ {file = "pydantic-1.10.5-cp37-cp37m-win_amd64.whl", hash = "sha256:261f357f0aecda005934e413dfd7aa4077004a174dafe414a8325e6098a8e419"},
+ {file = "pydantic-1.10.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b429f7c457aebb7fbe7cd69c418d1cd7c6fdc4d3c8697f45af78b8d5a7955760"},
+ {file = "pydantic-1.10.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:663d2dd78596c5fa3eb996bc3f34b8c2a592648ad10008f98d1348be7ae212fb"},
+ {file = "pydantic-1.10.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:51782fd81f09edcf265823c3bf43ff36d00db246eca39ee765ef58dc8421a642"},
+ {file = "pydantic-1.10.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c428c0f64a86661fb4873495c4fac430ec7a7cef2b8c1c28f3d1a7277f9ea5ab"},
+ {file = "pydantic-1.10.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:76c930ad0746c70f0368c4596020b736ab65b473c1f9b3872310a835d852eb19"},
+ {file = "pydantic-1.10.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:3257bd714de9db2102b742570a56bf7978e90441193acac109b1f500290f5718"},
+ {file = "pydantic-1.10.5-cp38-cp38-win_amd64.whl", hash = "sha256:f5bee6c523d13944a1fdc6f0525bc86dbbd94372f17b83fa6331aabacc8fd08e"},
+ {file = "pydantic-1.10.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:532e97c35719f137ee5405bd3eeddc5c06eb91a032bc755a44e34a712420daf3"},
+ {file = "pydantic-1.10.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:ca9075ab3de9e48b75fa8ccb897c34ccc1519177ad8841d99f7fd74cf43be5bf"},
+ {file = "pydantic-1.10.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bd46a0e6296346c477e59a954da57beaf9c538da37b9df482e50f836e4a7d4bb"},
+ {file = "pydantic-1.10.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3353072625ea2a9a6c81ad01b91e5c07fa70deb06368c71307529abf70d23325"},
+ {file = "pydantic-1.10.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:3f9d9b2be177c3cb6027cd67fbf323586417868c06c3c85d0d101703136e6b31"},
+ {file = "pydantic-1.10.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:b473d00ccd5c2061fd896ac127b7755baad233f8d996ea288af14ae09f8e0d1e"},
+ {file = "pydantic-1.10.5-cp39-cp39-win_amd64.whl", hash = "sha256:5f3bc8f103b56a8c88021d481410874b1f13edf6e838da607dcb57ecff9b4594"},
+ {file = "pydantic-1.10.5-py3-none-any.whl", hash = "sha256:7c5b94d598c90f2f46b3a983ffb46ab806a67099d118ae0da7ef21a2a4033b28"},
+ {file = "pydantic-1.10.5.tar.gz", hash = "sha256:9e337ac83686645a46db0e825acceea8e02fca4062483f40e9ae178e8bd1103a"},
@@ -2097 +2097 @@ name = "zipp"
-version = "3.13.0"
+version = "3.14.0"
@@ -2103,2 +2103,2 @@ files = [
- {file = "zipp-3.13.0-py3-none-any.whl", hash = "sha256:e8b2a36ea17df80ffe9e2c4fda3f693c3dad6df1697d3cd3af232db680950b0b"},
- {file = "zipp-3.13.0.tar.gz", hash = "sha256:23f70e964bc11a34cef175bc90ba2914e1e4545ea1e3e2f67c079671883f9cb6"},
+ {file = "zipp-3.14.0-py3-none-any.whl", hash = "sha256:188834565033387710d046e3fe96acfc9b5e86cbca7f39ff69cf21a4128198b7"},
+ {file = "zipp-3.14.0.tar.gz", hash = "sha256:9e5421e176ef5ab4c0ad896624e87a7b2f07aca746c9b2aa305952800cb8eecb"},
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index e5d9eef4..d5902509 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -253 +253 @@ name = "coverage"
-version = "7.1.0"
+version = "7.2.0"
@@ -259,51 +259,51 @@ files = [
- {file = "coverage-7.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3b946bbcd5a8231383450b195cfb58cb01cbe7f8949f5758566b881df4b33baf"},
- {file = "coverage-7.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ec8e767f13be637d056f7e07e61d089e555f719b387a7070154ad80a0ff31801"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4a5a5879a939cb84959d86869132b00176197ca561c664fc21478c1eee60d75"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b643cb30821e7570c0aaf54feaf0bfb630b79059f85741843e9dc23f33aaca2c"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:32df215215f3af2c1617a55dbdfb403b772d463d54d219985ac7cd3bf124cada"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:33d1ae9d4079e05ac4cc1ef9e20c648f5afabf1a92adfaf2ccf509c50b85717f"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:29571503c37f2ef2138a306d23e7270687c0efb9cab4bd8038d609b5c2393a3a"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:63ffd21aa133ff48c4dff7adcc46b7ec8b565491bfc371212122dd999812ea1c"},
- {file = "coverage-7.1.0-cp310-cp310-win32.whl", hash = "sha256:4b14d5e09c656de5038a3f9bfe5228f53439282abcab87317c9f7f1acb280352"},
- {file = "coverage-7.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:8361be1c2c073919500b6601220a6f2f98ea0b6d2fec5014c1d9cfa23dd07038"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:da9b41d4539eefd408c46725fb76ecba3a50a3367cafb7dea5f250d0653c1040"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c5b15ed7644ae4bee0ecf74fee95808dcc34ba6ace87e8dfbf5cb0dc20eab45a"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d12d076582507ea460ea2a89a8c85cb558f83406c8a41dd641d7be9a32e1274f"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2617759031dae1bf183c16cef8fcfb3de7617f394c813fa5e8e46e9b82d4222"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4e4881fa9e9667afcc742f0c244d9364d197490fbc91d12ac3b5de0bf2df146"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:9d58885215094ab4a86a6aef044e42994a2bd76a446dc59b352622655ba6621b"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:ffeeb38ee4a80a30a6877c5c4c359e5498eec095878f1581453202bfacc8fbc2"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3baf5f126f30781b5e93dbefcc8271cb2491647f8283f20ac54d12161dff080e"},
- {file = "coverage-7.1.0-cp311-cp311-win32.whl", hash = "sha256:ded59300d6330be27bc6cf0b74b89ada58069ced87c48eaf9344e5e84b0072f7"},
- {file = "coverage-7.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:6a43c7823cd7427b4ed763aa7fb63901ca8288591323b58c9cd6ec31ad910f3c"},
- {file = "coverage-7.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:7a726d742816cb3a8973c8c9a97539c734b3a309345236cd533c4883dda05b8d"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc7c85a150501286f8b56bd8ed3aa4093f4b88fb68c0843d21ff9656f0009d6a"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f5b4198d85a3755d27e64c52f8c95d6333119e49fd001ae5798dac872c95e0f8"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ddb726cb861c3117a553f940372a495fe1078249ff5f8a5478c0576c7be12050"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:51b236e764840a6df0661b67e50697aaa0e7d4124ca95e5058fa3d7cbc240b7c"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:7ee5c9bb51695f80878faaa5598040dd6c9e172ddcf490382e8aedb8ec3fec8d"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c31b75ae466c053a98bf26843563b3b3517b8f37da4d47b1c582fdc703112bc3"},
- {file = "coverage-7.1.0-cp37-cp37m-win32.whl", hash = "sha256:3b155caf3760408d1cb903b21e6a97ad4e2bdad43cbc265e3ce0afb8e0057e73"},
- {file = "coverage-7.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2a60d6513781e87047c3e630b33b4d1e89f39836dac6e069ffee28c4786715f5"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f2cba5c6db29ce991029b5e4ac51eb36774458f0a3b8d3137241b32d1bb91f06"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:beeb129cacea34490ffd4d6153af70509aa3cda20fdda2ea1a2be870dfec8d52"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c45948f613d5d18c9ec5eaa203ce06a653334cf1bd47c783a12d0dd4fd9c851"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef382417db92ba23dfb5864a3fc9be27ea4894e86620d342a116b243ade5d35d"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7c7c0d0827e853315c9bbd43c1162c006dd808dbbe297db7ae66cd17b07830f0"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e5cdbb5cafcedea04924568d990e20ce7f1945a1dd54b560f879ee2d57226912"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:9817733f0d3ea91bea80de0f79ef971ae94f81ca52f9b66500c6a2fea8e4b4f8"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:218fe982371ac7387304153ecd51205f14e9d731b34fb0568181abaf7b443ba0"},
- {file = "coverage-7.1.0-cp38-cp38-win32.whl", hash = "sha256:04481245ef966fbd24ae9b9e537ce899ae584d521dfbe78f89cad003c38ca2ab"},
- {file = "coverage-7.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ae125d1134bf236acba8b83e74c603d1b30e207266121e76484562bc816344c"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2bf1d5f2084c3932b56b962a683074a3692bce7cabd3aa023c987a2a8e7612f6"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:98b85dd86514d889a2e3dd22ab3c18c9d0019e696478391d86708b805f4ea0fa"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38da2db80cc505a611938d8624801158e409928b136c8916cd2e203970dde4dc"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3164d31078fa9efe406e198aecd2a02d32a62fecbdef74f76dad6a46c7e48311"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db61a79c07331e88b9a9974815c075fbd812bc9dbc4dc44b366b5368a2936063"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9ccb092c9ede70b2517a57382a601619d20981f56f440eae7e4d7eaafd1d1d09"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:33ff26d0f6cc3ca8de13d14fde1ff8efe1456b53e3f0273e63cc8b3c84a063d8"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d47dd659a4ee952e90dc56c97d78132573dc5c7b09d61b416a9deef4ebe01a0c"},
- {file = "coverage-7.1.0-cp39-cp39-win32.whl", hash = "sha256:d248cd4a92065a4d4543b8331660121b31c4148dd00a691bfb7a5cdc7483cfa4"},
- {file = "coverage-7.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:7ed681b0f8e8bcbbffa58ba26fcf5dbc8f79e7997595bf071ed5430d8c08d6f3"},
- {file = "coverage-7.1.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:755e89e32376c850f826c425ece2c35a4fc266c081490eb0a841e7c1cb0d3bda"},
- {file = "coverage-7.1.0.tar.gz", hash = "sha256:10188fe543560ec4874f974b5305cd1a8bdcfa885ee00ea3a03733464c4ca265"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
+ {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
+ {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
+ {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
+ {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
+ {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
+ {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
+ {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
+ {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
+ {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
+ {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
+ {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
+ {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
+ {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
@@ -436,2 +436,2 @@ name = "gitpython"
-version = "3.1.30"
-description = "GitPython is a python library used to interact with Git repositories"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
@@ -442,2 +442,2 @@ files = [
- {file = "GitPython-3.1.30-py3-none-any.whl", hash = "sha256:cd455b0000615c60e286208ba540271af9fe531fa6a87cc590a7298785ab2882"},
- {file = "GitPython-3.1.30.tar.gz", hash = "sha256:769c2d83e13f5d938b7688479da374c4e3d49f71549aaf462b646db9602ea6f8"},
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
@@ -473 +473 @@ name = "huggingface-hub"
-version = "0.12.0"
+version = "0.12.1"
@@ -479,2 +479,2 @@ files = [
- {file = "huggingface_hub-0.12.0-py3-none-any.whl", hash = "sha256:93809eabbfb2058a808bddf8b2a70f645de3f9df73ce87ddf5163d4c74b71c0c"},
- {file = "huggingface_hub-0.12.0.tar.gz", hash = "sha256:da82c9ec8f9d8f976ffd3fd8249d20bb35c2dd3145a9f7ca1106f0ebefd9afa0"},
+ {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"},
+ {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"},
@@ -583 +583 @@ name = "markdown-it-py"
-version = "2.1.0"
+version = "2.2.0"
@@ -589,2 +589,2 @@ files = [
- {file = "markdown-it-py-2.1.0.tar.gz", hash = "sha256:cf7e59fed14b5ae17c0006eff14a2d9a00ed5f3a846148153899a0224e2c07da"},
- {file = "markdown_it_py-2.1.0-py3-none-any.whl", hash = "sha256:93de681e5c021a432c63147656fe21790bc01231e0cd2da73626f1aa3ac0fe27"},
+ {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"},
+ {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"},
@@ -597,4 +597,4 @@ mdurl = ">=0.1,<1.0"
-benchmarking = ["psutil", "pytest", "pytest-benchmark (>=3.2,<4.0)"]
-code-style = ["pre-commit (==2.6)"]
-compare = ["commonmark (>=0.9.1,<0.10.0)", "markdown (>=3.3.6,<3.4.0)", "mistletoe (>=0.8.1,<0.9.0)", "mistune (>=2.0.2,<2.1.0)", "panflute (>=2.1.3,<2.2.0)"]
-linkify = ["linkify-it-py (>=1.0,<2.0)"]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
@@ -742 +742 @@ name = "mypy"
-version = "1.0.0"
+version = "1.0.1"
@@ -748,26 +748,26 @@ files = [
- {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
- {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
- {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
- {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
- {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
- {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
- {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
- {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
- {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
- {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
- {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
- {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
- {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
- {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
- {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
- {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
- {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
- {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
- {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
- {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"},
+ {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"},
+ {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"},
+ {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"},
+ {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"},
+ {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"},
+ {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"},
+ {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"},
+ {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"},
+ {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"},
+ {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"},
+ {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"},
+ {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"},
+ {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"},
+ {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"},
+ {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"},
+ {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"},
+ {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"},
+ {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"},
@@ -907 +907 @@ name = "pip"
-version = "23.0"
+version = "23.0.1"
@@ -913,2 +913,2 @@ files = [
- {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
- {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
+ {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"},
+ {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"},
@@ -1272 +1272 @@ name = "python-dotenv"
-version = "0.21.1"
+version = "1.0.0"
@@ -1276 +1276 @@ optional = false
-python-versions = ">=3.7"
+python-versions = ">=3.8"
@@ -1278,2 +1278,2 @@ files = [
- {file = "python-dotenv-0.21.1.tar.gz", hash = "sha256:1c93de8f636cde3ce377292818d0e440b6e45a82f215c3744979151fa8151c49"},
- {file = "python_dotenv-0.21.1-py3-none-any.whl", hash = "sha256:41e12e0318bebc859fcc4d97d4db8d20ad21721a6aa5047dd59f090391cb549a"},
+ {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"},
+ {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"},
@@ -1396 +1396 @@ name = "setuptools"
-version = "67.3.1"
+version = "67.4.0"
@@ -1402,2 +1402,2 @@ files = [
- {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
- {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
+ {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"},
+ {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"},
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index bcd413ac..72864289 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -253 +253 @@ name = "coverage"
-version = "7.1.0"
+version = "7.2.0"
@@ -259,51 +259,51 @@ files = [
- {file = "coverage-7.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3b946bbcd5a8231383450b195cfb58cb01cbe7f8949f5758566b881df4b33baf"},
- {file = "coverage-7.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ec8e767f13be637d056f7e07e61d089e555f719b387a7070154ad80a0ff31801"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4a5a5879a939cb84959d86869132b00176197ca561c664fc21478c1eee60d75"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b643cb30821e7570c0aaf54feaf0bfb630b79059f85741843e9dc23f33aaca2c"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:32df215215f3af2c1617a55dbdfb403b772d463d54d219985ac7cd3bf124cada"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:33d1ae9d4079e05ac4cc1ef9e20c648f5afabf1a92adfaf2ccf509c50b85717f"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:29571503c37f2ef2138a306d23e7270687c0efb9cab4bd8038d609b5c2393a3a"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:63ffd21aa133ff48c4dff7adcc46b7ec8b565491bfc371212122dd999812ea1c"},
- {file = "coverage-7.1.0-cp310-cp310-win32.whl", hash = "sha256:4b14d5e09c656de5038a3f9bfe5228f53439282abcab87317c9f7f1acb280352"},
- {file = "coverage-7.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:8361be1c2c073919500b6601220a6f2f98ea0b6d2fec5014c1d9cfa23dd07038"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:da9b41d4539eefd408c46725fb76ecba3a50a3367cafb7dea5f250d0653c1040"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c5b15ed7644ae4bee0ecf74fee95808dcc34ba6ace87e8dfbf5cb0dc20eab45a"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d12d076582507ea460ea2a89a8c85cb558f83406c8a41dd641d7be9a32e1274f"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2617759031dae1bf183c16cef8fcfb3de7617f394c813fa5e8e46e9b82d4222"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4e4881fa9e9667afcc742f0c244d9364d197490fbc91d12ac3b5de0bf2df146"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:9d58885215094ab4a86a6aef044e42994a2bd76a446dc59b352622655ba6621b"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:ffeeb38ee4a80a30a6877c5c4c359e5498eec095878f1581453202bfacc8fbc2"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3baf5f126f30781b5e93dbefcc8271cb2491647f8283f20ac54d12161dff080e"},
- {file = "coverage-7.1.0-cp311-cp311-win32.whl", hash = "sha256:ded59300d6330be27bc6cf0b74b89ada58069ced87c48eaf9344e5e84b0072f7"},
- {file = "coverage-7.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:6a43c7823cd7427b4ed763aa7fb63901ca8288591323b58c9cd6ec31ad910f3c"},
- {file = "coverage-7.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:7a726d742816cb3a8973c8c9a97539c734b3a309345236cd533c4883dda05b8d"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc7c85a150501286f8b56bd8ed3aa4093f4b88fb68c0843d21ff9656f0009d6a"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f5b4198d85a3755d27e64c52f8c95d6333119e49fd001ae5798dac872c95e0f8"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ddb726cb861c3117a553f940372a495fe1078249ff5f8a5478c0576c7be12050"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:51b236e764840a6df0661b67e50697aaa0e7d4124ca95e5058fa3d7cbc240b7c"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:7ee5c9bb51695f80878faaa5598040dd6c9e172ddcf490382e8aedb8ec3fec8d"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c31b75ae466c053a98bf26843563b3b3517b8f37da4d47b1c582fdc703112bc3"},
- {file = "coverage-7.1.0-cp37-cp37m-win32.whl", hash = "sha256:3b155caf3760408d1cb903b21e6a97ad4e2bdad43cbc265e3ce0afb8e0057e73"},
- {file = "coverage-7.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2a60d6513781e87047c3e630b33b4d1e89f39836dac6e069ffee28c4786715f5"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f2cba5c6db29ce991029b5e4ac51eb36774458f0a3b8d3137241b32d1bb91f06"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:beeb129cacea34490ffd4d6153af70509aa3cda20fdda2ea1a2be870dfec8d52"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c45948f613d5d18c9ec5eaa203ce06a653334cf1bd47c783a12d0dd4fd9c851"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef382417db92ba23dfb5864a3fc9be27ea4894e86620d342a116b243ade5d35d"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7c7c0d0827e853315c9bbd43c1162c006dd808dbbe297db7ae66cd17b07830f0"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e5cdbb5cafcedea04924568d990e20ce7f1945a1dd54b560f879ee2d57226912"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:9817733f0d3ea91bea80de0f79ef971ae94f81ca52f9b66500c6a2fea8e4b4f8"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:218fe982371ac7387304153ecd51205f14e9d731b34fb0568181abaf7b443ba0"},
- {file = "coverage-7.1.0-cp38-cp38-win32.whl", hash = "sha256:04481245ef966fbd24ae9b9e537ce899ae584d521dfbe78f89cad003c38ca2ab"},
- {file = "coverage-7.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ae125d1134bf236acba8b83e74c603d1b30e207266121e76484562bc816344c"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2bf1d5f2084c3932b56b962a683074a3692bce7cabd3aa023c987a2a8e7612f6"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:98b85dd86514d889a2e3dd22ab3c18c9d0019e696478391d86708b805f4ea0fa"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38da2db80cc505a611938d8624801158e409928b136c8916cd2e203970dde4dc"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3164d31078fa9efe406e198aecd2a02d32a62fecbdef74f76dad6a46c7e48311"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db61a79c07331e88b9a9974815c075fbd812bc9dbc4dc44b366b5368a2936063"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9ccb092c9ede70b2517a57382a601619d20981f56f440eae7e4d7eaafd1d1d09"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:33ff26d0f6cc3ca8de13d14fde1ff8efe1456b53e3f0273e63cc8b3c84a063d8"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d47dd659a4ee952e90dc56c97d78132573dc5c7b09d61b416a9deef4ebe01a0c"},
- {file = "coverage-7.1.0-cp39-cp39-win32.whl", hash = "sha256:d248cd4a92065a4d4543b8331660121b31c4148dd00a691bfb7a5cdc7483cfa4"},
- {file = "coverage-7.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:7ed681b0f8e8bcbbffa58ba26fcf5dbc8f79e7997595bf071ed5430d8c08d6f3"},
- {file = "coverage-7.1.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:755e89e32376c850f826c425ece2c35a4fc266c081490eb0a841e7c1cb0d3bda"},
- {file = "coverage-7.1.0.tar.gz", hash = "sha256:10188fe543560ec4874f974b5305cd1a8bdcfa885ee00ea3a03733464c4ca265"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
+ {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
+ {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
+ {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
+ {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
+ {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
+ {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
+ {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
+ {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
+ {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
+ {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
+ {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
+ {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
+ {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
@@ -436,2 +436,2 @@ name = "gitpython"
-version = "3.1.30"
-description = "GitPython is a python library used to interact with Git repositories"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
@@ -442,2 +442,2 @@ files = [
- {file = "GitPython-3.1.30-py3-none-any.whl", hash = "sha256:cd455b0000615c60e286208ba540271af9fe531fa6a87cc590a7298785ab2882"},
- {file = "GitPython-3.1.30.tar.gz", hash = "sha256:769c2d83e13f5d938b7688479da374c4e3d49f71549aaf462b646db9602ea6f8"},
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
@@ -473 +473 @@ name = "huggingface-hub"
-version = "0.12.0"
+version = "0.12.1"
@@ -479,2 +479,2 @@ files = [
- {file = "huggingface_hub-0.12.0-py3-none-any.whl", hash = "sha256:93809eabbfb2058a808bddf8b2a70f645de3f9df73ce87ddf5163d4c74b71c0c"},
- {file = "huggingface_hub-0.12.0.tar.gz", hash = "sha256:da82c9ec8f9d8f976ffd3fd8249d20bb35c2dd3145a9f7ca1106f0ebefd9afa0"},
+ {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"},
+ {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"},
@@ -558 +558 @@ name = "markdown-it-py"
-version = "2.1.0"
+version = "2.2.0"
@@ -564,2 +564,2 @@ files = [
- {file = "markdown-it-py-2.1.0.tar.gz", hash = "sha256:cf7e59fed14b5ae17c0006eff14a2d9a00ed5f3a846148153899a0224e2c07da"},
- {file = "markdown_it_py-2.1.0-py3-none-any.whl", hash = "sha256:93de681e5c021a432c63147656fe21790bc01231e0cd2da73626f1aa3ac0fe27"},
+ {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"},
+ {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"},
@@ -572,4 +572,4 @@ mdurl = ">=0.1,<1.0"
-benchmarking = ["psutil", "pytest", "pytest-benchmark (>=3.2,<4.0)"]
-code-style = ["pre-commit (==2.6)"]
-compare = ["commonmark (>=0.9.1,<0.10.0)", "markdown (>=3.3.6,<3.4.0)", "mistletoe (>=0.8.1,<0.9.0)", "mistune (>=2.0.2,<2.1.0)", "panflute (>=2.1.3,<2.2.0)"]
-linkify = ["linkify-it-py (>=1.0,<2.0)"]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
@@ -717 +717 @@ name = "mypy"
-version = "1.0.0"
+version = "1.0.1"
@@ -723,26 +723,26 @@ files = [
- {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
- {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
- {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
- {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
- {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
- {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
- {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
- {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
- {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
- {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
- {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
- {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
- {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
- {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
- {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
- {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
- {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
- {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
- {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
- {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"},
+ {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"},
+ {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"},
+ {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"},
+ {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"},
+ {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"},
+ {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"},
+ {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"},
+ {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"},
+ {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"},
+ {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"},
+ {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"},
+ {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"},
+ {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"},
+ {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"},
+ {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"},
+ {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"},
+ {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"},
+ {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"},
@@ -882 +882 @@ name = "pip"
-version = "23.0"
+version = "23.0.1"
@@ -888,2 +888,2 @@ files = [
- {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
- {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
+ {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"},
+ {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"},
@@ -1247 +1247 @@ name = "python-dotenv"
-version = "0.21.1"
+version = "1.0.0"
@@ -1251 +1251 @@ optional = false
-python-versions = ">=3.7"
+python-versions = ">=3.8"
@@ -1253,2 +1253,2 @@ files = [
- {file = "python-dotenv-0.21.1.tar.gz", hash = "sha256:1c93de8f636cde3ce377292818d0e440b6e45a82f215c3744979151fa8151c49"},
- {file = "python_dotenv-0.21.1-py3-none-any.whl", hash = "sha256:41e12e0318bebc859fcc4d97d4db8d20ad21721a6aa5047dd59f090391cb549a"},
+ {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"},
+ {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"},
@@ -1371 +1371 @@ name = "setuptools"
-version = "67.3.1"
+version = "67.4.0"
@@ -1377,2 +1377,2 @@ files = [
- {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
- {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
+ {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"},
+ {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"},
@@ -1484 +1484 @@ name = "types-psutil"
-version = "5.9.5.6"
+version = "5.9.5.8"
@@ -1490,2 +1490,2 @@ files = [
- {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
- {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+ {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
+ {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
@@ -1496 +1496 @@ name = "types-requests"
-version = "2.28.11.12"
+version = "2.28.11.14"
@@ -1502,2 +1502,2 @@ files = [
- {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
- {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+ {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
+ {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
@@ -1511 +1511 @@ name = "types-urllib3"
-version = "1.26.25.5"
+version = "1.26.25.7"
@@ -1517,2 +1517,2 @@ files = [
- {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
- {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
+ {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
+ {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index 95625f12..2a0769e7 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -274 +274 @@ name = "coverage"
-version = "7.1.0"
+version = "7.2.0"
@@ -280,51 +280,51 @@ files = [
- {file = "coverage-7.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3b946bbcd5a8231383450b195cfb58cb01cbe7f8949f5758566b881df4b33baf"},
- {file = "coverage-7.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ec8e767f13be637d056f7e07e61d089e555f719b387a7070154ad80a0ff31801"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4a5a5879a939cb84959d86869132b00176197ca561c664fc21478c1eee60d75"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b643cb30821e7570c0aaf54feaf0bfb630b79059f85741843e9dc23f33aaca2c"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:32df215215f3af2c1617a55dbdfb403b772d463d54d219985ac7cd3bf124cada"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:33d1ae9d4079e05ac4cc1ef9e20c648f5afabf1a92adfaf2ccf509c50b85717f"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:29571503c37f2ef2138a306d23e7270687c0efb9cab4bd8038d609b5c2393a3a"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:63ffd21aa133ff48c4dff7adcc46b7ec8b565491bfc371212122dd999812ea1c"},
- {file = "coverage-7.1.0-cp310-cp310-win32.whl", hash = "sha256:4b14d5e09c656de5038a3f9bfe5228f53439282abcab87317c9f7f1acb280352"},
- {file = "coverage-7.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:8361be1c2c073919500b6601220a6f2f98ea0b6d2fec5014c1d9cfa23dd07038"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:da9b41d4539eefd408c46725fb76ecba3a50a3367cafb7dea5f250d0653c1040"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c5b15ed7644ae4bee0ecf74fee95808dcc34ba6ace87e8dfbf5cb0dc20eab45a"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d12d076582507ea460ea2a89a8c85cb558f83406c8a41dd641d7be9a32e1274f"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2617759031dae1bf183c16cef8fcfb3de7617f394c813fa5e8e46e9b82d4222"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4e4881fa9e9667afcc742f0c244d9364d197490fbc91d12ac3b5de0bf2df146"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:9d58885215094ab4a86a6aef044e42994a2bd76a446dc59b352622655ba6621b"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:ffeeb38ee4a80a30a6877c5c4c359e5498eec095878f1581453202bfacc8fbc2"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3baf5f126f30781b5e93dbefcc8271cb2491647f8283f20ac54d12161dff080e"},
- {file = "coverage-7.1.0-cp311-cp311-win32.whl", hash = "sha256:ded59300d6330be27bc6cf0b74b89ada58069ced87c48eaf9344e5e84b0072f7"},
- {file = "coverage-7.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:6a43c7823cd7427b4ed763aa7fb63901ca8288591323b58c9cd6ec31ad910f3c"},
- {file = "coverage-7.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:7a726d742816cb3a8973c8c9a97539c734b3a309345236cd533c4883dda05b8d"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc7c85a150501286f8b56bd8ed3aa4093f4b88fb68c0843d21ff9656f0009d6a"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f5b4198d85a3755d27e64c52f8c95d6333119e49fd001ae5798dac872c95e0f8"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ddb726cb861c3117a553f940372a495fe1078249ff5f8a5478c0576c7be12050"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:51b236e764840a6df0661b67e50697aaa0e7d4124ca95e5058fa3d7cbc240b7c"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:7ee5c9bb51695f80878faaa5598040dd6c9e172ddcf490382e8aedb8ec3fec8d"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c31b75ae466c053a98bf26843563b3b3517b8f37da4d47b1c582fdc703112bc3"},
- {file = "coverage-7.1.0-cp37-cp37m-win32.whl", hash = "sha256:3b155caf3760408d1cb903b21e6a97ad4e2bdad43cbc265e3ce0afb8e0057e73"},
- {file = "coverage-7.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2a60d6513781e87047c3e630b33b4d1e89f39836dac6e069ffee28c4786715f5"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f2cba5c6db29ce991029b5e4ac51eb36774458f0a3b8d3137241b32d1bb91f06"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:beeb129cacea34490ffd4d6153af70509aa3cda20fdda2ea1a2be870dfec8d52"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c45948f613d5d18c9ec5eaa203ce06a653334cf1bd47c783a12d0dd4fd9c851"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef382417db92ba23dfb5864a3fc9be27ea4894e86620d342a116b243ade5d35d"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7c7c0d0827e853315c9bbd43c1162c006dd808dbbe297db7ae66cd17b07830f0"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e5cdbb5cafcedea04924568d990e20ce7f1945a1dd54b560f879ee2d57226912"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:9817733f0d3ea91bea80de0f79ef971ae94f81ca52f9b66500c6a2fea8e4b4f8"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:218fe982371ac7387304153ecd51205f14e9d731b34fb0568181abaf7b443ba0"},
- {file = "coverage-7.1.0-cp38-cp38-win32.whl", hash = "sha256:04481245ef966fbd24ae9b9e537ce899ae584d521dfbe78f89cad003c38ca2ab"},
- {file = "coverage-7.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ae125d1134bf236acba8b83e74c603d1b30e207266121e76484562bc816344c"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2bf1d5f2084c3932b56b962a683074a3692bce7cabd3aa023c987a2a8e7612f6"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:98b85dd86514d889a2e3dd22ab3c18c9d0019e696478391d86708b805f4ea0fa"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38da2db80cc505a611938d8624801158e409928b136c8916cd2e203970dde4dc"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3164d31078fa9efe406e198aecd2a02d32a62fecbdef74f76dad6a46c7e48311"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db61a79c07331e88b9a9974815c075fbd812bc9dbc4dc44b366b5368a2936063"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9ccb092c9ede70b2517a57382a601619d20981f56f440eae7e4d7eaafd1d1d09"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:33ff26d0f6cc3ca8de13d14fde1ff8efe1456b53e3f0273e63cc8b3c84a063d8"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d47dd659a4ee952e90dc56c97d78132573dc5c7b09d61b416a9deef4ebe01a0c"},
- {file = "coverage-7.1.0-cp39-cp39-win32.whl", hash = "sha256:d248cd4a92065a4d4543b8331660121b31c4148dd00a691bfb7a5cdc7483cfa4"},
- {file = "coverage-7.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:7ed681b0f8e8bcbbffa58ba26fcf5dbc8f79e7997595bf071ed5430d8c08d6f3"},
- {file = "coverage-7.1.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:755e89e32376c850f826c425ece2c35a4fc266c081490eb0a841e7c1cb0d3bda"},
- {file = "coverage-7.1.0.tar.gz", hash = "sha256:10188fe543560ec4874f974b5305cd1a8bdcfa885ee00ea3a03733464c4ca265"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
+ {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
+ {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
+ {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
+ {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
+ {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
+ {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
+ {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
+ {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
+ {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
+ {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
+ {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
+ {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
+ {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
@@ -457,2 +457,2 @@ name = "gitpython"
-version = "3.1.30"
-description = "GitPython is a python library used to interact with Git repositories"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
@@ -463,2 +463,2 @@ files = [
- {file = "GitPython-3.1.30-py3-none-any.whl", hash = "sha256:cd455b0000615c60e286208ba540271af9fe531fa6a87cc590a7298785ab2882"},
- {file = "GitPython-3.1.30.tar.gz", hash = "sha256:769c2d83e13f5d938b7688479da374c4e3d49f71549aaf462b646db9602ea6f8"},
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
@@ -552 +552 @@ name = "huggingface-hub"
-version = "0.12.0"
+version = "0.12.1"
@@ -558,2 +558,2 @@ files = [
- {file = "huggingface_hub-0.12.0-py3-none-any.whl", hash = "sha256:93809eabbfb2058a808bddf8b2a70f645de3f9df73ce87ddf5163d4c74b71c0c"},
- {file = "huggingface_hub-0.12.0.tar.gz", hash = "sha256:da82c9ec8f9d8f976ffd3fd8249d20bb35c2dd3145a9f7ca1106f0ebefd9afa0"},
+ {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"},
+ {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"},
@@ -662 +662 @@ name = "markdown-it-py"
-version = "2.1.0"
+version = "2.2.0"
@@ -668,2 +668,2 @@ files = [
- {file = "markdown-it-py-2.1.0.tar.gz", hash = "sha256:cf7e59fed14b5ae17c0006eff14a2d9a00ed5f3a846148153899a0224e2c07da"},
- {file = "markdown_it_py-2.1.0-py3-none-any.whl", hash = "sha256:93de681e5c021a432c63147656fe21790bc01231e0cd2da73626f1aa3ac0fe27"},
+ {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"},
+ {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"},
@@ -676,4 +676,4 @@ mdurl = ">=0.1,<1.0"
-benchmarking = ["psutil", "pytest", "pytest-benchmark (>=3.2,<4.0)"]
-code-style = ["pre-commit (==2.6)"]
-compare = ["commonmark (>=0.9.1,<0.10.0)", "markdown (>=3.3.6,<3.4.0)", "mistletoe (>=0.8.1,<0.9.0)", "mistune (>=2.0.2,<2.1.0)", "panflute (>=2.1.3,<2.2.0)"]
-linkify = ["linkify-it-py (>=1.0,<2.0)"]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
@@ -821 +821 @@ name = "mypy"
-version = "1.0.0"
+version = "1.0.1"
@@ -827,26 +827,26 @@ files = [
- {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
- {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
- {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
- {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
- {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
- {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
- {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
- {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
- {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
- {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
- {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
- {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
- {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
- {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
- {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
- {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
- {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
- {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
- {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
- {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"},
+ {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"},
+ {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"},
+ {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"},
+ {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"},
+ {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"},
+ {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"},
+ {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"},
+ {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"},
+ {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"},
+ {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"},
+ {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"},
+ {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"},
+ {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"},
+ {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"},
+ {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"},
+ {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"},
+ {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"},
+ {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"},
@@ -986 +986 @@ name = "pip"
-version = "23.0"
+version = "23.0.1"
@@ -992,2 +992,2 @@ files = [
- {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
- {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
+ {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"},
+ {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"},
@@ -1366 +1366 @@ name = "python-dotenv"
-version = "0.21.1"
+version = "1.0.0"
@@ -1370 +1370 @@ optional = false
-python-versions = ">=3.7"
+python-versions = ">=3.8"
@@ -1372,2 +1372,2 @@ files = [
- {file = "python-dotenv-0.21.1.tar.gz", hash = "sha256:1c93de8f636cde3ce377292818d0e440b6e45a82f215c3744979151fa8151c49"},
- {file = "python_dotenv-0.21.1-py3-none-any.whl", hash = "sha256:41e12e0318bebc859fcc4d97d4db8d20ad21721a6aa5047dd59f090391cb549a"},
+ {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"},
+ {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"},
@@ -1529 +1529 @@ name = "setuptools"
-version = "67.3.1"
+version = "67.4.0"
@@ -1535,2 +1535,2 @@ files = [
- {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
- {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
+ {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"},
+ {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"},
@@ -1689 +1689 @@ name = "types-psutil"
-version = "5.9.5.6"
+version = "5.9.5.8"
@@ -1695,2 +1695,2 @@ files = [
- {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
- {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+ {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
+ {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
@@ -1701 +1701 @@ name = "types-requests"
-version = "2.28.11.12"
+version = "2.28.11.14"
@@ -1707,2 +1707,2 @@ files = [
- {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
- {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+ {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
+ {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
@@ -1716 +1716 @@ name = "types-toml"
-version = "0.10.8.3"
+version = "0.10.8.5"
@@ -1722,2 +1722,2 @@ files = [
- {file = "types-toml-0.10.8.3.tar.gz", hash = "sha256:f37244eff4cd7eace9cb70d0bac54d3eba77973aa4ef26c271ac3d1c6503a48e"},
- {file = "types_toml-0.10.8.3-py3-none-any.whl", hash = "sha256:a2286a053aea6ab6ff814659272b1d4a05d86a1dd52b807a87b23511993b46c5"},
+ {file = "types-toml-0.10.8.5.tar.gz", hash = "sha256:bf80fce7d2d74be91148f47b88d9ae5adeb1024abef22aa2fdbabc036d6b8b3c"},
+ {file = "types_toml-0.10.8.5-py3-none-any.whl", hash = "sha256:2432017febe43174af0f3c65f03116e3d3cf43e7e1406b8200e106da8cf98992"},
@@ -1728 +1728 @@ name = "types-urllib3"
-version = "1.26.25.5"
+version = "1.26.25.7"
@@ -1734,2 +1734,2 @@ files = [
- {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
- {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
+ {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
+ {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
@@ -1788 +1788 @@ name = "watchdog"
-version = "2.2.1"
+version = "2.3.0"
@@ -1794,28 +1794,28 @@ files = [
- {file = "watchdog-2.2.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:a09483249d25cbdb4c268e020cb861c51baab2d1affd9a6affc68ffe6a231260"},
- {file = "watchdog-2.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:5100eae58133355d3ca6c1083a33b81355c4f452afa474c2633bd2fbbba398b3"},
- {file = "watchdog-2.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e618a4863726bc7a3c64f95c218437f3349fb9d909eb9ea3a1ed3b567417c661"},
- {file = "watchdog-2.2.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:102a60093090fc3ff76c983367b19849b7cc24ec414a43c0333680106e62aae1"},
- {file = "watchdog-2.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:748ca797ff59962e83cc8e4b233f87113f3cf247c23e6be58b8a2885c7337aa3"},
- {file = "watchdog-2.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:6ccd8d84b9490a82b51b230740468116b8205822ea5fdc700a553d92661253a3"},
- {file = "watchdog-2.2.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:6e01d699cd260d59b84da6bda019dce0a3353e3fcc774408ae767fe88ee096b7"},
- {file = "watchdog-2.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8586d98c494690482c963ffb24c49bf9c8c2fe0589cec4dc2f753b78d1ec301d"},
- {file = "watchdog-2.2.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:adaf2ece15f3afa33a6b45f76b333a7da9256e1360003032524d61bdb4c422ae"},
- {file = "watchdog-2.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:83a7cead445008e880dbde833cb9e5cc7b9a0958edb697a96b936621975f15b9"},
- {file = "watchdog-2.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:f8ac23ff2c2df4471a61af6490f847633024e5aa120567e08d07af5718c9d092"},
- {file = "watchdog-2.2.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:d0f29fd9f3f149a5277929de33b4f121a04cf84bb494634707cfa8ea8ae106a8"},
- {file = "watchdog-2.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:967636031fa4c4955f0f3f22da3c5c418aa65d50908d31b73b3b3ffd66d60640"},
- {file = "watchdog-2.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:96cbeb494e6cbe3ae6aacc430e678ce4b4dd3ae5125035f72b6eb4e5e9eb4f4e"},
- {file = "watchdog-2.2.1-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:61fdb8e9c57baf625e27e1420e7ca17f7d2023929cd0065eb79c83da1dfbeacd"},
- {file = "watchdog-2.2.1-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4cb5ecc332112017fbdb19ede78d92e29a8165c46b68a0b8ccbd0a154f196d5e"},
- {file = "watchdog-2.2.1-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:a480d122740debf0afac4ddd583c6c0bb519c24f817b42ed6f850e2f6f9d64a8"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_aarch64.whl", hash = "sha256:978a1aed55de0b807913b7482d09943b23a2d634040b112bdf31811a422f6344"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_armv7l.whl", hash = "sha256:8c28c23972ec9c524967895ccb1954bc6f6d4a557d36e681a36e84368660c4ce"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_i686.whl", hash = "sha256:c27d8c1535fd4474e40a4b5e01f4ba6720bac58e6751c667895cbc5c8a7af33c"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_ppc64.whl", hash = "sha256:d6b87477752bd86ac5392ecb9eeed92b416898c30bd40c7e2dd03c3146105646"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:cece1aa596027ff56369f0b50a9de209920e1df9ac6d02c7f9e5d8162eb4f02b"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_s390x.whl", hash = "sha256:8b5cde14e5c72b2df5d074774bdff69e9b55da77e102a91f36ef26ca35f9819c"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_x86_64.whl", hash = "sha256:e038be858425c4f621900b8ff1a3a1330d9edcfeaa1c0468aeb7e330fb87693e"},
- {file = "watchdog-2.2.1-py3-none-win32.whl", hash = "sha256:bc43c1b24d2f86b6e1cc15f68635a959388219426109233e606517ff7d0a5a73"},
- {file = "watchdog-2.2.1-py3-none-win_amd64.whl", hash = "sha256:17f1708f7410af92ddf591e94ae71a27a13974559e72f7e9fde3ec174b26ba2e"},
- {file = "watchdog-2.2.1-py3-none-win_ia64.whl", hash = "sha256:195ab1d9d611a4c1e5311cbf42273bc541e18ea8c32712f2fb703cfc6ff006f9"},
- {file = "watchdog-2.2.1.tar.gz", hash = "sha256:cdcc23c9528601a8a293eb4369cbd14f6b4f34f07ae8769421252e9c22718b6f"},
+ {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:c1b3962e5463a848ba2a342cb66c80251dca27a102933b8f38d231d2a9e5a543"},
+ {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9e651b4874477c1bf239417d43818bbfd047aaf641b029fa60d6f5109ede0db0"},
+ {file = "watchdog-2.3.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d04662017efd00a014cff9068708e085d67f2fac43f48bbbb95a7f97490487f3"},
+ {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0f7d759299ce21a3d2a77e18d430c24811369c3432453701790acc6ff45a7101"},
+ {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:a4b9bece40d46bf6fb8621817ea7d903eae2b9b3ebac55a51ed50354a79061a8"},
+ {file = "watchdog-2.3.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:242e57253e84a736e6777ba756c48cf6a68d3d90cb9e01bd6bfd371a949ace3a"},
+ {file = "watchdog-2.3.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:3fa74b0ef4825f9112932675a002296cb2d3d3e400d7a44c32fafd1ecc83ada0"},
+ {file = "watchdog-2.3.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:15bf5b165d7a6b48265411dad74fb0d33053f8270eb6575faad0e016035cf9f7"},
+ {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:139262f678b4e6a7013261c772059bca358441de04fb0e0087489a34db9e3db0"},
+ {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8a214955769d2ef0f7aaa82f31863e3bdf6b083ce1b5f1c2e85cab0f66fba024"},
+ {file = "watchdog-2.3.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4e648df44a4c6ea6da4d9eb6722745c986b9d70268f25ae60f140082d7c8908e"},
+ {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:473164a2de473f708ca194a992466eeefff73b58273bbb88e089c5a5a98fcda1"},
+ {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebe756f788cb130fdc5c150ea8a4fda39cb4ee3a5873a345607c8b84fecf018b"},
+ {file = "watchdog-2.3.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a623de186477e9e05f8461087f856412eae5cd005cc4bcb232ed5c6f9a8709f5"},
+ {file = "watchdog-2.3.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:43d76d7888b26850b908208bb82383a193e8b0f25d0abaa84452f191b4acdea4"},
+ {file = "watchdog-2.3.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:5ddbbe87f9ed726940d174076da030cd01ec45433ef2b1b2e6094c84f2af17f1"},
+ {file = "watchdog-2.3.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3fa1572f5a2f6d17d4d860edbc04488fef31b007c25c2f3b11203fb8179b7c67"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_aarch64.whl", hash = "sha256:1d9c656495172873bf1ddc7e39e80055fcdd21c4608cf68f23a28116dcba0b43"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_armv7l.whl", hash = "sha256:00f93782c67042d9525ec51628330b5faf5fb84bcb7ebaac05ea8528cfb20bba"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_i686.whl", hash = "sha256:f1a655f4a49f9232311b9967f42cc2eaf43fd4903f3bed850dd4570fda5d5eff"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64.whl", hash = "sha256:aa4773160b9cb21ba369cb42d59a947087330b3a02480173033a6a6cc137a510"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:982f5416a2817003172994d865285dd6a2b3836f033cd3fa87d1a62096a162cc"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_s390x.whl", hash = "sha256:45c13e7e6eea1013da419bf9aa9a8f5df7bbf3e5edce40bc6df84130febf39d5"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_x86_64.whl", hash = "sha256:7767a3da3307d9cf597832f692702441a97c259e5d0d560f2e57c43ad0d191d2"},
+ {file = "watchdog-2.3.0-py3-none-win32.whl", hash = "sha256:8863913ea2c3f256d18c33d84546518636e391cd8f50d209b9a31221e0f7d3fd"},
+ {file = "watchdog-2.3.0-py3-none-win_amd64.whl", hash = "sha256:6d79b5954db8f41d6a7f5763042b988f7a4afd40b7d141456061fa7c5b7f2159"},
+ {file = "watchdog-2.3.0-py3-none-win_ia64.whl", hash = "sha256:a3559ee82a10976de1ec544b6ebe3b4aa398d491860a283d80ec0f550076d068"},
+ {file = "watchdog-2.3.0.tar.gz", hash = "sha256:9d39effe6909be898ba3e7286a9e9b17a6a9f734fb1ef9dde3e9bb68715fca39"},
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index 99a74d80..900cf425 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -274 +274 @@ name = "coverage"
-version = "7.1.0"
+version = "7.2.0"
@@ -280,51 +280,51 @@ files = [
- {file = "coverage-7.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3b946bbcd5a8231383450b195cfb58cb01cbe7f8949f5758566b881df4b33baf"},
- {file = "coverage-7.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ec8e767f13be637d056f7e07e61d089e555f719b387a7070154ad80a0ff31801"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4a5a5879a939cb84959d86869132b00176197ca561c664fc21478c1eee60d75"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b643cb30821e7570c0aaf54feaf0bfb630b79059f85741843e9dc23f33aaca2c"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:32df215215f3af2c1617a55dbdfb403b772d463d54d219985ac7cd3bf124cada"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:33d1ae9d4079e05ac4cc1ef9e20c648f5afabf1a92adfaf2ccf509c50b85717f"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:29571503c37f2ef2138a306d23e7270687c0efb9cab4bd8038d609b5c2393a3a"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:63ffd21aa133ff48c4dff7adcc46b7ec8b565491bfc371212122dd999812ea1c"},
- {file = "coverage-7.1.0-cp310-cp310-win32.whl", hash = "sha256:4b14d5e09c656de5038a3f9bfe5228f53439282abcab87317c9f7f1acb280352"},
- {file = "coverage-7.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:8361be1c2c073919500b6601220a6f2f98ea0b6d2fec5014c1d9cfa23dd07038"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:da9b41d4539eefd408c46725fb76ecba3a50a3367cafb7dea5f250d0653c1040"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c5b15ed7644ae4bee0ecf74fee95808dcc34ba6ace87e8dfbf5cb0dc20eab45a"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d12d076582507ea460ea2a89a8c85cb558f83406c8a41dd641d7be9a32e1274f"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2617759031dae1bf183c16cef8fcfb3de7617f394c813fa5e8e46e9b82d4222"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4e4881fa9e9667afcc742f0c244d9364d197490fbc91d12ac3b5de0bf2df146"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:9d58885215094ab4a86a6aef044e42994a2bd76a446dc59b352622655ba6621b"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:ffeeb38ee4a80a30a6877c5c4c359e5498eec095878f1581453202bfacc8fbc2"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3baf5f126f30781b5e93dbefcc8271cb2491647f8283f20ac54d12161dff080e"},
- {file = "coverage-7.1.0-cp311-cp311-win32.whl", hash = "sha256:ded59300d6330be27bc6cf0b74b89ada58069ced87c48eaf9344e5e84b0072f7"},
- {file = "coverage-7.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:6a43c7823cd7427b4ed763aa7fb63901ca8288591323b58c9cd6ec31ad910f3c"},
- {file = "coverage-7.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:7a726d742816cb3a8973c8c9a97539c734b3a309345236cd533c4883dda05b8d"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc7c85a150501286f8b56bd8ed3aa4093f4b88fb68c0843d21ff9656f0009d6a"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f5b4198d85a3755d27e64c52f8c95d6333119e49fd001ae5798dac872c95e0f8"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ddb726cb861c3117a553f940372a495fe1078249ff5f8a5478c0576c7be12050"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:51b236e764840a6df0661b67e50697aaa0e7d4124ca95e5058fa3d7cbc240b7c"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:7ee5c9bb51695f80878faaa5598040dd6c9e172ddcf490382e8aedb8ec3fec8d"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c31b75ae466c053a98bf26843563b3b3517b8f37da4d47b1c582fdc703112bc3"},
- {file = "coverage-7.1.0-cp37-cp37m-win32.whl", hash = "sha256:3b155caf3760408d1cb903b21e6a97ad4e2bdad43cbc265e3ce0afb8e0057e73"},
- {file = "coverage-7.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2a60d6513781e87047c3e630b33b4d1e89f39836dac6e069ffee28c4786715f5"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f2cba5c6db29ce991029b5e4ac51eb36774458f0a3b8d3137241b32d1bb91f06"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:beeb129cacea34490ffd4d6153af70509aa3cda20fdda2ea1a2be870dfec8d52"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c45948f613d5d18c9ec5eaa203ce06a653334cf1bd47c783a12d0dd4fd9c851"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef382417db92ba23dfb5864a3fc9be27ea4894e86620d342a116b243ade5d35d"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7c7c0d0827e853315c9bbd43c1162c006dd808dbbe297db7ae66cd17b07830f0"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e5cdbb5cafcedea04924568d990e20ce7f1945a1dd54b560f879ee2d57226912"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:9817733f0d3ea91bea80de0f79ef971ae94f81ca52f9b66500c6a2fea8e4b4f8"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:218fe982371ac7387304153ecd51205f14e9d731b34fb0568181abaf7b443ba0"},
- {file = "coverage-7.1.0-cp38-cp38-win32.whl", hash = "sha256:04481245ef966fbd24ae9b9e537ce899ae584d521dfbe78f89cad003c38ca2ab"},
- {file = "coverage-7.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ae125d1134bf236acba8b83e74c603d1b30e207266121e76484562bc816344c"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2bf1d5f2084c3932b56b962a683074a3692bce7cabd3aa023c987a2a8e7612f6"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:98b85dd86514d889a2e3dd22ab3c18c9d0019e696478391d86708b805f4ea0fa"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38da2db80cc505a611938d8624801158e409928b136c8916cd2e203970dde4dc"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3164d31078fa9efe406e198aecd2a02d32a62fecbdef74f76dad6a46c7e48311"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db61a79c07331e88b9a9974815c075fbd812bc9dbc4dc44b366b5368a2936063"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9ccb092c9ede70b2517a57382a601619d20981f56f440eae7e4d7eaafd1d1d09"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:33ff26d0f6cc3ca8de13d14fde1ff8efe1456b53e3f0273e63cc8b3c84a063d8"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d47dd659a4ee952e90dc56c97d78132573dc5c7b09d61b416a9deef4ebe01a0c"},
- {file = "coverage-7.1.0-cp39-cp39-win32.whl", hash = "sha256:d248cd4a92065a4d4543b8331660121b31c4148dd00a691bfb7a5cdc7483cfa4"},
- {file = "coverage-7.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:7ed681b0f8e8bcbbffa58ba26fcf5dbc8f79e7997595bf071ed5430d8c08d6f3"},
- {file = "coverage-7.1.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:755e89e32376c850f826c425ece2c35a4fc266c081490eb0a841e7c1cb0d3bda"},
- {file = "coverage-7.1.0.tar.gz", hash = "sha256:10188fe543560ec4874f974b5305cd1a8bdcfa885ee00ea3a03733464c4ca265"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
+ {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
+ {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
+ {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
+ {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
+ {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
+ {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
+ {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
+ {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
+ {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
+ {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
+ {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
+ {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
+ {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
@@ -457,2 +457,2 @@ name = "gitpython"
-version = "3.1.30"
-description = "GitPython is a python library used to interact with Git repositories"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
@@ -463,2 +463,2 @@ files = [
- {file = "GitPython-3.1.30-py3-none-any.whl", hash = "sha256:cd455b0000615c60e286208ba540271af9fe531fa6a87cc590a7298785ab2882"},
- {file = "GitPython-3.1.30.tar.gz", hash = "sha256:769c2d83e13f5d938b7688479da374c4e3d49f71549aaf462b646db9602ea6f8"},
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
@@ -552 +552 @@ name = "huggingface-hub"
-version = "0.12.0"
+version = "0.12.1"
@@ -558,2 +558,2 @@ files = [
- {file = "huggingface_hub-0.12.0-py3-none-any.whl", hash = "sha256:93809eabbfb2058a808bddf8b2a70f645de3f9df73ce87ddf5163d4c74b71c0c"},
- {file = "huggingface_hub-0.12.0.tar.gz", hash = "sha256:da82c9ec8f9d8f976ffd3fd8249d20bb35c2dd3145a9f7ca1106f0ebefd9afa0"},
+ {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"},
+ {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"},
@@ -682 +682 @@ name = "markdown-it-py"
-version = "2.1.0"
+version = "2.2.0"
@@ -688,2 +688,2 @@ files = [
- {file = "markdown-it-py-2.1.0.tar.gz", hash = "sha256:cf7e59fed14b5ae17c0006eff14a2d9a00ed5f3a846148153899a0224e2c07da"},
- {file = "markdown_it_py-2.1.0-py3-none-any.whl", hash = "sha256:93de681e5c021a432c63147656fe21790bc01231e0cd2da73626f1aa3ac0fe27"},
+ {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"},
+ {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"},
@@ -696,4 +696,4 @@ mdurl = ">=0.1,<1.0"
-benchmarking = ["psutil", "pytest", "pytest-benchmark (>=3.2,<4.0)"]
-code-style = ["pre-commit (==2.6)"]
-compare = ["commonmark (>=0.9.1,<0.10.0)", "markdown (>=3.3.6,<3.4.0)", "mistletoe (>=0.8.1,<0.9.0)", "mistune (>=2.0.2,<2.1.0)", "panflute (>=2.1.3,<2.2.0)"]
-linkify = ["linkify-it-py (>=1.0,<2.0)"]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
@@ -901 +901 @@ name = "mypy"
-version = "1.0.0"
+version = "1.0.1"
@@ -907,26 +907,26 @@ files = [
- {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
- {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
- {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
- {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
- {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
- {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
- {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
- {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
- {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
- {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
- {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
- {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
- {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
- {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
- {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
- {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
- {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
- {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
- {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
- {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"},
+ {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"},
+ {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"},
+ {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"},
+ {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"},
+ {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"},
+ {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"},
+ {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"},
+ {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"},
+ {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"},
+ {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"},
+ {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"},
+ {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"},
+ {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"},
+ {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"},
+ {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"},
+ {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"},
+ {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"},
+ {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"},
@@ -1066 +1066 @@ name = "pip"
-version = "23.0"
+version = "23.0.1"
@@ -1072,2 +1072,2 @@ files = [
- {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
- {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
+ {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"},
+ {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"},
@@ -1498 +1498 @@ name = "python-dotenv"
-version = "0.21.1"
+version = "1.0.0"
@@ -1502 +1502 @@ optional = false
-python-versions = ">=3.7"
+python-versions = ">=3.8"
@@ -1504,2 +1504,2 @@ files = [
- {file = "python-dotenv-0.21.1.tar.gz", hash = "sha256:1c93de8f636cde3ce377292818d0e440b6e45a82f215c3744979151fa8151c49"},
- {file = "python_dotenv-0.21.1-py3-none-any.whl", hash = "sha256:41e12e0318bebc859fcc4d97d4db8d20ad21721a6aa5047dd59f090391cb549a"},
+ {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"},
+ {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"},
@@ -1640 +1640 @@ name = "setuptools"
-version = "67.3.1"
+version = "67.4.0"
@@ -1646,2 +1646,2 @@ files = [
- {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
- {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
+ {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"},
+ {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"},
@@ -1800 +1800 @@ name = "types-jsonschema"
-version = "4.17.0.4"
+version = "4.17.0.6"
@@ -1806,2 +1806,2 @@ files = [
- {file = "types-jsonschema-4.17.0.4.tar.gz", hash = "sha256:f8e82ef9b618c7183ddbf604684b7a656119f388deea8fe8d4238a388bd919c2"},
- {file = "types_jsonschema-4.17.0.4-py3-none-any.whl", hash = "sha256:2f5e7848c8f210a92973a92c08f661cbfe7d1b33ea4cbc5d4c34a49b4adc8cd7"},
+ {file = "types-jsonschema-4.17.0.6.tar.gz", hash = "sha256:e9b15e34b4f2fd5587bd68530fa0eb2a17c73ead212f4471d71eea032d231c46"},
+ {file = "types_jsonschema-4.17.0.6-py3-none-any.whl", hash = "sha256:ecef99bc64848f3798ad18922dfb2b40da25f17796fafcee50da984a21c5d6e6"},
@@ -1812 +1812 @@ name = "types-psutil"
-version = "5.9.5.6"
+version = "5.9.5.8"
@@ -1818,2 +1818,2 @@ files = [
- {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
- {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+ {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
+ {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
@@ -1824 +1824 @@ name = "types-requests"
-version = "2.28.11.12"
+version = "2.28.11.14"
@@ -1830,2 +1830,2 @@ files = [
- {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
- {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+ {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
+ {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
@@ -1839 +1839 @@ name = "types-urllib3"
-version = "1.26.25.5"
+version = "1.26.25.7"
@@ -1845,2 +1845,2 @@ files = [
- {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
- {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
+ {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
+ {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
@@ -1899 +1899 @@ name = "watchdog"
-version = "2.2.1"
+version = "2.3.0"
@@ -1905,28 +1905,28 @@ files = [
- {file = "watchdog-2.2.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:a09483249d25cbdb4c268e020cb861c51baab2d1affd9a6affc68ffe6a231260"},
- {file = "watchdog-2.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:5100eae58133355d3ca6c1083a33b81355c4f452afa474c2633bd2fbbba398b3"},
- {file = "watchdog-2.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e618a4863726bc7a3c64f95c218437f3349fb9d909eb9ea3a1ed3b567417c661"},
- {file = "watchdog-2.2.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:102a60093090fc3ff76c983367b19849b7cc24ec414a43c0333680106e62aae1"},
- {file = "watchdog-2.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:748ca797ff59962e83cc8e4b233f87113f3cf247c23e6be58b8a2885c7337aa3"},
- {file = "watchdog-2.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:6ccd8d84b9490a82b51b230740468116b8205822ea5fdc700a553d92661253a3"},
- {file = "watchdog-2.2.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:6e01d699cd260d59b84da6bda019dce0a3353e3fcc774408ae767fe88ee096b7"},
- {file = "watchdog-2.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8586d98c494690482c963ffb24c49bf9c8c2fe0589cec4dc2f753b78d1ec301d"},
- {file = "watchdog-2.2.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:adaf2ece15f3afa33a6b45f76b333a7da9256e1360003032524d61bdb4c422ae"},
- {file = "watchdog-2.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:83a7cead445008e880dbde833cb9e5cc7b9a0958edb697a96b936621975f15b9"},
- {file = "watchdog-2.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:f8ac23ff2c2df4471a61af6490f847633024e5aa120567e08d07af5718c9d092"},
- {file = "watchdog-2.2.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:d0f29fd9f3f149a5277929de33b4f121a04cf84bb494634707cfa8ea8ae106a8"},
- {file = "watchdog-2.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:967636031fa4c4955f0f3f22da3c5c418aa65d50908d31b73b3b3ffd66d60640"},
- {file = "watchdog-2.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:96cbeb494e6cbe3ae6aacc430e678ce4b4dd3ae5125035f72b6eb4e5e9eb4f4e"},
- {file = "watchdog-2.2.1-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:61fdb8e9c57baf625e27e1420e7ca17f7d2023929cd0065eb79c83da1dfbeacd"},
- {file = "watchdog-2.2.1-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4cb5ecc332112017fbdb19ede78d92e29a8165c46b68a0b8ccbd0a154f196d5e"},
- {file = "watchdog-2.2.1-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:a480d122740debf0afac4ddd583c6c0bb519c24f817b42ed6f850e2f6f9d64a8"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_aarch64.whl", hash = "sha256:978a1aed55de0b807913b7482d09943b23a2d634040b112bdf31811a422f6344"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_armv7l.whl", hash = "sha256:8c28c23972ec9c524967895ccb1954bc6f6d4a557d36e681a36e84368660c4ce"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_i686.whl", hash = "sha256:c27d8c1535fd4474e40a4b5e01f4ba6720bac58e6751c667895cbc5c8a7af33c"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_ppc64.whl", hash = "sha256:d6b87477752bd86ac5392ecb9eeed92b416898c30bd40c7e2dd03c3146105646"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:cece1aa596027ff56369f0b50a9de209920e1df9ac6d02c7f9e5d8162eb4f02b"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_s390x.whl", hash = "sha256:8b5cde14e5c72b2df5d074774bdff69e9b55da77e102a91f36ef26ca35f9819c"},
- {file = "watchdog-2.2.1-py3-none-manylinux2014_x86_64.whl", hash = "sha256:e038be858425c4f621900b8ff1a3a1330d9edcfeaa1c0468aeb7e330fb87693e"},
- {file = "watchdog-2.2.1-py3-none-win32.whl", hash = "sha256:bc43c1b24d2f86b6e1cc15f68635a959388219426109233e606517ff7d0a5a73"},
- {file = "watchdog-2.2.1-py3-none-win_amd64.whl", hash = "sha256:17f1708f7410af92ddf591e94ae71a27a13974559e72f7e9fde3ec174b26ba2e"},
- {file = "watchdog-2.2.1-py3-none-win_ia64.whl", hash = "sha256:195ab1d9d611a4c1e5311cbf42273bc541e18ea8c32712f2fb703cfc6ff006f9"},
- {file = "watchdog-2.2.1.tar.gz", hash = "sha256:cdcc23c9528601a8a293eb4369cbd14f6b4f34f07ae8769421252e9c22718b6f"},
+ {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:c1b3962e5463a848ba2a342cb66c80251dca27a102933b8f38d231d2a9e5a543"},
+ {file = "watchdog-2.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9e651b4874477c1bf239417d43818bbfd047aaf641b029fa60d6f5109ede0db0"},
+ {file = "watchdog-2.3.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d04662017efd00a014cff9068708e085d67f2fac43f48bbbb95a7f97490487f3"},
+ {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0f7d759299ce21a3d2a77e18d430c24811369c3432453701790acc6ff45a7101"},
+ {file = "watchdog-2.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:a4b9bece40d46bf6fb8621817ea7d903eae2b9b3ebac55a51ed50354a79061a8"},
+ {file = "watchdog-2.3.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:242e57253e84a736e6777ba756c48cf6a68d3d90cb9e01bd6bfd371a949ace3a"},
+ {file = "watchdog-2.3.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:3fa74b0ef4825f9112932675a002296cb2d3d3e400d7a44c32fafd1ecc83ada0"},
+ {file = "watchdog-2.3.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:15bf5b165d7a6b48265411dad74fb0d33053f8270eb6575faad0e016035cf9f7"},
+ {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:139262f678b4e6a7013261c772059bca358441de04fb0e0087489a34db9e3db0"},
+ {file = "watchdog-2.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8a214955769d2ef0f7aaa82f31863e3bdf6b083ce1b5f1c2e85cab0f66fba024"},
+ {file = "watchdog-2.3.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4e648df44a4c6ea6da4d9eb6722745c986b9d70268f25ae60f140082d7c8908e"},
+ {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:473164a2de473f708ca194a992466eeefff73b58273bbb88e089c5a5a98fcda1"},
+ {file = "watchdog-2.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebe756f788cb130fdc5c150ea8a4fda39cb4ee3a5873a345607c8b84fecf018b"},
+ {file = "watchdog-2.3.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a623de186477e9e05f8461087f856412eae5cd005cc4bcb232ed5c6f9a8709f5"},
+ {file = "watchdog-2.3.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:43d76d7888b26850b908208bb82383a193e8b0f25d0abaa84452f191b4acdea4"},
+ {file = "watchdog-2.3.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:5ddbbe87f9ed726940d174076da030cd01ec45433ef2b1b2e6094c84f2af17f1"},
+ {file = "watchdog-2.3.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3fa1572f5a2f6d17d4d860edbc04488fef31b007c25c2f3b11203fb8179b7c67"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_aarch64.whl", hash = "sha256:1d9c656495172873bf1ddc7e39e80055fcdd21c4608cf68f23a28116dcba0b43"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_armv7l.whl", hash = "sha256:00f93782c67042d9525ec51628330b5faf5fb84bcb7ebaac05ea8528cfb20bba"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_i686.whl", hash = "sha256:f1a655f4a49f9232311b9967f42cc2eaf43fd4903f3bed850dd4570fda5d5eff"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64.whl", hash = "sha256:aa4773160b9cb21ba369cb42d59a947087330b3a02480173033a6a6cc137a510"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:982f5416a2817003172994d865285dd6a2b3836f033cd3fa87d1a62096a162cc"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_s390x.whl", hash = "sha256:45c13e7e6eea1013da419bf9aa9a8f5df7bbf3e5edce40bc6df84130febf39d5"},
+ {file = "watchdog-2.3.0-py3-none-manylinux2014_x86_64.whl", hash = "sha256:7767a3da3307d9cf597832f692702441a97c259e5d0d560f2e57c43ad0d191d2"},
+ {file = "watchdog-2.3.0-py3-none-win32.whl", hash = "sha256:8863913ea2c3f256d18c33d84546518636e391cd8f50d209b9a31221e0f7d3fd"},
+ {file = "watchdog-2.3.0-py3-none-win_amd64.whl", hash = "sha256:6d79b5954db8f41d6a7f5763042b988f7a4afd40b7d141456061fa7c5b7f2159"},
+ {file = "watchdog-2.3.0-py3-none-win_ia64.whl", hash = "sha256:a3559ee82a10976de1ec544b6ebe3b4aa398d491860a283d80ec0f550076d068"},
+ {file = "watchdog-2.3.0.tar.gz", hash = "sha256:9d39effe6909be898ba3e7286a9e9b17a6a9f734fb1ef9dde3e9bb68715fca39"},
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 72d57d9f..56a2060f 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -785 +785 @@ name = "coverage"
-version = "7.1.0"
+version = "7.2.0"
@@ -791,51 +791,51 @@ files = [
- {file = "coverage-7.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3b946bbcd5a8231383450b195cfb58cb01cbe7f8949f5758566b881df4b33baf"},
- {file = "coverage-7.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ec8e767f13be637d056f7e07e61d089e555f719b387a7070154ad80a0ff31801"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4a5a5879a939cb84959d86869132b00176197ca561c664fc21478c1eee60d75"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b643cb30821e7570c0aaf54feaf0bfb630b79059f85741843e9dc23f33aaca2c"},
- {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:32df215215f3af2c1617a55dbdfb403b772d463d54d219985ac7cd3bf124cada"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:33d1ae9d4079e05ac4cc1ef9e20c648f5afabf1a92adfaf2ccf509c50b85717f"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:29571503c37f2ef2138a306d23e7270687c0efb9cab4bd8038d609b5c2393a3a"},
- {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:63ffd21aa133ff48c4dff7adcc46b7ec8b565491bfc371212122dd999812ea1c"},
- {file = "coverage-7.1.0-cp310-cp310-win32.whl", hash = "sha256:4b14d5e09c656de5038a3f9bfe5228f53439282abcab87317c9f7f1acb280352"},
- {file = "coverage-7.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:8361be1c2c073919500b6601220a6f2f98ea0b6d2fec5014c1d9cfa23dd07038"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:da9b41d4539eefd408c46725fb76ecba3a50a3367cafb7dea5f250d0653c1040"},
- {file = "coverage-7.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c5b15ed7644ae4bee0ecf74fee95808dcc34ba6ace87e8dfbf5cb0dc20eab45a"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d12d076582507ea460ea2a89a8c85cb558f83406c8a41dd641d7be9a32e1274f"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2617759031dae1bf183c16cef8fcfb3de7617f394c813fa5e8e46e9b82d4222"},
- {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4e4881fa9e9667afcc742f0c244d9364d197490fbc91d12ac3b5de0bf2df146"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:9d58885215094ab4a86a6aef044e42994a2bd76a446dc59b352622655ba6621b"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:ffeeb38ee4a80a30a6877c5c4c359e5498eec095878f1581453202bfacc8fbc2"},
- {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3baf5f126f30781b5e93dbefcc8271cb2491647f8283f20ac54d12161dff080e"},
- {file = "coverage-7.1.0-cp311-cp311-win32.whl", hash = "sha256:ded59300d6330be27bc6cf0b74b89ada58069ced87c48eaf9344e5e84b0072f7"},
- {file = "coverage-7.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:6a43c7823cd7427b4ed763aa7fb63901ca8288591323b58c9cd6ec31ad910f3c"},
- {file = "coverage-7.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:7a726d742816cb3a8973c8c9a97539c734b3a309345236cd533c4883dda05b8d"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc7c85a150501286f8b56bd8ed3aa4093f4b88fb68c0843d21ff9656f0009d6a"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f5b4198d85a3755d27e64c52f8c95d6333119e49fd001ae5798dac872c95e0f8"},
- {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ddb726cb861c3117a553f940372a495fe1078249ff5f8a5478c0576c7be12050"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:51b236e764840a6df0661b67e50697aaa0e7d4124ca95e5058fa3d7cbc240b7c"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:7ee5c9bb51695f80878faaa5598040dd6c9e172ddcf490382e8aedb8ec3fec8d"},
- {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c31b75ae466c053a98bf26843563b3b3517b8f37da4d47b1c582fdc703112bc3"},
- {file = "coverage-7.1.0-cp37-cp37m-win32.whl", hash = "sha256:3b155caf3760408d1cb903b21e6a97ad4e2bdad43cbc265e3ce0afb8e0057e73"},
- {file = "coverage-7.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2a60d6513781e87047c3e630b33b4d1e89f39836dac6e069ffee28c4786715f5"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f2cba5c6db29ce991029b5e4ac51eb36774458f0a3b8d3137241b32d1bb91f06"},
- {file = "coverage-7.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:beeb129cacea34490ffd4d6153af70509aa3cda20fdda2ea1a2be870dfec8d52"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c45948f613d5d18c9ec5eaa203ce06a653334cf1bd47c783a12d0dd4fd9c851"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef382417db92ba23dfb5864a3fc9be27ea4894e86620d342a116b243ade5d35d"},
- {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7c7c0d0827e853315c9bbd43c1162c006dd808dbbe297db7ae66cd17b07830f0"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e5cdbb5cafcedea04924568d990e20ce7f1945a1dd54b560f879ee2d57226912"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:9817733f0d3ea91bea80de0f79ef971ae94f81ca52f9b66500c6a2fea8e4b4f8"},
- {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:218fe982371ac7387304153ecd51205f14e9d731b34fb0568181abaf7b443ba0"},
- {file = "coverage-7.1.0-cp38-cp38-win32.whl", hash = "sha256:04481245ef966fbd24ae9b9e537ce899ae584d521dfbe78f89cad003c38ca2ab"},
- {file = "coverage-7.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ae125d1134bf236acba8b83e74c603d1b30e207266121e76484562bc816344c"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2bf1d5f2084c3932b56b962a683074a3692bce7cabd3aa023c987a2a8e7612f6"},
- {file = "coverage-7.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:98b85dd86514d889a2e3dd22ab3c18c9d0019e696478391d86708b805f4ea0fa"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38da2db80cc505a611938d8624801158e409928b136c8916cd2e203970dde4dc"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3164d31078fa9efe406e198aecd2a02d32a62fecbdef74f76dad6a46c7e48311"},
- {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db61a79c07331e88b9a9974815c075fbd812bc9dbc4dc44b366b5368a2936063"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9ccb092c9ede70b2517a57382a601619d20981f56f440eae7e4d7eaafd1d1d09"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:33ff26d0f6cc3ca8de13d14fde1ff8efe1456b53e3f0273e63cc8b3c84a063d8"},
- {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d47dd659a4ee952e90dc56c97d78132573dc5c7b09d61b416a9deef4ebe01a0c"},
- {file = "coverage-7.1.0-cp39-cp39-win32.whl", hash = "sha256:d248cd4a92065a4d4543b8331660121b31c4148dd00a691bfb7a5cdc7483cfa4"},
- {file = "coverage-7.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:7ed681b0f8e8bcbbffa58ba26fcf5dbc8f79e7997595bf071ed5430d8c08d6f3"},
- {file = "coverage-7.1.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:755e89e32376c850f826c425ece2c35a4fc266c081490eb0a841e7c1cb0d3bda"},
- {file = "coverage-7.1.0.tar.gz", hash = "sha256:10188fe543560ec4874f974b5305cd1a8bdcfa885ee00ea3a03733464c4ca265"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90e7a4cbbb7b1916937d380beb1315b12957b8e895d7d9fb032e2038ac367525"},
+ {file = "coverage-7.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:34d7211be69b215ad92298a962b2cd5a4ef4b17c7871d85e15d3d1b6dc8d8c96"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:971b49dbf713044c3e5f6451b39f65615d4d1c1d9a19948fa0f41b0245a98765"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f0557289260125a6c453ad5673ba79e5b6841d9a20c9e101f758bfbedf928a77"},
+ {file = "coverage-7.2.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:049806ae2df69468c130f04f0fab4212c46b34ba5590296281423bb1ae379df2"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:875b03d92ac939fbfa8ae74a35b2c468fc4f070f613d5b1692f9980099a3a210"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:c160e34e388277f10c50dc2c7b5e78abe6d07357d9fe7fcb2f3c156713fd647e"},
+ {file = "coverage-7.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32e6a730fd18b2556716039ab93278ccebbefa1af81e6aa0c8dba888cf659e6e"},
+ {file = "coverage-7.2.0-cp310-cp310-win32.whl", hash = "sha256:f3ff4205aff999164834792a3949f82435bc7c7655c849226d5836c3242d7451"},
+ {file = "coverage-7.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:93db11da6e728587e943dff8ae1b739002311f035831b6ecdb15e308224a4247"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cd38140b56538855d3d5722c6d1b752b35237e7ea3f360047ce57f3fade82d98"},
+ {file = "coverage-7.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9dbb21561b0e04acabe62d2c274f02df0d715e8769485353ddf3cf84727e31ce"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:171dd3aa71a49274a7e4fc26f5bc167bfae5a4421a668bc074e21a0522a0af4b"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4655ecd813f4ba44857af3e9cffd133ab409774e9d2a7d8fdaf4fdfd2941b789"},
+ {file = "coverage-7.2.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1856a8c4aa77eb7ca0d42c996d0ca395ecafae658c1432b9da4528c429f2575c"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd67df6b48db18c10790635060858e2ea4109601e84a1e9bfdd92e898dc7dc79"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:2d7daf3da9c7e0ed742b3e6b4de6cc464552e787b8a6449d16517b31bbdaddf5"},
+ {file = "coverage-7.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf9e02bc3dee792b9d145af30db8686f328e781bd212fdef499db5e9e4dd8377"},
+ {file = "coverage-7.2.0-cp311-cp311-win32.whl", hash = "sha256:3713a8ec18781fda408f0e853bf8c85963e2d3327c99a82a22e5c91baffcb934"},
+ {file = "coverage-7.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:88ae5929f0ef668b582fd7cad09b5e7277f50f912183cf969b36e82a1c26e49a"},
+ {file = "coverage-7.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5e29a64e9586194ea271048bc80c83cdd4587830110d1e07b109e6ff435e5dbc"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d5302eb84c61e758c9d68b8a2f93a398b272073a046d07da83d77b0edc8d76b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2c9fffbc39dc4a6277e1525cab06c161d11ee3995bbc97543dc74fcec33e045b"},
+ {file = "coverage-7.2.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a6ceeab5fca62bca072eba6865a12d881f281c74231d2990f8a398226e1a5d96"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:28563a35ef4a82b5bc5160a01853ce62b9fceee00760e583ffc8acf9e3413753"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:bfa065307667f1c6e1f4c3e13f415b0925e34e56441f5fda2c84110a4a1d8bda"},
+ {file = "coverage-7.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7f992b32286c86c38f07a8b5c3fc88384199e82434040a729ec06b067ee0d52c"},
+ {file = "coverage-7.2.0-cp37-cp37m-win32.whl", hash = "sha256:2c15bd09fd5009f3a79c8b3682b52973df29761030b692043f9834fc780947c4"},
+ {file = "coverage-7.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f332d61fbff353e2ef0f3130a166f499c3fad3a196e7f7ae72076d41a6bfb259"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:577a8bc40c01ad88bb9ab1b3a1814f2f860ff5c5099827da2a3cafc5522dadea"},
+ {file = "coverage-7.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:9240a0335365c29c968131bdf624bb25a8a653a9c0d8c5dbfcabf80b59c1973c"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:358d3bce1468f298b19a3e35183bdb13c06cdda029643537a0cc37e55e74e8f1"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:932048364ff9c39030c6ba360c31bf4500036d4e15c02a2afc5a76e7623140d4"},
+ {file = "coverage-7.2.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7efa21611ffc91156e6f053997285c6fe88cfef3fb7533692d0692d2cb30c846"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:465ea431c3b78a87e32d7d9ea6d081a1003c43a442982375cf2c247a19971961"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0f03c229f1453b936916f68a47b3dfb5e84e7ad48e160488168a5e35115320c8"},
+ {file = "coverage-7.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:40785553d68c61e61100262b73f665024fd2bb3c6f0f8e2cd5b13e10e4df027b"},
+ {file = "coverage-7.2.0-cp38-cp38-win32.whl", hash = "sha256:b09dd7bef59448c66e6b490cc3f3c25c14bc85d4e3c193b81a6204be8dd355de"},
+ {file = "coverage-7.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:dc4f9a89c82faf6254d646180b2e3aa4daf5ff75bdb2c296b9f6a6cf547e26a7"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c243b25051440386179591a8d5a5caff4484f92c980fb6e061b9559da7cc3f64"},
+ {file = "coverage-7.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4b8fd32f85b256fc096deeb4872aeb8137474da0c0351236f93cbedc359353d6"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d7f2a7df523791e6a63b40360afa6792a11869651307031160dc10802df9a252"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:da32526326e8da0effb452dc32a21ffad282c485a85a02aeff2393156f69c1c3"},
+ {file = "coverage-7.2.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c1153a6156715db9d6ae8283480ae67fb67452aa693a56d7dae9ffe8f7a80da"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:74cd60fa00f46f28bd40048d6ca26bd58e9bee61d2b0eb4ec18cea13493c003f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:59a427f8a005aa7254074719441acb25ac2c2f60c1f1026d43f846d4254c1c2f"},
+ {file = "coverage-7.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c3c4beddee01c8125a75cde3b71be273995e2e9ec08fbc260dd206b46bb99969"},
+ {file = "coverage-7.2.0-cp39-cp39-win32.whl", hash = "sha256:08e3dd256b8d3e07bb230896c8c96ec6c5dffbe5a133ba21f8be82b275b900e8"},
+ {file = "coverage-7.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:ad12c74c6ce53a027f5a5ecbac9be20758a41c85425c1bbab7078441794b04ee"},
+ {file = "coverage-7.2.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:ffa637a2d5883298449a5434b699b22ef98dd8e2ef8a1d9e60fa9cfe79813411"},
+ {file = "coverage-7.2.0.tar.gz", hash = "sha256:9cc9c41aa5af16d845b53287051340c363dd03b7ef408e45eec3af52be77810d"},
@@ -1113 +1113 @@ name = "fastavro"
-version = "1.7.1"
+version = "1.7.2"
@@ -1119,21 +1119,21 @@ files = [
- {file = "fastavro-1.7.1-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:f8f7f11af8c2c074341217d6247b7ba09cadcd55f899e046c14e3a44afa5fc95"},
- {file = "fastavro-1.7.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:11159cf23f5ff4b752b028a77bd2a7941599932527e8a6512779e25b0503f037"},
- {file = "fastavro-1.7.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:86ce89743711355eee7d3fec34847de1ab574f4567aa4a667e966711bb2e0cd9"},
- {file = "fastavro-1.7.1-cp310-cp310-win_amd64.whl", hash = "sha256:2e3f5ae42e033dbb6e0efa788c4b8a4e5a59bc2b9cb83f717b6d85176727faed"},
- {file = "fastavro-1.7.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:8b671a97e864e4c024061c0d6f93f2768ba0595f917317ca4fe3e99dca6fcf3d"},
- {file = "fastavro-1.7.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e9be910ca1645e0e5c4192d667cfdfa58dff4f8690db5af1ae33602643d41a78"},
- {file = "fastavro-1.7.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6f233a1d0d95265f81a25c274f55017bd39d9b8f7f1387a4235bf8e01841a9ff"},
- {file = "fastavro-1.7.1-cp311-cp311-win_amd64.whl", hash = "sha256:31c924ed2fdad56182b543941cdec9cc384443cc3ca9462e0580afeb4dc64f4b"},
- {file = "fastavro-1.7.1-cp37-cp37m-macosx_10_15_x86_64.whl", hash = "sha256:8ad522c7f2c7791cfe54a7e87af8ac09634872f4fdeef28deec97fab8de38b24"},
- {file = "fastavro-1.7.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:74fe0fb24a489e2fd567775935eb55a461fc6c4da8b5e3467245752ac2098284"},
- {file = "fastavro-1.7.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f7a744569b4b44ea1f4f87e7ea8298e1e2bf23779aa6ef255b95f9f38faad48"},
- {file = "fastavro-1.7.1-cp37-cp37m-win_amd64.whl", hash = "sha256:b793252a6325341890bbbbdca2804d4db3d5d29ff7f15a58fcf84dda440808fa"},
- {file = "fastavro-1.7.1-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:a107a8b333628c0f8a2ece5d5a57c69923719a139b106ace4050206250df4b13"},
- {file = "fastavro-1.7.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:182671595683054ec234beec03f485b5c889c21c08e429577ae7929480703409"},
- {file = "fastavro-1.7.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:980ce7fdccf328d287e8b789f4e3b422f64c84ed9cd81c05dd7c560c3d8076b1"},
- {file = "fastavro-1.7.1-cp38-cp38-win_amd64.whl", hash = "sha256:a5f104bc5b4986bbbcab170918c4d8ffa4f8efa3ebe8ec190954178630074d5a"},
- {file = "fastavro-1.7.1-cp39-cp39-macosx_11_0_x86_64.whl", hash = "sha256:5ff701d6b228218a3d9c09b392205dd0899afa501a4f14724bef0ce13a719700"},
- {file = "fastavro-1.7.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb2d8ea7b21493bb18ff7c68401edbc663bbd8a57016d6cf2c4b0a2dc4464e7"},
- {file = "fastavro-1.7.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e15e8163581983f07a3156902761cf746bbe1e646abd9553cc9a1cede6e23ae9"},
- {file = "fastavro-1.7.1-cp39-cp39-win_amd64.whl", hash = "sha256:4d56cc05ceb1c9f4de19c1f330a4f140af3b944834d63cd0e11517deaea21ee1"},
- {file = "fastavro-1.7.1.tar.gz", hash = "sha256:4b8bcae4ed6343af186e638061cdfbc5331cdb5e026d055099c91d4f07be838c"},
+ {file = "fastavro-1.7.2-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:20a94269bc0910c41f06106424429289e5c8cdce65b006e48d600c9bb6c442a5"},
+ {file = "fastavro-1.7.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b34e9cd755e4d9d12377fe151ad95b9bff08215f0a3756601b21ed31f4d127ab"},
+ {file = "fastavro-1.7.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8993e2f34e9ae195f417105d495fdd3d7f6d1532bacf57dc5dea600082f9b1a0"},
+ {file = "fastavro-1.7.2-cp310-cp310-win_amd64.whl", hash = "sha256:e83069f971f307175ab806e9929a15826e586ef05a620fcfc7768ceaa3ec19c8"},
+ {file = "fastavro-1.7.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:4c44f72a1a7b2d28cc21c95686916b782354dcc1f1aa1c8ca044c66d5ada3e14"},
+ {file = "fastavro-1.7.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:88005d5c6fe436e80f756399a52aaa6d0cdc506c1ee195277287a09c7617aec3"},
+ {file = "fastavro-1.7.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a5c5d466631f455213b8bfd54888d2d8691cfc240cd6dbb40eb3694bf05e562e"},
+ {file = "fastavro-1.7.2-cp311-cp311-win_amd64.whl", hash = "sha256:4c2523c5e2279debb5fa788bf665a5abbd16a1e889338b973e3b318d50fcce77"},
+ {file = "fastavro-1.7.2-cp37-cp37m-macosx_10_15_x86_64.whl", hash = "sha256:db3d9b6237934be5001fc18c07d25c09d3008ce6ab3448daf4b30608b84488e0"},
+ {file = "fastavro-1.7.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e82862cb7d3bf99fdca65048b8b09db295795d7fb1eb7c5dae29c0a4d7dcf1f2"},
+ {file = "fastavro-1.7.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5417dfbf99ac5bf963a929f2541c643d35e1a5a22876ba5ae0dca7b86326d9c2"},
+ {file = "fastavro-1.7.2-cp37-cp37m-win_amd64.whl", hash = "sha256:b8cd8bd9a8ad4b687bd8dbf12effc2996bdf41ec04960875c19a7a90c949e8a8"},
+ {file = "fastavro-1.7.2-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:f07d1f2ebdcc18938908e710a79ab3731061936fbea27534e436f9e12342450b"},
+ {file = "fastavro-1.7.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9919d866319b1e123374fced84afb4728802872bd5938ba7e12a691c54593542"},
+ {file = "fastavro-1.7.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8da1a9d4f0b871e7c968c8413caef19e19002100036b2714e037543c36550b8d"},
+ {file = "fastavro-1.7.2-cp38-cp38-win_amd64.whl", hash = "sha256:e842beff23b0b207c45f99fa79d50cfc0d84ac137d84a7e283e9e59a8368b115"},
+ {file = "fastavro-1.7.2-cp39-cp39-macosx_11_0_x86_64.whl", hash = "sha256:c1d5805c50a9507c99b814cb014da2812e6e0ee2770f4b793ce56b82a5b0af21"},
+ {file = "fastavro-1.7.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:22cb5d2bf87c1ed9a1dcc97b6b13dc6b7ee4395dbf1fd5903dac01cdde0291c2"},
+ {file = "fastavro-1.7.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:06bd51962df9b3f28c7081c0a3c8663d840df05d964b7817d85bd37c1d101236"},
+ {file = "fastavro-1.7.2-cp39-cp39-win_amd64.whl", hash = "sha256:d787e7739d93db59f378c0e44a4eca8e56ff979e5b43dea6378e87c48548818c"},
+ {file = "fastavro-1.7.2.tar.gz", hash = "sha256:3c524c65a5d578c020997800fce992cd90cc88ef13148b70f01a3429de34c8a4"},
@@ -1330 +1330 @@ name = "gdown"
-version = "4.6.3"
+version = "4.6.4"
@@ -1336,2 +1336,2 @@ files = [
- {file = "gdown-4.6.3-py3-none-any.whl", hash = "sha256:36bbb962b373f9634af1b9a10c73e231d3ad9b084659efd93bf3c62d57c25a97"},
- {file = "gdown-4.6.3.tar.gz", hash = "sha256:ec6ea922ef4e3616a054073656d9aa2cdc717fbdbd0a94d2f77d52d323943307"},
+ {file = "gdown-4.6.4-py3-none-any.whl", hash = "sha256:03eb456d43611d7f8531a68662cae0d5c65cb74c7159365f2d1fbe882b39ce6b"},
+ {file = "gdown-4.6.4.tar.gz", hash = "sha256:d33210b1b3d7b12fef05d6b9ea7f22a513739909ca511d5d9d236dc676667f79"},
@@ -1364,2 +1364,2 @@ name = "gitpython"
-version = "3.1.30"
-description = "GitPython is a python library used to interact with Git repositories"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
@@ -1370,2 +1370,2 @@ files = [
- {file = "GitPython-3.1.30-py3-none-any.whl", hash = "sha256:cd455b0000615c60e286208ba540271af9fe531fa6a87cc590a7298785ab2882"},
- {file = "GitPython-3.1.30.tar.gz", hash = "sha256:769c2d83e13f5d938b7688479da374c4e3d49f71549aaf462b646db9602ea6f8"},
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
@@ -1379 +1379 @@ name = "google-auth"
-version = "2.16.0"
+version = "2.16.1"
@@ -1385,2 +1385,2 @@ files = [
- {file = "google-auth-2.16.0.tar.gz", hash = "sha256:ed7057a101af1146f0554a769930ac9de506aeca4fd5af6543ebe791851a9fbd"},
- {file = "google_auth-2.16.0-py2.py3-none-any.whl", hash = "sha256:5045648c821fb72384cdc0e82cc326df195f113a33049d9b62b74589243d2acc"},
+ {file = "google-auth-2.16.1.tar.gz", hash = "sha256:5fd170986bce6bfd7bb5c845c4b8362edb1e0cba901e062196e83f8bb5d5d32c"},
+ {file = "google_auth-2.16.1-py2.py3-none-any.whl", hash = "sha256:75d76ea857df65938e1f71dcbcd7d0cd48e3f80b34b8870ba229c9292081f7ef"},
@@ -1439 +1439 @@ name = "grpcio"
-version = "1.51.1"
+version = "1.51.3"
@@ -1445,45 +1445,45 @@ files = [
- {file = "grpcio-1.51.1-cp310-cp310-linux_armv7l.whl", hash = "sha256:cc2bece1737b44d878cc1510ea04469a8073dbbcdd762175168937ae4742dfb3"},
- {file = "grpcio-1.51.1-cp310-cp310-macosx_12_0_x86_64.whl", hash = "sha256:e223a9793522680beae44671b9ed8f6d25bbe5ddf8887e66aebad5e0686049ef"},
- {file = "grpcio-1.51.1-cp310-cp310-manylinux_2_17_aarch64.whl", hash = "sha256:24ac1154c4b2ab4a0c5326a76161547e70664cd2c39ba75f00fc8a2170964ea2"},
- {file = "grpcio-1.51.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4ef09f8997c4be5f3504cefa6b5c6cc3cf648274ce3cede84d4342a35d76db6"},
- {file = "grpcio-1.51.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a8a0b77e992c64880e6efbe0086fe54dfc0bbd56f72a92d9e48264dcd2a3db98"},
- {file = "grpcio-1.51.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:eacad297ea60c72dd280d3353d93fb1dcca952ec11de6bb3c49d12a572ba31dd"},
- {file = "grpcio-1.51.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:16c71740640ba3a882f50b01bf58154681d44b51f09a5728180a8fdc66c67bd5"},
- {file = "grpcio-1.51.1-cp310-cp310-win32.whl", hash = "sha256:29cb97d41a4ead83b7bcad23bdb25bdd170b1e2cba16db6d3acbb090bc2de43c"},
- {file = "grpcio-1.51.1-cp310-cp310-win_amd64.whl", hash = "sha256:9ff42c5620b4e4530609e11afefa4a62ca91fa0abb045a8957e509ef84e54d30"},
- {file = "grpcio-1.51.1-cp311-cp311-linux_armv7l.whl", hash = "sha256:bc59f7ba87972ab236f8669d8ca7400f02a0eadf273ca00e02af64d588046f02"},
- {file = "grpcio-1.51.1-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:3c2b3842dcf870912da31a503454a33a697392f60c5e2697c91d133130c2c85d"},
- {file = "grpcio-1.51.1-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:22b011674090594f1f3245960ced7386f6af35485a38901f8afee8ad01541dbd"},
- {file = "grpcio-1.51.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:49d680356a975d9c66a678eb2dde192d5dc427a7994fb977363634e781614f7c"},
- {file = "grpcio-1.51.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:094e64236253590d9d4075665c77b329d707b6fca864dd62b144255e199b4f87"},
- {file = "grpcio-1.51.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:257478300735ce3c98d65a930bbda3db172bd4e00968ba743e6a1154ea6edf10"},
- {file = "grpcio-1.51.1-cp311-cp311-win32.whl", hash = "sha256:5a6ebcdef0ef12005d56d38be30f5156d1cb3373b52e96f147f4a24b0ddb3a9d"},
- {file = "grpcio-1.51.1-cp311-cp311-win_amd64.whl", hash = "sha256:3f9b0023c2c92bebd1be72cdfca23004ea748be1813a66d684d49d67d836adde"},
- {file = "grpcio-1.51.1-cp37-cp37m-linux_armv7l.whl", hash = "sha256:cd3baccea2bc5c38aeb14e5b00167bd4e2373a373a5e4d8d850bd193edad150c"},
- {file = "grpcio-1.51.1-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:17ec9b13cec4a286b9e606b48191e560ca2f3bbdf3986f91e480a95d1582e1a7"},
- {file = "grpcio-1.51.1-cp37-cp37m-manylinux_2_17_aarch64.whl", hash = "sha256:fbdbe9a849854fe484c00823f45b7baab159bdd4a46075302281998cb8719df5"},
- {file = "grpcio-1.51.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:31bb6bc7ff145e2771c9baf612f4b9ebbc9605ccdc5f3ff3d5553de7fc0e0d79"},
- {file = "grpcio-1.51.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e473525c28251558337b5c1ad3fa969511e42304524a4e404065e165b084c9e4"},
- {file = "grpcio-1.51.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:6f0b89967ee11f2b654c23b27086d88ad7bf08c0b3c2a280362f28c3698b2896"},
- {file = "grpcio-1.51.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:7942b32a291421460d6a07883033e392167d30724aa84987e6956cd15f1a21b9"},
- {file = "grpcio-1.51.1-cp37-cp37m-win32.whl", hash = "sha256:f96ace1540223f26fbe7c4ebbf8a98e3929a6aa0290c8033d12526847b291c0f"},
- {file = "grpcio-1.51.1-cp37-cp37m-win_amd64.whl", hash = "sha256:f1fec3abaf274cdb85bf3878167cfde5ad4a4d97c68421afda95174de85ba813"},
- {file = "grpcio-1.51.1-cp38-cp38-linux_armv7l.whl", hash = "sha256:0e1a9e1b4a23808f1132aa35f968cd8e659f60af3ffd6fb00bcf9a65e7db279f"},
- {file = "grpcio-1.51.1-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:6df3b63538c362312bc5fa95fb965069c65c3ea91d7ce78ad9c47cab57226f54"},
- {file = "grpcio-1.51.1-cp38-cp38-manylinux_2_17_aarch64.whl", hash = "sha256:172405ca6bdfedd6054c74c62085946e45ad4d9cec9f3c42b4c9a02546c4c7e9"},
- {file = "grpcio-1.51.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:506b9b7a4cede87d7219bfb31014d7b471cfc77157da9e820a737ec1ea4b0663"},
- {file = "grpcio-1.51.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0fb93051331acbb75b49a2a0fd9239c6ba9528f6bdc1dd400ad1cb66cf864292"},
- {file = "grpcio-1.51.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:5dca372268c6ab6372d37d6b9f9343e7e5b4bc09779f819f9470cd88b2ece3c3"},
- {file = "grpcio-1.51.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:471d39d3370ca923a316d49c8aac66356cea708a11e647e3bdc3d0b5de4f0a40"},
- {file = "grpcio-1.51.1-cp38-cp38-win32.whl", hash = "sha256:75e29a90dc319f0ad4d87ba6d20083615a00d8276b51512e04ad7452b5c23b04"},
- {file = "grpcio-1.51.1-cp38-cp38-win_amd64.whl", hash = "sha256:f1158bccbb919da42544a4d3af5d9296a3358539ffa01018307337365a9a0c64"},
- {file = "grpcio-1.51.1-cp39-cp39-linux_armv7l.whl", hash = "sha256:59dffade859f157bcc55243714d57b286da6ae16469bf1ac0614d281b5f49b67"},
- {file = "grpcio-1.51.1-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:dad6533411d033b77f5369eafe87af8583178efd4039c41d7515d3336c53b4f1"},
- {file = "grpcio-1.51.1-cp39-cp39-manylinux_2_17_aarch64.whl", hash = "sha256:4c4423ea38a7825b8fed8934d6d9aeebdf646c97e3c608c3b0bcf23616f33877"},
- {file = "grpcio-1.51.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0dc5354e38e5adf2498312f7241b14c7ce3484eefa0082db4297189dcbe272e6"},
- {file = "grpcio-1.51.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:97d67983189e2e45550eac194d6234fc38b8c3b5396c153821f2d906ed46e0ce"},
- {file = "grpcio-1.51.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:538d981818e49b6ed1e9c8d5e5adf29f71c4e334e7d459bf47e9b7abb3c30e09"},
- {file = "grpcio-1.51.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:9235dcd5144a83f9ca6f431bd0eccc46b90e2c22fe27b7f7d77cabb2fb515595"},
- {file = "grpcio-1.51.1-cp39-cp39-win32.whl", hash = "sha256:aacb54f7789ede5cbf1d007637f792d3e87f1c9841f57dd51abf89337d1b8472"},
- {file = "grpcio-1.51.1-cp39-cp39-win_amd64.whl", hash = "sha256:2b170eaf51518275c9b6b22ccb59450537c5a8555326fd96ff7391b5dd75303c"},
- {file = "grpcio-1.51.1.tar.gz", hash = "sha256:e6dfc2b6567b1c261739b43d9c59d201c1b89e017afd9e684d85aa7a186c9f7a"},
+ {file = "grpcio-1.51.3-cp310-cp310-linux_armv7l.whl", hash = "sha256:f601aaeae18dab81930fb8d4f916b0da21e89bb4b5f7367ef793f46b4a76b7b0"},
+ {file = "grpcio-1.51.3-cp310-cp310-macosx_12_0_universal2.whl", hash = "sha256:eef0450a4b5ed11feab639bf3eb1b6e23d0efa9b911bf7b06fb60e14f5f8a585"},
+ {file = "grpcio-1.51.3-cp310-cp310-manylinux_2_17_aarch64.whl", hash = "sha256:82b0ad8ac825d4bb31bff9f638557c045f4a6d824d84b21e893968286f88246b"},
+ {file = "grpcio-1.51.3-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3667c06e37d6cd461afdd51cefe6537702f3d1dc5ff4cac07e88d8b4795dc16f"},
+ {file = "grpcio-1.51.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3709048fe0aa23dda09b3e69849a12055790171dab9e399a72ea8f9dfbf9ac80"},
+ {file = "grpcio-1.51.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:200d69857f9910f7458b39b9bcf83ee4a180591b40146ba9e49314e3a7419313"},
+ {file = "grpcio-1.51.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:cd9a5e68e79c5f031500e67793048a90209711e0854a9ddee8a3ce51728de4e5"},
+ {file = "grpcio-1.51.3-cp310-cp310-win32.whl", hash = "sha256:6604f614016127ae10969176bbf12eb0e03d2fb3d643f050b3b69e160d144fb4"},
+ {file = "grpcio-1.51.3-cp310-cp310-win_amd64.whl", hash = "sha256:e95c7ccd4c5807adef1602005513bf7c7d14e5a41daebcf9d8d30d8bf51b8f81"},
+ {file = "grpcio-1.51.3-cp311-cp311-linux_armv7l.whl", hash = "sha256:5e77ee138100f0bb55cbd147840f87ee6241dbd25f09ea7cd8afe7efff323449"},
+ {file = "grpcio-1.51.3-cp311-cp311-macosx_10_10_universal2.whl", hash = "sha256:68a7514b754e38e8de9075f7bb4dee919919515ec68628c43a894027e40ddec4"},
+ {file = "grpcio-1.51.3-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3c1b9f8afa62ff265d86a4747a2990ec5a96e4efce5d5888f245a682d66eca47"},
+ {file = "grpcio-1.51.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8de30f0b417744288cec65ec8cf84b8a57995cf7f1e84ccad2704d93f05d0aae"},
+ {file = "grpcio-1.51.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b69c7adc7ed60da1cb1b502853db61f453fc745f940cbcc25eb97c99965d8f41"},
+ {file = "grpcio-1.51.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d81528ffe0e973dc840ec73a4132fd18b8203ad129d7410155d951a0a7e4f5d0"},
+ {file = "grpcio-1.51.3-cp311-cp311-win32.whl", hash = "sha256:040eb421613b57c696063abde405916dd830203c184c9000fc8c3b3b3c950325"},
+ {file = "grpcio-1.51.3-cp311-cp311-win_amd64.whl", hash = "sha256:2a8e17286c4240137d933b8ca506465472248b4ce0fe46f3404459e708b65b68"},
+ {file = "grpcio-1.51.3-cp37-cp37m-linux_armv7l.whl", hash = "sha256:d5cd1389669a847555df54177b911d9ff6f17345b2a6f19388707b7a9f724c88"},
+ {file = "grpcio-1.51.3-cp37-cp37m-macosx_10_10_universal2.whl", hash = "sha256:be1bf35ce82cdbcac14e39d5102d8de4079a1c1a6a06b68e41fcd9ef64f9dd28"},
+ {file = "grpcio-1.51.3-cp37-cp37m-manylinux_2_17_aarch64.whl", hash = "sha256:5eed34994c095e2bf7194ffac7381c6068b057ef1e69f8f08db77771350a7566"},
+ {file = "grpcio-1.51.3-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3f9a7d88082b2a17ae7bd3c2354d13bab0453899e0851733f6afa6918373f476"},
+ {file = "grpcio-1.51.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c8abbc5f837111e7bd619612eedc223c290b0903b952ce0c7b00840ea70f14"},
+ {file = "grpcio-1.51.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:165b05af77e6aecb4210ae7663e25acf234ba78a7c1c157fa5f2efeb0d6ec53c"},
+ {file = "grpcio-1.51.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:54e36c2ee304ff15f2bfbdc43d2b56c63331c52d818c364e5b5214e5bc2ad9f6"},
+ {file = "grpcio-1.51.3-cp37-cp37m-win32.whl", hash = "sha256:cd0daac21d9ef5e033a5100c1d3aa055bbed28bfcf070b12d8058045c4e821b1"},
+ {file = "grpcio-1.51.3-cp37-cp37m-win_amd64.whl", hash = "sha256:2fdd6333ce96435408565a9dbbd446212cd5d62e4d26f6a3c0feb1e3c35f1cc8"},
+ {file = "grpcio-1.51.3-cp38-cp38-linux_armv7l.whl", hash = "sha256:54b0c29bdd9a3b1e1b61443ab152f060fc719f1c083127ab08d03fac5efd51be"},
+ {file = "grpcio-1.51.3-cp38-cp38-macosx_10_10_universal2.whl", hash = "sha256:ffaaf7e93fcb437356b5a4b23bf36e8a3d0221399ff77fd057e4bc77776a24be"},
+ {file = "grpcio-1.51.3-cp38-cp38-manylinux_2_17_aarch64.whl", hash = "sha256:eafbe7501a3268d05f2e450e1ddaffb950d842a8620c13ec328b501d25d2e2c3"},
+ {file = "grpcio-1.51.3-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:881ecb34feabf31c6b3b9bbbddd1a5b57e69f805041e5a2c6c562a28574f71c4"},
+ {file = "grpcio-1.51.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e860a3222139b41d430939bbec2ec9c3f6c740938bf7a04471a9a8caaa965a2e"},
+ {file = "grpcio-1.51.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:49ede0528e9dac7e8a9fe30b16c73b630ddd9a576bf4b675eb6b0c53ee5ca00f"},
+ {file = "grpcio-1.51.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6972b009638b40a448d10e1bc18e2223143b8a7aa20d7def0d78dd4af4126d12"},
+ {file = "grpcio-1.51.3-cp38-cp38-win32.whl", hash = "sha256:5694448256e3cdfe5bd358f1574a3f2f51afa20cc834713c4b9788d60b7cc646"},
+ {file = "grpcio-1.51.3-cp38-cp38-win_amd64.whl", hash = "sha256:3ea4341efe603b049e8c9a5f13c696ca37fcdf8a23ca35f650428ad3606381d9"},
+ {file = "grpcio-1.51.3-cp39-cp39-linux_armv7l.whl", hash = "sha256:6c677581ce129f5fa228b8f418cee10bd28dd449f3a544ea73c8ba590ee49d0b"},
+ {file = "grpcio-1.51.3-cp39-cp39-macosx_10_10_universal2.whl", hash = "sha256:30e09b5e0531685e176f49679b6a3b190762cc225f4565e55a899f5e14b3aa62"},
+ {file = "grpcio-1.51.3-cp39-cp39-manylinux_2_17_aarch64.whl", hash = "sha256:c831f31336e81243f85b6daff3e5e8a123302ce0ea1f2726ad752fd7a59f3aee"},
+ {file = "grpcio-1.51.3-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2cd2e4cefb724cab1ba2df4b7535a9980531b9ec51b4dbb5f137a1f3a3754ef0"},
+ {file = "grpcio-1.51.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f7a0d0bf44438869d307f85a54f25a896ad6b4b0ca12370f76892ad732928d87"},
+ {file = "grpcio-1.51.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:c02abd55409bfb293371554adf6a4401197ec2133dd97727c01180889014ba4d"},
+ {file = "grpcio-1.51.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:2f8ff75e61e1227ba7a3f16b2eadbcc11d0a54096d52ab75a6b88cfbe56f55d1"},
+ {file = "grpcio-1.51.3-cp39-cp39-win32.whl", hash = "sha256:6c99a73a6260bdf844b2e5ddad02dcd530310f80e1fa72c300fa19c1c7496962"},
+ {file = "grpcio-1.51.3-cp39-cp39-win_amd64.whl", hash = "sha256:22bdfac4f7f27acdd4da359b5e7e1973dc74bf1ed406729b07d0759fde2f064b"},
+ {file = "grpcio-1.51.3.tar.gz", hash = "sha256:be7b2265b7527bb12109a7727581e274170766d5b3c9258d4e466f4872522d7a"},
@@ -1493 +1493 @@ files = [
-protobuf = ["grpcio-tools (>=1.51.1)"]
+protobuf = ["grpcio-tools (>=1.51.3)"]
@@ -1594 +1594 @@ name = "huggingface-hub"
-version = "0.12.0"
+version = "0.12.1"
@@ -1600,2 +1600,2 @@ files = [
- {file = "huggingface_hub-0.12.0-py3-none-any.whl", hash = "sha256:93809eabbfb2058a808bddf8b2a70f645de3f9df73ce87ddf5163d4c74b71c0c"},
- {file = "huggingface_hub-0.12.0.tar.gz", hash = "sha256:da82c9ec8f9d8f976ffd3fd8249d20bb35c2dd3145a9f7ca1106f0ebefd9afa0"},
+ {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"},
+ {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"},
@@ -1810 +1810 @@ reference = "master"
-resolved_reference = "ba17d213a27bbc7263cc30d4f293967aa1021cff"
+resolved_reference = "9af679c38477b564c26917a5dcf52d2c86177fb9"
@@ -1833,0 +1834,16 @@ files = [
+[[package]]
+name = "lazy-loader"
+version = "0.1"
+description = "lazy_loader"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "lazy_loader-0.1-py3-none-any.whl", hash = "sha256:623bd4831a40ce659d74472af40a58d016f2a5a047685409affbc2ba5c044641"},
+ {file = "lazy_loader-0.1.tar.gz", hash = "sha256:77ce7f2737ebabf9c0ff73b4a99c947876d74d24c2f026544e32246ecca5feca"},
+]
+
+[package.extras]
+lint = ["pre-commit (>=2.20)"]
+test = ["codecov (>=2.1)", "pytest (>=7.1)", "pytest-cov (>=3.0)"]
+
@@ -1879 +1895 @@ name = "librosa"
-version = "0.9.2"
+version = "0.10.0"
@@ -1883 +1899 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -1885,2 +1901,2 @@ files = [
- {file = "librosa-0.9.2-py3-none-any.whl", hash = "sha256:322a813e6d37af9fbc369e6a637dcf5fdc5c6925ce806a0d27c68de61a81350f"},
- {file = "librosa-0.9.2.tar.gz", hash = "sha256:5b576b5efdce428e90bc988bdd5a953d12a727e5f931f30d74c53b63abbe3c89"},
+ {file = "librosa-0.10.0-py3-none-any.whl", hash = "sha256:6db29c1467168da21313203dcef405a73a678d3aad0fbc67607250b2f08a3f5a"},
+ {file = "librosa-0.10.0.tar.gz", hash = "sha256:8e8669e5084002d1a87f6c82b732f370784a368d0e55c2dd7d7aef3fa02fd058"},
@@ -1891 +1907 @@ audioread = ">=2.1.9"
-decorator = ">=4.0.10"
+decorator = ">=4.3.0"
@@ -1893,3 +1909,4 @@ joblib = ">=0.14"
-numba = ">=0.45.1"
-numpy = ">=1.17.0"
-packaging = ">=20.0"
+lazy-loader = ">=0.1"
+msgpack = ">=1.0"
+numba = ">=0.51.0"
+numpy = ">=1.20.3"
@@ -1897,2 +1914 @@ pooch = ">=1.0"
-resampy = ">=0.2.2"
-scikit-learn = ">=0.19.1"
+scikit-learn = ">=0.20.0"
@@ -1900 +1916,3 @@ scipy = ">=1.2.0"
-soundfile = ">=0.10.2"
+soundfile = ">=0.12.1"
+soxr = ">=0.3.2"
+typing-extensions = ">=4.1.1"
@@ -1904,2 +1922,2 @@ display = ["matplotlib (>=3.3.0)"]
-docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (<0.50)", "numpydoc", "presets", "sphinx (!=1.3.1)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"]
-tests = ["contextlib2", "matplotlib (>=3.3.0)", "pytest", "pytest-cov", "pytest-mpl", "samplerate", "soxr"]
+docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"]
+tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"]
@@ -2087 +2105 @@ name = "markdown-it-py"
-version = "2.1.0"
+version = "2.2.0"
@@ -2093,2 +2111,2 @@ files = [
- {file = "markdown-it-py-2.1.0.tar.gz", hash = "sha256:cf7e59fed14b5ae17c0006eff14a2d9a00ed5f3a846148153899a0224e2c07da"},
- {file = "markdown_it_py-2.1.0-py3-none-any.whl", hash = "sha256:93de681e5c021a432c63147656fe21790bc01231e0cd2da73626f1aa3ac0fe27"},
+ {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"},
+ {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"},
@@ -2101,4 +2119,4 @@ mdurl = ">=0.1,<1.0"
-benchmarking = ["psutil", "pytest", "pytest-benchmark (>=3.2,<4.0)"]
-code-style = ["pre-commit (==2.6)"]
-compare = ["commonmark (>=0.9.1,<0.10.0)", "markdown (>=3.3.6,<3.4.0)", "mistletoe (>=0.8.1,<0.9.0)", "mistune (>=2.0.2,<2.1.0)", "panflute (>=2.1.3,<2.2.0)"]
-linkify = ["linkify-it-py (>=1.0,<2.0)"]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
@@ -2217 +2235 @@ name = "mirakuru"
-version = "2.4.2"
+version = "2.5.0"
@@ -2223,2 +2241,2 @@ files = [
- {file = "mirakuru-2.4.2-py3-none-any.whl", hash = "sha256:fdb67d141cc9f7abd485a515d618daf3272c3e6ff48380749997ff8e8c5f2cb2"},
- {file = "mirakuru-2.4.2.tar.gz", hash = "sha256:ec84d4d81b4bca96cb0e598c6b3d198a92f036a0c1223c881482c02a98508226"},
+ {file = "mirakuru-2.5.0-py3-none-any.whl", hash = "sha256:20c440f07e11589735e2634bcf5ab9df0e9f901a1a18e323e15e0200112499fc"},
+ {file = "mirakuru-2.5.0.tar.gz", hash = "sha256:4484f9405610886de3eb1bb1405b3ccd9b0d88b3ea4de68ee7873a873c11956b"},
@@ -2230,3 +2247,0 @@ psutil = {version = ">=4.0.0", markers = "sys_platform != \"cygwin\""}
-[package.extras]
-tests = ["pytest", "pytest-cov", "python-daemon"]
-
@@ -2264 +2279 @@ description = "MessagePack serializer"
-category = "dev"
+category = "main"
@@ -2441 +2456 @@ name = "mypy"
-version = "1.0.0"
+version = "1.0.1"
@@ -2447,26 +2462,26 @@ files = [
- {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
- {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
- {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
- {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
- {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
- {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
- {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
- {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
- {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
- {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
- {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
- {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
- {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
- {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
- {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
- {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
- {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
- {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
- {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
- {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
- {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
- {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
- {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"},
+ {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"},
+ {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"},
+ {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"},
+ {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"},
+ {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"},
+ {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"},
+ {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"},
+ {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"},
+ {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"},
+ {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"},
+ {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"},
+ {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"},
+ {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"},
+ {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"},
+ {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"},
+ {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"},
+ {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"},
+ {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"},
+ {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"},
+ {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"},
+ {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"},
+ {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"},
@@ -2878,7 +2892,0 @@ files = [
- {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
- {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
- {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
- {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
- {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
- {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
- {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -2956 +2964 @@ name = "pip"
-version = "23.0"
+version = "23.0.1"
@@ -2962,2 +2970,2 @@ files = [
- {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
- {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
+ {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"},
+ {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"},
@@ -3826 +3834 @@ name = "python-dotenv"
-version = "0.21.1"
+version = "1.0.0"
@@ -3830 +3838 @@ optional = false
-python-versions = ">=3.7"
+python-versions = ">=3.8"
@@ -3832,2 +3840,2 @@ files = [
- {file = "python-dotenv-0.21.1.tar.gz", hash = "sha256:1c93de8f636cde3ce377292818d0e440b6e45a82f215c3744979151fa8151c49"},
- {file = "python_dotenv-0.21.1-py3-none-any.whl", hash = "sha256:41e12e0318bebc859fcc4d97d4db8d20ad21721a6aa5047dd59f090391cb549a"},
+ {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"},
+ {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"},
@@ -4168,21 +4175,0 @@ rsa = ["oauthlib[signedtoken] (>=3.0.0)"]
-[[package]]
-name = "resampy"
-version = "0.4.2"
-description = "Efficient signal resampling"
-category = "main"
-optional = false
-python-versions = "*"
-files = [
- {file = "resampy-0.4.2-py3-none-any.whl", hash = "sha256:4340b6c4e685a865621dfcf016e2a3dd49d865446b6025e30fe88567f22e052e"},
- {file = "resampy-0.4.2.tar.gz", hash = "sha256:0a469e6ddb89956f4fd6c88728300e4bbd186fae569dd4fd17dae51a91cbaa15"},
-]
-
-[package.dependencies]
-numba = ">=0.53"
-numpy = ">=1.17"
-
-[package.extras]
-design = ["optuna (>=2.10.0)"]
-docs = ["numpydoc", "sphinx (!=1.3.1)"]
-tests = ["pytest (<8)", "pytest-cov", "scipy (>=1.0)"]
-
@@ -4305 +4292 @@ name = "scipy"
-version = "1.10.0"
+version = "1.10.1"
@@ -4311,21 +4298,21 @@ files = [
- {file = "scipy-1.10.0-cp310-cp310-macosx_10_15_x86_64.whl", hash = "sha256:b901b423c91281a974f6cd1c36f5c6c523e665b5a6d5e80fcb2334e14670eefd"},
- {file = "scipy-1.10.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:16ba05d3d1b9f2141004f3f36888e05894a525960b07f4c2bfc0456b955a00be"},
- {file = "scipy-1.10.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:151f066fe7d6653c3ffefd489497b8fa66d7316e3e0d0c0f7ff6acca1b802809"},
- {file = "scipy-1.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2f9ea0a37aca111a407cb98aa4e8dfde6e5d9333bae06dfa5d938d14c80bb5c3"},
- {file = "scipy-1.10.0-cp310-cp310-win_amd64.whl", hash = "sha256:27e548276b5a88b51212b61f6dda49a24acf5d770dff940bd372b3f7ced8c6c2"},
- {file = "scipy-1.10.0-cp311-cp311-macosx_10_15_x86_64.whl", hash = "sha256:42ab8b9e7dc1ebe248e55f54eea5307b6ab15011a7883367af48dd781d1312e4"},
- {file = "scipy-1.10.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:e096b062d2efdea57f972d232358cb068413dc54eec4f24158bcbb5cb8bddfd8"},
- {file = "scipy-1.10.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4df25a28bd22c990b22129d3c637fd5c3be4b7c94f975dca909d8bab3309b694"},
- {file = "scipy-1.10.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2ad449db4e0820e4b42baccefc98ec772ad7818dcbc9e28b85aa05a536b0f1a2"},
- {file = "scipy-1.10.0-cp311-cp311-win_amd64.whl", hash = "sha256:6faf86ef7717891195ae0537e48da7524d30bc3b828b30c9b115d04ea42f076f"},
- {file = "scipy-1.10.0-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:4bd0e3278126bc882d10414436e58fa3f1eca0aa88b534fcbf80ed47e854f46c"},
- {file = "scipy-1.10.0-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:38bfbd18dcc69eeb589811e77fae552fa923067fdfbb2e171c9eac749885f210"},
- {file = "scipy-1.10.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0ab2a58064836632e2cec31ca197d3695c86b066bc4818052b3f5381bfd2a728"},
- {file = "scipy-1.10.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd7a30970c29d9768a7164f564d1fbf2842bfc77b7d114a99bc32703ce0bf48"},
- {file = "scipy-1.10.0-cp38-cp38-win_amd64.whl", hash = "sha256:9b878c671655864af59c108c20e4da1e796154bd78c0ed6bb02bc41c84625686"},
- {file = "scipy-1.10.0-cp39-cp39-macosx_10_15_x86_64.whl", hash = "sha256:3afcbddb4488ac950ce1147e7580178b333a29cd43524c689b2e3543a080a2c8"},
- {file = "scipy-1.10.0-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:6e4497e5142f325a5423ff5fda2fff5b5d953da028637ff7c704378c8c284ea7"},
- {file = "scipy-1.10.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:441cab2166607c82e6d7a8683779cb89ba0f475b983c7e4ab88f3668e268c143"},
- {file = "scipy-1.10.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0490dc499fe23e4be35b8b6dd1e60a4a34f0c4adb30ac671e6332446b3cbbb5a"},
- {file = "scipy-1.10.0-cp39-cp39-win_amd64.whl", hash = "sha256:954ff69d2d1bf666b794c1d7216e0a746c9d9289096a64ab3355a17c7c59db54"},
- {file = "scipy-1.10.0.tar.gz", hash = "sha256:c8b3cbc636a87a89b770c6afc999baa6bcbb01691b5ccbbc1b1791c7c0a07540"},
+ {file = "scipy-1.10.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e7354fd7527a4b0377ce55f286805b34e8c54b91be865bac273f527e1b839019"},
+ {file = "scipy-1.10.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:4b3f429188c66603a1a5c549fb414e4d3bdc2a24792e061ffbd607d3d75fd84e"},
+ {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1553b5dcddd64ba9a0d95355e63fe6c3fc303a8fd77c7bc91e77d61363f7433f"},
+ {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c0ff64b06b10e35215abce517252b375e580a6125fd5fdf6421b98efbefb2d2"},
+ {file = "scipy-1.10.1-cp310-cp310-win_amd64.whl", hash = "sha256:fae8a7b898c42dffe3f7361c40d5952b6bf32d10c4569098d276b4c547905ee1"},
+ {file = "scipy-1.10.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:0f1564ea217e82c1bbe75ddf7285ba0709ecd503f048cb1236ae9995f64217bd"},
+ {file = "scipy-1.10.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d925fa1c81b772882aa55bcc10bf88324dadb66ff85d548c71515f6689c6dac5"},
+ {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaea0a6be54462ec027de54fca511540980d1e9eea68b2d5c1dbfe084797be35"},
+ {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15a35c4242ec5f292c3dd364a7c71a61be87a3d4ddcc693372813c0b73c9af1d"},
+ {file = "scipy-1.10.1-cp311-cp311-win_amd64.whl", hash = "sha256:43b8e0bcb877faf0abfb613d51026cd5cc78918e9530e375727bf0625c82788f"},
+ {file = "scipy-1.10.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5678f88c68ea866ed9ebe3a989091088553ba12c6090244fdae3e467b1139c35"},
+ {file = "scipy-1.10.1-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:39becb03541f9e58243f4197584286e339029e8908c46f7221abeea4b749fa88"},
+ {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bce5869c8d68cf383ce240e44c1d9ae7c06078a9396df68ce88a1230f93a30c1"},
+ {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:07c3457ce0b3ad5124f98a86533106b643dd811dd61b548e78cf4c8786652f6f"},
+ {file = "scipy-1.10.1-cp38-cp38-win_amd64.whl", hash = "sha256:049a8bbf0ad95277ffba9b3b7d23e5369cc39e66406d60422c8cfef40ccc8415"},
+ {file = "scipy-1.10.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cd9f1027ff30d90618914a64ca9b1a77a431159df0e2a195d8a9e8a04c78abf9"},
+ {file = "scipy-1.10.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:79c8e5a6c6ffaf3a2262ef1be1e108a035cf4f05c14df56057b64acc5bebffb6"},
+ {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:51af417a000d2dbe1ec6c372dfe688e041a7084da4fdd350aeb139bd3fb55353"},
+ {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b4735d6c28aad3cdcf52117e0e91d6b39acd4272f3f5cd9907c24ee931ad601"},
+ {file = "scipy-1.10.1-cp39-cp39-win_amd64.whl", hash = "sha256:7ff7f37b1bf4417baca958d254e8e2875d0cc23aaadbe65b3d5b3077b0eb23ea"},
+ {file = "scipy-1.10.1.tar.gz", hash = "sha256:2cf9dfb80a7b4589ba4c40ce7588986d6d5cebc5457cad2c2880f6bc2d42f3a5"},
@@ -4344 +4331 @@ name = "setuptools"
-version = "67.3.1"
+version = "67.4.0"
@@ -4350,2 +4337,2 @@ files = [
- {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
- {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
+ {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"},
+ {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"},
@@ -4397 +4384 @@ name = "soundfile"
-version = "0.12.0"
+version = "0.12.1"
@@ -4403,7 +4390,7 @@ files = [
- {file = "soundfile-0.12.0-py2.py3-none-any.whl", hash = "sha256:9e7dc313f4f1e835f3b3bf35fd7c2908bba31bd6080727f075799b7a03f8d426"},
- {file = "soundfile-0.12.0-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:ff687a6db726229adc93e7df93e6bd80ffb041a0c4d84fdf6659f14fa32854bd"},
- {file = "soundfile-0.12.0-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:29623eae59f22d8c1b18d8cab90d39de1c834270081e3e50bdcafc95d6dd7bc9"},
- {file = "soundfile-0.12.0-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:a90a27ce766dbc2fbdee91bf549d4e60ebbc2b94a2c2a7f12ca48b8bbe7b0123"},
- {file = "soundfile-0.12.0-py2.py3-none-win32.whl", hash = "sha256:63f320b284781a346f1244f1bc8505567f1340d0ceb8841b4e3b598c9eec9d0c"},
- {file = "soundfile-0.12.0-py2.py3-none-win_amd64.whl", hash = "sha256:5e34870f8e8aef60be350fd4c0aa7f294497b615e39e4bbbe5e87181af93aad5"},
- {file = "soundfile-0.12.0.tar.gz", hash = "sha256:e50c42733ff5396e49a6a689722fa362387b2c403273bcc195994bf4a8e2df3f"},
+ {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"},
+ {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"},
+ {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"},
+ {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"},
+ {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"},
+ {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"},
+ {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"},
@@ -4429,0 +4417,45 @@ files = [
+[[package]]
+name = "soxr"
+version = "0.3.3"
+description = "High quality, one-dimensional sample-rate conversion library"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "soxr-0.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:086fb3e3837216cdac4673530af641bc9e7a65ff34b012d81d6b594690a4c5c6"},
+ {file = "soxr-0.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:65431e8ab1b8cc86e8cab2fe0b0a284046814b4aad42ee6611d1ca1bc7e0bbf0"},
+ {file = "soxr-0.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b7402e59294d3a39696990c63d45f24cbf8cb011a167da21f37b820d65a6f20"},
+ {file = "soxr-0.3.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4637508a34b577ffb55207da7e8ec6e55a813b904fc56e87819a04cd16cf4bd4"},
+ {file = "soxr-0.3.3-cp310-cp310-win32.whl", hash = "sha256:45e5c5bbfc605be646b944aa044bc3315bd81cac1935c220c4814f0659ceb8fb"},
+ {file = "soxr-0.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:5b51eb310529bf19c57a02d3fbadf9d45896bb9981e94d013327ffb27a3b00b0"},
+ {file = "soxr-0.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:52cf5b58abc82f0263b9c2d40da63073a57b4a70b14e8d4549b1aa09cdfef3c9"},
+ {file = "soxr-0.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ccf76f0487d0db186f57a2a616feca0a11d37477776e64d260daf4c01013aa39"},
+ {file = "soxr-0.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2202eb8807a834621b61f13230afaa3643eafb822e37c9cf19a50b20c6f6310"},
+ {file = "soxr-0.3.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d7687274daf541125c508dd6509310927180a0de0aa0c268f5084d4dab70501b"},
+ {file = "soxr-0.3.3-cp311-cp311-win32.whl", hash = "sha256:56fe880b97b8466166256b443ec45a1ec97f90cc1cd3167bc60922501f415b2c"},
+ {file = "soxr-0.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:f6673054a609d862cc39cc90e1b1a05de6b7fd51fb18e45ebdd127d171656b02"},
+ {file = "soxr-0.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:65c353979cb57f38ee3a24f68ea766706570075db216e079c05dd7d38c1df4ca"},
+ {file = "soxr-0.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:da548494b2f5ae84d7c551165b442834db6c42807e959dbf742146cf9ee4d97f"},
+ {file = "soxr-0.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a472c2380f624a45ea3d7c9f4c6db6a2a0dbeb610eafe1a57ec731ccb29cdd76"},
+ {file = "soxr-0.3.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e0a4bbdc2246d05201ed1e080ddf3fcc1856ca604c9565fdb8f69477589f2c17"},
+ {file = "soxr-0.3.3-cp38-cp38-win32.whl", hash = "sha256:a0996e682f6d8b796702823c21aca53707b2a8667c1154f8a12dc1ca97bfba1c"},
+ {file = "soxr-0.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:a7ddd6db851530ca356d5db5f257a38a12e7484899c209c9bed10eceb8d8edeb"},
+ {file = "soxr-0.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cb5d66b3b3a493f237ea80add2176a097a118488926f3f2fae369fe5caafe976"},
+ {file = "soxr-0.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:54a9934b98b151562a08ff396d2c0ea65e7a40398319b622e31afc34a510a523"},
+ {file = "soxr-0.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c41c5f1a75397e0255d62dc03e8759e011128d24385c1a7de54edd3f4bca2e36"},
+ {file = "soxr-0.3.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2081e5534218d30d1fa88c1ef4a1b5399b439e2ee396cd210ef4fd5dbe3659b5"},
+ {file = "soxr-0.3.3-cp39-cp39-win32.whl", hash = "sha256:54a6d7b17cad791ed91041abe28a645d36869d0410fac20cb97337a08b21a904"},
+ {file = "soxr-0.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:9e179d8fe35cbf8584f025bee1516fdaa6a69056a1ca7ccc600022d0d3f8a6bc"},
+ {file = "soxr-0.3.3-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:61bb9cd20f8aa52a39f7677b8edf587e017a1d97147a13a39bc836d989347760"},
+ {file = "soxr-0.3.3-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d3b9e692be2adb959ef114ef632ffc78eb83f6f0a7fbc0be0c49e87fbb151d81"},
+ {file = "soxr-0.3.3-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:e0b2334101798fa79e75218e94da47e0e5c056690559df84d5f321c9bb3893cf"},
+ {file = "soxr-0.3.3.tar.gz", hash = "sha256:af874e9f14c4c261c22ce55972a1231c085a538c2d8fec78580b331370fdb551"},
+]
+
+[package.dependencies]
+numpy = "*"
+
+[package.extras]
+docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"]
+test = ["pytest"]
+
@@ -4697,2 +4728,0 @@ files = [
- {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"},
- {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"},
@@ -4704 +4733,0 @@ files = [
- {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"},
@@ -4713 +4741,0 @@ files = [
- {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"},
@@ -4780 +4807,0 @@ url = "https://download.pytorch.org/whl/cpu/torch-1.13.1%2Bcpu-cp39-cp39-linux_x
-
@@ -4798 +4824,0 @@ url = "https://download.pytorch.org/whl/cpu/torchaudio-0.13.1%2Bcpu-cp39-cp39-li
-
@@ -4928 +4954 @@ name = "types-psutil"
-version = "5.9.5.7"
+version = "5.9.5.8"
@@ -4934,2 +4960,2 @@ files = [
- {file = "types-psutil-5.9.5.7.tar.gz", hash = "sha256:b38e376ba0d53b9a08d0eb77c3ccf621f9e0e0f90f9bdfb2ecc473d62c6828a3"},
- {file = "types_psutil-5.9.5.7-py3-none-any.whl", hash = "sha256:de57e5610fa568354b7bc0df05b2392a05fd0c1bf7a804f32f988126a5767a00"},
+ {file = "types-psutil-5.9.5.8.tar.gz", hash = "sha256:26e2841996fdfc66542ef74eb27ed3fbd8199bfa84e9aac0c36e177b8a74341e"},
+ {file = "types_psutil-5.9.5.8-py3-none-any.whl", hash = "sha256:21dc3aa224554b81226785708443c8b1f19d8c52a350f1a75eb9f82dfb35fffd"},
@@ -4940 +4966 @@ name = "types-pytz"
-version = "2022.7.1.0"
+version = "2022.7.1.2"
@@ -4946,2 +4972,2 @@ files = [
- {file = "types-pytz-2022.7.1.0.tar.gz", hash = "sha256:918f9c3e7a950ba7e7d6f84b18a7cacabc8886cb7125fb1927ff1c752b4b59de"},
- {file = "types_pytz-2022.7.1.0-py3-none-any.whl", hash = "sha256:10ec7d009a02340f1cecd654ac03f0c29b6088a03b63d164401fc52df45936b2"},
+ {file = "types-pytz-2022.7.1.2.tar.gz", hash = "sha256:487d3e8e9f4071eec8081746d53fa982bbc05812e719dcbf2ebf3d55a1a4cd28"},
+ {file = "types_pytz-2022.7.1.2-py3-none-any.whl", hash = "sha256:40ca448a928d566f7d44ddfde0066e384f7ffbd4da2778e42a4570eaca572446"},
@@ -4952 +4978 @@ name = "types-requests"
-version = "2.28.11.13"
+version = "2.28.11.14"
@@ -4958,2 +4984,2 @@ files = [
- {file = "types-requests-2.28.11.13.tar.gz", hash = "sha256:3fd332842e8759ea5f7eb7789df8aa772ba155216ccf10ef4aa3b0e5b42e1b46"},
- {file = "types_requests-2.28.11.13-py3-none-any.whl", hash = "sha256:94896f6f8e9f3db11e422c6e3e4abbc5d7ccace853eac74b23bdd65eeee3cdee"},
+ {file = "types-requests-2.28.11.14.tar.gz", hash = "sha256:232792870b60adb07d23175451ab4e6190021b0c584edf052d92d9b993118f06"},
+ {file = "types_requests-2.28.11.14-py3-none-any.whl", hash = "sha256:f84613b0d4c5d0eeb7879dfa05e14a3702b9c1f7a4ee81dfe9b4321b13fe93a1"},
@@ -4967 +4993 @@ name = "types-urllib3"
-version = "1.26.25.6"
+version = "1.26.25.7"
@@ -4973,2 +4999,2 @@ files = [
- {file = "types-urllib3-1.26.25.6.tar.gz", hash = "sha256:35586727cbd7751acccf2c0f34a88baffc092f435ab62458f10776466590f2d5"},
- {file = "types_urllib3-1.26.25.6-py3-none-any.whl", hash = "sha256:a6c23c41bd03e542eaee5423a018f833077b51c4bf9ceb5aa544e12b812d5604"},
+ {file = "types-urllib3-1.26.25.7.tar.gz", hash = "sha256:df4d3e5472bf8830bd74eac12d56e659f88662ba040c7d106bf3a5bee26fff28"},
+ {file = "types_urllib3-1.26.25.7-py3-none-any.whl", hash = "sha256:28d2d7f5c31ff8ed4d9d2e396ce906c49d37523c3ec207d03d3b1695755a7199"},
@@ -5409 +5435 @@ name = "zipp"
-version = "3.13.0"
+version = "3.14.0"
@@ -5415,2 +5441,2 @@ files = [
- {file = "zipp-3.13.0-py3-none-any.whl", hash = "sha256:e8b2a36ea17df80ffe9e2c4fda3f693c3dad6df1697d3cd3af232db680950b0b"},
- {file = "zipp-3.13.0.tar.gz", hash = "sha256:23f70e964bc11a34cef175bc90ba2914e1e4545ea1e3e2f67c079671883f9cb6"},
+ {file = "zipp-3.14.0-py3-none-any.whl", hash = "sha256:188834565033387710d046e3fe96acfc9b5e86cbca7f39ff69cf21a4128198b7"},
+ {file = "zipp-3.14.0.tar.gz", hash = "sha256:9e5421e176ef5ab4c0ad896624e87a7b2f07aca746c9b2aa305952800cb8eecb"},
@@ -5425 +5451 @@ name = "zstandard"
-version = "0.19.0"
+version = "0.20.0"
@@ -5431,51 +5457,51 @@ files = [
- {file = "zstandard-0.19.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:a65e0119ad39e855427520f7829618f78eb2824aa05e63ff19b466080cd99210"},
- {file = "zstandard-0.19.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:4fa496d2d674c6e9cffc561639d17009d29adee84a27cf1e12d3c9be14aa8feb"},
- {file = "zstandard-0.19.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8f7c68de4f362c1b2f426395fe4e05028c56d0782b2ec3ae18a5416eaf775576"},
- {file = "zstandard-0.19.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d1a7a716bb04b1c3c4a707e38e2dee46ac544fff931e66d7ae944f3019fc55b8"},
- {file = "zstandard-0.19.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:72758c9f785831d9d744af282d54c3e0f9db34f7eae521c33798695464993da2"},
- {file = "zstandard-0.19.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:04c298d381a3b6274b0a8001f0da0ec7819d052ad9c3b0863fe8c7f154061f76"},
- {file = "zstandard-0.19.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:aef0889417eda2db000d791f9739f5cecb9ccdd45c98f82c6be531bdc67ff0f2"},
- {file = "zstandard-0.19.0-cp310-cp310-win32.whl", hash = "sha256:9d97c713433087ba5cee61a3e8edb54029753d45a4288ad61a176fa4718033ce"},
- {file = "zstandard-0.19.0-cp310-cp310-win_amd64.whl", hash = "sha256:81ab21d03e3b0351847a86a0b298b297fde1e152752614138021d6d16a476ea6"},
- {file = "zstandard-0.19.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:593f96718ad906e24d6534187fdade28b611f8ed06e27ba972ba48aecec45fc6"},
- {file = "zstandard-0.19.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5e21032efe673b887464667d09406bab6e16d96b09ad87e80859e3a20b6745b6"},
- {file = "zstandard-0.19.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:876567136b0359f6581ecd892bdb4ca03a0eead0265db73206c78cff03bcdb0f"},
- {file = "zstandard-0.19.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aa9087571729c968cd853d54b3f6e9d0ec61e45cd2c31e0eb8a0d4bdbbe6da2f"},
- {file = "zstandard-0.19.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8371217dff635cfc0220db2720fc3ce728cd47e72bb7572cca035332823dbdfc"},
- {file = "zstandard-0.19.0-cp311-cp311-win32.whl", hash = "sha256:126aa8433773efad0871f624339c7984a9c43913952f77d5abeee7f95a0c0860"},
- {file = "zstandard-0.19.0-cp311-cp311-win_amd64.whl", hash = "sha256:0fde1c56ec118940974e726c2a27e5b54e71e16c6f81d0b4722112b91d2d9009"},
- {file = "zstandard-0.19.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:898500957ae5e7f31b7271ace4e6f3625b38c0ac84e8cedde8de3a77a7fdae5e"},
- {file = "zstandard-0.19.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:660b91eca10ee1b44c47843894abe3e6cfd80e50c90dee3123befbf7ca486bd3"},
- {file = "zstandard-0.19.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:55b3187e0bed004533149882ef8c24e954321f3be81f8a9ceffe35099b82a0d0"},
- {file = "zstandard-0.19.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:6d2182e648e79213b3881998b30225b3f4b1f3e681f1c1eaf4cacf19bde1040d"},
- {file = "zstandard-0.19.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8ec2c146e10b59c376b6bc0369929647fcd95404a503a7aa0990f21c16462248"},
- {file = "zstandard-0.19.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:67710d220af405f5ce22712fa741d85e8b3ada7a457ea419b038469ba379837c"},
- {file = "zstandard-0.19.0-cp36-cp36m-win32.whl", hash = "sha256:f097dda5d4f9b9b01b3c9fa2069f9c02929365f48f341feddf3d6b32510a2f93"},
- {file = "zstandard-0.19.0-cp36-cp36m-win_amd64.whl", hash = "sha256:f4ebfe03cbae821ef994b2e58e4df6a087470cc522aca502614e82a143365d45"},
- {file = "zstandard-0.19.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:b80f6f6478f9d4ca26daee6c61584499493bf97950cfaa1a02b16bb5c2c17e70"},
- {file = "zstandard-0.19.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:909bdd4e19ea437eb9b45d6695d722f6f0fd9d8f493e837d70f92062b9f39faf"},
- {file = "zstandard-0.19.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e9c90a44470f2999779057aeaf33461cbd8bb59d8f15e983150d10bb260e16e0"},
- {file = "zstandard-0.19.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:401508efe02341ae681752a87e8ac9ef76df85ef1a238a7a21786a489d2c983d"},
- {file = "zstandard-0.19.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:47dfa52bed3097c705451bafd56dac26535545a987b6759fa39da1602349d7ba"},
- {file = "zstandard-0.19.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:1a4fb8b4ac6772e4d656103ccaf2e43e45bd16b5da324b963d58ef360d09eb73"},
- {file = "zstandard-0.19.0-cp37-cp37m-win32.whl", hash = "sha256:d63b04e16df8ea21dfcedbf5a60e11cbba9d835d44cb3cbff233cfd037a916d5"},
- {file = "zstandard-0.19.0-cp37-cp37m-win_amd64.whl", hash = "sha256:74c2637d12eaacb503b0b06efdf55199a11b1d7c580bd3dd9dfe84cac97ef2f6"},
- {file = "zstandard-0.19.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:2e4812720582d0803e84aefa2ac48ce1e1e6e200ca3ce1ae2be6d410c1d637ae"},
- {file = "zstandard-0.19.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4514b19abe6dbd36d6c5d75c54faca24b1ceb3999193c5b1f4b685abeabde3d0"},
- {file = "zstandard-0.19.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6caed86cd47ae93915d9031dc04be5283c275e1a2af2ceff33932071f3eeff4d"},
- {file = "zstandard-0.19.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ccc4727300f223184520a6064c161a90b5d0283accd72d1455bcd85ec44dd0d"},
- {file = "zstandard-0.19.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:879411d04068bd489db57dcf6b82ffad3c5fb2a1fdd30817c566d8b7bedee442"},
- {file = "zstandard-0.19.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8c9ca56345b0c5574db47560603de9d05f63cce5dfeb3a456eb60f3fec737ff2"},
- {file = "zstandard-0.19.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:d777d239036815e9b3a093fa9208ad314c040c26d7246617e70e23025b60083a"},
- {file = "zstandard-0.19.0-cp38-cp38-win32.whl", hash = "sha256:be6329b5ba18ec5d32dc26181e0148e423347ed936dda48bf49fb243895d1566"},
- {file = "zstandard-0.19.0-cp38-cp38-win_amd64.whl", hash = "sha256:3d5bb598963ac1f1f5b72dd006adb46ca6203e4fb7269a5b6e1f99e85b07ad38"},
- {file = "zstandard-0.19.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:619f9bf37cdb4c3dc9d4120d2a1003f5db9446f3618a323219f408f6a9df6725"},
- {file = "zstandard-0.19.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:b253d0c53c8ee12c3e53d181fb9ef6ce2cd9c41cbca1c56a535e4fc8ec41e241"},
- {file = "zstandard-0.19.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3c927b6aa682c6d96225e1c797f4a5d0b9f777b327dea912b23471aaf5385376"},
- {file = "zstandard-0.19.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2f01b27d0b453f07cbcff01405cdd007e71f5d6410eb01303a16ba19213e58e4"},
- {file = "zstandard-0.19.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:c7560f622e3849cc8f3e999791a915addd08fafe80b47fcf3ffbda5b5151047c"},
- {file = "zstandard-0.19.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e892d3177380ec080550b56a7ffeab680af25575d291766bdd875147ba246a91"},
- {file = "zstandard-0.19.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:60a86b7b2b1c300779167cf595e019e61afcc0e20c4838692983a921db9006ac"},
- {file = "zstandard-0.19.0-cp39-cp39-win32.whl", hash = "sha256:755020d5aeb1b10bffd93d119e7709a2a7475b6ad79c8d5226cea3f76d152ce0"},
- {file = "zstandard-0.19.0-cp39-cp39-win_amd64.whl", hash = "sha256:55a513ec67e85abd8b8b83af8813368036f03e2d29a50fc94033504918273980"},
- {file = "zstandard-0.19.0.tar.gz", hash = "sha256:31d12fcd942dd8dbf52ca5f6b1bbe287f44e5d551a081a983ff3ea2082867863"},
+ {file = "zstandard-0.20.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c4efa051799703dc37c072e22af1f0e4c77069a78fb37caf70e26414c738ca1d"},
+ {file = "zstandard-0.20.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f847701d77371d90783c0ce6cfdb7ebde4053882c2aaba7255c70ae3c3eb7af0"},
+ {file = "zstandard-0.20.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0aa4d178560d7ee32092ddfd415c2cdc6ab5ddce9554985c75f1a019a0ff4c55"},
+ {file = "zstandard-0.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0488f2a238b4560828b3a595f3337daac4d3725c2a1637ffe2a0d187c091da59"},
+ {file = "zstandard-0.20.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:cd0aa9a043c38901925ae1bba49e1e638f2d9c3cdf1b8000868993c642deb7f2"},
+ {file = "zstandard-0.20.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cdd769da7add8498658d881ce0eeb4c35ea1baac62e24c5a030c50f859f29724"},
+ {file = "zstandard-0.20.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:9aea3c7bab4276212e5ac63d28e6bd72a79ff058d57e06926dfe30a52451d943"},
+ {file = "zstandard-0.20.0-cp310-cp310-win32.whl", hash = "sha256:0d213353d58ad37fb5070314b156fb983b4d680ed5f3fce76ab013484cf3cf12"},
+ {file = "zstandard-0.20.0-cp310-cp310-win_amd64.whl", hash = "sha256:d08459f7f7748398a6cc65eb7f88aa7ef5731097be2ddfba544be4b558acd900"},
+ {file = "zstandard-0.20.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:c1929afea64da48ec59eca9055d7ec7e5955801489ac40ac2a19dde19e7edad9"},
+ {file = "zstandard-0.20.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:b6d718f1b7cd30adb02c2a46dde0f25a84a9de8865126e0fff7d0162332d6b92"},
+ {file = "zstandard-0.20.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5499d65d4a1978dccf0a9c2c0d12415e16d4995ffad7a0bc4f72cc66691cf9f2"},
+ {file = "zstandard-0.20.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:302a31400de0280f17c4ce67a73444a7a069f228db64048e4ce555cd0c02fbc4"},
+ {file = "zstandard-0.20.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:39ae788dcdc404c07ef7aac9b11925185ea0831b985db0bbc43f95acdbd1c2ce"},
+ {file = "zstandard-0.20.0-cp311-cp311-win32.whl", hash = "sha256:e3f6887d2bdfb5752d5544860bd6b778e53ebfaf4ab6c3f9d7fd388445429d41"},
+ {file = "zstandard-0.20.0-cp311-cp311-win_amd64.whl", hash = "sha256:4abf9a9e0841b844736d1ae8ead2b583d2cd212815eab15391b702bde17477a7"},
+ {file = "zstandard-0.20.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:dc47cec184e66953f635254e5381df8a22012a2308168c069230b1a95079ccd0"},
+ {file = "zstandard-0.20.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84c1dae0c0a21eea245b5691286fe6470dc797d5e86e0c26b57a3afd1e750b48"},
+ {file = "zstandard-0.20.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:059316f07e39b7214cd9eed565d26ab239035d2c76835deeff381995f7a27ba8"},
+ {file = "zstandard-0.20.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:9aca916724d0802d3e70dc68adeff893efece01dffe7252ee3ae0053f1f1990f"},
+ {file = "zstandard-0.20.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b07f391fd85e3d07514c05fb40c5573b398d0063ab2bada6eb09949ec6004772"},
+ {file = "zstandard-0.20.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:2adf65cfce73ce94ef4c482f6cc01f08ddf5e1ca0c1ec95f2b63840f9e4c226c"},
+ {file = "zstandard-0.20.0-cp36-cp36m-win32.whl", hash = "sha256:ee2a1510e06dfc7706ea9afad363efe222818a1eafa59abc32d9bbcd8465fba7"},
+ {file = "zstandard-0.20.0-cp36-cp36m-win_amd64.whl", hash = "sha256:29699746fae2760d3963a4ffb603968e77da55150ee0a3326c0569f4e35f319f"},
+ {file = "zstandard-0.20.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:78fb35d07423f25efd0fc90d0d4710ae83cfc86443a32192b0c6cb8475ec79a5"},
+ {file = "zstandard-0.20.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:40466adfa071f58bfa448d90f9623d6aff67c6d86de6fc60be47a26388f6c74d"},
+ {file = "zstandard-0.20.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ba86f931bf925e9561ccd6cb978acb163e38c425990927feb38be10c894fa937"},
+ {file = "zstandard-0.20.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:b671b75ae88139b1dd022fa4aa66ba419abd66f98869af55a342cb9257a1831e"},
+ {file = "zstandard-0.20.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cc98c8bcaa07150d3f5d7c4bd264eaa4fdd4a4dfb8fd3f9d62565ae5c4aba227"},
+ {file = "zstandard-0.20.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:0b815dec62e2d5a1bf7a373388f2616f21a27047b9b999de328bca7462033708"},
+ {file = "zstandard-0.20.0-cp37-cp37m-win32.whl", hash = "sha256:5a3578b182c21b8af3c49619eb4cd0b9127fa60791e621b34217d65209722002"},
+ {file = "zstandard-0.20.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f1ba6bbd28ad926d130f0af8016f3a2930baa013c2128cfff46ca76432f50669"},
+ {file = "zstandard-0.20.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b0f556c74c6f0f481b61d917e48c341cdfbb80cc3391511345aed4ce6fb52fdc"},
+ {file = "zstandard-0.20.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:862ad0a5c94670f2bd6f64fff671bd2045af5f4ed428a3f2f69fa5e52483f86a"},
+ {file = "zstandard-0.20.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a56036c08645aa6041d435a50103428f0682effdc67f5038de47cea5e4221d6f"},
+ {file = "zstandard-0.20.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4af5d1891eebef430038ea4981957d31b1eb70aca14b906660c3ac1c3e7a8612"},
+ {file = "zstandard-0.20.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:489959e2d52f7f1fe8ea275fecde6911d454df465265bf3ec51b3e755e769a5e"},
+ {file = "zstandard-0.20.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7041efe3a93d0975d2ad16451720932e8a3d164be8521bfd0873b27ac917b77a"},
+ {file = "zstandard-0.20.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:c28c7441638c472bfb794f424bd560a22c7afce764cd99196e8d70fbc4d14e85"},
+ {file = "zstandard-0.20.0-cp38-cp38-win32.whl", hash = "sha256:ba4bb4c5a0cac802ff485fa1e57f7763df5efa0ad4ee10c2693ecc5a018d2c1a"},
+ {file = "zstandard-0.20.0-cp38-cp38-win_amd64.whl", hash = "sha256:a5efe366bf0545a1a5a917787659b445ba16442ae4093f102204f42a9da1ecbc"},
+ {file = "zstandard-0.20.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:79c3058ccbe1fa37356a73c9d3c0475ec935ab528f5b76d56fc002a5a23407c7"},
+ {file = "zstandard-0.20.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:39cbaf8fe3fa3515d35fb790465db4dc1ff45e58e1e00cbaf8b714e85437f039"},
+ {file = "zstandard-0.20.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f199d58f3fd7dfa0d447bc255ff22571f2e4e5e5748bfd1c41370454723cb053"},
+ {file = "zstandard-0.20.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0f32a8f3a697ef87e67c0d0c0673b245babee6682b2c95e46eb30208ffb720bd"},
+ {file = "zstandard-0.20.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:4a3c36284c219a4d2694e52b2582fe5d5f0ecaf94a22cf0ea959b527dbd8a2a6"},
+ {file = "zstandard-0.20.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2eeb9e1ecd48ac1d352608bfe0dc1ed78a397698035a1796cf72f0c9d905d219"},
+ {file = "zstandard-0.20.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:6179808ebd1ebc42b1e2f221a23c28a22d3bc8f79209ae4a3cc114693c380bff"},
+ {file = "zstandard-0.20.0-cp39-cp39-win32.whl", hash = "sha256:afbcd2ed0c1145e24dd3df8440a429688a1614b83424bc871371b176bed429f9"},
+ {file = "zstandard-0.20.0-cp39-cp39-win_amd64.whl", hash = "sha256:e6b4de1ba2f3028fafa0d82222d1e91b729334c8d65fbf04290c65c09d7457e1"},
+ {file = "zstandard-0.20.0.tar.gz", hash = "sha256:613daadd72c71b1488742cafb2c3b381c39d0c9bb8c6cc157aa2d5ea45cc2efc"},
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index 76791302..4edc2342 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -85,0 +86,5 @@ module = "datasets"
+
+[[tool.mypy.overrides]]
+ignore_missing_imports = true
+module = "mirakuru.*"
+# ^ until https://github.com/ClearcodeHQ/mirakuru/pull/633 is merged and released
|
|
af839dd68c70197123949a11e47bc72ac73edeb3
|
Albert Villanova del Moral
| 2023-02-24T13:22:28 |
Improve error messages (#754)
|
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json
index d27575e6..356e24a8 100644
--- a/chart/static-files/openapi.json
+++ b/chart/static-files/openapi.json
@@ -43 +43 @@
- "summary": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication.",
+ "summary": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication.",
@@ -61 +61 @@
- "summary": "The dataset does not exist, or is not accessible with the current credentials (private or gated).",
+ "summary": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials.",
@@ -101 +101 @@
- "summary": "The list of splits is not ready yet. Please retry later.",
+ "summary": "The server is busier than usual and the list of splits is not ready yet. Please retry later.",
@@ -123 +123 @@
- "summary": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication.",
+ "summary": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication.",
@@ -143 +143 @@
- "summary": "The dataset does not exist, or is not accessible with the current credentials (private or gated).",
+ "summary": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials.",
@@ -199 +199 @@
- "summary": "The info cannot be fetched for the dataset config.",
+ "summary": "The info cannot be fetched for the config of the dataset.",
@@ -203 +203 @@
- "summary": "The split features (columns) cannot be extracted.",
+ "summary": "Cannot extract the features (columns) for the split of the config of the dataset.",
@@ -247 +247 @@
- "summary": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication.",
+ "summary": "Cannot access the route. Please retry with authentication.",
@@ -261 +261 @@
- "summary": "The dataset does not exist, or is not accessible with the current credentials (private or gated).",
+ "summary": "Cannot access the route with the current credentials. Please retry with other authentication credentials.",
@@ -1021 +1021 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -1027 +1027 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -1033 +1033 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -1062 +1062 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -1068 +1068 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -1074 +1074 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -1171 +1171 @@
- "error": "The list of splits is not ready yet. Please retry later."
+ "error": "The server is busier than usual and the list of splits is not ready yet. Please retry later."
@@ -1784 +1784 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -1790 +1790 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -1796 +1796 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -1825 +1825 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -1831 +1831 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -1837 +1837 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -2279 +2279 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -2285 +2285 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -2291 +2291 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -2320 +2320 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -2326 +2326 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -2332 +2332 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -2392 +2392 @@
- "error": "The response is not ready yet. Please retry later."
+ "error": "The server is busier than usual and the response is not ready yet. Please retry later."
@@ -2598 +2598 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -2604 +2604 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -2610 +2610 @@
- "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry with authentication."
+ "error": "The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication."
@@ -2639 +2639 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -2645 +2645 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
@@ -2651 +2651 @@
- "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "error": "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please check the spelling of the dataset name or retry with other authentication credentials."
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py
index 7ef405e1..d03d2e39 100644
--- a/libs/libcommon/src/libcommon/dataset.py
+++ b/libs/libcommon/src/libcommon/dataset.py
@@ -111 +111,3 @@ def ask_access(dataset: str, hf_endpoint: str, hf_token: Optional[str]) -> None:
- raise DatasetNotFoundError("The dataset does not exist on the Hub, or is private.") from err
+ raise DatasetNotFoundError(
+ "The dataset does not exist on the Hub, or is private. Private datasets are not yet supported."
+ ) from err
@@ -149 +151 @@ def get_dataset_info_for_supported_datasets(
- raise DatasetNotFoundError("The dataset does not exist on the Hub, or is private.")
+ raise DatasetNotFoundError("The dataset is private and private datasets are not yet supported.")
diff --git a/services/admin/src/admin/authentication.py b/services/admin/src/admin/authentication.py
index 6218b52c..68f5ba50 100644
--- a/services/admin/src/admin/authentication.py
+++ b/services/admin/src/admin/authentication.py
@@ -60 +60,4 @@ def auth_check(
- raise ExternalAuthenticatedError("Cannot access the route with the current credentials.") from err
+ raise ExternalAuthenticatedError(
+ "Cannot access the route with the current credentials. Please retry with other authentication"
+ " credentials."
+ ) from err
@@ -64 +67,3 @@ def auth_check(
- raise ExternalAuthenticatedError("Cannot access the route with the current credentials.")
+ raise ExternalAuthenticatedError(
+ "Cannot access the route with the current credentials. Please retry with other authentication credentials."
+ )
diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py
index 7e0f4ed2..3c4d81b8 100644
--- a/services/api/src/api/authentication.py
+++ b/services/api/src/api/authentication.py
@@ -65,2 +65,2 @@ def auth_check(
- "The dataset does not exist, or is not accessible without authentication (private or gated). Please retry"
- " with authentication."
+ "The dataset does not exist, or is not accessible without authentication (private or gated). Please check"
+ " the spelling of the dataset name or retry with authentication."
@@ -70 +70,2 @@ def auth_check(
- "The dataset does not exist, or is not accessible with the current credentials (private or gated)."
+ "The dataset does not exist, or is not accessible with the current credentials (private or gated). Please"
+ " check the spelling of the dataset name or retry with other authentication credentials."
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 8b672ea3..9936dc62 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -105 +105,3 @@ def create_endpoint(
- raise ResponseNotReadyError("The response is not ready yet. Please retry later.") from e
+ raise ResponseNotReadyError(
+ "The server is busier than usual and the response is not ready yet. Please retry later."
+ ) from e
diff --git a/services/worker/src/worker/job_runners/dataset_info.py b/services/worker/src/worker/job_runners/dataset_info.py
index b0bf602f..7093ca57 100644
--- a/services/worker/src/worker/job_runners/dataset_info.py
+++ b/services/worker/src/worker/job_runners/dataset_info.py
@@ -65 +65 @@ def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse:
- If the the previous step gave an error.
+ If the previous step gave an error.
@@ -75 +75,3 @@ def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse:
- raise DatasetNotFoundError("No response found in previous step for this dataset.", e) from e
+ raise DatasetNotFoundError(
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info' endpoint.", e
+ ) from e
@@ -82 +84 @@ def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse:
- raise PreviousStepFormatError("Previous step did not return the expected content.")
+ raise PreviousStepFormatError("Previous step did not return the expected content: 'dataset_info'.")
diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py
index 221da844..020d0feb 100644
--- a/services/worker/src/worker/job_runners/first_rows.py
+++ b/services/worker/src/worker/job_runners/first_rows.py
@@ -490 +490 @@ def compute_first_rows_response(
- raise ConfigNotFoundError(f"config {config} does not exist for dataset {dataset}")
+ raise ConfigNotFoundError(f"The config '{config}' does not exist for the dataset.'")
@@ -499 +499 @@ def compute_first_rows_response(
- raise SplitNotFoundError("The config or the split does not exist in the dataset")
+ raise SplitNotFoundError(f"The split '{split}' does not exist for the config '{config}' of the dataset.")
@@ -508 +508,4 @@ def compute_first_rows_response(
- raise InfoError("The info cannot be fetched for the dataset config.", cause=err) from err
+ raise InfoError(
+ f"The info cannot be fetched for the config '{config}' of the dataset.",
+ cause=err,
+ ) from err
@@ -520 +523 @@ def compute_first_rows_response(
- raise TypeError("load_dataset should return an IterableDataset")
+ raise TypeError("load_dataset should return an IterableDataset.")
@@ -523 +526 @@ def compute_first_rows_response(
- raise TypeError("load_dataset should return an IterableDataset")
+ raise TypeError("load_dataset should return an IterableDataset.")
@@ -526 +529,7 @@ def compute_first_rows_response(
- raise FeaturesError("The split features (columns) cannot be extracted.", cause=err) from err
+ raise FeaturesError(
+ (
+ f"Cannot extract the features (columns) for the split '{split}' of the config '{config}' of the"
+ " dataset."
+ ),
+ cause=err,
+ ) from err
@@ -532 +541,3 @@ def compute_first_rows_response(
- f"Too many columns. The maximum supported number of columns is {columns_max_number}."
+ f"The number of columns ({len(features)}) exceeds the maximum supported number of columns"
+ f" ({columns_max_number}). This is a current limitation of the datasets viewer. You can reduce the number"
+ " of columns if you want the viewer to work."
@@ -547 +558,4 @@ def compute_first_rows_response(
- raise TooBigContentError("The first rows content after truncation exceeds the maximum size.")
+ raise TooBigContentError(
+ f"The size of the content of the first rows ({surrounding_json_size} B) exceeds the maximum"
+ f" supported size ({rows_max_bytes} B) even after truncation. Please report the issue."
+ )
@@ -669 +683 @@ class FirstRowsJobRunner(DatasetsBasedJobRunner):
- """Get the set of new splits, from the content created by the compute."""
+ """Get the set of new splits, from the content created by compute."""
diff --git a/services/worker/src/worker/job_runners/parquet.py b/services/worker/src/worker/job_runners/parquet.py
index a29ec018..e5807e2a 100644
--- a/services/worker/src/worker/job_runners/parquet.py
+++ b/services/worker/src/worker/job_runners/parquet.py
@@ -66 +66 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- If the the previous step gave an error.
+ If the previous step gave an error.
@@ -78 +78,3 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- raise DatasetNotFoundError("No response found in previous step for this dataset.", e) from e
+ raise DatasetNotFoundError(
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info' endpoint.", e
+ ) from e
@@ -85 +87 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- raise PreviousStepFormatError("Previous step did not return the expected content.")
+ raise PreviousStepFormatError("Previous step did not return the expected content: 'parquet_files'.")
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index bd3f03e1..8f9b688d 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -335,2 +335,2 @@ def raise_if_too_big_from_datasets(
- "datasets.config.HF_ENDPOINT should have already been set to {hf_endpoint}. "
- f"Current value: {datasets.config.HF_ENDPOINT}. "
+ f"Invalid datasets.config.HF_ENDPOINT value: '{datasets.config.HF_ENDPOINT}'. Please set it to:"
+ f" '{hf_endpoint}'."
@@ -344,2 +344,3 @@ def raise_if_too_big_from_datasets(
- f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes. "
- f"Current size as given per the datasets library is {dataset_size} bytes."
+ f"The dataset is too big to be converted to Parquet. The size of the dataset ({dataset_size} B, as given"
+ f" per the datasets library) exceeds the maximum supported size ({max_dataset_size} B). Please report the"
+ " issue."
@@ -771 +772 @@ def compute_parquet_and_dataset_info_response(
- raise ConfigNamesError("Cannot get the configuration names for the dataset.", cause=err) from err
+ raise ConfigNamesError("Cannot get the config names for the dataset.", cause=err) from err
diff --git a/services/worker/src/worker/job_runners/sizes.py b/services/worker/src/worker/job_runners/sizes.py
index fa48abf7..a18ec238 100644
--- a/services/worker/src/worker/job_runners/sizes.py
+++ b/services/worker/src/worker/job_runners/sizes.py
@@ -99 +99 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- If the the previous step gave an error.
+ If the previous step gave an error.
@@ -109 +109,3 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- raise DatasetNotFoundError("No response found in previous step for this dataset.", e) from e
+ raise DatasetNotFoundError(
+ "No response found in previous step for this dataset: '/parquet-and-dataset-info' endpoint.", e
+ ) from e
diff --git a/services/worker/src/worker/job_runners/split_names_from_streaming.py b/services/worker/src/worker/job_runners/split_names_from_streaming.py
index b7cfdc84..309c2422 100644
--- a/services/worker/src/worker/job_runners/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split_names_from_streaming.py
@@ -109 +109,2 @@ def compute_split_names_from_streaming_response(
- "Cannot get the split names for the dataset and config.", cause=err
+ f"Cannot get the split names for the config '{config}' of the dataset.",
+ cause=err,
|
|
e5d7abed2d7568723c2eb7a8310bc13a5f4c92da
|
Adrien
| 2023-02-24T10:21:20 |
Add pdb to avoid disruption when we update kubernetes (#843)
|
diff --git a/chart/templates/reverse-proxy/pdb.yaml b/chart/templates/reverse-proxy/pdb.yaml
new file mode 100644
index 00000000..84a5aa30
--- /dev/null
+++ b/chart/templates/reverse-proxy/pdb.yaml
@@ -0,0 +1,10 @@
+apiVersion: policy/v1
+kind: PodDisruptionBudget
+metadata:
+ labels: {{ include "labels.reverseProxy" . | nindent 4 }}
+ name: "{{ include "name" . }}-reverse-proxy"
+ namespace: {{ .Release.Namespace }}
+spec:
+ maxUnavailable: 1
+ selector:
+ matchLabels: {{ include "labels.reverseProxy" . | nindent 6 }}
diff --git a/chart/templates/services/admin/pdb.yaml b/chart/templates/services/admin/pdb.yaml
new file mode 100644
index 00000000..cb4a6861
--- /dev/null
+++ b/chart/templates/services/admin/pdb.yaml
@@ -0,0 +1,10 @@
+apiVersion: policy/v1
+kind: PodDisruptionBudget
+metadata:
+ labels: {{ include "labels.admin" . | nindent 4 }}
+ name: "{{ include "name" . }}-admin"
+ namespace: {{ .Release.Namespace }}
+spec:
+ maxUnavailable: 1
+ selector:
+ matchLabels: {{ include "labels.admin" . | nindent 6 }}
diff --git a/chart/templates/services/api/pdb.yaml b/chart/templates/services/api/pdb.yaml
new file mode 100644
index 00000000..c7f47693
--- /dev/null
+++ b/chart/templates/services/api/pdb.yaml
@@ -0,0 +1,10 @@
+apiVersion: policy/v1
+kind: PodDisruptionBudget
+metadata:
+ labels: {{ include "labels.api" . | nindent 4 }}
+ name: "{{ include "name" . }}-api"
+ namespace: {{ .Release.Namespace }}
+spec:
+ maxUnavailable: 1
+ selector:
+ matchLabels: {{ include "labels.api" . | nindent 6 }}
|
|
ce7ebdfcecbe2980c9d04d72847ae0a1735d71a4
|
Polina Kazakova
| 2023-02-22T18:49:24 |
Delete extracted downloaded files of a dataset (#794)
|
diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py
index bc498e71..221da844 100644
--- a/services/worker/src/worker/job_runners/first_rows.py
+++ b/services/worker/src/worker/job_runners/first_rows.py
@@ -25,0 +26 @@ from datasets import (
+ DownloadConfig,
@@ -196,0 +198 @@ def get_rows(
+ download_config = DownloadConfig(delete_extracted=True)
@@ -202,0 +205 @@ def get_rows(
+ download_config=download_config,
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index 0d4d30a9..bd3f03e1 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -12 +12 @@ from pathlib import Path
-from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict
+from typing import Any, Dict, List, Literal, Mapping, Optional, Tuple, TypedDict
@@ -137 +137 @@ class ParquetAndDatasetInfoResponse(TypedDict):
- dataset_info: dict[str, Any]
+ dataset_info: Dict[str, Any]
@@ -775 +775,2 @@ def compute_parquet_and_dataset_info_response(
- dataset_info: dict[str, Any] = {}
+ dataset_info: Dict[str, Any] = {}
+ download_config = DownloadConfig(delete_extracted=True)
@@ -777 +778,7 @@ def compute_parquet_and_dataset_info_response(
- builder = load_dataset_builder(path=dataset, name=config, revision=source_revision, use_auth_token=hf_token)
+ builder = load_dataset_builder(
+ path=dataset,
+ name=config,
+ revision=source_revision,
+ use_auth_token=hf_token,
+ download_config=download_config,
+ )
|
|
176a81fdab7ddf58d6e79d65ab0abaef279c4010
|
Quentin Lhoest
| 2023-02-22T13:59:49 |
Set error response when zombie job is killed (#827)
|
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index 623ef954..7996438f 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -7,0 +8 @@ from datetime import timedelta
+from http import HTTPStatus
@@ -11,0 +13 @@ from filelock import FileLock
+from libcommon.exceptions import CustomError
@@ -13,2 +15,5 @@ from libcommon.log import init_logging
-from libcommon.queue import Job, Status, get_datetime
-from libcommon.resources import QueueMongoResource
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.queue import Job, JobInfo, Status, get_datetime
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import upsert_response
+from libcommon.storage import init_assets_dir
@@ -18,0 +24 @@ from worker.config import AppConfig
+from worker.job_runner_factory import JobRunnerFactory
@@ -19,0 +26 @@ from worker.loop import WorkerState
+from worker.resources import LibrariesResource
@@ -34,0 +42,5 @@ async def every(
+class WorkerCrashedError(CustomError):
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "WorkerCrashedError", cause, False)
+
+
@@ -36 +48 @@ class WorkerExecutor:
- def __init__(self, app_config: AppConfig) -> None:
+ def __init__(self, app_config: AppConfig, job_runner_factory: JobRunnerFactory) -> None:
@@ -37,0 +50 @@ class WorkerExecutor:
+ self.job_runner_factory = job_runner_factory
@@ -125,0 +139,29 @@ class WorkerExecutor:
+ for zombie in zombies:
+ job_info = JobInfo(
+ job_id=str(zombie.pk),
+ type=zombie.type,
+ dataset=zombie.dataset,
+ config=zombie.config,
+ split=zombie.split,
+ force=zombie.force,
+ priority=zombie.priority,
+ )
+ job_runner = self.job_runner_factory.create_job_runner(job_info)
+ error = WorkerCrashedError("Worker crashed while running this job.")
+ upsert_response(
+ kind=job_runner.processing_step.cache_kind,
+ dataset=job_runner.dataset,
+ config=job_runner.config,
+ split=job_runner.split,
+ content=dict(error.as_response()),
+ http_status=error.status_code,
+ error_code=error.code,
+ details=dict(error.as_response_with_cause()),
+ worker_version=job_runner.get_version(),
+ dataset_git_revision=job_runner.get_dataset_git_revision(),
+ )
+ logging.debug(
+ "response for"
+ f" dataset={job_runner.dataset} config={job_runner.config} split={job_runner.split} had an error,"
+ " cache updated"
+ )
@@ -143,3 +185,21 @@ if __name__ == "__main__":
- with QueueMongoResource(
- database=app_config.queue.mongo_database, host=app_config.queue.mongo_url
- ) as queue_resource:
+ init_logging(log_level=app_config.common.log_level)
+ # ^ set first to have logs as soon as possible
+ assets_directory = init_assets_dir(directory=app_config.assets.storage_directory)
+
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+
+ with (
+ LibrariesResource(
+ hf_endpoint=app_config.common.hf_endpoint,
+ init_hf_datasets_cache=app_config.datasets_based.hf_datasets_cache,
+ numba_path=app_config.numba.path,
+ ) as libraries_resource,
+ CacheMongoResource(
+ database=app_config.cache.mongo_database, host=app_config.cache.mongo_url
+ ) as cache_resource,
+ QueueMongoResource(
+ database=app_config.queue.mongo_database, host=app_config.queue.mongo_url
+ ) as queue_resource,
+ ):
+ if not cache_resource.is_available():
+ raise RuntimeError("The connection to the cache database could not be established. Exiting.")
@@ -148 +208,8 @@ if __name__ == "__main__":
- worker_executor = WorkerExecutor(app_config)
+
+ job_runner_factory = JobRunnerFactory(
+ app_config=app_config,
+ processing_graph=processing_graph,
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ assets_directory=assets_directory,
+ )
+ worker_executor = WorkerExecutor(app_config, job_runner_factory)
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 89cc1a79..33039aec 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -9 +9,11 @@ from pathlib import Path
-from typing import Any, Iterator, List, Literal, Mapping, Optional, Tuple, TypedDict
+from typing import (
+ Any,
+ Callable,
+ Iterator,
+ List,
+ Literal,
+ Mapping,
+ Optional,
+ Tuple,
+ TypedDict,
+)
@@ -133,0 +144,15 @@ def delete_hub_dataset_repo(repo_id: str) -> None:
[email protected]
+def tmp_dataset_repo_factory() -> Iterator[Callable[[str], str]]:
+ repo_ids: List[str] = []
+
+ def _tmp_dataset_repo(repo_id: str) -> str:
+ nonlocal repo_ids
+ hf_api.create_repo(repo_id=repo_id, token=CI_USER_TOKEN, repo_type=DATASET)
+ repo_ids.append(repo_id)
+ return repo_id
+
+ yield _tmp_dataset_repo
+ for repo_id in repo_ids:
+ delete_hub_dataset_repo(repo_id=repo_id)
+
+
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index e2aaee61..938856ae 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -5,0 +6 @@ from datetime import timedelta
+from http import HTTPStatus
@@ -7 +8 @@ from pathlib import Path
-from typing import Iterator
+from typing import Callable, Iterator
@@ -12,0 +14 @@ from filelock import FileLock
+from libcommon.processing_graph import ProcessingGraph
@@ -14 +16,3 @@ from libcommon.queue import Job, JobInfo, Priority, Status, get_datetime
-from libcommon.resources import QueueMongoResource
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import CachedResponse
+from libcommon.storage import StrPath
@@ -18,0 +23 @@ from worker.config import AppConfig
+from worker.job_runner_factory import JobRunnerFactory
@@ -20,0 +26,3 @@ from worker.main import WorkerExecutor
+from worker.resources import LibrariesResource
+
+_TIME = int(time.time() * 10e3)
@@ -28,2 +36,2 @@ def get_job_info(prefix: str = "base") -> JobInfo:
- type="bar",
- dataset=f"user/{prefix}_dataset",
+ type="/splits",
+ dataset=f"__DUMMY_DATASETS_SERVER_USER__/{prefix}_dataset_{_TIME}",
@@ -157,2 +165,19 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato
-def test_executor_get_state(app_config: AppConfig, set_worker_state: WorkerState) -> None:
- executor = WorkerExecutor(app_config)
+@fixture
+def job_runner_factory(
+ app_config: AppConfig, libraries_resource: LibrariesResource, assets_directory: StrPath
+) -> JobRunnerFactory:
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+ return JobRunnerFactory(
+ app_config=app_config,
+ processing_graph=processing_graph,
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ assets_directory=assets_directory,
+ )
+
+
+@fixture
+def executor(app_config: AppConfig, job_runner_factory: JobRunnerFactory) -> WorkerExecutor:
+ return WorkerExecutor(app_config, job_runner_factory)
+
+
+def test_executor_get_state(executor: WorkerExecutor, set_worker_state: WorkerState) -> None:
@@ -162,2 +187,3 @@ def test_executor_get_state(app_config: AppConfig, set_worker_state: WorkerState
-def test_executor_get_empty_state(app_config: AppConfig) -> None:
- executor = WorkerExecutor(app_config)
+def test_executor_get_empty_state(
+ executor: WorkerExecutor,
+) -> None:
@@ -168 +194 @@ def test_executor_get_current_job(
- app_config: AppConfig, set_just_started_job_in_queue: Job, set_worker_state: WorkerState
+ executor: WorkerExecutor, set_just_started_job_in_queue: Job, set_worker_state: WorkerState
@@ -170 +195,0 @@ def test_executor_get_current_job(
- executor = WorkerExecutor(app_config)
@@ -174,2 +199 @@ def test_executor_get_current_job(
-def test_executor_get_nonexisting_current_job(app_config: AppConfig) -> None:
- executor = WorkerExecutor(app_config)
+def test_executor_get_nonexisting_current_job(executor: WorkerExecutor) -> None:
@@ -180 +204 @@ def test_executor_get_zombies(
- app_config: AppConfig,
+ executor: WorkerExecutor,
@@ -186 +209,0 @@ def test_executor_get_zombies(
- executor = WorkerExecutor(app_config)
@@ -191 +214 @@ def test_executor_heartbeat(
- app_config: AppConfig,
+ executor: WorkerExecutor,
@@ -195 +217,0 @@ def test_executor_heartbeat(
- executor = WorkerExecutor(app_config)
@@ -208 +230 @@ def test_executor_kill_zombies(
- app_config: AppConfig,
+ executor: WorkerExecutor,
@@ -211,0 +234,3 @@ def test_executor_kill_zombies(
+ job_runner_factory: JobRunnerFactory,
+ tmp_dataset_repo_factory: Callable[[str], str],
+ cache_mongo_resource: CacheMongoResource,
@@ -214 +239 @@ def test_executor_kill_zombies(
- executor = WorkerExecutor(app_config)
+ tmp_dataset_repo_factory(zombie.dataset)
@@ -216,3 +241,17 @@ def test_executor_kill_zombies(
- executor.kill_zombies()
- assert executor.get_zombies() == []
- assert Job.objects.with_id(zombie.pk).status == Status.ERROR # type: ignore
+ try:
+ executor.kill_zombies()
+ assert executor.get_zombies() == []
+ assert Job.objects.with_id(zombie.pk).status == Status.ERROR # type: ignore
+ response = CachedResponse.objects()[0]
+ expected_error = {
+ "error": "Worker crashed while running this job.",
+ }
+ assert response.http_status == HTTPStatus.NOT_IMPLEMENTED
+ assert response.error_code == "WorkerCrashedError"
+ assert response.dataset == zombie.dataset
+ assert response.config == zombie.config
+ assert response.split == zombie.split
+ assert response.content == expected_error
+ assert response.details == expected_error
+ finally:
+ CachedResponse.objects().delete()
@@ -222 +261 @@ def test_executor_start(
- app_config: AppConfig,
+ executor: WorkerExecutor,
@@ -229 +267,0 @@ def test_executor_start(
- executor = WorkerExecutor(app_config)
@@ -247 +285 @@ def test_executor_raises_on_bad_worker(
- app_config: AppConfig, queue_mongo_resource: QueueMongoResource, tmp_path: Path, bad_worker_loop_type: str
+ executor: WorkerExecutor, queue_mongo_resource: QueueMongoResource, tmp_path: Path, bad_worker_loop_type: str
@@ -254 +291,0 @@ def test_executor_raises_on_bad_worker(
- executor = WorkerExecutor(app_config)
|
|
8128ec5b131511e7a0b4755f2e2455ce96f7d179
|
Quentin Lhoest
| 2023-02-17T18:15:05 |
Kill zombies (#826)
|
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index b7ae5cc1..53a36969 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -22,0 +23,2 @@ WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS = 60
+WORKER_MAX_MISSING_HEARTBEATS = 5
+WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS = 10 * 60
@@ -39,0 +42,2 @@ class WorkerConfig:
+ max_missing_heartbeats: int = WORKER_MAX_MISSING_HEARTBEATS
+ kill_zombies_time_interval_seconds: int = WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS
@@ -56,0 +61,4 @@ class WorkerConfig:
+ max_missing_heartbeats=env.int(name="MAX_MISSING_HEARTBEATS", default=WORKER_MAX_MISSING_HEARTBEATS),
+ kill_zombies_time_interval_seconds=env.int(
+ name="KILL_ZOMBIES_TIME_INTERVAL_SECONDS", default=WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS
+ ),
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index 439cefbe..623ef954 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -0,0 +1 @@
+import asyncio
@@ -6,2 +7,2 @@ import tempfile
-import time
-from typing import Optional
+from datetime import timedelta
+from typing import Any, Callable, List, Optional
@@ -8,0 +10 @@ from typing import Optional
+import pytz
@@ -22,0 +25,10 @@ START_WORKER_LOOP_PATH = start_worker_loop.__file__
+async def every(
+ func: Callable[..., Optional[Any]], *args: Any, seconds: int, stop_on: Optional[Any] = None, **kwargs: Any
+) -> None:
+ while True:
+ out = func(*args, **kwargs)
+ if stop_on is not None and out == stop_on:
+ break
+ await asyncio.sleep(seconds)
+
+
@@ -40,0 +53,2 @@ class WorkerExecutor:
+
+ loop = asyncio.get_event_loop()
@@ -42,4 +56,5 @@ class WorkerExecutor:
- while worker_loop_executor.running():
- self.heartbeat()
- time.sleep(self.app_config.worker.heartbeat_time_interval_seconds)
- worker_loop_executor.stop()
+ loop.create_task(every(self.heartbeat, seconds=self.app_config.worker.heartbeat_time_interval_seconds))
+ loop.create_task(every(self.kill_zombies, seconds=self.app_config.worker.kill_zombies_time_interval_seconds))
+ loop.run_until_complete(
+ every(self.is_worker_alive, worker_loop_executor=worker_loop_executor, seconds=1, stop_on=False)
+ )
@@ -74,0 +90,44 @@ class WorkerExecutor:
+ def get_zombies(self) -> List[Job]:
+ started_jobs = Job.objects(status=Status.STARTED)
+ max_missing_heartbeats = self.app_config.worker.max_missing_heartbeats
+ heartbeat_time_interval_seconds = self.app_config.worker.heartbeat_time_interval_seconds
+ if max_missing_heartbeats >= 1:
+ return [
+ job
+ for job in started_jobs
+ if (
+ job.last_heartbeat is not None
+ and get_datetime()
+ >= pytz.UTC.localize(job.last_heartbeat)
+ + timedelta(seconds=max_missing_heartbeats * heartbeat_time_interval_seconds)
+ )
+ or (
+ job.last_heartbeat is None
+ and job.started_at is not None
+ and get_datetime()
+ >= pytz.UTC.localize(job.started_at)
+ + timedelta(seconds=max_missing_heartbeats * heartbeat_time_interval_seconds)
+ )
+ ]
+ else:
+ return []
+
+ def kill_zombies(self) -> None:
+ zombies = self.get_zombies()
+ if zombies:
+ zombies_examples = [str(zombie.pk) for zombie in zombies[:10]]
+ zombies_examples_str = ", ".join(zombies_examples) + (
+ "..." if len(zombies_examples) != len(zombies) else ""
+ )
+ logging.info(f"Killing {len(zombies)} zombies. Job ids = " + zombies_examples_str)
+ Job.objects(pk__in=[zombie.pk for zombie in zombies]).update(
+ status=Status.ERROR, finished_at=get_datetime()
+ )
+
+ def is_worker_alive(self, worker_loop_executor: OutputExecutor) -> bool:
+ if not worker_loop_executor.running():
+ worker_loop_executor.stop() # raises an error if the worker returned exit code 1
+ return False
+ else:
+ return True
+
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index ed4039ee..77325b05 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -67,0 +68 @@ def set_env_vars(
+ mp.setenv("WORKER_KILL_ZOMBIES_TIME_INTERVAL_SECONDS", "1")
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index 46f9ad5d..e2aaee61 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -23 +23,3 @@ from worker.main import WorkerExecutor
-def get_job_info() -> JobInfo:
+def get_job_info(prefix: str = "base") -> JobInfo:
+ job_id = prefix.encode().hex()
+ assert len(job_id) <= 24, "please choose a smaller prefix"
@@ -25 +27 @@ def get_job_info() -> JobInfo:
- job_id="a" * 24,
+ job_id=job_id + "0" * (24 - len(job_id)),
@@ -27 +29 @@ def get_job_info() -> JobInfo:
- dataset="user/my_dataset",
+ dataset=f"user/{prefix}_dataset",
@@ -75 +77 @@ def set_worker_state(worker_state_path: Path) -> Iterator[WorkerState]:
-def set_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
+def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
@@ -80,0 +83 @@ def set_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterat
+ created_at = get_datetime()
@@ -91 +94,57 @@ def set_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterat
- created_at=get_datetime(),
+ created_at=created_at,
+ started_at=created_at + timedelta(microseconds=1),
+ )
+ job.save()
+ yield job
+ job.delete()
+
+
+@fixture
+def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
+ if not queue_mongo_resource.is_available():
+ raise RuntimeError("Mongo resource is not available")
+ job_info = get_job_info("long")
+ if Job.objects.with_id(job_info["job_id"]): # type: ignore
+ Job.objects.with_id(job_info["job_id"]).delete() # type: ignore
+ created_at = get_datetime() - timedelta(days=1)
+ last_heartbeat = get_datetime() - timedelta(seconds=app_config.worker.heartbeat_time_interval_seconds)
+ job = Job(
+ pk=job_info["job_id"],
+ type=job_info["type"],
+ dataset=job_info["dataset"],
+ config=job_info["config"],
+ split=job_info["split"],
+ unicity_id="unicity_id",
+ namespace="user",
+ priority=job_info["priority"],
+ status=Status.STARTED,
+ created_at=created_at,
+ started_at=created_at + timedelta(milliseconds=1),
+ last_heartbeat=last_heartbeat,
+ )
+ job.save()
+ yield job
+ job.delete()
+
+
+@fixture
+def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
+ if not queue_mongo_resource.is_available():
+ raise RuntimeError("Mongo resource is not available")
+ job_info = get_job_info("zombie")
+ if Job.objects.with_id(job_info["job_id"]): # type: ignore
+ Job.objects.with_id(job_info["job_id"]).delete() # type: ignore
+ created_at = get_datetime() - timedelta(days=1)
+ job = Job(
+ pk=job_info["job_id"],
+ type=job_info["type"],
+ dataset=job_info["dataset"],
+ config=job_info["config"],
+ split=job_info["split"],
+ unicity_id="unicity_id",
+ namespace="user",
+ priority=job_info["priority"],
+ status=Status.STARTED,
+ created_at=created_at,
+ started_at=created_at + timedelta(milliseconds=1),
+ last_heartbeat=created_at + timedelta(milliseconds=2),
@@ -109 +168 @@ def test_executor_get_current_job(
- app_config: AppConfig, set_started_job_in_queue: Job, set_worker_state: WorkerState
+ app_config: AppConfig, set_just_started_job_in_queue: Job, set_worker_state: WorkerState
@@ -112 +171 @@ def test_executor_get_current_job(
- assert executor.get_current_job() == set_started_job_in_queue
+ assert executor.get_current_job() == set_just_started_job_in_queue
@@ -119,0 +179,11 @@ def test_executor_get_nonexisting_current_job(app_config: AppConfig) -> None:
+def test_executor_get_zombies(
+ app_config: AppConfig,
+ set_just_started_job_in_queue: Job,
+ set_long_running_job_in_queue: Job,
+ set_zombie_job_in_queue: Job,
+) -> None:
+ zombie = set_zombie_job_in_queue
+ executor = WorkerExecutor(app_config)
+ assert executor.get_zombies() == [zombie]
+
+
@@ -122 +192 @@ def test_executor_heartbeat(
- set_started_job_in_queue: Job,
+ set_just_started_job_in_queue: Job,
@@ -124 +193,0 @@ def test_executor_heartbeat(
- queue_mongo_resource: QueueMongoResource,
@@ -126,2 +194,0 @@ def test_executor_heartbeat(
- if not queue_mongo_resource.is_available():
- raise RuntimeError("Mongo resource is not available")
@@ -139,0 +207,14 @@ def test_executor_heartbeat(
+def test_executor_kill_zombies(
+ app_config: AppConfig,
+ set_just_started_job_in_queue: Job,
+ set_long_running_job_in_queue: Job,
+ set_zombie_job_in_queue: Job,
+) -> None:
+ zombie = set_zombie_job_in_queue
+ executor = WorkerExecutor(app_config)
+ assert executor.get_zombies() == [zombie]
+ executor.kill_zombies()
+ assert executor.get_zombies() == []
+ assert Job.objects.with_id(zombie.pk).status == Status.ERROR # type: ignore
+
+
@@ -141 +222,4 @@ def test_executor_start(
- app_config: AppConfig, queue_mongo_resource: QueueMongoResource, set_started_job_in_queue: Job
+ app_config: AppConfig,
+ queue_mongo_resource: QueueMongoResource,
+ set_just_started_job_in_queue: Job,
+ set_zombie_job_in_queue: Job,
@@ -147,2 +231,3 @@ def test_executor_start(
- with patch("worker.main.START_WORKER_LOOP_PATH", __file__):
- executor.start()
+ with patch.object(executor, "kill_zombies", wraps=executor.kill_zombies) as kill_zombies_mock:
+ with patch("worker.main.START_WORKER_LOOP_PATH", __file__):
+ executor.start()
@@ -152,0 +238,3 @@ def test_executor_start(
+ assert Job.objects.with_id(set_just_started_job_in_queue.pk).last_heartbeat is not None # type: ignore
+ assert kill_zombies_mock.call_count > 0
+ assert Job.objects.with_id(set_zombie_job_in_queue.pk).status == Status.ERROR # type: ignore
|
|
61947cfe5bf92dde001d7345353315594db2b590
|
Andrea Francis Soria Jimenez
| 2023-02-17T15:38:40 |
Separate endpoint from processing step logic (#817)
|
diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py
index 43a2a6c4..db2f97a4 100644
--- a/libs/libcommon/src/libcommon/operations.py
+++ b/libs/libcommon/src/libcommon/operations.py
@@ -18 +18 @@ class PreviousStepError(LoggedError):
- f"Response for {step.endpoint} for dataset={dataset}, config={config}, split={split} is an error."
+ f"Response for {step.job_type} for dataset={dataset}, config={config}, split={split} is an error."
diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py
index 656c84b1..55d8adc4 100644
--- a/libs/libcommon/src/libcommon/processing_graph.py
+++ b/libs/libcommon/src/libcommon/processing_graph.py
@@ -5,0 +6 @@ from __future__ import annotations
+import warnings
@@ -26 +27 @@ class ProcessingStep:
- - the API endpoint
+ - the step name
@@ -35 +36 @@ class ProcessingStep:
- endpoint: str
+ name: str
@@ -42,0 +44,5 @@ class ProcessingStep:
+ @property
+ def endpoint(self) -> str:
+ warnings.warn("The use of endpoint is deprecated, name will be used instead.", category=DeprecationWarning)
+ return self.name
+
@@ -46 +52 @@ class ProcessingStep:
- return self.endpoint
+ return self.name
@@ -51 +57 @@ class ProcessingStep:
- return self.endpoint
+ return self.name
@@ -62 +68 @@ class ProcessingStep:
- raise ValueError(f"Cycle detected between {self.endpoint} and {self.parent.endpoint}")
+ raise ValueError(f"Cycle detected between {self.job_type} and {self.parent.job_type}")
@@ -78 +84 @@ class ProcessingGraph:
- - the index of all the steps, identified by their endpoint
+ - the index of all the steps, identified by their name
@@ -88 +93,0 @@ class ProcessingGraph:
- # TODO: validate the graph specification: endpoints must start with "/" and use only lowercase letters
@@ -90,2 +95,2 @@ class ProcessingGraph:
- endpoint: ProcessingStep(
- endpoint=endpoint,
+ name: ProcessingStep(
+ name=name,
@@ -99 +104 @@ class ProcessingGraph:
- for endpoint, specification in processing_graph_specification.items()
+ for name, specification in processing_graph_specification.items()
@@ -115,5 +120,5 @@ class ProcessingGraph:
- def get_step(self, endpoint: str) -> ProcessingStep:
- """Get a step by its endpoint."""
- if endpoint not in self.steps:
- raise ValueError(f"Unknown endpoint: {endpoint}")
- return self.steps[endpoint]
+ def get_step(self, name: str) -> ProcessingStep:
+ """Get a step by its name."""
+ if name not in self.steps:
+ raise ValueError(f"Unknown name: {name}")
+ return self.steps[name]
@@ -122 +127 @@ class ProcessingGraph:
- # for now: the job_type is just an alias for the endpoint
+ # for now: the job_type is just an alias for the step name
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 884e9474..60ae169a 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -420 +420 @@ class Queue:
- Returns: the job id, the type (endpoint), the input arguments: dataset, config and split and the force flag
+ Returns: the job id, the type, the input arguments: dataset, config and split and the force flag
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index 906bfdde..37ce1098 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -66 +66 @@ class SplitFullName(NamedTuple):
-# cache of any endpoint
+# cache of any job
@@ -68 +68 @@ class CachedResponse(Document):
- """A response to an endpoint request, cached in the mongoDB database
+ """A response computed for a job, cached in the mongoDB database
@@ -71 +71 @@ class CachedResponse(Document):
- kind (`str`): The kind of the cached response, identifies the endpoint
+ kind (`str`): The kind of the cached response, identifies the job type
diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py
index e19cb6b5..8007127a 100644
--- a/services/admin/src/admin/app.py
+++ b/services/admin/src/admin/app.py
@@ -87 +87 @@ def create_app() -> Starlette:
- f"/force-refresh{processing_step.endpoint}",
+ f"/force-refresh{processing_step.job_type}",
@@ -114 +114 @@ def create_app() -> Starlette:
- f"/cache-reports{processing_step.endpoint}",
+ f"/cache-reports{processing_step.job_type}",
@@ -127 +127 @@ def create_app() -> Starlette:
- f"/cache-reports-with-content{processing_step.endpoint}",
+ f"/cache-reports-with-content{processing_step.job_type}",
@@ -140 +140 @@ def create_app() -> Starlette:
- f"/cancel-jobs{processing_step.endpoint}",
+ f"/cancel-jobs{processing_step.job_type}",
@@ -152 +152 @@ def create_app() -> Starlette:
- f"/jobs-duration-per-dataset{processing_step.endpoint}",
+ f"/jobs-duration-per-dataset{processing_step.job_type}",
diff --git a/services/admin/src/admin/routes/cancel_jobs.py b/services/admin/src/admin/routes/cancel_jobs.py
index 22292cf3..628fbcc6 100644
--- a/services/admin/src/admin/routes/cancel_jobs.py
+++ b/services/admin/src/admin/routes/cancel_jobs.py
@@ -29 +29 @@ def create_cancel_jobs_endpoint(
- logging.info(f"/cancel-jobs{processing_step.endpoint}")
+ logging.info(f"/cancel-jobs{processing_step.job_type}")
diff --git a/services/admin/src/admin/routes/dataset_status.py b/services/admin/src/admin/routes/dataset_status.py
index 6ddadd53..2f95c483 100644
--- a/services/admin/src/admin/routes/dataset_status.py
+++ b/services/admin/src/admin/routes/dataset_status.py
@@ -43 +43 @@ def create_dataset_status_endpoint(
- processing_step.endpoint: {
+ processing_step.job_type: {
diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py
index 3010b407..a1199b92 100644
--- a/services/admin/src/admin/routes/force_refresh.py
+++ b/services/admin/src/admin/routes/force_refresh.py
@@ -51 +51 @@ def create_force_refresh_endpoint(
- f"/force-refresh{processing_step.endpoint}, dataset={dataset}, config={config}, split={split}"
+ f"/force-refresh{processing_step.job_type}, dataset={dataset}, config={config}, split={split}"
diff --git a/services/admin/src/admin/routes/pending_jobs.py b/services/admin/src/admin/routes/pending_jobs.py
index 1f2bbf3d..da185712 100644
--- a/services/admin/src/admin/routes/pending_jobs.py
+++ b/services/admin/src/admin/routes/pending_jobs.py
@@ -36 +36 @@ def create_pending_jobs_endpoint(
- processing_step.endpoint: queue.get_dump_by_pending_status(job_type=processing_step.job_type)
+ processing_step.job_type: queue.get_dump_by_pending_status(job_type=processing_step.job_type)
diff --git a/services/admin/tests/test_app.py b/services/admin/tests/test_app.py
index 8a3a23f3..e79a4d85 100644
--- a/services/admin/tests/test_app.py
+++ b/services/admin/tests/test_app.py
@@ -100 +100 @@ def test_cache_reports(
- path = processing_steps[0].endpoint
+ path = processing_steps[0].job_type
@@ -127 +127 @@ def test_cache_reports_with_content(
- path = processing_steps[0].endpoint
+ path = processing_steps[0].job_type
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 5e77f013..bb67a1ca 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -17 +17 @@ from starlette_prometheus import PrometheusMiddleware
-from api.config import AppConfig, UvicornConfig
+from api.config import AppConfig, EndpointConfig, UvicornConfig
@@ -18,0 +19 @@ from api.prometheus import Prometheus
+from api.routes.endpoint import EndpointsDefinition, create_endpoint
@@ -20 +20,0 @@ from api.routes.healthcheck import healthcheck_endpoint
-from api.routes.processing_step import create_processing_step_endpoint
@@ -34 +34 @@ def create_app() -> Starlette:
- processing_steps = list(processing_graph.steps.values())
+ endpoints_definition = EndpointsDefinition(processing_graph, EndpointConfig.from_env())
@@ -74 +74 @@ def create_app() -> Starlette:
- processing_step_endpoints: List[BaseRoute] = [
+ endpoints: List[BaseRoute] = [
@@ -76,3 +76,3 @@ def create_app() -> Starlette:
- processing_step.endpoint,
- endpoint=create_processing_step_endpoint(
- processing_step=processing_step,
+ endpoint_name,
+ endpoint=create_endpoint(
+ processing_steps=processing_steps,
@@ -87 +87 @@ def create_app() -> Starlette:
- for processing_step in processing_steps
+ for endpoint_name, processing_steps in endpoints_definition.definition.items()
@@ -106 +106 @@ def create_app() -> Starlette:
- routes: List[BaseRoute] = valid + processing_step_endpoints + to_protect + protected
+ routes: List[BaseRoute] = valid + endpoints + to_protect + protected
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 2cf5637c..39af9f83 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -5 +5 @@ from dataclasses import dataclass, field
-from typing import Optional
+from typing import List, Mapping, Optional
@@ -81,0 +82,24 @@ class AppConfig:
+
+
+EndpointProcessingStepNamesMapping = Mapping[str, List[str]]
+
+
+@dataclass(frozen=True)
+class EndpointConfig:
+ specification: EndpointProcessingStepNamesMapping = field(
+ default_factory=lambda: {
+ "/config-names": ["/config-names"],
+ "/split-names-from-streaming": ["/split-names-from-streaming"],
+ "/splits": ["/splits"],
+ "/first-rows": ["/first-rows"],
+ "/parquet-and-dataset-info": ["/parquet-and-dataset-info"],
+ "/parquet": ["/parquet"],
+ "/dataset-info": ["/dataset-info"],
+ "/sizes": ["/sizes"],
+ }
+ )
+
+ @classmethod
+ def from_env(cls) -> "EndpointConfig":
+ # TODO: allow passing the mapping between endpoint and processing steps via env vars
+ return cls()
diff --git a/services/api/src/api/routes/processing_step.py b/services/api/src/api/routes/endpoint.py
similarity index 78%
rename from services/api/src/api/routes/processing_step.py
rename to services/api/src/api/routes/endpoint.py
index 7468ba8a..8b672ea3 100644
--- a/services/api/src/api/routes/processing_step.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import List, Optional
+from typing import List, Mapping, Optional
@@ -10 +10 @@ from libcommon.operations import PreviousStepError, check_in_process
-from libcommon.processing_graph import ProcessingStep
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
@@ -15,0 +16 @@ from api.authentication import auth_check
+from api.config import EndpointConfig
@@ -18,0 +20 @@ from api.utils import (
+ MissingProcessingStepsError,
@@ -30,2 +32,14 @@ from api.utils import (
-def create_processing_step_endpoint(
- processing_step: ProcessingStep,
+class EndpointsDefinition:
+ """Definition of supported endpoints and its relation with processing steps."""
+
+ processing_steps_by_endpoint: Mapping[str, List[ProcessingStep]]
+
+ def __init__(self, graph: ProcessingGraph, endpoint_config: EndpointConfig):
+ self.definition = {
+ endpoint: [graph.get_step(step) for step in processing_steps]
+ for endpoint, processing_steps in endpoint_config.specification.items()
+ }
+
+
+def create_endpoint(
+ processing_steps: List[ProcessingStep],
@@ -39,0 +54,5 @@ def create_processing_step_endpoint(
+ if not processing_steps:
+ raise MissingProcessingStepsError("Missing processing steps or empty list")
+
+ processing_step = processing_steps[0]
+ # TODO: Add support to iterate over the processing steps list in case of failure
@@ -57 +76 @@ def create_processing_step_endpoint(
- logging.info(f"{processing_step.endpoint}, dataset={dataset}, config={config}, split={split}")
+ logging.info(f"{processing_step.job_type}, dataset={dataset}, config={config}, split={split}")
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index 2800aee1..dfbc1d05 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -19,0 +20 @@ ApiErrorCode = Literal[
+ "MissingProcessingStepsError",
@@ -82,0 +84,7 @@ class ExternalAuthenticatedError(ApiCustomError):
+class MissingProcessingStepsError(ApiCustomError):
+ """Raised when an endpoint does not have related processing steps."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "MissingProcessingStepsError")
+
+
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index 40850f49..7f22af4c 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -4 +4 @@
-from typing import Iterator, List
+from typing import Iterator, List, Mapping
@@ -12 +12,2 @@ from pytest import MonkeyPatch, fixture
-from api.config import AppConfig, UvicornConfig
+from api.config import AppConfig, EndpointConfig, UvicornConfig
+from api.routes.endpoint import EndpointsDefinition
@@ -39 +40 @@ def app_config(monkeypatch_session: MonkeyPatch) -> AppConfig:
-def processing_steps(app_config: AppConfig) -> List[ProcessingStep]:
+def endpoint_definition(app_config: AppConfig) -> Mapping[str, List[ProcessingStep]]:
@@ -41 +42,2 @@ def processing_steps(app_config: AppConfig) -> List[ProcessingStep]:
- return list(processing_graph.steps.values())
+ endpoint_specification = EndpointsDefinition(processing_graph, EndpointConfig.from_env())
+ return endpoint_specification.definition
@@ -45,2 +47,6 @@ def processing_steps(app_config: AppConfig) -> List[ProcessingStep]:
-def first_dataset_processing_step(processing_steps: List[ProcessingStep]) -> ProcessingStep:
- return next(step for step in processing_steps if step.input_type == "dataset")
+def first_dataset_endpoint(endpoint_definition: Mapping[str, List[ProcessingStep]]) -> str:
+ return next(
+ endpoint
+ for endpoint, processing_steps in endpoint_definition.items()
+ if next((step for step in processing_steps if step.input_type == "dataset"), None)
+ )
@@ -50,2 +56,6 @@ def first_dataset_processing_step(processing_steps: List[ProcessingStep]) -> Pro
-def first_config_processing_step(processing_steps: List[ProcessingStep]) -> ProcessingStep:
- return next(step for step in processing_steps if step.input_type == "config")
+def first_config_endoint(endpoint_definition: Mapping[str, List[ProcessingStep]]) -> str:
+ return next(
+ endpoint
+ for endpoint, processing_steps in endpoint_definition.items()
+ if next((step for step in processing_steps if step.input_type == "config"), None)
+ )
@@ -55,2 +65,6 @@ def first_config_processing_step(processing_steps: List[ProcessingStep]) -> Proc
-def first_split_processing_step(processing_steps: List[ProcessingStep]) -> ProcessingStep:
- return next(step for step in processing_steps if step.input_type == "split")
+def first_split_endpoint(endpoint_definition: Mapping[str, List[ProcessingStep]]) -> str:
+ return next(
+ endpoint
+ for endpoint, processing_steps in endpoint_definition.items()
+ if next((step for step in processing_steps if step.input_type == "split"), None)
+ )
diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py
index 6b93d9e0..6afb9357 100644
--- a/services/api/tests/routes/test_valid.py
+++ b/services/api/tests/routes/test_valid.py
@@ -12 +12 @@ dataset_step = ProcessingStep(
- endpoint="/dataset-step",
+ name="/dataset-step",
@@ -21 +21 @@ config_step = ProcessingStep(
- endpoint="/config-step",
+ name="/config-step",
@@ -30 +30 @@ split_step = ProcessingStep(
- endpoint="/split-step",
+ name="/split-step",
diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py
index 26feb3c0..8cacbda0 100644
--- a/services/api/tests/test_app.py
+++ b/services/api/tests/test_app.py
@@ -7 +6,0 @@ import pytest
-from libcommon.processing_graph import ProcessingStep
@@ -21 +20 @@ def client(monkeypatch_session: pytest.MonkeyPatch) -> TestClient:
-def test_cors(client: TestClient, first_dataset_processing_step: ProcessingStep) -> None:
+def test_cors(client: TestClient, first_dataset_endpoint: str) -> None:
@@ -26 +25 @@ def test_cors(client: TestClient, first_dataset_processing_step: ProcessingStep)
- f"{first_dataset_processing_step.endpoint}?dataset=dataset1",
+ f"{first_dataset_endpoint}?dataset=dataset1",
@@ -86 +85 @@ def test_get_healthcheck(client: TestClient) -> None:
-def test_get_step(client: TestClient, first_dataset_processing_step: ProcessingStep) -> None:
+def test_get_endpoint(client: TestClient, first_dataset_endpoint: str) -> None:
@@ -88 +87 @@ def test_get_step(client: TestClient, first_dataset_processing_step: ProcessingS
- response = client.get(first_dataset_processing_step.endpoint)
+ response = client.get(first_dataset_endpoint)
@@ -91 +90 @@ def test_get_step(client: TestClient, first_dataset_processing_step: ProcessingS
- response = client.get(f"{first_dataset_processing_step.endpoint}?dataset=")
+ response = client.get(f"{first_dataset_endpoint}?dataset=")
@@ -107 +106 @@ def test_get_config_missing_parameter(
- first_config_processing_step: ProcessingStep,
+ first_config_endoint: str,
@@ -109,3 +108 @@ def test_get_config_missing_parameter(
- response = client.get(
- first_config_processing_step.endpoint, params={"dataset": dataset, "config": config, "split": None}
- )
+ response = client.get(first_config_endoint, params={"dataset": dataset, "config": config, "split": None})
@@ -129 +126 @@ def test_get_split_missing_parameter(
- first_split_processing_step: ProcessingStep,
+ first_split_endpoint: str,
@@ -131,3 +128 @@ def test_get_split_missing_parameter(
- response = client.get(
- first_split_processing_step.endpoint, params={"dataset": dataset, "config": config, "split": split}
- )
+ response = client.get(first_split_endpoint, params={"dataset": dataset, "config": config, "split": split})
diff --git a/services/worker/README.md b/services/worker/README.md
index e2b979c9..38e6d783 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -51,2 +51,2 @@ Set environment variables to configure the first rows worker (`FIRST_ROWS_` pref
-- `FIRST_ROWS_MAX_BYTES`: the max size of the /first-rows endpoint response in bytes. Defaults to `1_000_000` (1 MB).
-- `FIRST_ROWS_MAX_NUMBER`: the max number of rows fetched by the worker for the split and provided in the /first-rows endpoint response. Defaults to `100`.
+- `FIRST_ROWS_MAX_BYTES`: the max size of the /first-rows response in bytes. Defaults to `1_000_000` (1 MB).
+- `FIRST_ROWS_MAX_NUMBER`: the max number of rows fetched by the worker for the split and provided in the /first-rows response. Defaults to `100`.
@@ -54,2 +54,2 @@ Set environment variables to configure the first rows worker (`FIRST_ROWS_` pref
-- `FIRST_ROWS_MIN_NUMBER`: the min number of rows fetched by the worker for the split and provided in the /first-rows endpoint response. Defaults to `10`.
-- `FIRST_ROWS_COLUMNS_MAX_NUMBER`: the max number of columns (features) provided in the /first-rows endpoint response. If the number of columns is greater than the limit, an error is returned. Defaults to `1_000`.
+- `FIRST_ROWS_MIN_NUMBER`: the min number of rows fetched by the worker for the split and provided in the /first-rows response. Defaults to `10`.
+- `FIRST_ROWS_COLUMNS_MAX_NUMBER`: the max number of columns (features) provided in the /first-rows response. If the number of columns is greater than the limit, an error is returned. Defaults to `1_000`.
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index ac5d5693..df0d5f2c 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -212 +212 @@ class JobRunner(ABC):
- logging.log(level=level, msg=f"[{self.processing_step.endpoint}] {msg}")
+ logging.log(level=level, msg=f"[{self.processing_step.job_type}] {msg}")
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index eed12a27..ed4039ee 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -117 +117 @@ def test_processing_step() -> ProcessingStep:
- endpoint="/dummy",
+ name="/dummy",
diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py
index ba24c4c5..8f1a5996 100644
--- a/services/worker/tests/job_runners/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__datasets_based_worker.py
@@ -28 +28 @@ class DummyJobRunner(DatasetsBasedJobRunner):
- # ^ borrowing the endpoint, so that the processing step exists and the job runner can be initialized
+ # ^ borrowing the type, so that the processing step exists and the job runner can be initialized
@@ -70 +70 @@ def get_job_runner(
- endpoint=DummyJobRunner.get_job_type(),
+ name=DummyJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py
index a82a6912..b41fafc5 100644
--- a/services/worker/tests/job_runners/test_config_names.py
+++ b/services/worker/tests/job_runners/test_config_names.py
@@ -47 +47 @@ def get_job_runner(
- endpoint=ConfigNamesJobRunner.get_job_type(),
+ name=ConfigNamesJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_dataset_info.py b/services/worker/tests/job_runners/test_dataset_info.py
index 5674de6a..37fae457 100644
--- a/services/worker/tests/job_runners/test_dataset_info.py
+++ b/services/worker/tests/job_runners/test_dataset_info.py
@@ -54 +54 @@ def get_job_runner(
- endpoint=DatasetInfoJobRunner.get_job_type(),
+ name=DatasetInfoJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py
index 66adf01d..23add41e 100644
--- a/services/worker/tests/job_runners/test_first_rows.py
+++ b/services/worker/tests/job_runners/test_first_rows.py
@@ -53 +53 @@ def get_job_runner(
- endpoint=FirstRowsJobRunner.get_job_type(),
+ name=FirstRowsJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_parquet.py b/services/worker/tests/job_runners/test_parquet.py
index 12dc805b..6e0d71c2 100644
--- a/services/worker/tests/job_runners/test_parquet.py
+++ b/services/worker/tests/job_runners/test_parquet.py
@@ -54 +54 @@ def get_job_runner(
- endpoint=ParquetJobRunner.get_job_type(),
+ name=ParquetJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index 5706f2ef..7192d30b 100644
--- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -89 +89 @@ def get_job_runner(
- endpoint=ParquetAndDatasetInfoJobRunner.get_job_type(),
+ name=ParquetAndDatasetInfoJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_sizes.py b/services/worker/tests/job_runners/test_sizes.py
index 5e272fff..6b731071 100644
--- a/services/worker/tests/job_runners/test_sizes.py
+++ b/services/worker/tests/job_runners/test_sizes.py
@@ -54 +54 @@ def get_job_runner(
- endpoint=SizesJobRunner.get_job_type(),
+ name=SizesJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
index 31253213..2870218c 100644
--- a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
+++ b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
@@ -50 +50 @@ def get_job_runner(
- endpoint=SplitNamesFromDatasetInfoJobRunner.get_job_type(),
+ name=SplitNamesFromDatasetInfoJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_split_names_from_streaming.py b/services/worker/tests/job_runners/test_split_names_from_streaming.py
index 8ace3ce6..04493ba3 100644
--- a/services/worker/tests/job_runners/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/test_split_names_from_streaming.py
@@ -50 +50 @@ def get_job_runner(
- endpoint=SplitNamesFromStreamingJobRunner.get_job_type(),
+ name=SplitNamesFromStreamingJobRunner.get_job_type(),
diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py
index a3e2d93a..b69c7022 100644
--- a/services/worker/tests/job_runners/test_splits.py
+++ b/services/worker/tests/job_runners/test_splits.py
@@ -47 +47 @@ def get_job_runner(
- endpoint=SplitsJobRunner.get_job_type(),
+ name=SplitsJobRunner.get_job_type(),
diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py
index d79af11d..cd4cef03 100644
--- a/services/worker/tests/test_job_runner.py
+++ b/services/worker/tests/test_job_runner.py
@@ -244 +244 @@ def test_check_type(
- endpoint=f"not-{test_processing_step.endpoint}",
+ name=f"not-{test_processing_step.name}",
|
|
ec6bbedca73b1a24b8502fb55c611b0ffd214c93
|
Andrea Francis Soria Jimenez
| 2023-02-17T14:33:40 |
Renaming split-names to split-names-from-streaming (#825)
|
diff --git a/e2e/tests/test_11_auth.py b/e2e/tests/test_11_auth.py
index aff84ea2..f4e0df3b 100644
--- a/e2e/tests/test_11_auth.py
+++ b/e2e/tests/test_11_auth.py
@@ -43 +43,2 @@ def test_auth_e2e(
- f"/split-names?dataset={dataset}&config={config}",
+ f"/split-names-from-streaming?dataset={dataset}&config={config}",
+ # TODO: ^ this will be changed to /splits in an uncoming PR
diff --git a/jobs/mongodb_migration/src/mongodb_migration/check.py b/jobs/mongodb_migration/src/mongodb_migration/check.py
index df84c992..a554f0e9 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/check.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/check.py
@@ -19,2 +19,2 @@ def get_random_oids(collection: Collection, sample_size: int) -> List[int]:
- pipeline = [{"$project": {"pk": 1}}, {"$sample": {"size": sample_size}}]
- return [s["pk"] for s in collection.aggregate(pipeline)]
+ pipeline = [{"$project": {"_id": 1}}, {"$sample": {"size": sample_size}}]
+ return [s["_id"] for s in collection.aggregate(pipeline)]
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 71cada64..7ba24295 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -16,0 +17,6 @@ from mongodb_migration.migrations._20230126164900_queue_job_add_priority import
+from mongodb_migration.migrations._20230216112500_cache_split_names_from_streaming import (
+ MigrationCacheUpdateSplitNames,
+)
+from mongodb_migration.migrations._20230216141000_queue_split_names_from_streaming import (
+ MigrationQueueUpdateSplitNames,
+)
@@ -34,0 +41,8 @@ class MigrationsCollector:
+ MigrationCacheUpdateSplitNames(
+ version="20230216112500",
+ description="update 'kind' field in cache from /split-names to /split-names-streaming",
+ ),
+ MigrationQueueUpdateSplitNames(
+ version="20230216141000",
+ description="update 'type' and 'unicity_id' fields in job from /split-names to /split-names-streaming",
+ ),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
new file mode 100644
index 00000000..0083817d
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py
@@ -0,0 +1,30 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.simple_cache import CachedResponse
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import Migration
+
+db_name = "cache"
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationCacheUpdateSplitNames(Migration):
+ def up(self) -> None:
+ logging.info("Rename cache_kind field from /split-name to /split-names-streaming")
+ db = get_db(db_name)
+ db["cachedResponsesBlue"].update_many({"kind": "/split-names"}, {"$set": {"kind": "/split-names-streaming"}})
+
+ def down(self) -> None:
+ logging.info("Rollback cache_kind field from /split-name-streaming to /split-names")
+ db = get_db(db_name)
+ db["cachedResponsesBlue"].update_many({"kind": "/split-names-streaming"}, {"$set": {"kind": "/split-names"}})
+
+ def validate(self) -> None:
+ logging.info("Validate modified documents")
+
+ check_documents(DocCls=CachedResponse, sample_size=10)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
new file mode 100644
index 00000000..b9c2cc4d
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py
@@ -0,0 +1,48 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.queue import Job
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import Migration
+
+split_names = "/split-names"
+split_names_from_streaming = "/split-names-from-streaming"
+db_name = "queue"
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationQueueUpdateSplitNames(Migration):
+ def up(self) -> None:
+ logging.info(
+ f"Rename unicity_id field from Job[{split_names}][<dataset>][<config>][<split>] to"
+ f" Job[{split_names_from_streaming}][<dataset>][<config>][<split>]"
+ )
+
+ for job in Job.objects(type=split_names):
+ job.update(unicity_id=f"Job[{split_names_from_streaming}][{job.dataset}][{job.config}][{job.split}]")
+
+ logging.info(f"Rename type field from {split_names} to {split_names_from_streaming}")
+ db = get_db("queue")
+ db["jobsBlue"].update_many({"type": split_names}, {"$set": {"type": split_names_from_streaming}})
+
+ def down(self) -> None:
+ logging.info(
+ f"Rename unicity_id field from Job[{split_names_from_streaming}][<dataset>][<config>][<split>] to"
+ f" Job[{split_names}][<dataset>][<config>][<split>]"
+ )
+
+ for job in Job.objects(type=split_names_from_streaming):
+ job.update(unicity_id=f"Job[{split_names}][{job.dataset}][{job.config}][{job.split}]")
+
+ logging.info(f"Rename type field from {split_names_from_streaming} to {split_names}")
+ db = get_db("queue")
+ db["jobsBlue"].update_many({"type": split_names_from_streaming}, {"$set": {"type": split_names}})
+
+ def validate(self) -> None:
+ logging.info("Validate modified documents")
+
+ check_documents(DocCls=Job, sample_size=10)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 251f8c68..75a99799 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -100 +100 @@ class ProcessingGraphConfig:
- "/split-names": {"input_type": "config", "requires": "/config-names"},
+ "/split-names-from-streaming": {"input_type": "config", "requires": "/config-names"},
@@ -102 +102,5 @@ class ProcessingGraphConfig:
- "/first-rows": {"input_type": "split", "requires": "/split-names", "required_by_dataset_viewer": True},
+ "/first-rows": {
+ "input_type": "split",
+ "requires": "/split-names-from-streaming",
+ "required_by_dataset_viewer": True,
+ },
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index af2b57e8..02974524 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -13 +13 @@ def test_default_graph() -> None:
- split_names = graph.get_step("/split-names")
+ split_names_from_streaming = graph.get_step("/split-names-from-streaming")
@@ -24 +24 @@ def test_default_graph() -> None:
- assert config_names.children == [split_names]
+ assert config_names.children == [split_names_from_streaming]
@@ -27,4 +27,4 @@ def test_default_graph() -> None:
- assert split_names is not None
- assert split_names.parent is config_names
- assert split_names.children == [first_rows]
- assert split_names.get_ancestors() == [config_names]
+ assert split_names_from_streaming is not None
+ assert split_names_from_streaming.parent is config_names
+ assert split_names_from_streaming.children == [first_rows]
+ assert split_names_from_streaming.get_ancestors() == [config_names]
@@ -38 +38 @@ def test_default_graph() -> None:
- assert first_rows.parent is split_names
+ assert first_rows.parent is split_names_from_streaming
@@ -40 +40 @@ def test_default_graph() -> None:
- assert first_rows.get_ancestors() == [config_names, split_names]
+ assert first_rows.get_ancestors() == [config_names, split_names_from_streaming]
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 7d7be466..1207e9de 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -20 +19,0 @@ from worker.job_runners.sizes import SizesJobRunner
-from worker.job_runners.split_names import SplitNamesJobRunner
@@ -23,0 +23,3 @@ from worker.job_runners.split_names_from_dataset_info import (
+from worker.job_runners.split_names_from_streaming import (
+ SplitNamesFromStreamingJobRunner,
+)
@@ -67,2 +69,2 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- if job_type == SplitNamesJobRunner.get_job_type():
- return SplitNamesJobRunner(
+ if job_type == SplitNamesFromStreamingJobRunner.get_job_type():
+ return SplitNamesFromStreamingJobRunner(
@@ -129 +131 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- SplitNamesJobRunner.get_job_type(),
+ SplitNamesFromStreamingJobRunner.get_job_type(),
diff --git a/services/worker/src/worker/job_runners/split_names.py b/services/worker/src/worker/job_runners/split_names_from_streaming.py
similarity index 75%
rename from services/worker/src/worker/job_runners/split_names.py
rename to services/worker/src/worker/job_runners/split_names_from_streaming.py
index c485f640..b7cfdc84 100644
--- a/services/worker/src/worker/job_runners/split_names.py
+++ b/services/worker/src/worker/job_runners/split_names_from_streaming.py
@@ -15 +15 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
-SplitNamesJobRunnerErrorCode = Literal[
+SplitNamesFromStreamingJobRunnerErrorCode = Literal[
@@ -17 +17 @@ SplitNamesJobRunnerErrorCode = Literal[
- "SplitNamesError",
+ "SplitNamesFromStreamingError",
@@ -21 +21 @@ SplitNamesJobRunnerErrorCode = Literal[
-class SplitNamesJobRunnerError(JobRunnerError):
+class SplitNamesFromStreamingJobRunnerError(JobRunnerError):
@@ -28 +28 @@ class SplitNamesJobRunnerError(JobRunnerError):
- code: SplitNamesJobRunnerErrorCode,
+ code: SplitNamesFromStreamingJobRunnerErrorCode,
@@ -37 +37 @@ class SplitNamesJobRunnerError(JobRunnerError):
-class SplitNamesError(SplitNamesJobRunnerError):
+class SplitNamesFromStreamingError(SplitNamesFromStreamingJobRunnerError):
@@ -41 +41 @@ class SplitNamesError(SplitNamesJobRunnerError):
- super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "SplitNamesError", cause, True)
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "SplitNamesFromStreamingError", cause, True)
@@ -44 +44 @@ class SplitNamesError(SplitNamesJobRunnerError):
-class EmptyDatasetError(SplitNamesJobRunnerError):
+class EmptyDatasetError(SplitNamesFromStreamingJobRunnerError):
@@ -57 +57 @@ class SplitNameItem(TypedDict):
-class SplitNamesResponseContent(TypedDict):
+class SplitNamesFromStreamingResponseContent(TypedDict):
@@ -61 +61 @@ class SplitNamesResponseContent(TypedDict):
-def compute_split_names_response(
+def compute_split_names_from_streaming_response(
@@ -65 +65 @@ def compute_split_names_response(
-) -> SplitNamesResponseContent:
+) -> SplitNamesFromStreamingResponseContent:
@@ -67 +67 @@ def compute_split_names_response(
- Get the response of /split-names for one specific dataset and config on huggingface.co.
+ Get the response of /split-names-from-streaming for one specific dataset and config on huggingface.co.
@@ -76,2 +76,2 @@ def compute_split_names_response(
- The /split-names response generated by this function does not include stats about the split, like the size or
- number of samples. See /dataset-info or /sizes for that.
+ The /split-names-from-streaming response generated by this function does not include stats about the split,
+ like the size or number of samples. See /dataset-info or /sizes for that.
@@ -88 +88 @@ def compute_split_names_response(
- `SplitNamesResponseContent`: An object with the list of split names for the dataset and config.
+ `SplitNamesFromStreamingResponseContent`: An object with the list of split names for the dataset and config.
@@ -99 +99 @@ def compute_split_names_response(
- # get the list of splits in streaming mode
+
@@ -108 +108,3 @@ def compute_split_names_response(
- raise SplitNamesError("Cannot get the split names for the dataset and config.", cause=err) from err
+ raise SplitNamesFromStreamingError(
+ "Cannot get the split names for the dataset and config.", cause=err
+ ) from err
@@ -112 +114 @@ def compute_split_names_response(
-class SplitNamesJobRunner(DatasetsBasedJobRunner):
+class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner):
@@ -115 +117 @@ class SplitNamesJobRunner(DatasetsBasedJobRunner):
- return "/split-names"
+ return "/split-names-from-streaming"
@@ -124 +126 @@ class SplitNamesJobRunner(DatasetsBasedJobRunner):
- return compute_split_names_response(
+ return compute_split_names_from_streaming_response(
diff --git a/services/worker/tests/job_runners/test_split_names.py b/services/worker/tests/job_runners/test_split_names_from_streaming.py
similarity index 87%
rename from services/worker/tests/job_runners/test_split_names.py
rename to services/worker/tests/job_runners/test_split_names_from_streaming.py
index 4d45e534..8ace3ce6 100644
--- a/services/worker/tests/job_runners/test_split_names.py
+++ b/services/worker/tests/job_runners/test_split_names_from_streaming.py
@@ -16 +16,3 @@ from worker.config import AppConfig
-from worker.job_runners.split_names import SplitNamesJobRunner
+from worker.job_runners.split_names_from_streaming import (
+ SplitNamesFromStreamingJobRunner,
+)
@@ -21 +23 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
-GetJobRunner = Callable[[str, str, AppConfig, bool], SplitNamesJobRunner]
+GetJobRunner = Callable[[str, str, AppConfig, bool], SplitNamesFromStreamingJobRunner]
@@ -35,2 +37,2 @@ def get_job_runner(
- ) -> SplitNamesJobRunner:
- return SplitNamesJobRunner(
+ ) -> SplitNamesFromStreamingJobRunner:
+ return SplitNamesFromStreamingJobRunner(
@@ -38 +40 @@ def get_job_runner(
- "type": SplitNamesJobRunner.get_job_type(),
+ "type": SplitNamesFromStreamingJobRunner.get_job_type(),
@@ -48 +50 @@ def get_job_runner(
- endpoint=SplitNamesJobRunner.get_job_type(),
+ endpoint=SplitNamesFromStreamingJobRunner.get_job_type(),
@@ -95,3 +97,3 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No
- ("does_not_exist", False, "SplitNamesError", "FileNotFoundError"),
- ("gated", False, "SplitNamesError", "FileNotFoundError"),
- ("private", False, "SplitNamesError", "FileNotFoundError"),
+ ("does_not_exist", False, "SplitNamesFromStreamingError", "FileNotFoundError"),
+ ("gated", False, "SplitNamesFromStreamingError", "FileNotFoundError"),
+ ("private", False, "SplitNamesFromStreamingError", "FileNotFoundError"),
|
|
3a27146498c67ea9dc6dd8d8092005f55c6c961f
|
Quentin Lhoest
| 2023-02-16T22:54:14 |
Add heartbeat (#824)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 10c6abb3..884e9474 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -40 +40 @@ class QuerySetManager(Generic[U]):
-class Status(enum.Enum):
+class Status(str, enum.Enum):
@@ -49 +49 @@ class Status(enum.Enum):
-class Priority(enum.Enum):
+class Priority(str, enum.Enum):
@@ -66,0 +67 @@ class JobDict(TypedDict):
+ last_heartbeat: Optional[datetime]
@@ -123,0 +125 @@ class Job(Document):
+ last_heartbeat (`datetime`, optional): Last time the running job got a heartbeat from the worker.
@@ -150,0 +153 @@ class Job(Document):
+ last_heartbeat = DateTimeField()
@@ -165,0 +169 @@ class Job(Document):
+ "last_heartbeat": self.last_heartbeat,
diff --git a/services/worker/dev.Dockerfile b/services/worker/dev.Dockerfile
index 4fd090fd..2468dd10 100644
--- a/services/worker/dev.Dockerfile
+++ b/services/worker/dev.Dockerfile
@@ -33 +33,3 @@ WORKDIR /src/services/worker/
-RUN poetry install --no-cache
+RUN --mount=type=cache,target=/home/.cache/pypoetry/cache \
+ --mount=type=cache,target=/home/.cache/pypoetry/artifacts \
+ poetry install --no-root
@@ -41 +43 @@ RUN poetry install --no-cache
-ENTRYPOINT ["/bin/sh", "-c" , "poetry install && poetry run python src/worker/main.py"]
+ENTRYPOINT ["/bin/sh", "-c" , "poetry install --only-root && poetry run python src/worker/main.py"]
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 58eb2dcf..72d57d9f 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -2214,0 +2215,18 @@ files = [
+[[package]]
+name = "mirakuru"
+version = "2.4.2"
+description = "Process executor (not only) for tests."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "mirakuru-2.4.2-py3-none-any.whl", hash = "sha256:fdb67d141cc9f7abd485a515d618daf3272c3e6ff48380749997ff8e8c5f2cb2"},
+ {file = "mirakuru-2.4.2.tar.gz", hash = "sha256:ec84d4d81b4bca96cb0e598c6b3d198a92f036a0c1223c881482c02a98508226"},
+]
+
+[package.dependencies]
+psutil = {version = ">=4.0.0", markers = "sys_platform != \"cygwin\""}
+
+[package.extras]
+tests = ["pytest", "pytest-cov", "python-daemon"]
+
@@ -2859,0 +2878,7 @@ files = [
+ {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
+ {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
+ {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
+ {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
+ {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
+ {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
+ {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -4671,0 +4697,2 @@ files = [
+ {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"},
+ {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"},
@@ -4676,0 +4704 @@ files = [
+ {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"},
@@ -4684,0 +4713 @@ files = [
+ {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"},
@@ -4750,0 +4780 @@ url = "https://download.pytorch.org/whl/cpu/torch-1.13.1%2Bcpu-cp39-cp39-linux_x
+
@@ -4767,0 +4798 @@ url = "https://download.pytorch.org/whl/cpu/torchaudio-0.13.1%2Bcpu-cp39-cp39-li
+
@@ -5462 +5493 @@ python-versions = "3.9.15"
-content-hash = "4b1fde55056862f335ccae9046704cced7bf1d9599e54a00e5b35c5dde787528"
+content-hash = "c20f6820064117295d9db223142b9ccdafd534b12c3304e3fa51ccf3dbcd622f"
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index b885443c..76791302 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -42,0 +43 @@ wget = "^3.2"
+mirakuru = "^2.4.2"
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index b0ae7cda..b7ae5cc1 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -21,0 +22 @@ WORKER_SLEEP_SECONDS = 15
+WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS = 60
@@ -36,0 +38,2 @@ class WorkerConfig:
+ state_path: Optional[str] = None
+ heartbeat_time_interval_seconds: int = WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS
@@ -49,0 +53,4 @@ class WorkerConfig:
+ state_path=env.str(name="STATE_PATH", default=None),
+ heartbeat_time_interval_seconds=env.int(
+ name="HEARTBEAT_TIME_INTERVAL_SECONDS", default=WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS
+ ),
diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py
index e4cffbf6..24b013f1 100644
--- a/services/worker/src/worker/loop.py
+++ b/services/worker/src/worker/loop.py
@@ -3,0 +4 @@
+import json
@@ -7,0 +9 @@ from dataclasses import dataclass, field
+from typing import Optional, TypedDict
@@ -9 +11,2 @@ from dataclasses import dataclass, field
-from libcommon.queue import EmptyQueueError, Queue
+from filelock import FileLock
+from libcommon.queue import EmptyQueueError, JobInfo, Queue
@@ -19,0 +23,4 @@ class UnknownJobTypeError(Exception):
+class WorkerState(TypedDict):
+ current_job_info: Optional[JobInfo]
+
+
@@ -99 +106 @@ class Loop:
- logging.info("Worker started")
+ logging.info("Worker loop started")
@@ -121,0 +129 @@ class Loop:
+ self.set_worker_state(current_job_info=job_info)
@@ -123,0 +132 @@ class Loop:
+ self.set_worker_state(current_job_info=None)
@@ -129,0 +139 @@ class Loop:
+ self.set_worker_state(current_job_info=None)
@@ -131,0 +142,7 @@ class Loop:
+
+ def set_worker_state(self, current_job_info: Optional[JobInfo]) -> None:
+ worker_state: WorkerState = {"current_job_info": current_job_info}
+ if self.worker_config.state_path:
+ with FileLock(self.worker_config.state_path + ".lock"):
+ with open(self.worker_config.state_path, "w") as worker_state_f:
+ json.dump(worker_state, worker_state_f)
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index a7c1816d..439cefbe 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -1,2 +1,7 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
+import json
+import logging
+import os
+import sys
+import tempfile
+import time
+from typing import Optional
@@ -3,0 +9 @@
+from filelock import FileLock
@@ -5,3 +11,3 @@ from libcommon.log import init_logging
-from libcommon.processing_graph import ProcessingGraph
-from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.storage import init_assets_dir
+from libcommon.queue import Job, Status, get_datetime
+from libcommon.resources import QueueMongoResource
+from mirakuru import OutputExecutor
@@ -8,0 +15 @@ from libcommon.storage import init_assets_dir
+from worker import start_worker_loop
@@ -10,3 +17,58 @@ from worker.config import AppConfig
-from worker.job_runner_factory import JobRunnerFactory
-from worker.loop import Loop
-from worker.resources import LibrariesResource
+from worker.loop import WorkerState
+
+WORKER_STATE_FILE_NAME = "worker_state.json"
+START_WORKER_LOOP_PATH = start_worker_loop.__file__
+
+
+class WorkerExecutor:
+ def __init__(self, app_config: AppConfig) -> None:
+ self.app_config = app_config
+
+ def _create_worker_loop_executor(self) -> OutputExecutor:
+ banner = self.app_config.worker.state_path
+ if not banner:
+ raise ValueError("Failed to create the executor because WORKER_STATE_PATH is missing.")
+ start_worker_loop_command = [
+ sys.executable,
+ START_WORKER_LOOP_PATH,
+ "--print-worker-state-path",
+ ]
+ return OutputExecutor(start_worker_loop_command, banner, timeout=10)
+
+ def start(self) -> None:
+ worker_loop_executor = self._create_worker_loop_executor()
+ worker_loop_executor.start() # blocking until the banner is printed
+ logging.info("Starting heartbeat.")
+ while worker_loop_executor.running():
+ self.heartbeat()
+ time.sleep(self.app_config.worker.heartbeat_time_interval_seconds)
+ worker_loop_executor.stop()
+
+ def get_state(self) -> WorkerState:
+ worker_state_path = self.app_config.worker.state_path
+ if not worker_state_path:
+ raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
+ if os.path.exists(worker_state_path):
+ with FileLock(worker_state_path + ".lock"):
+ try:
+ with open(worker_state_path, "r") as worker_state_f:
+ worker_state = json.load(worker_state_f)
+ return WorkerState(current_job_info=worker_state.get("current_job_info"))
+ except json.JSONDecodeError:
+ return WorkerState(current_job_info=None)
+ else:
+ return WorkerState(current_job_info=None)
+
+ def get_current_job(self) -> Optional[Job]:
+ worker_state = self.get_state()
+ if worker_state["current_job_info"]:
+ job = Job.objects.with_id(worker_state["current_job_info"]["job_id"]) # type: ignore
+ if job and isinstance(job, Job) and job.status == Status.STARTED:
+ return job
+ return None
+
+ def heartbeat(self) -> None:
+ current_job = self.get_current_job()
+ if current_job:
+ current_job.update(last_heartbeat=get_datetime())
+
@@ -15,18 +77,8 @@ if __name__ == "__main__":
- app_config = AppConfig.from_env()
-
- init_logging(log_level=app_config.common.log_level)
- # ^ set first to have logs as soon as possible
- assets_directory = init_assets_dir(directory=app_config.assets.storage_directory)
-
- processing_graph = ProcessingGraph(app_config.processing_graph.specification)
-
- with (
- LibrariesResource(
- hf_endpoint=app_config.common.hf_endpoint,
- init_hf_datasets_cache=app_config.datasets_based.hf_datasets_cache,
- numba_path=app_config.numba.path,
- ) as libraries_resource,
- CacheMongoResource(
- database=app_config.cache.mongo_database, host=app_config.cache.mongo_url
- ) as cache_resource,
- QueueMongoResource(
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ if "WORKER_STATE_PATH" not in os.environ:
+ os.environ["WORKER_STATE_PATH"] = os.path.join(tmp_dir, WORKER_STATE_FILE_NAME)
+
+ app_config = AppConfig.from_env()
+ init_logging(log_level=app_config.common.log_level)
+
+ with QueueMongoResource(
@@ -34,20 +86,5 @@ if __name__ == "__main__":
- ) as queue_resource,
- ):
- if not cache_resource.is_available():
- raise RuntimeError("The connection to the cache database could not be established. Exiting.")
- if not queue_resource.is_available():
- raise RuntimeError("The connection to the queue database could not be established. Exiting.")
-
- job_runner_factory = JobRunnerFactory(
- app_config=app_config,
- processing_graph=processing_graph,
- hf_datasets_cache=libraries_resource.hf_datasets_cache,
- assets_directory=assets_directory,
- )
- loop = Loop(
- library_cache_paths=libraries_resource.storage_paths,
- job_runner_factory=job_runner_factory,
- max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace,
- worker_config=app_config.worker,
- )
- loop.run()
+ ) as queue_resource:
+ if not queue_resource.is_available():
+ raise RuntimeError("The connection to the queue database could not be established. Exiting.")
+ worker_executor = WorkerExecutor(app_config)
+ worker_executor.start()
diff --git a/services/worker/src/worker/start_worker_loop.py b/services/worker/src/worker/start_worker_loop.py
new file mode 100644
index 00000000..311971ee
--- /dev/null
+++ b/services/worker/src/worker/start_worker_loop.py
@@ -0,0 +1,57 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import sys
+
+from libcommon.log import init_logging
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.storage import init_assets_dir
+
+from worker.config import AppConfig
+from worker.job_runner_factory import JobRunnerFactory
+from worker.loop import Loop
+from worker.resources import LibrariesResource
+
+if __name__ == "__main__":
+ app_config = AppConfig.from_env()
+ if "--print-worker-state-path" in sys.argv:
+ print(app_config.worker.state_path, flush=True)
+
+ init_logging(log_level=app_config.common.log_level)
+ # ^ set first to have logs as soon as possible
+ assets_directory = init_assets_dir(directory=app_config.assets.storage_directory)
+
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+
+ with (
+ LibrariesResource(
+ hf_endpoint=app_config.common.hf_endpoint,
+ init_hf_datasets_cache=app_config.datasets_based.hf_datasets_cache,
+ numba_path=app_config.numba.path,
+ ) as libraries_resource,
+ CacheMongoResource(
+ database=app_config.cache.mongo_database, host=app_config.cache.mongo_url
+ ) as cache_resource,
+ QueueMongoResource(
+ database=app_config.queue.mongo_database, host=app_config.queue.mongo_url
+ ) as queue_resource,
+ ):
+ if not cache_resource.is_available():
+ raise RuntimeError("The connection to the cache database could not be established. Exiting.")
+ if not queue_resource.is_available():
+ raise RuntimeError("The connection to the queue database could not be established. Exiting.")
+
+ job_runner_factory = JobRunnerFactory(
+ app_config=app_config,
+ processing_graph=processing_graph,
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ assets_directory=assets_directory,
+ )
+ loop = Loop(
+ library_cache_paths=libraries_resource.storage_paths,
+ job_runner_factory=job_runner_factory,
+ max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace,
+ worker_config=app_config.worker,
+ )
+ loop.run()
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index 279ae7d6..eed12a27 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -14,0 +15 @@ from worker.config import AppConfig, FirstRowsConfig
+from worker.main import WORKER_STATE_FILE_NAME
@@ -29,0 +31,5 @@ def modules_cache_directory(tmp_path: Path) -> Path:
+@fixture
+def worker_state_path(tmp_path: Path) -> Path:
+ return tmp_path / WORKER_STATE_FILE_NAME
+
+
@@ -44 +50,3 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]:
-def set_env_vars(datasets_cache_directory: Path, modules_cache_directory: Path) -> Iterator[MonkeyPatch]:
+def set_env_vars(
+ datasets_cache_directory: Path, modules_cache_directory: Path, worker_state_path: Path
+) -> Iterator[MonkeyPatch]:
@@ -57,0 +66,2 @@ def set_env_vars(datasets_cache_directory: Path, modules_cache_directory: Path)
+ mp.setenv("WORKER_STATE_PATH", str(worker_state_path))
+ mp.setenv("WORKER_HEARTBEAT_TIME_INTERVAL_SECONDS", "1")
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
new file mode 100644
index 00000000..46f9ad5d
--- /dev/null
+++ b/services/worker/tests/test_executor.py
@@ -0,0 +1,180 @@
+import json
+import os
+import sys
+import time
+from datetime import timedelta
+from pathlib import Path
+from typing import Iterator
+from unittest.mock import patch
+
+import pytest
+import pytz
+from filelock import FileLock
+from libcommon.queue import Job, JobInfo, Priority, Status, get_datetime
+from libcommon.resources import QueueMongoResource
+from mirakuru import ProcessExitedWithError, TimeoutExpired
+from pytest import fixture
+
+from worker.config import AppConfig
+from worker.loop import WorkerState
+from worker.main import WorkerExecutor
+
+
+def get_job_info() -> JobInfo:
+ return JobInfo(
+ job_id="a" * 24,
+ type="bar",
+ dataset="user/my_dataset",
+ config="default",
+ split="train",
+ force=False,
+ priority=Priority.LOW,
+ )
+
+
+def write_worker_state(worker_state: WorkerState, worker_state_path: str) -> None:
+ with FileLock(worker_state_path + ".lock"):
+ with open(worker_state_path, "w") as worker_state_f:
+ json.dump(worker_state, worker_state_f)
+
+
+def start_worker_loop() -> None:
+ app_config = AppConfig.from_env()
+ if not app_config.worker.state_path:
+ raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
+ if "--print-worker-state-path" in sys.argv:
+ print(app_config.worker.state_path, flush=True)
+ current_job_info = get_job_info()
+ worker_state = WorkerState(current_job_info=current_job_info)
+ write_worker_state(worker_state, app_config.worker.state_path)
+
+
+def start_worker_loop_that_crashes() -> None:
+ app_config = AppConfig.from_env()
+ if not app_config.worker.state_path:
+ raise ValueError("Failed to get worker state because WORKER_STATE_PATH is missing.")
+ if "--print-worker-state-path" in sys.argv:
+ print(app_config.worker.state_path, flush=True)
+ raise RuntimeError("Tried to run a bad worker loop")
+
+
+def start_worker_loop_that_times_out() -> None:
+ time.sleep(20)
+
+
+@fixture
+def set_worker_state(worker_state_path: Path) -> Iterator[WorkerState]:
+ job_info = get_job_info()
+ worker_state = WorkerState(current_job_info=job_info)
+ write_worker_state(worker_state, str(worker_state_path))
+ yield worker_state
+ os.remove(worker_state_path)
+
+
+@fixture
+def set_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
+ if not queue_mongo_resource.is_available():
+ raise RuntimeError("Mongo resource is not available")
+ job_info = get_job_info()
+ if Job.objects.with_id(job_info["job_id"]): # type: ignore
+ Job.objects.with_id(job_info["job_id"]).delete() # type: ignore
+ job = Job(
+ pk=job_info["job_id"],
+ type=job_info["type"],
+ dataset=job_info["dataset"],
+ config=job_info["config"],
+ split=job_info["split"],
+ unicity_id="unicity_id",
+ namespace="user",
+ priority=job_info["priority"],
+ status=Status.STARTED,
+ created_at=get_datetime(),
+ )
+ job.save()
+ yield job
+ job.delete()
+
+
+def test_executor_get_state(app_config: AppConfig, set_worker_state: WorkerState) -> None:
+ executor = WorkerExecutor(app_config)
+ assert executor.get_state() == set_worker_state
+
+
+def test_executor_get_empty_state(app_config: AppConfig) -> None:
+ executor = WorkerExecutor(app_config)
+ assert executor.get_state() == WorkerState(current_job_info=None)
+
+
+def test_executor_get_current_job(
+ app_config: AppConfig, set_started_job_in_queue: Job, set_worker_state: WorkerState
+) -> None:
+ executor = WorkerExecutor(app_config)
+ assert executor.get_current_job() == set_started_job_in_queue
+
+
+def test_executor_get_nonexisting_current_job(app_config: AppConfig) -> None:
+ executor = WorkerExecutor(app_config)
+ assert executor.get_current_job() is None
+
+
+def test_executor_heartbeat(
+ app_config: AppConfig,
+ set_started_job_in_queue: Job,
+ set_worker_state: WorkerState,
+ queue_mongo_resource: QueueMongoResource,
+) -> None:
+ if not queue_mongo_resource.is_available():
+ raise RuntimeError("Mongo resource is not available")
+ executor = WorkerExecutor(app_config)
+ current_job = executor.get_current_job()
+ assert current_job is not None
+ assert current_job.last_heartbeat is None
+ executor.heartbeat()
+ current_job = executor.get_current_job()
+ assert current_job is not None
+ assert current_job.last_heartbeat is not None
+ last_heartbeat_datetime = pytz.UTC.localize(current_job.last_heartbeat)
+ assert last_heartbeat_datetime >= get_datetime() - timedelta(seconds=1)
+
+
+def test_executor_start(
+ app_config: AppConfig, queue_mongo_resource: QueueMongoResource, set_started_job_in_queue: Job
+) -> None:
+ if not queue_mongo_resource.is_available():
+ raise RuntimeError("Mongo resource is not available")
+ executor = WorkerExecutor(app_config)
+ with patch.object(executor, "heartbeat", wraps=executor.heartbeat) as heartbeat_mock:
+ with patch("worker.main.START_WORKER_LOOP_PATH", __file__):
+ executor.start()
+ current_job = executor.get_current_job()
+ assert current_job is not None
+ assert str(current_job.pk) == get_job_info()["job_id"]
+ assert heartbeat_mock.call_count > 0
+
+
[email protected](
+ "bad_worker_loop_type", ["start_worker_loop_that_crashes", "start_worker_loop_that_times_out"]
+)
+def test_executor_raises_on_bad_worker(
+ app_config: AppConfig, queue_mongo_resource: QueueMongoResource, tmp_path: Path, bad_worker_loop_type: str
+) -> None:
+ if not queue_mongo_resource.is_available():
+ raise RuntimeError("Mongo resource is not available")
+ bad_start_worker_loop_path = tmp_path / "bad_start_worker_loop.py"
+ with bad_start_worker_loop_path.open("w") as bad_start_worker_loop_f:
+ bad_start_worker_loop_f.write("raise RuntimeError('Tried to start a bad worker loop.')")
+ executor = WorkerExecutor(app_config)
+ with patch.dict(os.environ, {"WORKER_LOOP_TYPE": bad_worker_loop_type}):
+ with patch("worker.main.START_WORKER_LOOP_PATH", __file__):
+ with pytest.raises((ProcessExitedWithError, TimeoutExpired)):
+ executor.start()
+
+
+if __name__ == "__main__":
+ worker_loop_type = os.environ.get("WORKER_LOOP_TYPE", "start_worker_loop")
+ if worker_loop_type == "start_worker_loop_that_crashes":
+ start_worker_loop_that_crashes()
+ elif worker_loop_type == "start_worker_loop_that_times_out":
+ start_worker_loop_that_times_out()
+ else:
+ start_worker_loop()
|
|
502386edbf6edd63465cc8a6759e4239d35b7cad
|
Andrea Francis Soria Jimenez
| 2023-02-15T17:03:32 |
Adding new job runner for split names based on dataset info cached response (#820)
|
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index d13466d7..251f8c68 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -105,0 +106 @@ class ProcessingGraphConfig:
+ "/split-names-from-dataset-info": {"input_type": "config", "requires": "/dataset-info"},
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index e51e0d49..af2b57e8 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -19,0 +20 @@ def test_default_graph() -> None:
+ split_names_from_dataset_info = graph.get_step("/split-names-from-dataset-info")
@@ -53 +54 @@ def test_default_graph() -> None:
- assert dataset_info.children == []
+ assert dataset_info.children == [split_names_from_dataset_info]
@@ -55,0 +57,5 @@ def test_default_graph() -> None:
+ assert split_names_from_dataset_info is not None
+ assert split_names_from_dataset_info.parent is dataset_info
+ assert split_names_from_dataset_info.children == []
+ assert split_names_from_dataset_info.get_ancestors() == [parquet_and_dataset_info, dataset_info]
+
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index ec516d5e..7d7be466 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -20,0 +21,3 @@ from worker.job_runners.split_names import SplitNamesJobRunner
+from worker.job_runners.split_names_from_dataset_info import (
+ SplitNamesFromDatasetInfoJobRunner,
+)
@@ -116,0 +120,7 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ if job_type == SplitNamesFromDatasetInfoJobRunner.get_job_type():
+ return SplitNamesFromDatasetInfoJobRunner(
+ job_info=job_info,
+ app_config=self.app_config,
+ processing_step=processing_step,
+ hf_datasets_cache=self.hf_datasets_cache,
+ )
@@ -125,0 +136 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ SplitNamesFromDatasetInfoJobRunner.get_job_type(),
diff --git a/services/worker/src/worker/job_runners/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
new file mode 100644
index 00000000..be24011f
--- /dev/null
+++ b/services/worker/src/worker/job_runners/split_names_from_dataset_info.py
@@ -0,0 +1,125 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+from http import HTTPStatus
+from typing import Any, List, Literal, Mapping, Optional, TypedDict
+
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.simple_cache import DoesNotExist, get_response
+
+from worker.job_runner import JobRunnerError
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+
+SplitNamesFromDatasetInfoJobRunnerErrorCode = Literal[
+ "PreviousStepStatusError",
+ "PreviousStepFormatError",
+]
+
+
+class SplitNamesFromDatasetInfoJobRunnerError(JobRunnerError):
+ """Base class for split names job runner exceptions."""
+
+ def __init__(
+ self,
+ message: str,
+ status_code: HTTPStatus,
+ code: SplitNamesFromDatasetInfoJobRunnerErrorCode,
+ cause: Optional[BaseException] = None,
+ disclose_cause: bool = False,
+ ):
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
+
+
+class PreviousStepStatusError(SplitNamesFromDatasetInfoJobRunnerError):
+ """Raised when the previous step gave an error. The job should not have been created."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False)
+
+
+class PreviousStepFormatError(SplitNamesFromDatasetInfoJobRunnerError):
+ """Raised when the content of the previous step has not the expected format."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False)
+
+
+class SplitNameItem(TypedDict):
+ dataset: str
+ config: str
+ split: str
+
+
+class SplitNamesFromDatasetInfoResponseContent(TypedDict):
+ split_names: List[SplitNameItem]
+
+
+def compute_split_names_from_dataset_info_response(
+ dataset: str,
+ config: str,
+) -> SplitNamesFromDatasetInfoResponseContent:
+ """
+ Get the response of /split-names-from-dataset-info for one specific dataset and config on huggingface.co
+ computed from cached response in /dataset-info step.
+
+ The /split-names-from-dataset-info response generated by this function does not include stats about the split,
+ like the size or number of samples. See /dataset-info or /sizes for that.
+
+ Args:
+ dataset (`str`):
+ A namespace (user or an organization) and a repo name separated
+ by a `/`.
+ config (`str`):
+ A configuration name.
+ Returns:
+ `SplitNamesFromDatasetInfoResponseContent`: An object with the list of split names for the dataset and config.
+ <Tip>
+ Raises the following errors:
+ - [`~job_runners.split_names_from_dataset_info.PreviousStepStatusError`]
+ If the the previous step gave an error.
+ - [`~job_runners.split_names_from_dataset_info.PreviousStepFormatError`]
+ If the content of the previous step has not the expected format
+ - [`~libcommon.dataset.DatasetNotFoundError`]
+ If previous step content was not found for the dataset
+ </Tip>
+ """
+ logging.info(f"get split names from dataset info for dataset={dataset}, config={config}")
+ try:
+ response = get_response(kind="/dataset-info", dataset=dataset)
+ except DoesNotExist as e:
+ raise DatasetNotFoundError("No response found in previous step for this dataset.", e) from e
+ if response["http_status"] != HTTPStatus.OK:
+ raise PreviousStepStatusError(
+ f"Previous step gave an error: {response['http_status']}. This job should not have been created."
+ )
+
+ try:
+ splits_content = response["content"]["dataset_info"][config]["splits"]
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.") from e
+
+ split_name_items: List[SplitNameItem] = [
+ {"dataset": dataset, "config": config, "split": str(split)} for split in splits_content
+ ]
+
+ return {"split_names": split_name_items}
+
+
+class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner):
+ @staticmethod
+ def get_job_type() -> str:
+ return "/split-names-from-dataset-info"
+
+ @staticmethod
+ def get_version() -> str:
+ return "1.0.0"
+
+ def compute(self) -> Mapping[str, Any]:
+ if self.dataset is None:
+ raise ValueError("dataset is required")
+ if self.config is None:
+ raise ValueError("config is required")
+ return compute_split_names_from_dataset_info_response(dataset=self.dataset, config=self.config)
diff --git a/services/worker/tests/job_runners/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
new file mode 100644
index 00000000..31253213
--- /dev/null
+++ b/services/worker/tests/job_runners/test_split_names_from_dataset_info.py
@@ -0,0 +1,146 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from http import HTTPStatus
+from typing import Any, Callable
+
+import pytest
+from libcommon.dataset import DatasetNotFoundError
+from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import upsert_response
+
+from worker.config import AppConfig
+from worker.job_runners.split_names_from_dataset_info import (
+ PreviousStepFormatError,
+ PreviousStepStatusError,
+ SplitNamesFromDatasetInfoJobRunner,
+)
+from worker.resources import LibrariesResource
+
+GetJobRunner = Callable[[str, str, AppConfig, bool], SplitNamesFromDatasetInfoJobRunner]
+
+
[email protected]
+def get_job_runner(
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ config: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> SplitNamesFromDatasetInfoJobRunner:
+ return SplitNamesFromDatasetInfoJobRunner(
+ job_info={
+ "type": SplitNamesFromDatasetInfoJobRunner.get_job_type(),
+ "dataset": dataset,
+ "config": config,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ endpoint=SplitNamesFromDatasetInfoJobRunner.get_job_type(),
+ input_type="config",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_job_runner
+
+
[email protected](
+ "dataset,upstream_status,upstream_content,error_code,content",
+ [
+ (
+ "ok",
+ HTTPStatus.OK,
+ {
+ "dataset_info": {
+ "config_name": {
+ "splits": {
+ "train": {"name": "train", "dataset_name": "ok"},
+ "validation": {"name": "validation", "dataset_name": "ok"},
+ "test": {"name": "test", "dataset_name": "ok"},
+ },
+ }
+ }
+ },
+ None,
+ {
+ "split_names": [
+ {"dataset": "ok", "config": "config_name", "split": "train"},
+ {"dataset": "ok", "config": "config_name", "split": "validation"},
+ {"dataset": "ok", "config": "config_name", "split": "test"},
+ ]
+ },
+ ),
+ (
+ "upstream_fail",
+ HTTPStatus.INTERNAL_SERVER_ERROR,
+ {"error": "error"},
+ PreviousStepStatusError.__name__,
+ None,
+ ),
+ (
+ "without_dataset_info",
+ HTTPStatus.OK,
+ {"some_column": "wrong_format"},
+ PreviousStepFormatError.__name__,
+ None,
+ ),
+ (
+ "without_config_name",
+ HTTPStatus.OK,
+ {"dataset_info": "wrong_format"},
+ PreviousStepFormatError.__name__,
+ None,
+ ),
+ (
+ "without_splits",
+ HTTPStatus.OK,
+ {"dataset_info": {"config_name": "wrong_format"}},
+ PreviousStepFormatError.__name__,
+ None,
+ ),
+ ],
+)
+def test_compute(
+ app_config: AppConfig,
+ get_job_runner: GetJobRunner,
+ dataset: str,
+ upstream_status: HTTPStatus,
+ upstream_content: Any,
+ error_code: str,
+ content: Any,
+) -> None:
+ upsert_response(kind="/dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status)
+ job_runner = get_job_runner(dataset, "config_name", app_config, False)
+ if error_code:
+ with pytest.raises(Exception) as e:
+ job_runner.compute()
+ assert e.type.__name__ == error_code
+ else:
+ assert job_runner.compute() == content
+
+
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset = "non_existent"
+ config = "non_existent"
+ worker = get_job_runner(dataset, config, app_config, False)
+ with pytest.raises(CustomError) as exc_info:
+ worker.compute()
+ assert exc_info.value.status_code == HTTPStatus.NOT_FOUND
+ assert exc_info.value.code == DatasetNotFoundError.__name__
|
|
c12637e5eb8ee2765ff7d0da651a22a00ccadfee
|
Sylvain Lesage
| 2023-02-15T15:32:56 |
Update dependencies (#823)
|
diff --git a/e2e/poetry.lock b/e2e/poetry.lock
index c426e46f..a611e9bd 100644
--- a/e2e/poetry.lock
+++ b/e2e/poetry.lock
@@ -238,0 +239,64 @@ files = [
+[[package]]
+name = "coverage"
+version = "7.1.0"
+description = "Code coverage measurement for Python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "coverage-7.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:3b946bbcd5a8231383450b195cfb58cb01cbe7f8949f5758566b881df4b33baf"},
+ {file = "coverage-7.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ec8e767f13be637d056f7e07e61d089e555f719b387a7070154ad80a0ff31801"},
+ {file = "coverage-7.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4a5a5879a939cb84959d86869132b00176197ca561c664fc21478c1eee60d75"},
+ {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b643cb30821e7570c0aaf54feaf0bfb630b79059f85741843e9dc23f33aaca2c"},
+ {file = "coverage-7.1.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:32df215215f3af2c1617a55dbdfb403b772d463d54d219985ac7cd3bf124cada"},
+ {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:33d1ae9d4079e05ac4cc1ef9e20c648f5afabf1a92adfaf2ccf509c50b85717f"},
+ {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:29571503c37f2ef2138a306d23e7270687c0efb9cab4bd8038d609b5c2393a3a"},
+ {file = "coverage-7.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:63ffd21aa133ff48c4dff7adcc46b7ec8b565491bfc371212122dd999812ea1c"},
+ {file = "coverage-7.1.0-cp310-cp310-win32.whl", hash = "sha256:4b14d5e09c656de5038a3f9bfe5228f53439282abcab87317c9f7f1acb280352"},
+ {file = "coverage-7.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:8361be1c2c073919500b6601220a6f2f98ea0b6d2fec5014c1d9cfa23dd07038"},
+ {file = "coverage-7.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:da9b41d4539eefd408c46725fb76ecba3a50a3367cafb7dea5f250d0653c1040"},
+ {file = "coverage-7.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:c5b15ed7644ae4bee0ecf74fee95808dcc34ba6ace87e8dfbf5cb0dc20eab45a"},
+ {file = "coverage-7.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d12d076582507ea460ea2a89a8c85cb558f83406c8a41dd641d7be9a32e1274f"},
+ {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2617759031dae1bf183c16cef8fcfb3de7617f394c813fa5e8e46e9b82d4222"},
+ {file = "coverage-7.1.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4e4881fa9e9667afcc742f0c244d9364d197490fbc91d12ac3b5de0bf2df146"},
+ {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:9d58885215094ab4a86a6aef044e42994a2bd76a446dc59b352622655ba6621b"},
+ {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:ffeeb38ee4a80a30a6877c5c4c359e5498eec095878f1581453202bfacc8fbc2"},
+ {file = "coverage-7.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3baf5f126f30781b5e93dbefcc8271cb2491647f8283f20ac54d12161dff080e"},
+ {file = "coverage-7.1.0-cp311-cp311-win32.whl", hash = "sha256:ded59300d6330be27bc6cf0b74b89ada58069ced87c48eaf9344e5e84b0072f7"},
+ {file = "coverage-7.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:6a43c7823cd7427b4ed763aa7fb63901ca8288591323b58c9cd6ec31ad910f3c"},
+ {file = "coverage-7.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:7a726d742816cb3a8973c8c9a97539c734b3a309345236cd533c4883dda05b8d"},
+ {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bc7c85a150501286f8b56bd8ed3aa4093f4b88fb68c0843d21ff9656f0009d6a"},
+ {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f5b4198d85a3755d27e64c52f8c95d6333119e49fd001ae5798dac872c95e0f8"},
+ {file = "coverage-7.1.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ddb726cb861c3117a553f940372a495fe1078249ff5f8a5478c0576c7be12050"},
+ {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:51b236e764840a6df0661b67e50697aaa0e7d4124ca95e5058fa3d7cbc240b7c"},
+ {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:7ee5c9bb51695f80878faaa5598040dd6c9e172ddcf490382e8aedb8ec3fec8d"},
+ {file = "coverage-7.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:c31b75ae466c053a98bf26843563b3b3517b8f37da4d47b1c582fdc703112bc3"},
+ {file = "coverage-7.1.0-cp37-cp37m-win32.whl", hash = "sha256:3b155caf3760408d1cb903b21e6a97ad4e2bdad43cbc265e3ce0afb8e0057e73"},
+ {file = "coverage-7.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2a60d6513781e87047c3e630b33b4d1e89f39836dac6e069ffee28c4786715f5"},
+ {file = "coverage-7.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f2cba5c6db29ce991029b5e4ac51eb36774458f0a3b8d3137241b32d1bb91f06"},
+ {file = "coverage-7.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:beeb129cacea34490ffd4d6153af70509aa3cda20fdda2ea1a2be870dfec8d52"},
+ {file = "coverage-7.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c45948f613d5d18c9ec5eaa203ce06a653334cf1bd47c783a12d0dd4fd9c851"},
+ {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef382417db92ba23dfb5864a3fc9be27ea4894e86620d342a116b243ade5d35d"},
+ {file = "coverage-7.1.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7c7c0d0827e853315c9bbd43c1162c006dd808dbbe297db7ae66cd17b07830f0"},
+ {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e5cdbb5cafcedea04924568d990e20ce7f1945a1dd54b560f879ee2d57226912"},
+ {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:9817733f0d3ea91bea80de0f79ef971ae94f81ca52f9b66500c6a2fea8e4b4f8"},
+ {file = "coverage-7.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:218fe982371ac7387304153ecd51205f14e9d731b34fb0568181abaf7b443ba0"},
+ {file = "coverage-7.1.0-cp38-cp38-win32.whl", hash = "sha256:04481245ef966fbd24ae9b9e537ce899ae584d521dfbe78f89cad003c38ca2ab"},
+ {file = "coverage-7.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ae125d1134bf236acba8b83e74c603d1b30e207266121e76484562bc816344c"},
+ {file = "coverage-7.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2bf1d5f2084c3932b56b962a683074a3692bce7cabd3aa023c987a2a8e7612f6"},
+ {file = "coverage-7.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:98b85dd86514d889a2e3dd22ab3c18c9d0019e696478391d86708b805f4ea0fa"},
+ {file = "coverage-7.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38da2db80cc505a611938d8624801158e409928b136c8916cd2e203970dde4dc"},
+ {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3164d31078fa9efe406e198aecd2a02d32a62fecbdef74f76dad6a46c7e48311"},
+ {file = "coverage-7.1.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db61a79c07331e88b9a9974815c075fbd812bc9dbc4dc44b366b5368a2936063"},
+ {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9ccb092c9ede70b2517a57382a601619d20981f56f440eae7e4d7eaafd1d1d09"},
+ {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:33ff26d0f6cc3ca8de13d14fde1ff8efe1456b53e3f0273e63cc8b3c84a063d8"},
+ {file = "coverage-7.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d47dd659a4ee952e90dc56c97d78132573dc5c7b09d61b416a9deef4ebe01a0c"},
+ {file = "coverage-7.1.0-cp39-cp39-win32.whl", hash = "sha256:d248cd4a92065a4d4543b8331660121b31c4148dd00a691bfb7a5cdc7483cfa4"},
+ {file = "coverage-7.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:7ed681b0f8e8bcbbffa58ba26fcf5dbc8f79e7997595bf071ed5430d8c08d6f3"},
+ {file = "coverage-7.1.0-pp37.pp38.pp39-none-any.whl", hash = "sha256:755e89e32376c850f826c425ece2c35a4fc266c081490eb0a841e7c1cb0d3bda"},
+ {file = "coverage-7.1.0.tar.gz", hash = "sha256:10188fe543560ec4874f974b5305cd1a8bdcfa885ee00ea3a03733464c4ca265"},
+]
+
+[package.extras]
+toml = ["tomli"]
+
@@ -414 +478 @@ name = "isort"
-version = "5.11.4"
+version = "5.12.0"
@@ -418 +482 @@ optional = false
-python-versions = ">=3.7.0"
+python-versions = ">=3.8.0"
@@ -420,2 +484,2 @@ files = [
- {file = "isort-5.11.4-py3-none-any.whl", hash = "sha256:c033fd0edb91000a7f09527fe5c75321878f98322a77ddcc81adbd83724afb7b"},
- {file = "isort-5.11.4.tar.gz", hash = "sha256:6db30c5ded9815d813932c04c2f85a360bcdd35fed496f4d8f35495ef0a261b6"},
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
@@ -425,2 +489,2 @@ files = [
-colors = ["colorama (>=0.4.3,<0.5.0)"]
-pipfile-deprecated-finder = ["pipreqs", "requirementslib"]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
@@ -449,0 +514,64 @@ format-nongpl = ["fqdn", "idna", "isoduration", "jsonpointer (>1.13)", "rfc3339-
+[[package]]
+name = "jsonschema-spec"
+version = "0.1.3"
+description = "JSONSchema Spec with object-oriented paths"
+category = "main"
+optional = false
+python-versions = ">=3.7.0,<4.0.0"
+files = [
+ {file = "jsonschema_spec-0.1.3-py3-none-any.whl", hash = "sha256:b3cde007ad65c2e631e2f8653cf187124a2c714d02d9fafbab68ad64bf5745d6"},
+ {file = "jsonschema_spec-0.1.3.tar.gz", hash = "sha256:8d8db7c255e524fab1016a952a9143e5b6e3c074f4ed25d1878f8e97806caec0"},
+]
+
+[package.dependencies]
+jsonschema = ">=4.0.0,<5.0.0"
+pathable = ">=0.4.1,<0.5.0"
+PyYAML = ">=5.1"
+typing-extensions = ">=4.3.0,<5.0.0"
+
+[[package]]
+name = "lazy-object-proxy"
+version = "1.9.0"
+description = "A fast and thorough lazy object proxy."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "lazy-object-proxy-1.9.0.tar.gz", hash = "sha256:659fb5809fa4629b8a1ac5106f669cfc7bef26fbb389dda53b3e010d1ac4ebae"},
+ {file = "lazy_object_proxy-1.9.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b40387277b0ed2d0602b8293b94d7257e17d1479e257b4de114ea11a8cb7f2d7"},
+ {file = "lazy_object_proxy-1.9.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8c6cfb338b133fbdbc5cfaa10fe3c6aeea827db80c978dbd13bc9dd8526b7d4"},
+ {file = "lazy_object_proxy-1.9.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:721532711daa7db0d8b779b0bb0318fa87af1c10d7fe5e52ef30f8eff254d0cd"},
+ {file = "lazy_object_proxy-1.9.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:66a3de4a3ec06cd8af3f61b8e1ec67614fbb7c995d02fa224813cb7afefee701"},
+ {file = "lazy_object_proxy-1.9.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:1aa3de4088c89a1b69f8ec0dcc169aa725b0ff017899ac568fe44ddc1396df46"},
+ {file = "lazy_object_proxy-1.9.0-cp310-cp310-win32.whl", hash = "sha256:f0705c376533ed2a9e5e97aacdbfe04cecd71e0aa84c7c0595d02ef93b6e4455"},
+ {file = "lazy_object_proxy-1.9.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea806fd4c37bf7e7ad82537b0757999264d5f70c45468447bb2b91afdbe73a6e"},
+ {file = "lazy_object_proxy-1.9.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:946d27deaff6cf8452ed0dba83ba38839a87f4f7a9732e8f9fd4107b21e6ff07"},
+ {file = "lazy_object_proxy-1.9.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:79a31b086e7e68b24b99b23d57723ef7e2c6d81ed21007b6281ebcd1688acb0a"},
+ {file = "lazy_object_proxy-1.9.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f699ac1c768270c9e384e4cbd268d6e67aebcfae6cd623b4d7c3bfde5a35db59"},
+ {file = "lazy_object_proxy-1.9.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bfb38f9ffb53b942f2b5954e0f610f1e721ccebe9cce9025a38c8ccf4a5183a4"},
+ {file = "lazy_object_proxy-1.9.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:189bbd5d41ae7a498397287c408617fe5c48633e7755287b21d741f7db2706a9"},
+ {file = "lazy_object_proxy-1.9.0-cp311-cp311-win32.whl", hash = "sha256:81fc4d08b062b535d95c9ea70dbe8a335c45c04029878e62d744bdced5141586"},
+ {file = "lazy_object_proxy-1.9.0-cp311-cp311-win_amd64.whl", hash = "sha256:f2457189d8257dd41ae9b434ba33298aec198e30adf2dcdaaa3a28b9994f6adb"},
+ {file = "lazy_object_proxy-1.9.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:d9e25ef10a39e8afe59a5c348a4dbf29b4868ab76269f81ce1674494e2565a6e"},
+ {file = "lazy_object_proxy-1.9.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cbf9b082426036e19c6924a9ce90c740a9861e2bdc27a4834fd0a910742ac1e8"},
+ {file = "lazy_object_proxy-1.9.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f5fa4a61ce2438267163891961cfd5e32ec97a2c444e5b842d574251ade27d2"},
+ {file = "lazy_object_proxy-1.9.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:8fa02eaab317b1e9e03f69aab1f91e120e7899b392c4fc19807a8278a07a97e8"},
+ {file = "lazy_object_proxy-1.9.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:e7c21c95cae3c05c14aafffe2865bbd5e377cfc1348c4f7751d9dc9a48ca4bda"},
+ {file = "lazy_object_proxy-1.9.0-cp37-cp37m-win32.whl", hash = "sha256:f12ad7126ae0c98d601a7ee504c1122bcef553d1d5e0c3bfa77b16b3968d2734"},
+ {file = "lazy_object_proxy-1.9.0-cp37-cp37m-win_amd64.whl", hash = "sha256:edd20c5a55acb67c7ed471fa2b5fb66cb17f61430b7a6b9c3b4a1e40293b1671"},
+ {file = "lazy_object_proxy-1.9.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:2d0daa332786cf3bb49e10dc6a17a52f6a8f9601b4cf5c295a4f85854d61de63"},
+ {file = "lazy_object_proxy-1.9.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9cd077f3d04a58e83d04b20e334f678c2b0ff9879b9375ed107d5d07ff160171"},
+ {file = "lazy_object_proxy-1.9.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:660c94ea760b3ce47d1855a30984c78327500493d396eac4dfd8bd82041b22be"},
+ {file = "lazy_object_proxy-1.9.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:212774e4dfa851e74d393a2370871e174d7ff0ebc980907723bb67d25c8a7c30"},
+ {file = "lazy_object_proxy-1.9.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:f0117049dd1d5635bbff65444496c90e0baa48ea405125c088e93d9cf4525b11"},
+ {file = "lazy_object_proxy-1.9.0-cp38-cp38-win32.whl", hash = "sha256:0a891e4e41b54fd5b8313b96399f8b0e173bbbfc03c7631f01efbe29bb0bcf82"},
+ {file = "lazy_object_proxy-1.9.0-cp38-cp38-win_amd64.whl", hash = "sha256:9990d8e71b9f6488e91ad25f322898c136b008d87bf852ff65391b004da5e17b"},
+ {file = "lazy_object_proxy-1.9.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:9e7551208b2aded9c1447453ee366f1c4070602b3d932ace044715d89666899b"},
+ {file = "lazy_object_proxy-1.9.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5f83ac4d83ef0ab017683d715ed356e30dd48a93746309c8f3517e1287523ef4"},
+ {file = "lazy_object_proxy-1.9.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7322c3d6f1766d4ef1e51a465f47955f1e8123caee67dd641e67d539a534d006"},
+ {file = "lazy_object_proxy-1.9.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:18b78ec83edbbeb69efdc0e9c1cb41a3b1b1ed11ddd8ded602464c3fc6020494"},
+ {file = "lazy_object_proxy-1.9.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:09763491ce220c0299688940f8dc2c5d05fd1f45af1e42e636b2e8b2303e4382"},
+ {file = "lazy_object_proxy-1.9.0-cp39-cp39-win32.whl", hash = "sha256:9090d8e53235aa280fc9239a86ae3ea8ac58eff66a705fa6aa2ec4968b95c821"},
+ {file = "lazy_object_proxy-1.9.0-cp39-cp39-win_amd64.whl", hash = "sha256:db1c1722726f47e10e0b5fdbf15ac3b8adb58c091d12b3ab713965795036985f"},
+]
+
@@ -575 +703 @@ name = "mypy"
-version = "0.812"
+version = "1.0.0"
@@ -579 +707 @@ optional = false
-python-versions = ">=3.5"
+python-versions = ">=3.7"
@@ -581,22 +709,26 @@ files = [
- {file = "mypy-0.812-cp35-cp35m-macosx_10_9_x86_64.whl", hash = "sha256:a26f8ec704e5a7423c8824d425086705e381b4f1dfdef6e3a1edab7ba174ec49"},
- {file = "mypy-0.812-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:28fb5479c494b1bab244620685e2eb3c3f988d71fd5d64cc753195e8ed53df7c"},
- {file = "mypy-0.812-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:9743c91088d396c1a5a3c9978354b61b0382b4e3c440ce83cf77994a43e8c521"},
- {file = "mypy-0.812-cp35-cp35m-win_amd64.whl", hash = "sha256:d7da2e1d5f558c37d6e8c1246f1aec1e7349e4913d8fb3cb289a35de573fe2eb"},
- {file = "mypy-0.812-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:4eec37370483331d13514c3f55f446fc5248d6373e7029a29ecb7b7494851e7a"},
- {file = "mypy-0.812-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:d65cc1df038ef55a99e617431f0553cd77763869eebdf9042403e16089fe746c"},
- {file = "mypy-0.812-cp36-cp36m-manylinux2010_x86_64.whl", hash = "sha256:61a3d5b97955422964be6b3baf05ff2ce7f26f52c85dd88db11d5e03e146a3a6"},
- {file = "mypy-0.812-cp36-cp36m-win_amd64.whl", hash = "sha256:25adde9b862f8f9aac9d2d11971f226bd4c8fbaa89fb76bdadb267ef22d10064"},
- {file = "mypy-0.812-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:552a815579aa1e995f39fd05dde6cd378e191b063f031f2acfe73ce9fb7f9e56"},
- {file = "mypy-0.812-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:499c798053cdebcaa916eef8cd733e5584b5909f789de856b482cd7d069bdad8"},
- {file = "mypy-0.812-cp37-cp37m-manylinux2010_x86_64.whl", hash = "sha256:5873888fff1c7cf5b71efbe80e0e73153fe9212fafdf8e44adfe4c20ec9f82d7"},
- {file = "mypy-0.812-cp37-cp37m-win_amd64.whl", hash = "sha256:9f94aac67a2045ec719ffe6111df543bac7874cee01f41928f6969756e030564"},
- {file = "mypy-0.812-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d23e0ea196702d918b60c8288561e722bf437d82cb7ef2edcd98cfa38905d506"},
- {file = "mypy-0.812-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:674e822aa665b9fd75130c6c5f5ed9564a38c6cea6a6432ce47eafb68ee578c5"},
- {file = "mypy-0.812-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:abf7e0c3cf117c44d9285cc6128856106183938c68fd4944763003decdcfeb66"},
- {file = "mypy-0.812-cp38-cp38-win_amd64.whl", hash = "sha256:0d0a87c0e7e3a9becdfbe936c981d32e5ee0ccda3e0f07e1ef2c3d1a817cf73e"},
- {file = "mypy-0.812-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:7ce3175801d0ae5fdfa79b4f0cfed08807af4d075b402b7e294e6aa72af9aa2a"},
- {file = "mypy-0.812-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:b09669bcda124e83708f34a94606e01b614fa71931d356c1f1a5297ba11f110a"},
- {file = "mypy-0.812-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:33f159443db0829d16f0a8d83d94df3109bb6dd801975fe86bacb9bf71628e97"},
- {file = "mypy-0.812-cp39-cp39-win_amd64.whl", hash = "sha256:3f2aca7f68580dc2508289c729bd49ee929a436208d2b2b6aab15745a70a57df"},
- {file = "mypy-0.812-py3-none-any.whl", hash = "sha256:2f9b3407c58347a452fc0736861593e105139b905cca7d097e413453a1d650b4"},
- {file = "mypy-0.812.tar.gz", hash = "sha256:cd07039aa5df222037005b08fbbfd69b3ab0b0bd7a07d7906de75ae52c4e3119"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
+ {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
+ {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
+ {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
+ {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
+ {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
+ {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
+ {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
+ {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
+ {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
+ {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
+ {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
+ {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
+ {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
+ {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
+ {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
+ {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
+ {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
+ {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
@@ -606,3 +738,3 @@ files = [
-mypy-extensions = ">=0.4.3,<0.5.0"
-typed-ast = ">=1.4.0,<1.5.0"
-typing-extensions = ">=3.7.4"
+mypy-extensions = ">=0.4.3"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=3.10"
@@ -611,0 +744,3 @@ dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
@@ -615,2 +750,2 @@ name = "mypy-extensions"
-version = "0.4.3"
-description = "Experimental type system extensions for programs checked with the mypy typechecker."
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
@@ -619 +754 @@ optional = false
-python-versions = "*"
+python-versions = ">=3.5"
@@ -621,2 +756,2 @@ files = [
- {file = "mypy_extensions-0.4.3-py2.py3-none-any.whl", hash = "sha256:090fedd75945a69ae91ce1303b5824f428daf5a028d2f6ab8a299250a846f15d"},
- {file = "mypy_extensions-0.4.3.tar.gz", hash = "sha256:2d82818f5bb3e369420cb3c4060a7970edba416647068eb4c5343488a6c604a8"},
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
@@ -627 +762 @@ name = "openapi-schema-validator"
-version = "0.2.3"
+version = "0.4.3"
@@ -633,2 +768,2 @@ files = [
- {file = "openapi-schema-validator-0.2.3.tar.gz", hash = "sha256:2c64907728c3ef78e23711c8840a423f0b241588c9ed929855e4b2d1bb0cf5f2"},
- {file = "openapi_schema_validator-0.2.3-py3-none-any.whl", hash = "sha256:9bae709212a19222892cabcc60cafd903cbf4b220223f48583afa3c0e3cc6fc4"},
+ {file = "openapi_schema_validator-0.4.3-py3-none-any.whl", hash = "sha256:f1eff2a7936546a3ce62b88a17d09de93c9bd229cbc43cb696c988a61a382548"},
+ {file = "openapi_schema_validator-0.4.3.tar.gz", hash = "sha256:6940dba9f4906c97078fea6fd9d5a3a3384207db368c4e32f6af6abd7c5c560b"},
@@ -638,6 +773,2 @@ files = [
-jsonschema = ">=3.0.0,<5.0.0"
-
-[package.extras]
-isodate = ["isodate"]
-rfc3339-validator = ["rfc3339-validator"]
-strict-rfc3339 = ["strict-rfc3339"]
+jsonschema = ">=4.0.0,<5.0.0"
+rfc3339-validator = "*"
@@ -647,2 +778,2 @@ name = "openapi-spec-validator"
-version = "0.4.0"
-description = "OpenAPI 2.0 (aka Swagger) and OpenAPI 3.0 spec validator"
+version = "0.5.5"
+description = "OpenAPI 2.0 (aka Swagger) and OpenAPI 3 spec validator"
@@ -653,2 +784,2 @@ files = [
- {file = "openapi-spec-validator-0.4.0.tar.gz", hash = "sha256:97f258850afc97b048f7c2653855e0f88fa66ac103c2be5077c7960aca2ad49a"},
- {file = "openapi_spec_validator-0.4.0-py3-none-any.whl", hash = "sha256:06900ac4d546a1df3642a779da0055be58869c598e3042a2fef067cfd99d04d0"},
+ {file = "openapi_spec_validator-0.5.5-py3-none-any.whl", hash = "sha256:93ba247f585e1447214b4207728a7cce3726d148238217be69e6b8725c118fbe"},
+ {file = "openapi_spec_validator-0.5.5.tar.gz", hash = "sha256:3010df5237748e25d7fac2b2aaf13457c1afd02735b2bd6f008a10079c8f443a"},
@@ -658,4 +789,4 @@ files = [
-jsonschema = ">=3.2.0,<5.0.0"
-openapi-schema-validator = ">=0.2.0,<0.3.0"
-PyYAML = ">=5.1"
-setuptools = "*"
+jsonschema = ">=4.0.0,<5.0.0"
+jsonschema-spec = ">=0.1.1,<0.2.0"
+lazy-object-proxy = ">=1.7.1,<2.0.0"
+openapi-schema-validator = ">=0.4.2,<0.5.0"
@@ -693,0 +825,12 @@ files = [
+[[package]]
+name = "pathable"
+version = "0.4.3"
+description = "Object-oriented paths"
+category = "main"
+optional = false
+python-versions = ">=3.7.0,<4.0.0"
+files = [
+ {file = "pathable-0.4.3-py3-none-any.whl", hash = "sha256:cdd7b1f9d7d5c8b8d3315dbf5a86b2596053ae845f056f57d97c0eefff84da14"},
+ {file = "pathable-0.4.3.tar.gz", hash = "sha256:5c869d315be50776cc8a993f3af43e0c60dc01506b399643f919034ebf4cdcab"},
+]
+
@@ -720 +863 @@ name = "pip"
-version = "22.3.1"
+version = "23.0"
@@ -726,2 +869,2 @@ files = [
- {file = "pip-22.3.1-py3-none-any.whl", hash = "sha256:908c78e6bc29b676ede1c4d57981d490cb892eb45cd8c214ab6298125119e077"},
- {file = "pip-22.3.1.tar.gz", hash = "sha256:65fd48317359f3af8e593943e6ae1506b66325085ea64b706a998c6e83eeaf38"},
+ {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
+ {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
@@ -796 +939 @@ name = "platformdirs"
-version = "2.6.2"
+version = "3.0.0"
@@ -802,2 +945,2 @@ files = [
- {file = "platformdirs-2.6.2-py3-none-any.whl", hash = "sha256:83c8f6d04389165de7c9b6f0c682439697887bca0aa2f1c87ef1826be3584490"},
- {file = "platformdirs-2.6.2.tar.gz", hash = "sha256:e1fea1fe471b9ff8332e229df3cb7de4f53eeea4998d3b6bfff542115e998bd2"},
+ {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"},
+ {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"},
@@ -807,2 +950,2 @@ files = [
-docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"]
-test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
+docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
@@ -826,15 +968,0 @@ testing = ["pytest", "pytest-benchmark"]
-[[package]]
-name = "poetryup"
-version = "0.3.15"
-description = "Update dependencies and bump their version in the pyproject.toml file"
-category = "dev"
-optional = false
-python-versions = ">=3.6,<4.0"
-files = [
- {file = "poetryup-0.3.15-py3-none-any.whl", hash = "sha256:db068f55d10c0f89c76ea2b62c6bb81c0b0512454f7a83bdc0a13c146e5fb13e"},
- {file = "poetryup-0.3.15.tar.gz", hash = "sha256:efa4e7bb0cd005db4aff3cc678c8bfba9474ef42d5759c0168f2a55fc0f17bc3"},
-]
-
-[package.dependencies]
-tomlkit = ">=0.7.2,<0.8.0"
-
@@ -955,0 +1084,20 @@ testing = ["argcomplete", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.
+[[package]]
+name = "pytest-cov"
+version = "2.12.1"
+description = "Pytest plugin for measuring coverage."
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"},
+ {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"},
+]
+
+[package.dependencies]
+coverage = ">=5.2.1"
+pytest = ">=4.6"
+toml = "*"
+
+[package.extras]
+testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"]
+
@@ -1045,0 +1194,15 @@ test = ["commentjson", "packaging", "pytest"]
+[[package]]
+name = "rfc3339-validator"
+version = "0.1.4"
+description = "A pure python RFC3339 validator"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "rfc3339_validator-0.1.4-py2.py3-none-any.whl", hash = "sha256:24f6ec1eda14ef823da9e36ec7113124b39c04d50a4d3d3a3c2859577e7791fa"},
+ {file = "rfc3339_validator-0.1.4.tar.gz", hash = "sha256:138a2abdf93304ad60530167e51d2dfb9549521a836871b88d7f4695d0022f6b"},
+]
+
+[package.dependencies]
+six = "*"
+
@@ -1048 +1211 @@ name = "rich"
-version = "13.2.0"
+version = "13.3.1"
@@ -1054,2 +1217,2 @@ files = [
- {file = "rich-13.2.0-py3-none-any.whl", hash = "sha256:7c963f0d03819221e9ac561e1bc866e3f95a02248c1234daa48954e6d381c003"},
- {file = "rich-13.2.0.tar.gz", hash = "sha256:f1a00cdd3eebf999a15d85ec498bfe0b1a77efe9b34f645768a54132ef444ac5"},
+ {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"},
+ {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"},
@@ -1060 +1223 @@ markdown-it-py = ">=2.1.0,<3.0.0"
-pygments = ">=2.6.0,<3.0.0"
+pygments = ">=2.14.0,<3.0.0"
@@ -1063 +1226 @@ pygments = ">=2.6.0,<3.0.0"
-jupyter = ["ipywidgets (>=7.5.1,<8.0.0)"]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
@@ -1067 +1230 @@ name = "setuptools"
-version = "66.1.1"
+version = "67.3.1"
@@ -1069 +1232 @@ description = "Easily download, build, install, upgrade, and uninstall Python pa
-category = "main"
+category = "dev"
@@ -1073,2 +1236,2 @@ files = [
- {file = "setuptools-66.1.1-py3-none-any.whl", hash = "sha256:6f590d76b713d5de4e49fe4fbca24474469f53c83632d5d0fd056f7ff7e8112b"},
- {file = "setuptools-66.1.1.tar.gz", hash = "sha256:ac4008d396bc9cd983ea483cb7139c0240a07bbc74ffb6232fceffedc6cf03a8"},
+ {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
+ {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
@@ -1086 +1249 @@ description = "Python 2 and 3 compatibility utilities"
-category = "dev"
+category = "main"
@@ -1120 +1283 @@ name = "stevedore"
-version = "4.1.1"
+version = "5.0.0"
@@ -1126,2 +1289,2 @@ files = [
- {file = "stevedore-4.1.1-py3-none-any.whl", hash = "sha256:aa6436565c069b2946fe4ebff07f5041e0c8bf18c7376dd29edf80cf7d524e4e"},
- {file = "stevedore-4.1.1.tar.gz", hash = "sha256:7f8aeb6e3f90f96832c301bff21a7eb5eefbe894c88c506483d355565d88cc1a"},
+ {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"},
+ {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"},
@@ -1157,12 +1319,0 @@ files = [
-[[package]]
-name = "tomlkit"
-version = "0.7.2"
-description = "Style preserving TOML library"
-category = "dev"
-optional = false
-python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
-files = [
- {file = "tomlkit-0.7.2-py2.py3-none-any.whl", hash = "sha256:173ad840fa5d2aac140528ca1933c29791b79a374a0861a80347f42ec9328117"},
- {file = "tomlkit-0.7.2.tar.gz", hash = "sha256:d7a454f319a7e9bd2e249f239168729327e4dd2d27b17dc68be264ad1ce36754"},
-]
-
@@ -1191,3 +1342,18 @@ telegram = ["requests"]
-name = "typed-ast"
-version = "1.4.3"
-description = "a fork of Python 2 and 3 ast modules with type comment support"
+name = "types-requests"
+version = "2.28.11.12"
+description = "Typing stubs for requests"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
+ {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+]
+
+[package.dependencies]
+types-urllib3 = "<1.27"
+
+[[package]]
+name = "types-urllib3"
+version = "1.26.25.5"
+description = "Typing stubs for urllib3"
@@ -1198,30 +1364,2 @@ files = [
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:2068531575a125b87a41802130fa7e29f26c09a2833fea68d9a40cf33902eba6"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:c907f561b1e83e93fad565bac5ba9c22d96a54e7ea0267c708bffe863cbe4075"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux2014_aarch64.whl", hash = "sha256:1b3ead4a96c9101bef08f9f7d1217c096f31667617b58de957f690c92378b528"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win32.whl", hash = "sha256:dde816ca9dac1d9c01dd504ea5967821606f02e510438120091b84e852367428"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win_amd64.whl", hash = "sha256:777a26c84bea6cd934422ac2e3b78863a37017618b6e5c08f92ef69853e765d3"},
- {file = "typed_ast-1.4.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:f8afcf15cc511ada719a88e013cec87c11aff7b91f019295eb4530f96fe5ef2f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:52b1eb8c83f178ab787f3a4283f68258525f8d70f778a2f6dd54d3b5e5fb4341"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:01ae5f73431d21eead5015997ab41afa53aa1fbe252f9da060be5dad2c730ace"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:c190f0899e9f9f8b6b7863debfb739abcb21a5c054f911ca3596d12b8a4c4c7f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win32.whl", hash = "sha256:398e44cd480f4d2b7ee8d98385ca104e35c81525dd98c519acff1b79bdaac363"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win_amd64.whl", hash = "sha256:bff6ad71c81b3bba8fa35f0f1921fb24ff4476235a6e94a26ada2e54370e6da7"},
- {file = "typed_ast-1.4.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0fb71b8c643187d7492c1f8352f2c15b4c4af3f6338f21681d3681b3dc31a266"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:760ad187b1041a154f0e4d0f6aae3e40fdb51d6de16e5c99aedadd9246450e9e"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:5feca99c17af94057417d744607b82dd0a664fd5e4ca98061480fd8b14b18d04"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:95431a26309a21874005845c21118c83991c63ea800dd44843e42a916aec5899"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win32.whl", hash = "sha256:aee0c1256be6c07bd3e1263ff920c325b59849dc95392a05f258bb9b259cf39c"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9ad2c92ec681e02baf81fdfa056fe0d818645efa9af1f1cd5fd6f1bd2bdfd805"},
- {file = "typed_ast-1.4.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b36b4f3920103a25e1d5d024d155c504080959582b928e91cb608a65c3a49e1a"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_i686.whl", hash = "sha256:067a74454df670dcaa4e59349a2e5c81e567d8d65458d480a5b3dfecec08c5ff"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:7538e495704e2ccda9b234b82423a4038f324f3a10c43bc088a1636180f11a41"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:af3d4a73793725138d6b334d9d247ce7e5f084d96284ed23f22ee626a7b88e39"},
- {file = "typed_ast-1.4.3-cp38-cp38-win32.whl", hash = "sha256:f2362f3cb0f3172c42938946dbc5b7843c2a28aec307c49100c8b38764eb6927"},
- {file = "typed_ast-1.4.3-cp38-cp38-win_amd64.whl", hash = "sha256:dd4a21253f42b8d2b48410cb31fe501d32f8b9fbeb1f55063ad102fe9c425e40"},
- {file = "typed_ast-1.4.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:f328adcfebed9f11301eaedfa48e15bdece9b519fb27e6a8c01aa52a17ec31b3"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2c726c276d09fc5c414693a2de063f521052d9ea7c240ce553316f70656c84d4"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:cae53c389825d3b46fb37538441f75d6aecc4174f615d048321b716df2757fb0"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:b9574c6f03f685070d859e75c7f9eeca02d6933273b5e69572e5ff9d5e3931c3"},
- {file = "typed_ast-1.4.3-cp39-cp39-win32.whl", hash = "sha256:209596a4ec71d990d71d5e0d312ac935d86930e6eecff6ccc7007fe54d703808"},
- {file = "typed_ast-1.4.3-cp39-cp39-win_amd64.whl", hash = "sha256:9c6d1a54552b5330bc657b7ef0eae25d00ba7ffe85d9ea8ae6540d2197a3788c"},
- {file = "typed_ast-1.4.3.tar.gz", hash = "sha256:fb1bbeac803adea29cedd70781399c99138358c26d05fcbd23c13016b7f5ec65"},
+ {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
+ {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
@@ -1232 +1370 @@ name = "typing-extensions"
-version = "4.4.0"
+version = "4.5.0"
@@ -1234 +1372 @@ description = "Backported and Experimental Type Hints for Python 3.7+"
-category = "dev"
+category = "main"
@@ -1238,2 +1376,2 @@ files = [
- {file = "typing_extensions-4.4.0-py3-none-any.whl", hash = "sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e"},
- {file = "typing_extensions-4.4.0.tar.gz", hash = "sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa"},
+ {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"},
+ {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"},
@@ -1274 +1412 @@ python-versions = "3.9.15"
-content-hash = "244cb47bc1d16906470419a4f8e2ff371df9c4c22e10f452d6d62e05aae9a98d"
+content-hash = "9d2cb8cda0199417c4c2d89abb558f93702f57571a0fadfae10f122b3774f827"
diff --git a/e2e/pyproject.toml b/e2e/pyproject.toml
index 00393c3f..066d81f7 100644
--- a/e2e/pyproject.toml
+++ b/e2e/pyproject.toml
@@ -9 +9 @@ license = "Apache-2.0"
-openapi-spec-validator = "^0.4.0"
+openapi-spec-validator = "^0.5.5"
@@ -13,2 +13,2 @@ python = "3.9.15"
-bandit = "^1.7.0"
-black = "^22.1.0"
+bandit = "^1.7.4"
+black = "^22.12.0"
@@ -17,6 +17,7 @@ huggingface-hub = "^0.12.0"
-isort = "^5.9.3"
-mypy = "0.812"
-pip-audit = "^2.4.6"
-poetryup = "^0.3.8"
-pytest = "^7.2.0"
-requests = "^2.27.1"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pip-audit = "^2.4.14"
+pytest = "^7.2.1"
+pytest-cov = "^2.12.1"
+requests = "^2.28.2"
+types-requests = "^2.28.11"
diff --git a/e2e/tests/fixtures/files.py b/e2e/tests/fixtures/files.py
index 1a4f812a..204b2081 100644
--- a/e2e/tests/fixtures/files.py
+++ b/e2e/tests/fixtures/files.py
@@ -6,0 +7 @@ import pytest
+from pytest import TempPathFactory
@@ -17 +18 @@ DATA = [
-def csv_path(tmp_path_factory):
+def csv_path(tmp_path_factory: TempPathFactory) -> str:
diff --git a/e2e/tests/fixtures/hub.py b/e2e/tests/fixtures/hub.py
index fd513b1b..5f6a3af8 100644
--- a/e2e/tests/fixtures/hub.py
+++ b/e2e/tests/fixtures/hub.py
@@ -8 +8 @@ from contextlib import contextmanager, suppress
-from typing import Iterable, Literal, Mapping, Optional, TypedDict
+from typing import Any, Callable, Iterator, Literal, Mapping, Optional, TypedDict
@@ -12,6 +12,3 @@ import requests
-from huggingface_hub.hf_api import (
- REPO_TYPES,
- REPO_TYPES_URL_PREFIXES,
- HfApi,
- hf_raise_for_status,
-)
+from huggingface_hub.constants import REPO_TYPES, REPO_TYPES_URL_PREFIXES
+from huggingface_hub.hf_api import HfApi
+from huggingface_hub.utils._errors import hf_raise_for_status
@@ -39,2 +36,2 @@ def update_repo_settings(
- name: str = None,
-) -> Mapping[str, bool]:
+ name: Optional[str] = None,
+) -> Any:
@@ -100 +97 @@ def update_repo_settings(
-def hf_api():
+def hf_api() -> HfApi:
@@ -110,2 +107,2 @@ def hf_token() -> str:
-def cleanup_repo(hf_api: HfApi):
- def _cleanup_repo(repo_id):
+def cleanup_repo(hf_api: HfApi) -> Callable[[str], None]:
+ def _cleanup_repo(repo_id: str) -> None:
@@ -118 +115 @@ def cleanup_repo(hf_api: HfApi):
-def temporary_repo(cleanup_repo):
+def temporary_repo(cleanup_repo: Callable[[str], None]) -> Callable[[str], Iterator[str]]:
@@ -120 +117 @@ def temporary_repo(cleanup_repo):
- def _temporary_repo(repo_id):
+ def _temporary_repo(repo_id: str) -> Iterator[str]:
@@ -126 +123 @@ def temporary_repo(cleanup_repo):
- return _temporary_repo
+ return _temporary_repo # type: ignore
@@ -135 +132 @@ def create_hf_dataset_repo_csv_data(
- hf_api: HfApi, hf_token: str, csv_path: str, *, private=False, gated=False, user=CI_HUB_USER
+ hf_api: HfApi, hf_token: str, csv_path: str, *, private: bool = False, gated: bool = False, user: str = CI_HUB_USER
@@ -153 +150 @@ def create_hf_dataset_repo_csv_data(
-def hf_public_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterable[str]:
+def hf_public_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterator[str]:
@@ -161 +158 @@ def hf_public_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str)
-def hf_public_2_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterable[str]:
+def hf_public_2_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterator[str]:
@@ -169 +166 @@ def hf_public_2_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: st
-def hf_private_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterable[str]:
+def hf_private_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterator[str]:
@@ -177 +174 @@ def hf_private_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str
-def hf_gated_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterable[str]:
+def hf_gated_dataset_repo_csv_data(hf_api: HfApi, hf_token: str, csv_path: str) -> Iterator[str]:
@@ -196,4 +193,4 @@ def hf_dataset_repos_csv_data(
- hf_public_dataset_repo_csv_data,
- hf_public_2_dataset_repo_csv_data,
- hf_private_dataset_repo_csv_data,
- hf_gated_dataset_repo_csv_data,
+ hf_public_dataset_repo_csv_data: str,
+ hf_public_2_dataset_repo_csv_data: str,
+ hf_private_dataset_repo_csv_data: str,
+ hf_gated_dataset_repo_csv_data: str,
diff --git a/e2e/tests/test_12_splits.py b/e2e/tests/test_12_splits.py
index 23481bad..97d7c476 100644
--- a/e2e/tests/test_12_splits.py
+++ b/e2e/tests/test_12_splits.py
@@ -40 +40 @@ from .utils import get, get_openapi_body_example, poll, poll_splits, post_refres
-def test_splits_using_openapi(status: int, name: str, dataset: str, error_code: str):
+def test_splits_using_openapi(status: int, name: str, dataset: str, error_code: str) -> None:
diff --git a/e2e/tests/test_13_first_rows.py b/e2e/tests/test_13_first_rows.py
index 3074f4a6..b59eb8a7 100644
--- a/e2e/tests/test_13_first_rows.py
+++ b/e2e/tests/test_13_first_rows.py
@@ -44 +44 @@ def prepare_json(response: requests.Response) -> Any:
-def test_first_rows(status: int, name: str, dataset: str, config: str, split: str, error_code: str):
+def test_first_rows(status: int, name: str, dataset: str, config: str, split: str, error_code: str) -> None:
diff --git a/e2e/tests/test_14_valid.py b/e2e/tests/test_14_valid.py
index 2d8a1f9b..c397b22b 100644
--- a/e2e/tests/test_14_valid.py
+++ b/e2e/tests/test_14_valid.py
@@ -8 +8 @@ from .utils import get
-def test_valid_after_datasets_processed(hf_dataset_repos_csv_data: DatasetRepos):
+def test_valid_after_datasets_processed(hf_dataset_repos_csv_data: DatasetRepos) -> None:
diff --git a/e2e/tests/test_15_is_valid.py b/e2e/tests/test_15_is_valid.py
index fe0fbcb6..7e5ded1f 100644
--- a/e2e/tests/test_15_is_valid.py
+++ b/e2e/tests/test_15_is_valid.py
@@ -8 +8 @@ from .utils import get
-def test_is_valid_after_datasets_processed(hf_dataset_repos_csv_data: DatasetRepos):
+def test_is_valid_after_datasets_processed(hf_dataset_repos_csv_data: DatasetRepos) -> None:
diff --git a/e2e/tests/test_20_api_healthcheck.py b/e2e/tests/test_20_api_healthcheck.py
index 7aa175de..f2272beb 100644
--- a/e2e/tests/test_20_api_healthcheck.py
+++ b/e2e/tests/test_20_api_healthcheck.py
@@ -7 +7 @@ from .utils import API_URL, poll
-def test_healthcheck():
+def test_healthcheck() -> None:
diff --git a/e2e/tests/test_21_api_metrics.py b/e2e/tests/test_21_api_metrics.py
index 4fb326df..67cadb1a 100644
--- a/e2e/tests/test_21_api_metrics.py
+++ b/e2e/tests/test_21_api_metrics.py
@@ -11 +11 @@ from .utils import API_URL, get
-def has_metric(name: str, labels: Mapping[str, str], metrics: set[str]) -> bool:
+def has_metric(name: str, labels: Mapping[str, str], metric_names: set[str]) -> bool:
@@ -14 +14 @@ def has_metric(name: str, labels: Mapping[str, str], metrics: set[str]) -> bool:
- return any(re.match(s, metric) is not None for metric in metrics)
+ return any(re.match(s, metric_name) is not None for metric_name in metric_names)
@@ -17 +17 @@ def has_metric(name: str, labels: Mapping[str, str], metrics: set[str]) -> bool:
-def test_metrics():
+def test_metrics() -> None:
@@ -32 +32 @@ def test_metrics():
- metrics = set(metrics.keys())
+ metric_names = set(metrics.keys())
@@ -36 +36,3 @@ def test_metrics():
- name="queue_jobs_total", labels={"pid": "[0-9]*", "queue": endpoint, "status": "started"}, metrics=metrics
+ name="queue_jobs_total",
+ labels={"pid": "[0-9]*", "queue": endpoint, "status": "started"},
+ metric_names=metric_names,
@@ -41 +43 @@ def test_metrics():
- metrics=metrics,
+ metric_names=metric_names,
diff --git a/e2e/tests/test_30_admin_healthcheck.py b/e2e/tests/test_30_admin_healthcheck.py
index 162e5bb3..c3e7bbf2 100644
--- a/e2e/tests/test_30_admin_healthcheck.py
+++ b/e2e/tests/test_30_admin_healthcheck.py
@@ -7 +7 @@ from .utils import ADMIN_URL, poll
-def test_healthcheck():
+def test_healthcheck() -> None:
diff --git a/e2e/tests/test_31_admin_metrics.py b/e2e/tests/test_31_admin_metrics.py
index 26f620b7..83677edb 100644
--- a/e2e/tests/test_31_admin_metrics.py
+++ b/e2e/tests/test_31_admin_metrics.py
@@ -11 +11 @@ from .utils import ADMIN_URL, get
-def has_metric(name: str, labels: Mapping[str, str], metrics: set[str]) -> bool:
+def has_metric(name: str, labels: Mapping[str, str], metric_names: set[str]) -> bool:
@@ -14 +14 @@ def has_metric(name: str, labels: Mapping[str, str], metrics: set[str]) -> bool:
- return any(re.match(s, metric) is not None for metric in metrics)
+ return any(re.match(s, metric_name) is not None for metric_name in metric_names)
@@ -17 +17 @@ def has_metric(name: str, labels: Mapping[str, str], metrics: set[str]) -> bool:
-def test_metrics():
+def test_metrics() -> None:
@@ -32 +32 @@ def test_metrics():
- metrics = set(metrics.keys())
+ metric_names = set(metrics.keys())
@@ -36 +36,3 @@ def test_metrics():
- name="queue_jobs_total", labels={"pid": "[0-9]*", "queue": queue, "status": "started"}, metrics=metrics
+ name="queue_jobs_total",
+ labels={"pid": "[0-9]*", "queue": queue, "status": "started"},
+ metric_names=metric_names,
@@ -44 +46 @@ def test_metrics():
- metrics=metrics,
+ metric_names=metric_names,
diff --git a/e2e/tests/utils.py b/e2e/tests/utils.py
index 41e5fb4b..176e60be 100644
--- a/e2e/tests/utils.py
+++ b/e2e/tests/utils.py
@@ -25 +25 @@ Headers = Mapping[str, str]
-def get(relative_url: str, headers: Headers = None, url: str = URL) -> Response:
+def get(relative_url: str, headers: Optional[Headers] = None, url: str = URL) -> Response:
@@ -31 +31 @@ def get(relative_url: str, headers: Headers = None, url: str = URL) -> Response:
-def post(relative_url: str, json: Optional[Any] = None, headers: Headers = None, url: str = URL) -> Response:
+def post(relative_url: str, json: Optional[Any] = None, headers: Optional[Headers] = None, url: str = URL) -> Response:
@@ -41 +41 @@ def poll(
- headers: Headers = None,
+ headers: Optional[Headers] = None,
@@ -73 +73 @@ def post_refresh(dataset: str) -> Response:
-def poll_parquet(dataset: str, headers: Headers = None) -> Response:
+def poll_parquet(dataset: str, headers: Optional[Headers] = None) -> Response:
@@ -77 +77 @@ def poll_parquet(dataset: str, headers: Headers = None) -> Response:
-def poll_splits(dataset: str, headers: Headers = None) -> Response:
+def poll_splits(dataset: str, headers: Optional[Headers] = None) -> Response:
@@ -81 +81 @@ def poll_splits(dataset: str, headers: Headers = None) -> Response:
-def poll_first_rows(dataset: str, config: str, split: str, headers: Headers = None) -> Response:
+def poll_first_rows(dataset: str, config: str, split: str, headers: Optional[Headers] = None) -> Response:
@@ -85 +85 @@ def poll_first_rows(dataset: str, config: str, split: str, headers: Headers = No
-def get_openapi_body_example(path, status, example_name):
+def get_openapi_body_example(path: str, status: int, example_name: str) -> Any:
@@ -118 +118 @@ def poll_until_ready_and_assert(
- headers: Headers = None,
+ headers: Optional[Headers] = None,
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index 94eba3e0..c5d13c02 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -17 +17 @@ name = "aiohttp"
-version = "3.7.4.post0"
+version = "3.8.4"
@@ -23,37 +23,87 @@ files = [
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-macosx_10_14_x86_64.whl", hash = "sha256:3cf75f7cdc2397ed4442594b935a11ed5569961333d49b7539ea741be2cc79d5"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:4b302b45040890cea949ad092479e01ba25911a15e648429c7c5aae9650c67a8"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:fe60131d21b31fd1a14bd43e6bb88256f69dfc3188b3a89d736d6c71ed43ec95"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:393f389841e8f2dfc86f774ad22f00923fdee66d238af89b70ea314c4aefd290"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:c6e9dcb4cb338d91a73f178d866d051efe7c62a7166653a91e7d9fb18274058f"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:5df68496d19f849921f05f14f31bd6ef53ad4b00245da3195048c69934521809"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:0563c1b3826945eecd62186f3f5c7d31abb7391fedc893b7e2b26303b5a9f3fe"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-win32.whl", hash = "sha256:3d78619672183be860b96ed96f533046ec97ca067fd46ac1f6a09cd9b7484287"},
- {file = "aiohttp-3.7.4.post0-cp36-cp36m-win_amd64.whl", hash = "sha256:f705e12750171c0ab4ef2a3c76b9a4024a62c4103e3a55dd6f99265b9bc6fcfc"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:230a8f7e24298dea47659251abc0fd8b3c4e38a664c59d4b89cca7f6c09c9e87"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:2e19413bf84934d651344783c9f5e22dee452e251cfd220ebadbed2d9931dbf0"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:e4b2b334e68b18ac9817d828ba44d8fcb391f6acb398bcc5062b14b2cbeac970"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:d012ad7911653a906425d8473a1465caa9f8dea7fcf07b6d870397b774ea7c0f"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:40eced07f07a9e60e825554a31f923e8d3997cfc7fb31dbc1328c70826e04cde"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:209b4a8ee987eccc91e2bd3ac36adee0e53a5970b8ac52c273f7f8fd4872c94c"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:14762875b22d0055f05d12abc7f7d61d5fd4fe4642ce1a249abdf8c700bf1fd8"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-win32.whl", hash = "sha256:7615dab56bb07bff74bc865307aeb89a8bfd9941d2ef9d817b9436da3a0ea54f"},
- {file = "aiohttp-3.7.4.post0-cp37-cp37m-win_amd64.whl", hash = "sha256:d9e13b33afd39ddeb377eff2c1c4f00544e191e1d1dee5b6c51ddee8ea6f0cf5"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:547da6cacac20666422d4882cfcd51298d45f7ccb60a04ec27424d2f36ba3eaf"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:af9aa9ef5ba1fd5b8c948bb11f44891968ab30356d65fd0cc6707d989cd521df"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:64322071e046020e8797117b3658b9c2f80e3267daec409b350b6a7a05041213"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:bb437315738aa441251214dad17428cafda9cdc9729499f1d6001748e1d432f4"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:e54962802d4b8b18b6207d4a927032826af39395a3bd9196a5af43fc4e60b009"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:a00bb73540af068ca7390e636c01cbc4f644961896fa9363154ff43fd37af2f5"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:79ebfc238612123a713a457d92afb4096e2148be17df6c50fb9bf7a81c2f8013"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-win32.whl", hash = "sha256:515dfef7f869a0feb2afee66b957cc7bbe9ad0cdee45aec7fdc623f4ecd4fb16"},
- {file = "aiohttp-3.7.4.post0-cp38-cp38-win_amd64.whl", hash = "sha256:114b281e4d68302a324dd33abb04778e8557d88947875cbf4e842c2c01a030c5"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:7b18b97cf8ee5452fa5f4e3af95d01d84d86d32c5e2bfa260cf041749d66360b"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:15492a6368d985b76a2a5fdd2166cddfea5d24e69eefed4630cbaae5c81d89bd"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:bdb230b4943891321e06fc7def63c7aace16095be7d9cf3b1e01be2f10fba439"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:cffe3ab27871bc3ea47df5d8f7013945712c46a3cc5a95b6bee15887f1675c22"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:f881853d2643a29e643609da57b96d5f9c9b93f62429dcc1cbb413c7d07f0e1a"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:a5ca29ee66f8343ed336816c553e82d6cade48a3ad702b9ffa6125d187e2dedb"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:17c073de315745a1510393a96e680d20af8e67e324f70b42accbd4cb3315c9fb"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-win32.whl", hash = "sha256:932bb1ea39a54e9ea27fc9232163059a0b8855256f4052e776357ad9add6f1c9"},
- {file = "aiohttp-3.7.4.post0-cp39-cp39-win_amd64.whl", hash = "sha256:02f46fc0e3c5ac58b80d4d56eb0a7c7d97fcef69ace9326289fb9f1955e65cfe"},
- {file = "aiohttp-3.7.4.post0.tar.gz", hash = "sha256:493d3299ebe5f5a7c66b9819eacdcfbbaaf1a8e84911ddffcdc48888497afecf"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"},
+ {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"},
@@ -63 +113,2 @@ files = [
-async-timeout = ">=3.0,<4.0"
+aiosignal = ">=1.1.2"
+async-timeout = ">=4.0.0a3,<5.0"
@@ -65 +116,2 @@ attrs = ">=17.3.0"
-chardet = ">=2.0,<5.0"
+charset-normalizer = ">=2.0,<4.0"
+frozenlist = ">=1.1.1"
@@ -67 +118,0 @@ multidict = ">=4.5,<7.0"
-typing-extensions = ">=3.6.5"
@@ -71 +122,16 @@ yarl = ">=1.0,<2.0"
-speedups = ["aiodns", "brotlipy", "cchardet"]
+speedups = ["Brotli", "aiodns", "cchardet"]
+
+[[package]]
+name = "aiosignal"
+version = "1.3.1"
+description = "aiosignal: a list of registered asynchronous callbacks"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"},
+ {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"},
+]
+
+[package.dependencies]
+frozenlist = ">=1.1.0"
@@ -119 +185 @@ name = "async-timeout"
-version = "3.0.1"
+version = "4.0.2"
@@ -123 +189 @@ optional = false
-python-versions = ">=3.5.3"
+python-versions = ">=3.6"
@@ -125,2 +191,2 @@ files = [
- {file = "async-timeout-3.0.1.tar.gz", hash = "sha256:0c3c816a028d47f659d6ff5c745cb2acf1f966da1fe5c19c77a70282b25f4c5f"},
- {file = "async_timeout-3.0.1-py3-none-any.whl", hash = "sha256:4291ca197d287d274d0b6cb5d6f8f8f82d434ed288f962539ff18cc9012f9ea3"},
+ {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"},
+ {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"},
@@ -160,12 +225,0 @@ files = [
-[[package]]
-name = "chardet"
-version = "4.0.0"
-description = "Universal encoding detector for Python 2 and 3"
-category = "main"
-optional = false
-python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
-files = [
- {file = "chardet-4.0.0-py2.py3-none-any.whl", hash = "sha256:f864054d66fd9118f2e67044ac8981a54775ec5b67aed0441892edb553d21da5"},
- {file = "chardet-4.0.0.tar.gz", hash = "sha256:0d6f53a15db4120f2b08c94f11e7d93d2c911ee118b6b30a04ec3ee8310179fa"},
-]
-
@@ -530,0 +585,84 @@ woff = ["brotli (>=1.0.1)", "brotlicffi (>=0.8.0)", "zopfli (>=0.1.4)"]
+[[package]]
+name = "frozenlist"
+version = "1.3.3"
+description = "A list-like structure which implements collections.abc.MutableSequence"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"},
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"},
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"},
+ {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"},
+ {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"},
+ {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"},
+ {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"},
+ {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"},
+ {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"},
+ {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"},
+ {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"},
+ {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"},
+]
+
@@ -568 +706 @@ name = "gradio"
-version = "3.16.2"
+version = "3.18.0"
@@ -574,2 +712,2 @@ files = [
- {file = "gradio-3.16.2-py3-none-any.whl", hash = "sha256:87f48d269003ee966ef30b0a4002db22d083265e46bf6cd48872fdce665f1a4b"},
- {file = "gradio-3.16.2.tar.gz", hash = "sha256:53fa92f3c286f42904b907ae6d852f5b50be7cd6d4bbb9c375c982d0f71a5afa"},
+ {file = "gradio-3.18.0-py3-none-any.whl", hash = "sha256:de608f310584d1b16c9554015352adfa4804e6319d20a6b66e271377d2bbb31d"},
+ {file = "gradio-3.18.0.tar.gz", hash = "sha256:f66d19c651c740da6cfa2b411b0e19942579532e0ffc5c41f71a2adbf0bc5c30"},
@@ -587 +725 @@ jinja2 = "*"
-markdown-it-py = {version = "*", extras = ["linkify", "plugins"]}
+markdown-it-py = {version = ">=2.0.0", extras = ["linkify", "plugins"]}
@@ -704,0 +843,19 @@ files = [
+[[package]]
+name = "importlib-resources"
+version = "5.10.2"
+description = "Read resources from Python packages"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "importlib_resources-5.10.2-py3-none-any.whl", hash = "sha256:7d543798b0beca10b6a01ac7cafda9f822c54db9e8376a6bf57e0cbd74d486b6"},
+ {file = "importlib_resources-5.10.2.tar.gz", hash = "sha256:e4a96c8cc0339647ff9a5e0550d9f276fc5a01ffa276012b58ec108cfd7b8484"},
+]
+
+[package.dependencies]
+zipp = {version = ">=3.1.0", markers = "python_version < \"3.10\""}
+
+[package.extras]
+docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-lint"]
+testing = ["flake8 (<5)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)"]
+
@@ -931 +1088 @@ name = "matplotlib"
-version = "3.6.3"
+version = "3.7.0"
@@ -937,41 +1094,41 @@ files = [
- {file = "matplotlib-3.6.3-cp310-cp310-macosx_10_12_universal2.whl", hash = "sha256:80c166a0e28512e26755f69040e6bf2f946a02ffdb7c00bf6158cca3d2b146e6"},
- {file = "matplotlib-3.6.3-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:eb9421c403ffd387fbe729de6d9a03005bf42faba5e8432f4e51e703215b49fc"},
- {file = "matplotlib-3.6.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:5223affa21050fb6118353c1380c15e23aedfb436bf3e162c26dc950617a7519"},
- {file = "matplotlib-3.6.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d00c248ab6b92bea3f8148714837937053a083ff03b4c5e30ed37e28fc0e7e56"},
- {file = "matplotlib-3.6.3-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca94f0362f6b6f424b555b956971dcb94b12d0368a6c3e07dc7a40d32d6d873d"},
- {file = "matplotlib-3.6.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59400cc9451094b7f08cc3f321972e6e1db4cd37a978d4e8a12824bf7fd2f03b"},
- {file = "matplotlib-3.6.3-cp310-cp310-win32.whl", hash = "sha256:57ad1aee29043163374bfa8990e1a2a10ff72c9a1bfaa92e9c46f6ea59269121"},
- {file = "matplotlib-3.6.3-cp310-cp310-win_amd64.whl", hash = "sha256:1fcc4cad498533d3c393a160975acc9b36ffa224d15a6b90ae579eacee5d8579"},
- {file = "matplotlib-3.6.3-cp311-cp311-macosx_10_12_universal2.whl", hash = "sha256:d2cfaa7fd62294d945b8843ea24228a27c8e7c5b48fa634f3c168153b825a21b"},
- {file = "matplotlib-3.6.3-cp311-cp311-macosx_10_12_x86_64.whl", hash = "sha256:c3f08df2ac4636249b8bc7a85b8b82c983bef1441595936f62c2918370ca7e1d"},
- {file = "matplotlib-3.6.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff2aa84e74f80891e6bcf292ebb1dd57714ffbe13177642d65fee25384a30894"},
- {file = "matplotlib-3.6.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:11011c97d62c1db7bc20509572557842dbb8c2a2ddd3dd7f20501aa1cde3e54e"},
- {file = "matplotlib-3.6.3-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c235bf9be052347373f589e018988cad177abb3f997ab1a2e2210c41562cc0c"},
- {file = "matplotlib-3.6.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bebcff4c3ed02c6399d47329f3554193abd824d3d53b5ca02cf583bcd94470e2"},
- {file = "matplotlib-3.6.3-cp311-cp311-win32.whl", hash = "sha256:d5f18430f5cfa5571ab8f4c72c89af52aa0618e864c60028f11a857d62200cba"},
- {file = "matplotlib-3.6.3-cp311-cp311-win_amd64.whl", hash = "sha256:dfba7057609ca9567b9704626756f0142e97ec8c5ba2c70c6e7bd1c25ef99f06"},
- {file = "matplotlib-3.6.3-cp38-cp38-macosx_10_12_universal2.whl", hash = "sha256:9fb8fb19d03abf3c5dab89a8677e62c4023632f919a62b6dd1d6d2dbf42cd9f5"},
- {file = "matplotlib-3.6.3-cp38-cp38-macosx_10_12_x86_64.whl", hash = "sha256:bbf269e1d24bc25247095d71c7a969813f7080e2a7c6fa28931a603f747ab012"},
- {file = "matplotlib-3.6.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:994637e2995b0342699b396a320698b07cd148bbcf2dd2fa2daba73f34dd19f2"},
- {file = "matplotlib-3.6.3-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:77b384cee7ab8cf75ffccbfea351a09b97564fc62d149827a5e864bec81526e5"},
- {file = "matplotlib-3.6.3-cp38-cp38-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:73b93af33634ed919e72811c9703e1105185cd3fb46d76f30b7f4cfbbd063f89"},
- {file = "matplotlib-3.6.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:debeab8e2ab07e5e3dac33e12456da79c7e104270d2b2d1df92b9e40347cca75"},
- {file = "matplotlib-3.6.3-cp38-cp38-win32.whl", hash = "sha256:acc3b1a4bddbf56fe461e36fb9ef94c2cb607fc90d24ccc650040bfcc7610de4"},
- {file = "matplotlib-3.6.3-cp38-cp38-win_amd64.whl", hash = "sha256:1183877d008c752d7d535396096c910f4663e4b74a18313adee1213328388e1e"},
- {file = "matplotlib-3.6.3-cp39-cp39-macosx_10_12_universal2.whl", hash = "sha256:6adc441b5b2098a4b904bbf9d9e92fb816fef50c55aa2ea6a823fc89b94bb838"},
- {file = "matplotlib-3.6.3-cp39-cp39-macosx_10_12_x86_64.whl", hash = "sha256:6d81b11ede69e3a751424b98dc869c96c10256b2206bfdf41f9c720eee86844c"},
- {file = "matplotlib-3.6.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:29f17b7f2e068dc346687cbdf80b430580bab42346625821c2d3abf3a1ec5417"},
- {file = "matplotlib-3.6.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3f56a7252eee8f3438447f75f5e1148a1896a2756a92285fe5d73bed6deebff4"},
- {file = "matplotlib-3.6.3-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bbddfeb1495484351fb5b30cf5bdf06b3de0bc4626a707d29e43dfd61af2a780"},
- {file = "matplotlib-3.6.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:809119d1cba3ece3c9742eb01827fe7a0e781ea3c5d89534655a75e07979344f"},
- {file = "matplotlib-3.6.3-cp39-cp39-win32.whl", hash = "sha256:e0a64d7cc336b52e90f59e6d638ae847b966f68582a7af041e063d568e814740"},
- {file = "matplotlib-3.6.3-cp39-cp39-win_amd64.whl", hash = "sha256:79e501eb847f4a489eb7065bb8d3187117f65a4c02d12ea3a19d6c5bef173bcc"},
- {file = "matplotlib-3.6.3-pp38-pypy38_pp73-macosx_10_12_x86_64.whl", hash = "sha256:2787a16df07370dcba385fe20cdd0cc3cfaabd3c873ddabca78c10514c799721"},
- {file = "matplotlib-3.6.3-pp38-pypy38_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:68d94a436f62b8a861bf3ace82067a71bafb724b4e4f9133521e4d8012420dd7"},
- {file = "matplotlib-3.6.3-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:81b409b2790cf8d7c1ef35920f01676d2ae7afa8241844e7aa5484fdf493a9a0"},
- {file = "matplotlib-3.6.3-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:faff486b36530a836a6b4395850322e74211cd81fc17f28b4904e1bd53668e3e"},
- {file = "matplotlib-3.6.3-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:38d38cb1ea1d80ee0f6351b65c6f76cad6060bbbead015720ba001348ae90f0c"},
- {file = "matplotlib-3.6.3-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:12f999661589981e74d793ee2f41b924b3b87d65fd929f6153bf0f30675c59b1"},
- {file = "matplotlib-3.6.3-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:01b7f521a9a73c383825813af255f8c4485d1706e4f3e2ed5ae771e4403a40ab"},
- {file = "matplotlib-3.6.3-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9ceebaf73f1a3444fa11014f38b9da37ff7ea328d6efa1652241fe3777bfdab9"},
- {file = "matplotlib-3.6.3.tar.gz", hash = "sha256:1f4d69707b1677560cd952544ee4962f68ff07952fb9069ff8c12b56353cb8c9"},
+ {file = "matplotlib-3.7.0-cp310-cp310-macosx_10_12_universal2.whl", hash = "sha256:3da8b9618188346239e51f1ea6c0f8f05c6e218cfcc30b399dd7dd7f52e8bceb"},
+ {file = "matplotlib-3.7.0-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:c0592ba57217c22987b7322df10f75ef95bc44dce781692b4b7524085de66019"},
+ {file = "matplotlib-3.7.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:21269450243d6928da81a9bed201f0909432a74e7d0d65db5545b9fa8a0d0223"},
+ {file = "matplotlib-3.7.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:eb2e76cd429058d8954121c334dddfcd11a6186c6975bca61f3f248c99031b05"},
+ {file = "matplotlib-3.7.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:de20eb1247725a2f889173d391a6d9e7e0f2540feda24030748283108b0478ec"},
+ {file = "matplotlib-3.7.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5465735eaaafd1cfaec3fed60aee776aeb3fd3992aa2e49f4635339c931d443"},
+ {file = "matplotlib-3.7.0-cp310-cp310-win32.whl", hash = "sha256:092e6abc80cdf8a95f7d1813e16c0e99ceda8d5b195a3ab859c680f3487b80a2"},
+ {file = "matplotlib-3.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:4f640534ec2760e270801056bc0d8a10777c48b30966eef78a7c35d8590915ba"},
+ {file = "matplotlib-3.7.0-cp311-cp311-macosx_10_12_universal2.whl", hash = "sha256:f336e7014889c38c59029ebacc35c59236a852e4b23836708cfd3f43d1eaeed5"},
+ {file = "matplotlib-3.7.0-cp311-cp311-macosx_10_12_x86_64.whl", hash = "sha256:3a10428d4f8d1a478ceabd652e61a175b2fdeed4175ab48da4a7b8deb561e3fa"},
+ {file = "matplotlib-3.7.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:46ca923e980f76d34c1c633343a72bb042d6ba690ecc649aababf5317997171d"},
+ {file = "matplotlib-3.7.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c849aa94ff2a70fb71f318f48a61076d1205c6013b9d3885ade7f992093ac434"},
+ {file = "matplotlib-3.7.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:827e78239292e561cfb70abf356a9d7eaf5bf6a85c97877f254009f20b892f89"},
+ {file = "matplotlib-3.7.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:691ef1f15360e439886186d0db77b5345b24da12cbc4fc57b26c4826db4d6cab"},
+ {file = "matplotlib-3.7.0-cp311-cp311-win32.whl", hash = "sha256:21a8aeac39b4a795e697265d800ce52ab59bdeb6bb23082e2d971f3041074f02"},
+ {file = "matplotlib-3.7.0-cp311-cp311-win_amd64.whl", hash = "sha256:01681566e95b9423021b49dea6a2395c16fa054604eacb87f0f4c439750f9114"},
+ {file = "matplotlib-3.7.0-cp38-cp38-macosx_10_12_universal2.whl", hash = "sha256:cf119eee4e57389fba5ac8b816934e95c256535e55f0b21628b4205737d1de85"},
+ {file = "matplotlib-3.7.0-cp38-cp38-macosx_10_12_x86_64.whl", hash = "sha256:21bd4033c40b95abd5b8453f036ed5aa70856e56ecbd887705c37dce007a4c21"},
+ {file = "matplotlib-3.7.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:111ef351f28fd823ed7177632070a6badd6f475607122bc9002a526f2502a0b5"},
+ {file = "matplotlib-3.7.0-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:f91d35b3ef51d29d9c661069b9e4ba431ce283ffc533b981506889e144b5b40e"},
+ {file = "matplotlib-3.7.0-cp38-cp38-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:0a776462a4a63c0bfc9df106c15a0897aa2dbab6795c693aa366e8e283958854"},
+ {file = "matplotlib-3.7.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dfd4a0cbd151f6439e6d7f8dca5292839ca311e7e650596d073774847ca2e4f"},
+ {file = "matplotlib-3.7.0-cp38-cp38-win32.whl", hash = "sha256:56b7b79488209041a9bf7ddc34f1b069274489ce69e34dc63ae241d0d6b4b736"},
+ {file = "matplotlib-3.7.0-cp38-cp38-win_amd64.whl", hash = "sha256:8665855f3919c80551f377bc16df618ceabf3ef65270bc14b60302dce88ca9ab"},
+ {file = "matplotlib-3.7.0-cp39-cp39-macosx_10_12_universal2.whl", hash = "sha256:f910d924da8b9fb066b5beae0b85e34ed1b6293014892baadcf2a51da1c65807"},
+ {file = "matplotlib-3.7.0-cp39-cp39-macosx_10_12_x86_64.whl", hash = "sha256:cf6346644e8fe234dc847e6232145dac199a650d3d8025b3ef65107221584ba4"},
+ {file = "matplotlib-3.7.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3d1e52365d8d5af699f04581ca191112e1d1220a9ce4386b57d807124d8b55e6"},
+ {file = "matplotlib-3.7.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c869b646489c6a94375714032e5cec08e3aa8d3f7d4e8ef2b0fb50a52b317ce6"},
+ {file = "matplotlib-3.7.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f4ddac5f59e78d04b20469bc43853a8e619bb6505c7eac8ffb343ff2c516d72f"},
+ {file = "matplotlib-3.7.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fb0304c1cd802e9a25743414c887e8a7cd51d96c9ec96d388625d2cd1c137ae3"},
+ {file = "matplotlib-3.7.0-cp39-cp39-win32.whl", hash = "sha256:a06a6c9822e80f323549c6bc9da96d4f233178212ad9a5f4ab87fd153077a507"},
+ {file = "matplotlib-3.7.0-cp39-cp39-win_amd64.whl", hash = "sha256:cb52aa97b92acdee090edfb65d1cb84ea60ab38e871ba8321a10bbcebc2a3540"},
+ {file = "matplotlib-3.7.0-pp38-pypy38_pp73-macosx_10_12_x86_64.whl", hash = "sha256:3493b48e56468c39bd9c1532566dff3b8062952721b7521e1f394eb6791495f4"},
+ {file = "matplotlib-3.7.0-pp38-pypy38_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d0dcd1a0bf8d56551e8617d6dc3881d8a1c7fb37d14e5ec12cbb293f3e6170a"},
+ {file = "matplotlib-3.7.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:51fb664c37714cbaac69c16d6b3719f517a13c96c3f76f4caadd5a0aa7ed0329"},
+ {file = "matplotlib-3.7.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:4497d88c559b76da320b7759d64db442178beeea06a52dc0c629086982082dcd"},
+ {file = "matplotlib-3.7.0-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:9d85355c48ef8b9994293eb7c00f44aa8a43cad7a297fbf0770a25cdb2244b91"},
+ {file = "matplotlib-3.7.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:03eb2c8ff8d85da679b71e14c7c95d16d014c48e0c0bfa14db85f6cdc5c92aad"},
+ {file = "matplotlib-3.7.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:71b751d06b2ed1fd017de512d7439c0259822864ea16731522b251a27c0b2ede"},
+ {file = "matplotlib-3.7.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:b51ab8a5d5d3bbd4527af633a638325f492e09e45e78afdf816ef55217a09664"},
+ {file = "matplotlib-3.7.0.tar.gz", hash = "sha256:8f6efd313430d7ef70a38a3276281cb2e8646b3a22b3b21eb227da20e15e6813"},
@@ -983,0 +1141 @@ fonttools = ">=4.22.0"
+importlib-resources = {version = ">=3.2.0", markers = "python_version < \"3.10\""}
@@ -985 +1143 @@ kiwisolver = ">=1.0.1"
-numpy = ">=1.19"
+numpy = ">=1.20"
@@ -988 +1146 @@ pillow = ">=6.2.0"
-pyparsing = ">=2.2.1"
+pyparsing = ">=2.3.1"
@@ -1271,7 +1428,0 @@ files = [
- {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
- {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
- {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
- {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
- {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
- {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
- {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -1716 +1867 @@ name = "typing-extensions"
-version = "4.4.0"
+version = "4.5.0"
@@ -1722,2 +1873,2 @@ files = [
- {file = "typing_extensions-4.4.0-py3-none-any.whl", hash = "sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e"},
- {file = "typing_extensions-4.4.0.tar.gz", hash = "sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa"},
+ {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"},
+ {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"},
@@ -1943,0 +2095,16 @@ multidict = ">=4.0"
+[[package]]
+name = "zipp"
+version = "3.13.0"
+description = "Backport of pathlib-compatible object wrapper for zip files"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "zipp-3.13.0-py3-none-any.whl", hash = "sha256:e8b2a36ea17df80ffe9e2c4fda3f693c3dad6df1697d3cd3af232db680950b0b"},
+ {file = "zipp-3.13.0.tar.gz", hash = "sha256:23f70e964bc11a34cef175bc90ba2914e1e4545ea1e3e2f67c079671883f9cb6"},
+]
+
+[package.extras]
+docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-lint"]
+testing = ["flake8 (<5)", "func-timeout", "jaraco.functools", "jaraco.itertools", "more-itertools", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)"]
+
@@ -1946,2 +2113,2 @@ lock-version = "2.0"
-python-versions = "~3.9"
-content-hash = "3cfe0b9594309e180b37d2405ae2f51be4d0ee50b56f874152bc914121724c29"
+python-versions = "~3.9.15"
+content-hash = "f5c77c5c5e2d965f1fe77cd0ab7f7299fc5cc788c3614b289bae08e8014348c1"
diff --git a/front/admin_ui/pyproject.toml b/front/admin_ui/pyproject.toml
index 807e15fd..f45088d0 100644
--- a/front/admin_ui/pyproject.toml
+++ b/front/admin_ui/pyproject.toml
@@ -8,3 +8,3 @@ authors = ["Quentin Lhoest <[email protected]>"]
-gradio = "~3.16.1"
-matplotlib = "^3.3.4"
-requests = "^2.26.0"
+gradio = "~3.18.0"
+matplotlib = "^3.7.0"
+requests = "^2.28.2"
diff --git a/front/admin_ui/requirements.txt b/front/admin_ui/requirements.txt
index a0054b99..05f8dd49 100644
--- a/front/admin_ui/requirements.txt
+++ b/front/admin_ui/requirements.txt
@@ -1,3 +1,3 @@
-gradio~=3.16.1
-matplotlib>=3.3.4
-requests>=2.26.0
+gradio~=3.18.0
+matplotlib>=3.7.0
+requests>=2.28.2
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index f98f4f4e..e5d9eef4 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -559,3 +559,3 @@ mongo-types = "0.15.1"
-mongoengine = "^0.24.1"
-orjson = "^3.6.4"
-psutil = "^5.9.2"
+mongoengine = "^0.24.2"
+orjson = "^3.8.6"
+psutil = "^5.9.4"
@@ -562,0 +563 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+requests = "^2.28.2"
@@ -741 +742 @@ name = "mypy"
-version = "0.812"
+version = "1.0.0"
@@ -745 +746 @@ optional = false
-python-versions = ">=3.5"
+python-versions = ">=3.7"
@@ -747,22 +748,26 @@ files = [
- {file = "mypy-0.812-cp35-cp35m-macosx_10_9_x86_64.whl", hash = "sha256:a26f8ec704e5a7423c8824d425086705e381b4f1dfdef6e3a1edab7ba174ec49"},
- {file = "mypy-0.812-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:28fb5479c494b1bab244620685e2eb3c3f988d71fd5d64cc753195e8ed53df7c"},
- {file = "mypy-0.812-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:9743c91088d396c1a5a3c9978354b61b0382b4e3c440ce83cf77994a43e8c521"},
- {file = "mypy-0.812-cp35-cp35m-win_amd64.whl", hash = "sha256:d7da2e1d5f558c37d6e8c1246f1aec1e7349e4913d8fb3cb289a35de573fe2eb"},
- {file = "mypy-0.812-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:4eec37370483331d13514c3f55f446fc5248d6373e7029a29ecb7b7494851e7a"},
- {file = "mypy-0.812-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:d65cc1df038ef55a99e617431f0553cd77763869eebdf9042403e16089fe746c"},
- {file = "mypy-0.812-cp36-cp36m-manylinux2010_x86_64.whl", hash = "sha256:61a3d5b97955422964be6b3baf05ff2ce7f26f52c85dd88db11d5e03e146a3a6"},
- {file = "mypy-0.812-cp36-cp36m-win_amd64.whl", hash = "sha256:25adde9b862f8f9aac9d2d11971f226bd4c8fbaa89fb76bdadb267ef22d10064"},
- {file = "mypy-0.812-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:552a815579aa1e995f39fd05dde6cd378e191b063f031f2acfe73ce9fb7f9e56"},
- {file = "mypy-0.812-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:499c798053cdebcaa916eef8cd733e5584b5909f789de856b482cd7d069bdad8"},
- {file = "mypy-0.812-cp37-cp37m-manylinux2010_x86_64.whl", hash = "sha256:5873888fff1c7cf5b71efbe80e0e73153fe9212fafdf8e44adfe4c20ec9f82d7"},
- {file = "mypy-0.812-cp37-cp37m-win_amd64.whl", hash = "sha256:9f94aac67a2045ec719ffe6111df543bac7874cee01f41928f6969756e030564"},
- {file = "mypy-0.812-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d23e0ea196702d918b60c8288561e722bf437d82cb7ef2edcd98cfa38905d506"},
- {file = "mypy-0.812-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:674e822aa665b9fd75130c6c5f5ed9564a38c6cea6a6432ce47eafb68ee578c5"},
- {file = "mypy-0.812-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:abf7e0c3cf117c44d9285cc6128856106183938c68fd4944763003decdcfeb66"},
- {file = "mypy-0.812-cp38-cp38-win_amd64.whl", hash = "sha256:0d0a87c0e7e3a9becdfbe936c981d32e5ee0ccda3e0f07e1ef2c3d1a817cf73e"},
- {file = "mypy-0.812-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:7ce3175801d0ae5fdfa79b4f0cfed08807af4d075b402b7e294e6aa72af9aa2a"},
- {file = "mypy-0.812-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:b09669bcda124e83708f34a94606e01b614fa71931d356c1f1a5297ba11f110a"},
- {file = "mypy-0.812-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:33f159443db0829d16f0a8d83d94df3109bb6dd801975fe86bacb9bf71628e97"},
- {file = "mypy-0.812-cp39-cp39-win_amd64.whl", hash = "sha256:3f2aca7f68580dc2508289c729bd49ee929a436208d2b2b6aab15745a70a57df"},
- {file = "mypy-0.812-py3-none-any.whl", hash = "sha256:2f9b3407c58347a452fc0736861593e105139b905cca7d097e413453a1d650b4"},
- {file = "mypy-0.812.tar.gz", hash = "sha256:cd07039aa5df222037005b08fbbfd69b3ab0b0bd7a07d7906de75ae52c4e3119"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
+ {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
+ {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
+ {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
+ {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
+ {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
+ {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
+ {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
+ {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
+ {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
+ {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
+ {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
+ {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
+ {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
+ {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
+ {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
+ {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
+ {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
+ {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
@@ -772,3 +777,3 @@ files = [
-mypy-extensions = ">=0.4.3,<0.5.0"
-typed-ast = ">=1.4.0,<1.5.0"
-typing-extensions = ">=3.7.4"
+mypy-extensions = ">=0.4.3"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=3.10"
@@ -777,0 +783,3 @@ dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
@@ -781,2 +789,2 @@ name = "mypy-extensions"
-version = "0.4.4"
-description = "Experimental type system extensions for programs checked with the mypy typechecker."
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
@@ -785 +793 @@ optional = false
-python-versions = ">=2.7"
+python-versions = ">=3.5"
@@ -787 +795,2 @@ files = [
- {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
@@ -1004,15 +1012,0 @@ testing = ["pytest", "pytest-benchmark"]
-[[package]]
-name = "poetryup"
-version = "0.3.15"
-description = "Update dependencies and bump their version in the pyproject.toml file"
-category = "dev"
-optional = false
-python-versions = ">=3.6,<4.0"
-files = [
- {file = "poetryup-0.3.15-py3-none-any.whl", hash = "sha256:db068f55d10c0f89c76ea2b62c6bb81c0b0512454f7a83bdc0a13c146e5fb13e"},
- {file = "poetryup-0.3.15.tar.gz", hash = "sha256:efa4e7bb0cd005db4aff3cc678c8bfba9474ef42d5759c0168f2a55fc0f17bc3"},
-]
-
-[package.dependencies]
-tomlkit = ">=0.7.2,<0.8.0"
-
@@ -1402 +1396 @@ name = "setuptools"
-version = "67.2.0"
+version = "67.3.1"
@@ -1408,2 +1402,2 @@ files = [
- {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
- {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
+ {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
+ {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
@@ -1455 +1449 @@ name = "stevedore"
-version = "4.1.1"
+version = "5.0.0"
@@ -1461,2 +1455,2 @@ files = [
- {file = "stevedore-4.1.1-py3-none-any.whl", hash = "sha256:aa6436565c069b2946fe4ebff07f5041e0c8bf18c7376dd29edf80cf7d524e4e"},
- {file = "stevedore-4.1.1.tar.gz", hash = "sha256:7f8aeb6e3f90f96832c301bff21a7eb5eefbe894c88c506483d355565d88cc1a"},
+ {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"},
+ {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"},
@@ -1492,12 +1485,0 @@ files = [
-[[package]]
-name = "tomlkit"
-version = "0.7.2"
-description = "Style preserving TOML library"
-category = "dev"
-optional = false
-python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
-files = [
- {file = "tomlkit-0.7.2-py2.py3-none-any.whl", hash = "sha256:173ad840fa5d2aac140528ca1933c29791b79a374a0861a80347f42ec9328117"},
- {file = "tomlkit-0.7.2.tar.gz", hash = "sha256:d7a454f319a7e9bd2e249f239168729327e4dd2d27b17dc68be264ad1ce36754"},
-]
-
@@ -1525,40 +1506,0 @@ telegram = ["requests"]
-[[package]]
-name = "typed-ast"
-version = "1.4.3"
-description = "a fork of Python 2 and 3 ast modules with type comment support"
-category = "dev"
-optional = false
-python-versions = "*"
-files = [
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:2068531575a125b87a41802130fa7e29f26c09a2833fea68d9a40cf33902eba6"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:c907f561b1e83e93fad565bac5ba9c22d96a54e7ea0267c708bffe863cbe4075"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux2014_aarch64.whl", hash = "sha256:1b3ead4a96c9101bef08f9f7d1217c096f31667617b58de957f690c92378b528"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win32.whl", hash = "sha256:dde816ca9dac1d9c01dd504ea5967821606f02e510438120091b84e852367428"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win_amd64.whl", hash = "sha256:777a26c84bea6cd934422ac2e3b78863a37017618b6e5c08f92ef69853e765d3"},
- {file = "typed_ast-1.4.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:f8afcf15cc511ada719a88e013cec87c11aff7b91f019295eb4530f96fe5ef2f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:52b1eb8c83f178ab787f3a4283f68258525f8d70f778a2f6dd54d3b5e5fb4341"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:01ae5f73431d21eead5015997ab41afa53aa1fbe252f9da060be5dad2c730ace"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:c190f0899e9f9f8b6b7863debfb739abcb21a5c054f911ca3596d12b8a4c4c7f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win32.whl", hash = "sha256:398e44cd480f4d2b7ee8d98385ca104e35c81525dd98c519acff1b79bdaac363"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win_amd64.whl", hash = "sha256:bff6ad71c81b3bba8fa35f0f1921fb24ff4476235a6e94a26ada2e54370e6da7"},
- {file = "typed_ast-1.4.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0fb71b8c643187d7492c1f8352f2c15b4c4af3f6338f21681d3681b3dc31a266"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:760ad187b1041a154f0e4d0f6aae3e40fdb51d6de16e5c99aedadd9246450e9e"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:5feca99c17af94057417d744607b82dd0a664fd5e4ca98061480fd8b14b18d04"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:95431a26309a21874005845c21118c83991c63ea800dd44843e42a916aec5899"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win32.whl", hash = "sha256:aee0c1256be6c07bd3e1263ff920c325b59849dc95392a05f258bb9b259cf39c"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9ad2c92ec681e02baf81fdfa056fe0d818645efa9af1f1cd5fd6f1bd2bdfd805"},
- {file = "typed_ast-1.4.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b36b4f3920103a25e1d5d024d155c504080959582b928e91cb608a65c3a49e1a"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_i686.whl", hash = "sha256:067a74454df670dcaa4e59349a2e5c81e567d8d65458d480a5b3dfecec08c5ff"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:7538e495704e2ccda9b234b82423a4038f324f3a10c43bc088a1636180f11a41"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:af3d4a73793725138d6b334d9d247ce7e5f084d96284ed23f22ee626a7b88e39"},
- {file = "typed_ast-1.4.3-cp38-cp38-win32.whl", hash = "sha256:f2362f3cb0f3172c42938946dbc5b7843c2a28aec307c49100c8b38764eb6927"},
- {file = "typed_ast-1.4.3-cp38-cp38-win_amd64.whl", hash = "sha256:dd4a21253f42b8d2b48410cb31fe501d32f8b9fbeb1f55063ad102fe9c425e40"},
- {file = "typed_ast-1.4.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:f328adcfebed9f11301eaedfa48e15bdece9b519fb27e6a8c01aa52a17ec31b3"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2c726c276d09fc5c414693a2de063f521052d9ea7c240ce553316f70656c84d4"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:cae53c389825d3b46fb37538441f75d6aecc4174f615d048321b716df2757fb0"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:b9574c6f03f685070d859e75c7f9eeca02d6933273b5e69572e5ff9d5e3931c3"},
- {file = "typed_ast-1.4.3-cp39-cp39-win32.whl", hash = "sha256:209596a4ec71d990d71d5e0d312ac935d86930e6eecff6ccc7007fe54d703808"},
- {file = "typed_ast-1.4.3-cp39-cp39-win_amd64.whl", hash = "sha256:9c6d1a54552b5330bc657b7ef0eae25d00ba7ffe85d9ea8ae6540d2197a3788c"},
- {file = "typed_ast-1.4.3.tar.gz", hash = "sha256:fb1bbeac803adea29cedd70781399c99138358c26d05fcbd23c13016b7f5ec65"},
-]
-
@@ -1567 +1509 @@ name = "typing-extensions"
-version = "4.4.0"
+version = "4.5.0"
@@ -1573,2 +1515,2 @@ files = [
- {file = "typing_extensions-4.4.0-py3-none-any.whl", hash = "sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e"},
- {file = "typing_extensions-4.4.0.tar.gz", hash = "sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa"},
+ {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"},
+ {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"},
@@ -1609 +1551 @@ python-versions = "3.9.15"
-content-hash = "fb2aa3c18a66fff5bdf26107bb6265120efb94a66e99ae8507aea1d6bc9def0c"
+content-hash = "67ae1eb17c3921f944b323ddcaa58919c1deff1e7cc25e63cf3a82ec23340c62"
diff --git a/jobs/mongodb_migration/pyproject.toml b/jobs/mongodb_migration/pyproject.toml
index ddc39bda..8f80140e 100644
--- a/jobs/mongodb_migration/pyproject.toml
+++ b/jobs/mongodb_migration/pyproject.toml
@@ -14,2 +14,2 @@ python = "3.9.15"
-bandit = "^1.7.0"
-black = "^22.1.0"
+bandit = "^1.7.4"
+black = "^22.12.0"
@@ -17,5 +17,4 @@ flake8 = "^3.9.2"
-isort = "^5.9.3"
-mypy = "0.812"
-pip-audit = "^2.4.6"
-poetryup = "^0.3.8"
-pytest = "^7.2.0"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pip-audit = "^2.4.14"
+pytest = "^7.2.1"
diff --git a/jobs/mongodb_migration/src/mongodb_migration/check.py b/jobs/mongodb_migration/src/mongodb_migration/check.py
index 1eb021be..df84c992 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/check.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/check.py
@@ -11 +10,0 @@ from pymongo.collection import Collection
-
@@ -13,5 +12 @@ from pymongo.collection import Collection
-class DocumentWithId(Document):
- id: str
-
-
-U = TypeVar("U", bound=DocumentWithId)
+U = TypeVar("U", bound=Document)
@@ -24,2 +19,2 @@ def get_random_oids(collection: Collection, sample_size: int) -> List[int]:
- pipeline = [{"$project": {"_id": 1}}, {"$sample": {"size": sample_size}}]
- return [s["_id"] for s in collection.aggregate(pipeline)]
+ pipeline = [{"$project": {"pk": 1}}, {"$sample": {"size": sample_size}}]
+ return [s["pk"] for s in collection.aggregate(pipeline)]
@@ -28 +23 @@ def get_random_oids(collection: Collection, sample_size: int) -> List[int]:
-def get_random_documents(DocCls: DocumentClass, sample_size: int) -> Iterator[DocumentWithId]:
+def get_random_documents(DocCls: DocumentClass[Document], sample_size: int) -> Iterator[Document]:
@@ -31 +26 @@ def get_random_documents(DocCls: DocumentClass, sample_size: int) -> Iterator[Do
- return DocCls.objects(id__in=random_oids) # type: ignore
+ return DocCls.objects(pk__in=random_oids) # type: ignore
@@ -34 +29,5 @@ def get_random_documents(DocCls: DocumentClass, sample_size: int) -> Iterator[Do
-def check_documents(DocCls: DocumentClass, sample_size: int, custom_validation: Optional[CustomValidation] = None):
+def check_documents(
+ DocCls: DocumentClass[Document],
+ sample_size: int,
+ custom_validation: Optional[CustomValidation[Document]] = None,
+) -> None:
@@ -47 +46 @@ def check_documents(DocCls: DocumentClass, sample_size: int, custom_validation:
- logging.error(f"Could not load field {field} in Document {doc.id}. Document: {doc.to_json()}")
+ logging.error(f"Could not load field {field} in Document {doc.pk}. Document: {doc.to_json()}")
@@ -54 +53 @@ def check_documents(DocCls: DocumentClass, sample_size: int, custom_validation:
- logging.error(f"Validation error on document {doc.id}: {e}. Document: {doc.to_json()}")
+ logging.error(f"Validation error on document {doc.pk}: {e}. Document: {doc.to_json()}")
diff --git a/jobs/mongodb_migration/src/mongodb_migration/plan.py b/jobs/mongodb_migration/src/mongodb_migration/plan.py
index 81707874..623bbbf5 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/plan.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/plan.py
@@ -50 +50 @@ class Plan:
- def execute(self):
+ def execute(self) -> None:
@@ -59 +59 @@ class Plan:
- def apply(self):
+ def apply(self) -> None:
@@ -73 +73 @@ class Plan:
- def rollback(self):
+ def rollback(self) -> None:
@@ -91 +91 @@ class Plan:
- def save(self, migration: Migration):
+ def save(self, migration: Migration) -> None:
diff --git a/jobs/mongodb_migration/tests/conftest.py b/jobs/mongodb_migration/tests/conftest.py
index 7312e3f5..e9358043 100644
--- a/jobs/mongodb_migration/tests/conftest.py
+++ b/jobs/mongodb_migration/tests/conftest.py
@@ -16 +16,4 @@ def mongo_host(env: Env) -> str:
- return env.str(name="DATABASE_MIGRATIONS_MONGO_URL")
+ url = env.str(name="DATABASE_MIGRATIONS_MONGO_URL")
+ if type(url) is not str:
+ raise ValueError("DATABASE_MIGRATIONS_MONGO_URL is not set")
+ return url
diff --git a/jobs/mongodb_migration/tests/test_collector.py b/jobs/mongodb_migration/tests/test_collector.py
index ae651857..a0a62c9d 100644
--- a/jobs/mongodb_migration/tests/test_collector.py
+++ b/jobs/mongodb_migration/tests/test_collector.py
@@ -7 +7 @@ from mongodb_migration.collector import MigrationsCollector
-def test_collector():
+def test_collector() -> None:
diff --git a/jobs/mongodb_migration/tests/test_migration.py b/jobs/mongodb_migration/tests/test_migration.py
index 8790bf80..e034b31c 100644
--- a/jobs/mongodb_migration/tests/test_migration.py
+++ b/jobs/mongodb_migration/tests/test_migration.py
@@ -40 +40 @@ version_too_short = "20221110"
-def test_migration(version: str, description: str, exception: Optional[Type[Exception]]):
+def test_migration(version: str, description: str, exception: Optional[Type[Exception]]) -> None:
diff --git a/jobs/mongodb_migration/tests/test_plan.py b/jobs/mongodb_migration/tests/test_plan.py
index ea66efcb..3a0fbc6c 100644
--- a/jobs/mongodb_migration/tests/test_plan.py
+++ b/jobs/mongodb_migration/tests/test_plan.py
@@ -82 +82 @@ class MigrationErrorIrreversible(Migration):
-def test_empty_plan():
+def test_empty_plan() -> None:
@@ -109 +109 @@ migration_error_irreversible = MigrationErrorIrreversible(
-def test_collected_migrations_order_dont_matter(collected_migrations: List[Migration]):
+def test_collected_migrations_order_dont_matter(collected_migrations: List[Migration]) -> None:
@@ -136 +136 @@ def test_errors_in_migration_steps(
-):
+) -> None:
@@ -169 +169 @@ def test_get_planned_migrations(
-):
+) -> None:
@@ -188 +188 @@ def test_get_planned_migrations(
-def test_internal_operations_are_idempotent():
+def test_internal_operations_are_idempotent() -> None:
@@ -202 +202 @@ def test_internal_operations_are_idempotent():
-def test_execute_is_idempotent():
+def test_execute_is_idempotent() -> None:
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index 29e1eb92..bcd413ac 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -717 +717 @@ name = "mypy"
-version = "0.812"
+version = "1.0.0"
@@ -721 +721 @@ optional = false
-python-versions = ">=3.5"
+python-versions = ">=3.7"
@@ -723,22 +723,26 @@ files = [
- {file = "mypy-0.812-cp35-cp35m-macosx_10_9_x86_64.whl", hash = "sha256:a26f8ec704e5a7423c8824d425086705e381b4f1dfdef6e3a1edab7ba174ec49"},
- {file = "mypy-0.812-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:28fb5479c494b1bab244620685e2eb3c3f988d71fd5d64cc753195e8ed53df7c"},
- {file = "mypy-0.812-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:9743c91088d396c1a5a3c9978354b61b0382b4e3c440ce83cf77994a43e8c521"},
- {file = "mypy-0.812-cp35-cp35m-win_amd64.whl", hash = "sha256:d7da2e1d5f558c37d6e8c1246f1aec1e7349e4913d8fb3cb289a35de573fe2eb"},
- {file = "mypy-0.812-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:4eec37370483331d13514c3f55f446fc5248d6373e7029a29ecb7b7494851e7a"},
- {file = "mypy-0.812-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:d65cc1df038ef55a99e617431f0553cd77763869eebdf9042403e16089fe746c"},
- {file = "mypy-0.812-cp36-cp36m-manylinux2010_x86_64.whl", hash = "sha256:61a3d5b97955422964be6b3baf05ff2ce7f26f52c85dd88db11d5e03e146a3a6"},
- {file = "mypy-0.812-cp36-cp36m-win_amd64.whl", hash = "sha256:25adde9b862f8f9aac9d2d11971f226bd4c8fbaa89fb76bdadb267ef22d10064"},
- {file = "mypy-0.812-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:552a815579aa1e995f39fd05dde6cd378e191b063f031f2acfe73ce9fb7f9e56"},
- {file = "mypy-0.812-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:499c798053cdebcaa916eef8cd733e5584b5909f789de856b482cd7d069bdad8"},
- {file = "mypy-0.812-cp37-cp37m-manylinux2010_x86_64.whl", hash = "sha256:5873888fff1c7cf5b71efbe80e0e73153fe9212fafdf8e44adfe4c20ec9f82d7"},
- {file = "mypy-0.812-cp37-cp37m-win_amd64.whl", hash = "sha256:9f94aac67a2045ec719ffe6111df543bac7874cee01f41928f6969756e030564"},
- {file = "mypy-0.812-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d23e0ea196702d918b60c8288561e722bf437d82cb7ef2edcd98cfa38905d506"},
- {file = "mypy-0.812-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:674e822aa665b9fd75130c6c5f5ed9564a38c6cea6a6432ce47eafb68ee578c5"},
- {file = "mypy-0.812-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:abf7e0c3cf117c44d9285cc6128856106183938c68fd4944763003decdcfeb66"},
- {file = "mypy-0.812-cp38-cp38-win_amd64.whl", hash = "sha256:0d0a87c0e7e3a9becdfbe936c981d32e5ee0ccda3e0f07e1ef2c3d1a817cf73e"},
- {file = "mypy-0.812-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:7ce3175801d0ae5fdfa79b4f0cfed08807af4d075b402b7e294e6aa72af9aa2a"},
- {file = "mypy-0.812-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:b09669bcda124e83708f34a94606e01b614fa71931d356c1f1a5297ba11f110a"},
- {file = "mypy-0.812-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:33f159443db0829d16f0a8d83d94df3109bb6dd801975fe86bacb9bf71628e97"},
- {file = "mypy-0.812-cp39-cp39-win_amd64.whl", hash = "sha256:3f2aca7f68580dc2508289c729bd49ee929a436208d2b2b6aab15745a70a57df"},
- {file = "mypy-0.812-py3-none-any.whl", hash = "sha256:2f9b3407c58347a452fc0736861593e105139b905cca7d097e413453a1d650b4"},
- {file = "mypy-0.812.tar.gz", hash = "sha256:cd07039aa5df222037005b08fbbfd69b3ab0b0bd7a07d7906de75ae52c4e3119"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
+ {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
+ {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
+ {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
+ {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
+ {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
+ {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
+ {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
+ {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
+ {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
+ {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
+ {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
+ {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
+ {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
+ {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
+ {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
+ {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
+ {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
+ {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
@@ -748,3 +752,3 @@ files = [
-mypy-extensions = ">=0.4.3,<0.5.0"
-typed-ast = ">=1.4.0,<1.5.0"
-typing-extensions = ">=3.7.4"
+mypy-extensions = ">=0.4.3"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=3.10"
@@ -753,0 +758,3 @@ dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
@@ -757,2 +764,2 @@ name = "mypy-extensions"
-version = "0.4.4"
-description = "Experimental type system extensions for programs checked with the mypy typechecker."
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
@@ -761 +768 @@ optional = false
-python-versions = ">=2.7"
+python-versions = ">=3.5"
@@ -763 +770,2 @@ files = [
- {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
@@ -980,15 +987,0 @@ testing = ["pytest", "pytest-benchmark"]
-[[package]]
-name = "poetryup"
-version = "0.3.15"
-description = "Update dependencies and bump their version in the pyproject.toml file"
-category = "dev"
-optional = false
-python-versions = ">=3.6,<4.0"
-files = [
- {file = "poetryup-0.3.15-py3-none-any.whl", hash = "sha256:db068f55d10c0f89c76ea2b62c6bb81c0b0512454f7a83bdc0a13c146e5fb13e"},
- {file = "poetryup-0.3.15.tar.gz", hash = "sha256:efa4e7bb0cd005db4aff3cc678c8bfba9474ef42d5759c0168f2a55fc0f17bc3"},
-]
-
-[package.dependencies]
-tomlkit = ">=0.7.2,<0.8.0"
-
@@ -1378 +1371 @@ name = "setuptools"
-version = "67.2.0"
+version = "67.3.1"
@@ -1384,2 +1377,2 @@ files = [
- {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
- {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
+ {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
+ {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
@@ -1431 +1424 @@ name = "stevedore"
-version = "4.1.1"
+version = "5.0.0"
@@ -1437,2 +1430,2 @@ files = [
- {file = "stevedore-4.1.1-py3-none-any.whl", hash = "sha256:aa6436565c069b2946fe4ebff07f5041e0c8bf18c7376dd29edf80cf7d524e4e"},
- {file = "stevedore-4.1.1.tar.gz", hash = "sha256:7f8aeb6e3f90f96832c301bff21a7eb5eefbe894c88c506483d355565d88cc1a"},
+ {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"},
+ {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"},
@@ -1468,12 +1460,0 @@ files = [
-[[package]]
-name = "tomlkit"
-version = "0.7.2"
-description = "Style preserving TOML library"
-category = "dev"
-optional = false
-python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
-files = [
- {file = "tomlkit-0.7.2-py2.py3-none-any.whl", hash = "sha256:173ad840fa5d2aac140528ca1933c29791b79a374a0861a80347f42ec9328117"},
- {file = "tomlkit-0.7.2.tar.gz", hash = "sha256:d7a454f319a7e9bd2e249f239168729327e4dd2d27b17dc68be264ad1ce36754"},
-]
-
@@ -1502,3 +1483,3 @@ telegram = ["requests"]
-name = "typed-ast"
-version = "1.4.3"
-description = "a fork of Python 2 and 3 ast modules with type comment support"
+name = "types-psutil"
+version = "5.9.5.6"
+description = "Typing stubs for psutil"
@@ -1509,30 +1490,2 @@ files = [
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:2068531575a125b87a41802130fa7e29f26c09a2833fea68d9a40cf33902eba6"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:c907f561b1e83e93fad565bac5ba9c22d96a54e7ea0267c708bffe863cbe4075"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux2014_aarch64.whl", hash = "sha256:1b3ead4a96c9101bef08f9f7d1217c096f31667617b58de957f690c92378b528"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win32.whl", hash = "sha256:dde816ca9dac1d9c01dd504ea5967821606f02e510438120091b84e852367428"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win_amd64.whl", hash = "sha256:777a26c84bea6cd934422ac2e3b78863a37017618b6e5c08f92ef69853e765d3"},
- {file = "typed_ast-1.4.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:f8afcf15cc511ada719a88e013cec87c11aff7b91f019295eb4530f96fe5ef2f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:52b1eb8c83f178ab787f3a4283f68258525f8d70f778a2f6dd54d3b5e5fb4341"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:01ae5f73431d21eead5015997ab41afa53aa1fbe252f9da060be5dad2c730ace"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:c190f0899e9f9f8b6b7863debfb739abcb21a5c054f911ca3596d12b8a4c4c7f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win32.whl", hash = "sha256:398e44cd480f4d2b7ee8d98385ca104e35c81525dd98c519acff1b79bdaac363"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win_amd64.whl", hash = "sha256:bff6ad71c81b3bba8fa35f0f1921fb24ff4476235a6e94a26ada2e54370e6da7"},
- {file = "typed_ast-1.4.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0fb71b8c643187d7492c1f8352f2c15b4c4af3f6338f21681d3681b3dc31a266"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:760ad187b1041a154f0e4d0f6aae3e40fdb51d6de16e5c99aedadd9246450e9e"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:5feca99c17af94057417d744607b82dd0a664fd5e4ca98061480fd8b14b18d04"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:95431a26309a21874005845c21118c83991c63ea800dd44843e42a916aec5899"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win32.whl", hash = "sha256:aee0c1256be6c07bd3e1263ff920c325b59849dc95392a05f258bb9b259cf39c"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9ad2c92ec681e02baf81fdfa056fe0d818645efa9af1f1cd5fd6f1bd2bdfd805"},
- {file = "typed_ast-1.4.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b36b4f3920103a25e1d5d024d155c504080959582b928e91cb608a65c3a49e1a"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_i686.whl", hash = "sha256:067a74454df670dcaa4e59349a2e5c81e567d8d65458d480a5b3dfecec08c5ff"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:7538e495704e2ccda9b234b82423a4038f324f3a10c43bc088a1636180f11a41"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:af3d4a73793725138d6b334d9d247ce7e5f084d96284ed23f22ee626a7b88e39"},
- {file = "typed_ast-1.4.3-cp38-cp38-win32.whl", hash = "sha256:f2362f3cb0f3172c42938946dbc5b7843c2a28aec307c49100c8b38764eb6927"},
- {file = "typed_ast-1.4.3-cp38-cp38-win_amd64.whl", hash = "sha256:dd4a21253f42b8d2b48410cb31fe501d32f8b9fbeb1f55063ad102fe9c425e40"},
- {file = "typed_ast-1.4.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:f328adcfebed9f11301eaedfa48e15bdece9b519fb27e6a8c01aa52a17ec31b3"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2c726c276d09fc5c414693a2de063f521052d9ea7c240ce553316f70656c84d4"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:cae53c389825d3b46fb37538441f75d6aecc4174f615d048321b716df2757fb0"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:b9574c6f03f685070d859e75c7f9eeca02d6933273b5e69572e5ff9d5e3931c3"},
- {file = "typed_ast-1.4.3-cp39-cp39-win32.whl", hash = "sha256:209596a4ec71d990d71d5e0d312ac935d86930e6eecff6ccc7007fe54d703808"},
- {file = "typed_ast-1.4.3-cp39-cp39-win_amd64.whl", hash = "sha256:9c6d1a54552b5330bc657b7ef0eae25d00ba7ffe85d9ea8ae6540d2197a3788c"},
- {file = "typed_ast-1.4.3.tar.gz", hash = "sha256:fb1bbeac803adea29cedd70781399c99138358c26d05fcbd23c13016b7f5ec65"},
+ {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
+ {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
@@ -1542,3 +1495,3 @@ files = [
-name = "types-psutil"
-version = "5.9.5.6"
-description = "Typing stubs for psutil"
+name = "types-requests"
+version = "2.28.11.12"
+description = "Typing stubs for requests"
@@ -1549,2 +1502,17 @@ files = [
- {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
- {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+ {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
+ {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+]
+
+[package.dependencies]
+types-urllib3 = "<1.27"
+
+[[package]]
+name = "types-urllib3"
+version = "1.26.25.5"
+description = "Typing stubs for urllib3"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
+ {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
@@ -1555 +1523 @@ name = "typing-extensions"
-version = "4.4.0"
+version = "4.5.0"
@@ -1561,2 +1529,2 @@ files = [
- {file = "typing_extensions-4.4.0-py3-none-any.whl", hash = "sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e"},
- {file = "typing_extensions-4.4.0.tar.gz", hash = "sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa"},
+ {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"},
+ {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"},
@@ -1597 +1565 @@ python-versions = "3.9.15"
-content-hash = "bcb7717006cb6327c70332dcf65e402f151fa0fdfd483c4a02a40380a87b544b"
+content-hash = "f56e77566d857fade5ef9b4a8a6ee7bf99c5bafbc6ec625894646eeb1221e655"
diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml
index d38c50ab..141d7a4d 100644
--- a/libs/libcommon/pyproject.toml
+++ b/libs/libcommon/pyproject.toml
@@ -13,3 +13,3 @@ mongo-types = "0.15.1"
-mongoengine = "^0.24.1"
-orjson = "^3.6.4"
-psutil = "^5.9.2"
+mongoengine = "^0.24.2"
+orjson = "^3.8.6"
+psutil = "^5.9.4"
@@ -17,0 +18 @@ python = "3.9.15"
+requests = "^2.28.2"
@@ -20,2 +21,2 @@ python = "3.9.15"
-bandit = "^1.7.0"
-black = "^22.1.0"
+bandit = "^1.7.4"
+black = "^22.12.0"
@@ -23,5 +24,4 @@ flake8 = "^3.9.2"
-isort = "^5.9.3"
-mypy = "0.812"
-pip-audit = "^2.4.6"
-poetryup = "^0.3.8"
-pytest = "^7.2.0"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pip-audit = "^2.4.14"
+pytest = "^7.2.1"
@@ -29,0 +30 @@ types-psutil = "^5.9.5"
+types-requests = "^2.28.11"
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py
index f7340a60..7ef405e1 100644
--- a/libs/libcommon/src/libcommon/dataset.py
+++ b/libs/libcommon/src/libcommon/dataset.py
@@ -8,6 +8,3 @@ import requests
-from huggingface_hub.hf_api import (
- DatasetInfo,
- HfApi,
- RepositoryNotFoundError,
- build_hf_headers,
-)
+from huggingface_hub.hf_api import DatasetInfo, HfApi
+from huggingface_hub.utils._errors import RepositoryNotFoundError
+from huggingface_hub.utils._headers import build_hf_headers
diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py
index 80ffd6cf..656c84b1 100644
--- a/libs/libcommon/src/libcommon/processing_graph.py
+++ b/libs/libcommon/src/libcommon/processing_graph.py
@@ -44 +44 @@ class ProcessingStep:
- def job_type(self):
+ def job_type(self) -> str:
@@ -49 +49 @@ class ProcessingStep:
- def cache_kind(self):
+ def cache_kind(self) -> str:
diff --git a/libs/libcommon/src/libcommon/resources.py b/libs/libcommon/src/libcommon/resources.py
index 375215b7..bb227a5a 100644
--- a/libs/libcommon/src/libcommon/resources.py
+++ b/libs/libcommon/src/libcommon/resources.py
@@ -4,0 +5,2 @@ from dataclasses import dataclass, field
+from types import TracebackType
+from typing import Optional, Type, TypeVar
@@ -7 +9 @@ from mongoengine.connection import ConnectionFailure, connect, disconnect
-from pymongo import MongoClient # type: ignore
+from pymongo import MongoClient
@@ -11,0 +14,2 @@ from libcommon.constants import CACHE_MONGOENGINE_ALIAS, QUEUE_MONGOENGINE_ALIAS
+T = TypeVar("T", bound="Resource")
+
@@ -30 +34 @@ class Resource:
- def __post_init__(self):
+ def __post_init__(self) -> None:
@@ -33 +37 @@ class Resource:
- def __enter__(self):
+ def __enter__(self: T) -> T:
@@ -36 +40,6 @@ class Resource:
- def __exit__(self, exc_type, exc_value, traceback):
+ def __exit__(
+ self,
+ exc_type: Optional[Type[BaseException]],
+ exc_value: Optional[BaseException],
+ traceback: Optional[TracebackType],
+ ) -> None:
@@ -39 +48 @@ class Resource:
- def allocate(self):
+ def allocate(self) -> None:
@@ -42 +51 @@ class Resource:
- def release(self):
+ def release(self) -> None:
@@ -72 +81 @@ class MongoResource(Resource):
- def allocate(self):
+ def allocate(self) -> None:
@@ -91 +100 @@ class MongoResource(Resource):
- def release(self):
+ def release(self) -> None:
diff --git a/libs/libcommon/tests/conftest.py b/libs/libcommon/tests/conftest.py
index 0226bc1f..628df4bd 100644
--- a/libs/libcommon/tests/conftest.py
+++ b/libs/libcommon/tests/conftest.py
@@ -16 +16,4 @@ def cache_mongo_host(env: Env) -> str:
- return env.str(name="CACHE_MONGO_URL")
+ url = env.str(name="CACHE_MONGO_URL")
+ if type(url) is not str:
+ raise ValueError("CACHE_MONGO_URL is not set")
+ return url
@@ -24 +27,4 @@ def queue_mongo_host(env: Env) -> str:
- return env.str(name="QUEUE_MONGO_URL")
+ url = env.str(name="QUEUE_MONGO_URL")
+ if type(url) is not str:
+ raise ValueError("QUEUE_MONGO_URL is not set")
+ return url
diff --git a/libs/libcommon/tests/test_config.py b/libs/libcommon/tests/test_config.py
index 8dab322e..83b06a16 100644
--- a/libs/libcommon/tests/test_config.py
+++ b/libs/libcommon/tests/test_config.py
@@ -7 +7 @@ from libcommon.config import CommonConfig
-def test_common_config():
+def test_common_config() -> None:
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index 5a3d5192..e51e0d49 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -8 +8 @@ from libcommon.processing_graph import ProcessingGraph
-def test_default_graph():
+def test_default_graph() -> None:
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index 3a1c59c2..1b7f8d2d 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -14,0 +15,2 @@ from libcommon.simple_cache import (
+ CacheReportsPage,
+ CacheReportsWithContentPage,
@@ -356 +358 @@ def test_count_by_status_and_error_code() -> None:
- assert "OK" not in get_responses_count_by_kind_status_and_error_code()
+ assert not get_responses_count_by_kind_status_and_error_code()
@@ -390,2 +392,3 @@ def test_get_cache_reports() -> None:
- assert get_cache_reports(kind=kind, cursor="", limit=2) == {"cache_reports": [], "next_cursor": ""}
- assert get_cache_reports_with_content(kind=kind, cursor="", limit=2) == {
+ expected_cache_reports: CacheReportsPage = {"cache_reports": [], "next_cursor": ""}
+ assert get_cache_reports(kind=kind, cursor="", limit=2) == expected_cache_reports
+ expected_cache_reports_with_content: CacheReportsWithContentPage = {
@@ -394,0 +398 @@ def test_get_cache_reports() -> None:
+ assert get_cache_reports_with_content(kind=kind, cursor="", limit=2) == expected_cache_reports_with_content
diff --git a/services/admin/Makefile b/services/admin/Makefile
index 1482c281..c177523c 100644
--- a/services/admin/Makefile
+++ b/services/admin/Makefile
@@ -7,0 +8 @@ DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml
+TEST_PATH ?= tests
@@ -24 +24,0 @@ test:
- $(MAKE) down
@@ -26,2 +26,2 @@ test:
- poetry run python -m pytest -vv -x tests
- PROMETHEUS_MULTIPROC_DIR=/tmp poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" tests
+ poetry run python -m pytest -vv -x ${ADDOPTS} $(TEST_PATH)
+ PROMETHEUS_MULTIPROC_DIR=/tmp poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
@@ -32 +31,0 @@ coverage:
- $(MAKE) down
@@ -34 +33 @@ coverage:
- poetry run python -m pytest -s --cov --cov-report xml:coverage.xml --cov-report=term tests
+ poetry run python -m pytest -s --cov --cov-report xml:coverage.xml --cov-report=term $(TEST_PATH)
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index 86171519..95625f12 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -36,15 +35,0 @@ files = [
-[[package]]
-name = "asgiref"
-version = "3.6.0"
-description = "ASGI specs, helper code, and adapters"
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "asgiref-3.6.0-py3-none-any.whl", hash = "sha256:71e68008da809b957b7ee4b43dbccff33d1b23519fb8344e33f049897077afac"},
- {file = "asgiref-3.6.0.tar.gz", hash = "sha256:9567dfe7bd8d3c8c892227827c41cce860b368104c3431da67a0c5a65a949506"},
-]
-
-[package.extras]
-tests = ["mypy (>=0.800)", "pytest", "pytest-asyncio"]
-
@@ -518,0 +504,46 @@ lxml = ["lxml"]
+[[package]]
+name = "httpcore"
+version = "0.16.3"
+description = "A minimal low-level HTTP client."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "httpcore-0.16.3-py3-none-any.whl", hash = "sha256:da1fb708784a938aa084bde4feb8317056c55037247c787bd7e19eb2c2949dc0"},
+ {file = "httpcore-0.16.3.tar.gz", hash = "sha256:c5d6f04e2fc530f39e0c077e6a30caa53f1451096120f1f38b954afd0b17c0cb"},
+]
+
+[package.dependencies]
+anyio = ">=3.0,<5.0"
+certifi = "*"
+h11 = ">=0.13,<0.15"
+sniffio = ">=1.0.0,<2.0.0"
+
+[package.extras]
+http2 = ["h2 (>=3,<5)"]
+socks = ["socksio (>=1.0.0,<2.0.0)"]
+
+[[package]]
+name = "httpx"
+version = "0.23.3"
+description = "The next generation HTTP client."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "httpx-0.23.3-py3-none-any.whl", hash = "sha256:a211fcce9b1254ea24f0cd6af9869b3d29aba40154e947d2a07bb499b3e310d6"},
+ {file = "httpx-0.23.3.tar.gz", hash = "sha256:9818458eb565bb54898ccb9b8b251a28785dd4a55afbc23d0eb410754fe7d0f9"},
+]
+
+[package.dependencies]
+certifi = "*"
+httpcore = ">=0.15.0,<0.17.0"
+rfc3986 = {version = ">=1.3,<2", extras = ["idna2008"]}
+sniffio = "*"
+
+[package.extras]
+brotli = ["brotli", "brotlicffi"]
+cli = ["click (>=8.0.0,<9.0.0)", "pygments (>=2.0.0,<3.0.0)", "rich (>=10,<13)"]
+http2 = ["h2 (>=3,<5)"]
+socks = ["socksio (>=1.0.0,<2.0.0)"]
+
@@ -607,3 +638,3 @@ mongo-types = "0.15.1"
-mongoengine = "^0.24.1"
-orjson = "^3.6.4"
-psutil = "^5.9.2"
+mongoengine = "^0.24.2"
+orjson = "^3.8.6"
+psutil = "^5.9.4"
@@ -610,0 +642 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+requests = "^2.28.2"
@@ -789 +821 @@ name = "mypy"
-version = "0.812"
+version = "1.0.0"
@@ -793 +825 @@ optional = false
-python-versions = ">=3.5"
+python-versions = ">=3.7"
@@ -795,22 +827,26 @@ files = [
- {file = "mypy-0.812-cp35-cp35m-macosx_10_9_x86_64.whl", hash = "sha256:a26f8ec704e5a7423c8824d425086705e381b4f1dfdef6e3a1edab7ba174ec49"},
- {file = "mypy-0.812-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:28fb5479c494b1bab244620685e2eb3c3f988d71fd5d64cc753195e8ed53df7c"},
- {file = "mypy-0.812-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:9743c91088d396c1a5a3c9978354b61b0382b4e3c440ce83cf77994a43e8c521"},
- {file = "mypy-0.812-cp35-cp35m-win_amd64.whl", hash = "sha256:d7da2e1d5f558c37d6e8c1246f1aec1e7349e4913d8fb3cb289a35de573fe2eb"},
- {file = "mypy-0.812-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:4eec37370483331d13514c3f55f446fc5248d6373e7029a29ecb7b7494851e7a"},
- {file = "mypy-0.812-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:d65cc1df038ef55a99e617431f0553cd77763869eebdf9042403e16089fe746c"},
- {file = "mypy-0.812-cp36-cp36m-manylinux2010_x86_64.whl", hash = "sha256:61a3d5b97955422964be6b3baf05ff2ce7f26f52c85dd88db11d5e03e146a3a6"},
- {file = "mypy-0.812-cp36-cp36m-win_amd64.whl", hash = "sha256:25adde9b862f8f9aac9d2d11971f226bd4c8fbaa89fb76bdadb267ef22d10064"},
- {file = "mypy-0.812-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:552a815579aa1e995f39fd05dde6cd378e191b063f031f2acfe73ce9fb7f9e56"},
- {file = "mypy-0.812-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:499c798053cdebcaa916eef8cd733e5584b5909f789de856b482cd7d069bdad8"},
- {file = "mypy-0.812-cp37-cp37m-manylinux2010_x86_64.whl", hash = "sha256:5873888fff1c7cf5b71efbe80e0e73153fe9212fafdf8e44adfe4c20ec9f82d7"},
- {file = "mypy-0.812-cp37-cp37m-win_amd64.whl", hash = "sha256:9f94aac67a2045ec719ffe6111df543bac7874cee01f41928f6969756e030564"},
- {file = "mypy-0.812-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d23e0ea196702d918b60c8288561e722bf437d82cb7ef2edcd98cfa38905d506"},
- {file = "mypy-0.812-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:674e822aa665b9fd75130c6c5f5ed9564a38c6cea6a6432ce47eafb68ee578c5"},
- {file = "mypy-0.812-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:abf7e0c3cf117c44d9285cc6128856106183938c68fd4944763003decdcfeb66"},
- {file = "mypy-0.812-cp38-cp38-win_amd64.whl", hash = "sha256:0d0a87c0e7e3a9becdfbe936c981d32e5ee0ccda3e0f07e1ef2c3d1a817cf73e"},
- {file = "mypy-0.812-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:7ce3175801d0ae5fdfa79b4f0cfed08807af4d075b402b7e294e6aa72af9aa2a"},
- {file = "mypy-0.812-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:b09669bcda124e83708f34a94606e01b614fa71931d356c1f1a5297ba11f110a"},
- {file = "mypy-0.812-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:33f159443db0829d16f0a8d83d94df3109bb6dd801975fe86bacb9bf71628e97"},
- {file = "mypy-0.812-cp39-cp39-win_amd64.whl", hash = "sha256:3f2aca7f68580dc2508289c729bd49ee929a436208d2b2b6aab15745a70a57df"},
- {file = "mypy-0.812-py3-none-any.whl", hash = "sha256:2f9b3407c58347a452fc0736861593e105139b905cca7d097e413453a1d650b4"},
- {file = "mypy-0.812.tar.gz", hash = "sha256:cd07039aa5df222037005b08fbbfd69b3ab0b0bd7a07d7906de75ae52c4e3119"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
+ {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
+ {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
+ {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
+ {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
+ {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
+ {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
+ {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
+ {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
+ {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
+ {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
+ {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
+ {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
+ {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
+ {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
+ {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
+ {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
+ {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
+ {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
@@ -820,3 +856,3 @@ files = [
-mypy-extensions = ">=0.4.3,<0.5.0"
-typed-ast = ">=1.4.0,<1.5.0"
-typing-extensions = ">=3.7.4"
+mypy-extensions = ">=0.4.3"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=3.10"
@@ -825,0 +862,3 @@ dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
@@ -829,2 +868,2 @@ name = "mypy-extensions"
-version = "0.4.4"
-description = "Experimental type system extensions for programs checked with the mypy typechecker."
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
@@ -833 +872 @@ optional = false
-python-versions = ">=2.7"
+python-versions = ">=3.5"
@@ -835 +874,2 @@ files = [
- {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
@@ -1052,15 +1091,0 @@ testing = ["pytest", "pytest-benchmark"]
-[[package]]
-name = "poetryup"
-version = "0.3.15"
-description = "Update dependencies and bump their version in the pyproject.toml file"
-category = "dev"
-optional = false
-python-versions = ">=3.6,<4.0"
-files = [
- {file = "poetryup-0.3.15-py3-none-any.whl", hash = "sha256:db068f55d10c0f89c76ea2b62c6bb81c0b0512454f7a83bdc0a13c146e5fb13e"},
- {file = "poetryup-0.3.15.tar.gz", hash = "sha256:efa4e7bb0cd005db4aff3cc678c8bfba9474ef42d5759c0168f2a55fc0f17bc3"},
-]
-
-[package.dependencies]
-tomlkit = ">=0.7.2,<0.8.0"
-
@@ -1446 +1471 @@ name = "responses"
-version = "0.21.0"
+version = "0.22.0"
@@ -1452,2 +1477,2 @@ files = [
- {file = "responses-0.21.0-py3-none-any.whl", hash = "sha256:2dcc863ba63963c0c3d9ee3fa9507cbe36b7d7b0fccb4f0bdfd9e96c539b1487"},
- {file = "responses-0.21.0.tar.gz", hash = "sha256:b82502eb5f09a0289d8e209e7bad71ef3978334f56d09b444253d5ad67bf5253"},
+ {file = "responses-0.22.0-py3-none-any.whl", hash = "sha256:dcf294d204d14c436fddcc74caefdbc5764795a40ff4e6a7740ed8ddbf3294be"},
+ {file = "responses-0.22.0.tar.gz", hash = "sha256:396acb2a13d25297789a5866b4881cf4e46ffd49cc26c43ab1117f40b973102e"},
@@ -1457 +1482,3 @@ files = [
-requests = ">=2.0,<3.0"
+requests = ">=2.22.0,<3.0"
+toml = "*"
+types-toml = "*"
@@ -1461 +1488,19 @@ urllib3 = ">=1.25.10"
-tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=7.0.0)", "pytest-asyncio", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"]
+tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=7.0.0)", "pytest-asyncio", "pytest-cov", "pytest-httpserver", "types-requests"]
+
+[[package]]
+name = "rfc3986"
+version = "1.5.0"
+description = "Validating URI References per RFC 3986"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "rfc3986-1.5.0-py2.py3-none-any.whl", hash = "sha256:a86d6e1f5b1dc238b218b012df0aa79409667bb209e58da56d0b94704e712a97"},
+ {file = "rfc3986-1.5.0.tar.gz", hash = "sha256:270aaf10d87d0d4e095063c65bf3ddbc6ee3d0b226328ce21e036f946e421835"},
+]
+
+[package.dependencies]
+idna = {version = "*", optional = true, markers = "extra == \"idna2008\""}
+
+[package.extras]
+idna2008 = ["idna"]
@@ -1484 +1529 @@ name = "setuptools"
-version = "67.2.0"
+version = "67.3.1"
@@ -1490,2 +1535,2 @@ files = [
- {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
- {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
+ {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
+ {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
@@ -1549 +1594 @@ name = "starlette"
-version = "0.16.0"
+version = "0.25.0"
@@ -1553 +1598 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -1555,2 +1600,2 @@ files = [
- {file = "starlette-0.16.0-py3-none-any.whl", hash = "sha256:38eb24bf705a2c317e15868e384c1b8a12ca396e5a3c3a003db7e667c43f939f"},
- {file = "starlette-0.16.0.tar.gz", hash = "sha256:e1904b5d0007aee24bdd3c43994be9b3b729f4f58e740200de1d623f8c3a8870"},
+ {file = "starlette-0.25.0-py3-none-any.whl", hash = "sha256:774f1df1983fd594b9b6fb3ded39c2aa1979d10ac45caac0f4255cbe2acb8628"},
+ {file = "starlette-0.25.0.tar.gz", hash = "sha256:854c71e73736c429c2bdb07801f2c76c9cba497e7c3cf4988fde5e95fe4cdb3c"},
@@ -1560 +1605,2 @@ files = [
-anyio = ">=3.0.0,<4"
+anyio = ">=3.4.0,<5"
+typing-extensions = {version = ">=3.10.0", markers = "python_version < \"3.10\""}
@@ -1563 +1609 @@ anyio = ">=3.0.0,<4"
-full = ["graphene", "itsdangerous", "jinja2", "python-multipart", "pyyaml", "requests"]
+full = ["httpx (>=0.22.0)", "itsdangerous", "jinja2", "python-multipart", "pyyaml"]
@@ -1583 +1629 @@ name = "stevedore"
-version = "4.1.1"
+version = "5.0.0"
@@ -1589,2 +1635,2 @@ files = [
- {file = "stevedore-4.1.1-py3-none-any.whl", hash = "sha256:aa6436565c069b2946fe4ebff07f5041e0c8bf18c7376dd29edf80cf7d524e4e"},
- {file = "stevedore-4.1.1.tar.gz", hash = "sha256:7f8aeb6e3f90f96832c301bff21a7eb5eefbe894c88c506483d355565d88cc1a"},
+ {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"},
+ {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"},
@@ -1620,12 +1665,0 @@ files = [
-[[package]]
-name = "tomlkit"
-version = "0.7.2"
-description = "Style preserving TOML library"
-category = "dev"
-optional = false
-python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
-files = [
- {file = "tomlkit-0.7.2-py2.py3-none-any.whl", hash = "sha256:173ad840fa5d2aac140528ca1933c29791b79a374a0861a80347f42ec9328117"},
- {file = "tomlkit-0.7.2.tar.gz", hash = "sha256:d7a454f319a7e9bd2e249f239168729327e4dd2d27b17dc68be264ad1ce36754"},
-]
-
@@ -1654,3 +1688,15 @@ telegram = ["requests"]
-name = "typed-ast"
-version = "1.4.3"
-description = "a fork of Python 2 and 3 ast modules with type comment support"
+name = "types-psutil"
+version = "5.9.5.6"
+description = "Typing stubs for psutil"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
+ {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+]
+
+[[package]]
+name = "types-requests"
+version = "2.28.11.12"
+description = "Typing stubs for requests"
@@ -1661,30 +1707,2 @@ files = [
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:2068531575a125b87a41802130fa7e29f26c09a2833fea68d9a40cf33902eba6"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:c907f561b1e83e93fad565bac5ba9c22d96a54e7ea0267c708bffe863cbe4075"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux2014_aarch64.whl", hash = "sha256:1b3ead4a96c9101bef08f9f7d1217c096f31667617b58de957f690c92378b528"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win32.whl", hash = "sha256:dde816ca9dac1d9c01dd504ea5967821606f02e510438120091b84e852367428"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win_amd64.whl", hash = "sha256:777a26c84bea6cd934422ac2e3b78863a37017618b6e5c08f92ef69853e765d3"},
- {file = "typed_ast-1.4.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:f8afcf15cc511ada719a88e013cec87c11aff7b91f019295eb4530f96fe5ef2f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:52b1eb8c83f178ab787f3a4283f68258525f8d70f778a2f6dd54d3b5e5fb4341"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:01ae5f73431d21eead5015997ab41afa53aa1fbe252f9da060be5dad2c730ace"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:c190f0899e9f9f8b6b7863debfb739abcb21a5c054f911ca3596d12b8a4c4c7f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win32.whl", hash = "sha256:398e44cd480f4d2b7ee8d98385ca104e35c81525dd98c519acff1b79bdaac363"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win_amd64.whl", hash = "sha256:bff6ad71c81b3bba8fa35f0f1921fb24ff4476235a6e94a26ada2e54370e6da7"},
- {file = "typed_ast-1.4.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0fb71b8c643187d7492c1f8352f2c15b4c4af3f6338f21681d3681b3dc31a266"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:760ad187b1041a154f0e4d0f6aae3e40fdb51d6de16e5c99aedadd9246450e9e"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:5feca99c17af94057417d744607b82dd0a664fd5e4ca98061480fd8b14b18d04"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:95431a26309a21874005845c21118c83991c63ea800dd44843e42a916aec5899"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win32.whl", hash = "sha256:aee0c1256be6c07bd3e1263ff920c325b59849dc95392a05f258bb9b259cf39c"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9ad2c92ec681e02baf81fdfa056fe0d818645efa9af1f1cd5fd6f1bd2bdfd805"},
- {file = "typed_ast-1.4.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b36b4f3920103a25e1d5d024d155c504080959582b928e91cb608a65c3a49e1a"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_i686.whl", hash = "sha256:067a74454df670dcaa4e59349a2e5c81e567d8d65458d480a5b3dfecec08c5ff"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:7538e495704e2ccda9b234b82423a4038f324f3a10c43bc088a1636180f11a41"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:af3d4a73793725138d6b334d9d247ce7e5f084d96284ed23f22ee626a7b88e39"},
- {file = "typed_ast-1.4.3-cp38-cp38-win32.whl", hash = "sha256:f2362f3cb0f3172c42938946dbc5b7843c2a28aec307c49100c8b38764eb6927"},
- {file = "typed_ast-1.4.3-cp38-cp38-win_amd64.whl", hash = "sha256:dd4a21253f42b8d2b48410cb31fe501d32f8b9fbeb1f55063ad102fe9c425e40"},
- {file = "typed_ast-1.4.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:f328adcfebed9f11301eaedfa48e15bdece9b519fb27e6a8c01aa52a17ec31b3"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2c726c276d09fc5c414693a2de063f521052d9ea7c240ce553316f70656c84d4"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:cae53c389825d3b46fb37538441f75d6aecc4174f615d048321b716df2757fb0"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:b9574c6f03f685070d859e75c7f9eeca02d6933273b5e69572e5ff9d5e3931c3"},
- {file = "typed_ast-1.4.3-cp39-cp39-win32.whl", hash = "sha256:209596a4ec71d990d71d5e0d312ac935d86930e6eecff6ccc7007fe54d703808"},
- {file = "typed_ast-1.4.3-cp39-cp39-win_amd64.whl", hash = "sha256:9c6d1a54552b5330bc657b7ef0eae25d00ba7ffe85d9ea8ae6540d2197a3788c"},
- {file = "typed_ast-1.4.3.tar.gz", hash = "sha256:fb1bbeac803adea29cedd70781399c99138358c26d05fcbd23c13016b7f5ec65"},
+ {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
+ {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
@@ -1692,0 +1711,3 @@ files = [
+[package.dependencies]
+types-urllib3 = "<1.27"
+
@@ -1694,3 +1715,3 @@ files = [
-name = "types-psutil"
-version = "5.9.5.6"
-description = "Typing stubs for psutil"
+name = "types-toml"
+version = "0.10.8.3"
+description = "Typing stubs for toml"
@@ -1701,2 +1722,14 @@ files = [
- {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
- {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+ {file = "types-toml-0.10.8.3.tar.gz", hash = "sha256:f37244eff4cd7eace9cb70d0bac54d3eba77973aa4ef26c271ac3d1c6503a48e"},
+ {file = "types_toml-0.10.8.3-py3-none-any.whl", hash = "sha256:a2286a053aea6ab6ff814659272b1d4a05d86a1dd52b807a87b23511993b46c5"},
+]
+
+[[package]]
+name = "types-urllib3"
+version = "1.26.25.5"
+description = "Typing stubs for urllib3"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
+ {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
@@ -1707 +1740 @@ name = "typing-extensions"
-version = "4.4.0"
+version = "4.5.0"
@@ -1713,2 +1746,2 @@ files = [
- {file = "typing_extensions-4.4.0-py3-none-any.whl", hash = "sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e"},
- {file = "typing_extensions-4.4.0.tar.gz", hash = "sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa"},
+ {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"},
+ {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"},
@@ -1736 +1769 @@ name = "uvicorn"
-version = "0.14.0"
+version = "0.20.0"
@@ -1740 +1773 @@ optional = false
-python-versions = "*"
+python-versions = ">=3.7"
@@ -1742,2 +1775,2 @@ files = [
- {file = "uvicorn-0.14.0-py3-none-any.whl", hash = "sha256:2a76bb359171a504b3d1c853409af3adbfa5cef374a4a59e5881945a97a93eae"},
- {file = "uvicorn-0.14.0.tar.gz", hash = "sha256:45ad7dfaaa7d55cab4cd1e85e03f27e9d60bc067ddc59db52a2b0aeca8870292"},
+ {file = "uvicorn-0.20.0-py3-none-any.whl", hash = "sha256:c3ed1598a5668208723f2bb49336f4509424ad198d6ab2615b7783db58d919fd"},
+ {file = "uvicorn-0.20.0.tar.gz", hash = "sha256:a4e12017b940247f836bc90b72e725d7dfd0c8ed1c51eb365f5ba30d9f5127d8"},
@@ -1747,2 +1780 @@ files = [
-asgiref = ">=3.3.4"
-click = ">=7"
+click = ">=7.0"
@@ -1752 +1784 @@ h11 = ">=0.8"
-standard = ["PyYAML (>=5.1)", "colorama (>=0.4)", "httptools (>=0.2.0,<0.3.0)", "python-dotenv (>=0.13)", "uvloop (>=0.14.0,!=0.15.0,!=0.15.1)", "watchgod (>=0.6)", "websockets (>=9.1)"]
+standard = ["colorama (>=0.4)", "httptools (>=0.5.0)", "python-dotenv (>=0.13)", "pyyaml (>=5.1)", "uvloop (>=0.14.0,!=0.15.0,!=0.15.1)", "watchfiles (>=0.13)", "websockets (>=10.4)"]
@@ -1813 +1845 @@ python-versions = "3.9.15"
-content-hash = "6758d4c8d9fc7cc871367f0f715dc44aade5b2e4c66ba38894ec08b2c698612b"
+content-hash = "78aebf9305fa94fab4f0ac3d225c25480c48e8cf74a4eee1c63af22e4f76686d"
diff --git a/services/admin/pyproject.toml b/services/admin/pyproject.toml
index ad77eeda..7680b5f8 100644
--- a/services/admin/pyproject.toml
+++ b/services/admin/pyproject.toml
@@ -12 +12,2 @@ python = "3.9.15"
-starlette = "^0.16.0"
+requests = "^2.28.2"
+starlette = "^0.25.0"
@@ -14,2 +15,2 @@ starlette-prometheus = "^0.9.0"
-uvicorn = "^0.14.0"
-watchdog = { extras = ["watchmedo"], version = "^2.1.3" }
+uvicorn = "^0.20.0"
+watchdog = { extras = ["watchmedo"], version = "^2.2.1" }
@@ -18,2 +19,2 @@ watchdog = { extras = ["watchmedo"], version = "^2.1.3" }
-bandit = "^1.7.0"
-black = "^22.1.0"
+bandit = "^1.7.4"
+black = "^22.12.0"
@@ -20,0 +22 @@ flake8 = "^3.9.2"
+httpx = "^0.23.3"
@@ -22,4 +24,4 @@ huggingface-hub = "^0.12.0"
-isort = "^5.9.3"
-mypy = "0.812"
-poetryup = "^0.3.8"
-pytest = "^7.2.0"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pip-audit = "^2.4.14"
+pytest = "^7.2.1"
@@ -27,2 +29 @@ pytest-cov = "^2.12.1"
-pip-audit = "^2.4.6"
-responses = "^0.21.0"
+responses = "^0.22.0"
@@ -29,0 +31 @@ types-psutil = "^5.9.5"
+types-requests = "^2.28.11"
diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py
index d28addc7..e19cb6b5 100644
--- a/services/admin/src/admin/app.py
+++ b/services/admin/src/admin/app.py
@@ -4 +4 @@
-import uvicorn # type: ignore
+import uvicorn
diff --git a/services/admin/src/admin/prometheus.py b/services/admin/src/admin/prometheus.py
index b668efa4..f54bbe2c 100644
--- a/services/admin/src/admin/prometheus.py
+++ b/services/admin/src/admin/prometheus.py
@@ -6 +6 @@ from dataclasses import dataclass
-from typing import List
+from typing import Any, List
@@ -12 +12 @@ from libcommon.storage import StrPath
-from prometheus_client import ( # type: ignore # https://github.com/prometheus/client_python/issues/491
+from prometheus_client import ( # type: ignore[import]
@@ -19,3 +19,4 @@ from prometheus_client import ( # type: ignore # https://github.com/prometheus/
-from prometheus_client.multiprocess import ( # type: ignore # https://github.com/prometheus/client_python/issues/491
- MultiProcessCollector,
-)
+from prometheus_client.multiprocess import MultiProcessCollector # type: ignore[import]
+
+# ^ type: ignore can be removed on next release:
+# https://github.com/prometheus/client_python/issues/491#issuecomment-1429287314
@@ -62 +63 @@ class Prometheus:
- def updateMetrics(self):
+ def updateMetrics(self) -> None:
@@ -80 +81,2 @@ class Prometheus:
- def getLatestContent(self) -> str:
+ def getLatestContent(self) -> Any:
+ # ^ returns Any because we cannot be sure latest are UTF8Bytes
@@ -82 +84,2 @@ class Prometheus:
- return generate_latest(self.getRegistry()).decode("utf-8")
+ latest = generate_latest(self.getRegistry())
+ return latest.decode("utf-8")
diff --git a/services/admin/src/admin/routes/dataset_status.py b/services/admin/src/admin/routes/dataset_status.py
index 602cb17a..6ddadd53 100644
--- a/services/admin/src/admin/routes/dataset_status.py
+++ b/services/admin/src/admin/routes/dataset_status.py
@@ -34 +34 @@ def create_dataset_status_endpoint(
- if not are_valid_parameters([dataset]):
+ if not are_valid_parameters([dataset]) or not dataset:
diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py
index 40b97558..3010b407 100644
--- a/services/admin/src/admin/routes/force_refresh.py
+++ b/services/admin/src/admin/routes/force_refresh.py
@@ -35 +35 @@ def create_force_refresh_endpoint(
- if not are_valid_parameters([dataset]):
+ if not are_valid_parameters([dataset]) or not dataset:
diff --git a/services/admin/tests/conftest.py b/services/admin/tests/conftest.py
index ccafee55..1e21b047 100644
--- a/services/admin/tests/conftest.py
+++ b/services/admin/tests/conftest.py
@@ -21 +21 @@ pytest_plugins = ["tests.fixtures.hub"]
-def monkeypatch_session(hf_endpoint: str, hf_token: str):
+def monkeypatch_session(hf_endpoint: str, hf_token: str) -> Iterator[MonkeyPatch]:
diff --git a/services/admin/tests/fixtures/hub.py b/services/admin/tests/fixtures/hub.py
index aa79005b..42aaba31 100644
--- a/services/admin/tests/fixtures/hub.py
+++ b/services/admin/tests/fixtures/hub.py
@@ -8 +8 @@ from contextlib import contextmanager, suppress
-from typing import Iterable, Literal, Mapping, Optional, TypedDict
+from typing import Any, Callable, Iterator, Literal, Optional, TypedDict
@@ -12,6 +12,3 @@ import requests
-from huggingface_hub.hf_api import (
- REPO_TYPES,
- REPO_TYPES_URL_PREFIXES,
- HfApi,
- hf_raise_for_status,
-)
+from huggingface_hub.constants import REPO_TYPES, REPO_TYPES_URL_PREFIXES
+from huggingface_hub.hf_api import HfApi
+from huggingface_hub.utils._errors import hf_raise_for_status
@@ -35,2 +32,2 @@ def update_repo_settings(
- name: str = None,
-) -> Mapping[str, bool]:
+ name: Optional[str] = None,
+) -> Any:
@@ -96 +93 @@ def update_repo_settings(
-def hf_api():
+def hf_api() -> HfApi:
@@ -111,2 +108,2 @@ def hf_token() -> str:
-def cleanup_repo(hf_api: HfApi):
- def _cleanup_repo(repo_id):
+def cleanup_repo(hf_api: HfApi) -> Callable[[str], None]:
+ def _cleanup_repo(repo_id: str) -> None:
@@ -119 +116 @@ def cleanup_repo(hf_api: HfApi):
-def temporary_repo(cleanup_repo):
+def temporary_repo(cleanup_repo: Callable[[str], None]) -> Callable[[str], Iterator[str]]:
@@ -121 +118 @@ def temporary_repo(cleanup_repo):
- def _temporary_repo(repo_id):
+ def _temporary_repo(repo_id: str) -> Iterator[str]:
@@ -127 +124 @@ def temporary_repo(cleanup_repo):
- return _temporary_repo
+ return _temporary_repo # type: ignore
@@ -136 +133 @@ def create_hf_dataset_repo(
- hf_api: HfApi, hf_token: str, prefix: str, *, private=False, gated=False, user=CI_HUB_USER
+ hf_api: HfApi, hf_token: str, prefix: str, *, private: bool = False, gated: bool = False, user: str = CI_HUB_USER
@@ -147 +144 @@ def create_hf_dataset_repo(
-def hf_public_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterable[str]:
+def hf_public_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterator[str]:
@@ -155 +152 @@ def hf_public_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterable[str]:
-def hf_gated_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterable[str]:
+def hf_gated_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterator[str]:
@@ -163 +160 @@ def hf_gated_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterable[str]:
-def hf_private_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterable[str]:
+def hf_private_dataset_repo_empty(hf_api: HfApi, hf_token: str) -> Iterator[str]:
@@ -181,3 +178,3 @@ def hf_dataset_repos_csv_data(
- hf_public_dataset_repo_empty,
- hf_gated_dataset_repo_empty,
- hf_private_dataset_repo_empty,
+ hf_public_dataset_repo_empty: str,
+ hf_gated_dataset_repo_empty: str,
+ hf_private_dataset_repo_empty: str,
diff --git a/services/admin/tests/test_app.py b/services/admin/tests/test_app.py
index afee6b0a..8a3a23f3 100644
--- a/services/admin/tests/test_app.py
+++ b/services/admin/tests/test_app.py
@@ -22 +22,2 @@ def test_cors(client: TestClient) -> None:
- response = client.options(
+ response = client.request(
+ "options",
@@ -47 +48 @@ def test_get_healthcheck(client: TestClient) -> None:
- response = client.get("/healthcheck")
+ response = client.request("get", "/healthcheck")
@@ -53 +54 @@ def test_metrics(client: TestClient) -> None:
- response = client.get("/metrics")
+ response = client.request("get", "/metrics")
@@ -66 +67 @@ def test_pending_jobs(client: TestClient, processing_steps: List[ProcessingStep]
- response = client.get("/pending-jobs")
+ response = client.request("get", "/pending-jobs")
@@ -74 +75 @@ def test_dataset_status(client: TestClient, processing_steps: List[ProcessingSte
- response = client.get("/dataset-status")
+ response = client.request("get", "/dataset-status")
@@ -76 +77 @@ def test_dataset_status(client: TestClient, processing_steps: List[ProcessingSte
- response = client.get("/dataset-status", params={"dataset": "test-dataset"})
+ response = client.request("get", "/dataset-status", params={"dataset": "test-dataset"})
@@ -101 +102 @@ def test_cache_reports(
- response = client.get(f"/cache-reports{path}{cursor_str}")
+ response = client.request("get", f"/cache-reports{path}{cursor_str}")
@@ -128 +129 @@ def test_cache_reports_with_content(
- response = client.get(f"/cache-reports-with-content{path}{cursor_str}")
+ response = client.request("get", f"/cache-reports-with-content{path}{cursor_str}")
diff --git a/services/admin/tests/test_app_real.py b/services/admin/tests/test_app_real.py
index a3183019..f41f733f 100644
--- a/services/admin/tests/test_app_real.py
+++ b/services/admin/tests/test_app_real.py
@@ -3,0 +4,2 @@
+from typing import Iterator
+
@@ -14 +16 @@ from admin.config import AppConfig
-def real_monkeypatch():
+def real_monkeypatch() -> Iterator[MonkeyPatch]:
@@ -47 +49 @@ def test_force_refresh(
- response = real_client.post(f"/force-refresh{path}?dataset={dataset}")
+ response = real_client.request("post", f"/force-refresh{path}?dataset={dataset}")
diff --git a/services/admin/tests/test_authentication.py b/services/admin/tests/test_authentication.py
index 3c285bb1..517f713a 100644
--- a/services/admin/tests/test_authentication.py
+++ b/services/admin/tests/test_authentication.py
@@ -8 +8,2 @@ import responses
-from starlette.requests import Headers, Request
+from starlette.datastructures import Headers
+from starlette.requests import Request
diff --git a/services/api/Makefile b/services/api/Makefile
index 98ba5614..004c8356 100644
--- a/services/api/Makefile
+++ b/services/api/Makefile
@@ -7,0 +8 @@ DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml
+TEST_PATH ?= tests
@@ -24 +24,0 @@ test:
- $(MAKE) down
@@ -26,2 +26,2 @@ test:
- poetry run python -m pytest -vv -x tests
- PROMETHEUS_MULTIPROC_DIR=/tmp poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" tests
+ poetry run python -m pytest -vv -x ${ADDOPTS} $(TEST_PATH)
+ PROMETHEUS_MULTIPROC_DIR=/tmp poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
@@ -32 +31,0 @@ coverage:
- $(MAKE) down
@@ -34 +33 @@ coverage:
- poetry run python -m pytest -s --cov --cov-report xml:coverage.xml --cov-report=term tests
+ poetry run python -m pytest -s --cov --cov-report xml:coverage.xml --cov-report=term $(TEST_PATH)
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index f9855ee0..99a74d80 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -36,15 +35,0 @@ files = [
-[[package]]
-name = "asgiref"
-version = "3.6.0"
-description = "ASGI specs, helper code, and adapters"
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "asgiref-3.6.0-py3-none-any.whl", hash = "sha256:71e68008da809b957b7ee4b43dbccff33d1b23519fb8344e33f049897077afac"},
- {file = "asgiref-3.6.0.tar.gz", hash = "sha256:9567dfe7bd8d3c8c892227827c41cce860b368104c3431da67a0c5a65a949506"},
-]
-
-[package.extras]
-tests = ["mypy (>=0.800)", "pytest", "pytest-asyncio"]
-
@@ -518,0 +504,46 @@ lxml = ["lxml"]
+[[package]]
+name = "httpcore"
+version = "0.16.3"
+description = "A minimal low-level HTTP client."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "httpcore-0.16.3-py3-none-any.whl", hash = "sha256:da1fb708784a938aa084bde4feb8317056c55037247c787bd7e19eb2c2949dc0"},
+ {file = "httpcore-0.16.3.tar.gz", hash = "sha256:c5d6f04e2fc530f39e0c077e6a30caa53f1451096120f1f38b954afd0b17c0cb"},
+]
+
+[package.dependencies]
+anyio = ">=3.0,<5.0"
+certifi = "*"
+h11 = ">=0.13,<0.15"
+sniffio = ">=1.0.0,<2.0.0"
+
+[package.extras]
+http2 = ["h2 (>=3,<5)"]
+socks = ["socksio (>=1.0.0,<2.0.0)"]
+
+[[package]]
+name = "httpx"
+version = "0.23.3"
+description = "The next generation HTTP client."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "httpx-0.23.3-py3-none-any.whl", hash = "sha256:a211fcce9b1254ea24f0cd6af9869b3d29aba40154e947d2a07bb499b3e310d6"},
+ {file = "httpx-0.23.3.tar.gz", hash = "sha256:9818458eb565bb54898ccb9b8b251a28785dd4a55afbc23d0eb410754fe7d0f9"},
+]
+
+[package.dependencies]
+certifi = "*"
+httpcore = ">=0.15.0,<0.17.0"
+rfc3986 = {version = ">=1.3,<2", extras = ["idna2008"]}
+sniffio = "*"
+
+[package.extras]
+brotli = ["brotli", "brotlicffi"]
+cli = ["click (>=8.0.0,<9.0.0)", "pygments (>=2.0.0,<3.0.0)", "rich (>=10,<13)"]
+http2 = ["h2 (>=3,<5)"]
+socks = ["socksio (>=1.0.0,<2.0.0)"]
+
@@ -627,3 +658,3 @@ mongo-types = "0.15.1"
-mongoengine = "^0.24.1"
-orjson = "^3.6.4"
-psutil = "^5.9.2"
+mongoengine = "^0.24.2"
+orjson = "^3.8.6"
+psutil = "^5.9.4"
@@ -630,0 +662 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+requests = "^2.28.2"
@@ -869 +901 @@ name = "mypy"
-version = "0.812"
+version = "1.0.0"
@@ -873 +905 @@ optional = false
-python-versions = ">=3.5"
+python-versions = ">=3.7"
@@ -875,22 +907,26 @@ files = [
- {file = "mypy-0.812-cp35-cp35m-macosx_10_9_x86_64.whl", hash = "sha256:a26f8ec704e5a7423c8824d425086705e381b4f1dfdef6e3a1edab7ba174ec49"},
- {file = "mypy-0.812-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:28fb5479c494b1bab244620685e2eb3c3f988d71fd5d64cc753195e8ed53df7c"},
- {file = "mypy-0.812-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:9743c91088d396c1a5a3c9978354b61b0382b4e3c440ce83cf77994a43e8c521"},
- {file = "mypy-0.812-cp35-cp35m-win_amd64.whl", hash = "sha256:d7da2e1d5f558c37d6e8c1246f1aec1e7349e4913d8fb3cb289a35de573fe2eb"},
- {file = "mypy-0.812-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:4eec37370483331d13514c3f55f446fc5248d6373e7029a29ecb7b7494851e7a"},
- {file = "mypy-0.812-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:d65cc1df038ef55a99e617431f0553cd77763869eebdf9042403e16089fe746c"},
- {file = "mypy-0.812-cp36-cp36m-manylinux2010_x86_64.whl", hash = "sha256:61a3d5b97955422964be6b3baf05ff2ce7f26f52c85dd88db11d5e03e146a3a6"},
- {file = "mypy-0.812-cp36-cp36m-win_amd64.whl", hash = "sha256:25adde9b862f8f9aac9d2d11971f226bd4c8fbaa89fb76bdadb267ef22d10064"},
- {file = "mypy-0.812-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:552a815579aa1e995f39fd05dde6cd378e191b063f031f2acfe73ce9fb7f9e56"},
- {file = "mypy-0.812-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:499c798053cdebcaa916eef8cd733e5584b5909f789de856b482cd7d069bdad8"},
- {file = "mypy-0.812-cp37-cp37m-manylinux2010_x86_64.whl", hash = "sha256:5873888fff1c7cf5b71efbe80e0e73153fe9212fafdf8e44adfe4c20ec9f82d7"},
- {file = "mypy-0.812-cp37-cp37m-win_amd64.whl", hash = "sha256:9f94aac67a2045ec719ffe6111df543bac7874cee01f41928f6969756e030564"},
- {file = "mypy-0.812-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d23e0ea196702d918b60c8288561e722bf437d82cb7ef2edcd98cfa38905d506"},
- {file = "mypy-0.812-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:674e822aa665b9fd75130c6c5f5ed9564a38c6cea6a6432ce47eafb68ee578c5"},
- {file = "mypy-0.812-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:abf7e0c3cf117c44d9285cc6128856106183938c68fd4944763003decdcfeb66"},
- {file = "mypy-0.812-cp38-cp38-win_amd64.whl", hash = "sha256:0d0a87c0e7e3a9becdfbe936c981d32e5ee0ccda3e0f07e1ef2c3d1a817cf73e"},
- {file = "mypy-0.812-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:7ce3175801d0ae5fdfa79b4f0cfed08807af4d075b402b7e294e6aa72af9aa2a"},
- {file = "mypy-0.812-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:b09669bcda124e83708f34a94606e01b614fa71931d356c1f1a5297ba11f110a"},
- {file = "mypy-0.812-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:33f159443db0829d16f0a8d83d94df3109bb6dd801975fe86bacb9bf71628e97"},
- {file = "mypy-0.812-cp39-cp39-win_amd64.whl", hash = "sha256:3f2aca7f68580dc2508289c729bd49ee929a436208d2b2b6aab15745a70a57df"},
- {file = "mypy-0.812-py3-none-any.whl", hash = "sha256:2f9b3407c58347a452fc0736861593e105139b905cca7d097e413453a1d650b4"},
- {file = "mypy-0.812.tar.gz", hash = "sha256:cd07039aa5df222037005b08fbbfd69b3ab0b0bd7a07d7906de75ae52c4e3119"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
+ {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
+ {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
+ {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
+ {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
+ {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
+ {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
+ {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
+ {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
+ {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
+ {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
+ {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
+ {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
+ {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
+ {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
+ {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
+ {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
+ {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
+ {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
@@ -900,3 +936,3 @@ files = [
-mypy-extensions = ">=0.4.3,<0.5.0"
-typed-ast = ">=1.4.0,<1.5.0"
-typing-extensions = ">=3.7.4"
+mypy-extensions = ">=0.4.3"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=3.10"
@@ -905,0 +942,3 @@ dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
@@ -909,2 +948,2 @@ name = "mypy-extensions"
-version = "0.4.4"
-description = "Experimental type system extensions for programs checked with the mypy typechecker."
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
@@ -913 +952 @@ optional = false
-python-versions = ">=2.7"
+python-versions = ">=3.5"
@@ -915 +954,2 @@ files = [
- {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
@@ -1132,15 +1171,0 @@ testing = ["pytest", "pytest-benchmark"]
-[[package]]
-name = "poetryup"
-version = "0.3.15"
-description = "Update dependencies and bump their version in the pyproject.toml file"
-category = "dev"
-optional = false
-python-versions = ">=3.6,<4.0"
-files = [
- {file = "poetryup-0.3.15-py3-none-any.whl", hash = "sha256:db068f55d10c0f89c76ea2b62c6bb81c0b0512454f7a83bdc0a13c146e5fb13e"},
- {file = "poetryup-0.3.15.tar.gz", hash = "sha256:efa4e7bb0cd005db4aff3cc678c8bfba9474ef42d5759c0168f2a55fc0f17bc3"},
-]
-
-[package.dependencies]
-tomlkit = ">=0.7.2,<0.8.0"
-
@@ -1575,0 +1601,18 @@ test = ["commentjson", "packaging", "pytest"]
+[[package]]
+name = "rfc3986"
+version = "1.5.0"
+description = "Validating URI References per RFC 3986"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "rfc3986-1.5.0-py2.py3-none-any.whl", hash = "sha256:a86d6e1f5b1dc238b218b012df0aa79409667bb209e58da56d0b94704e712a97"},
+ {file = "rfc3986-1.5.0.tar.gz", hash = "sha256:270aaf10d87d0d4e095063c65bf3ddbc6ee3d0b226328ce21e036f946e421835"},
+]
+
+[package.dependencies]
+idna = {version = "*", optional = true, markers = "extra == \"idna2008\""}
+
+[package.extras]
+idna2008 = ["idna"]
+
@@ -1597 +1640 @@ name = "setuptools"
-version = "67.2.0"
+version = "67.3.1"
@@ -1603,2 +1646,2 @@ files = [
- {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
- {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
+ {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
+ {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
@@ -1662 +1705 @@ name = "starlette"
-version = "0.16.0"
+version = "0.25.0"
@@ -1666 +1709 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -1668,2 +1711,2 @@ files = [
- {file = "starlette-0.16.0-py3-none-any.whl", hash = "sha256:38eb24bf705a2c317e15868e384c1b8a12ca396e5a3c3a003db7e667c43f939f"},
- {file = "starlette-0.16.0.tar.gz", hash = "sha256:e1904b5d0007aee24bdd3c43994be9b3b729f4f58e740200de1d623f8c3a8870"},
+ {file = "starlette-0.25.0-py3-none-any.whl", hash = "sha256:774f1df1983fd594b9b6fb3ded39c2aa1979d10ac45caac0f4255cbe2acb8628"},
+ {file = "starlette-0.25.0.tar.gz", hash = "sha256:854c71e73736c429c2bdb07801f2c76c9cba497e7c3cf4988fde5e95fe4cdb3c"},
@@ -1673 +1716,2 @@ files = [
-anyio = ">=3.0.0,<4"
+anyio = ">=3.4.0,<5"
+typing-extensions = {version = ">=3.10.0", markers = "python_version < \"3.10\""}
@@ -1676 +1720 @@ anyio = ">=3.0.0,<4"
-full = ["graphene", "itsdangerous", "jinja2", "python-multipart", "pyyaml", "requests"]
+full = ["httpx (>=0.22.0)", "itsdangerous", "jinja2", "python-multipart", "pyyaml"]
@@ -1696 +1740 @@ name = "stevedore"
-version = "4.1.1"
+version = "5.0.0"
@@ -1702,2 +1746,2 @@ files = [
- {file = "stevedore-4.1.1-py3-none-any.whl", hash = "sha256:aa6436565c069b2946fe4ebff07f5041e0c8bf18c7376dd29edf80cf7d524e4e"},
- {file = "stevedore-4.1.1.tar.gz", hash = "sha256:7f8aeb6e3f90f96832c301bff21a7eb5eefbe894c88c506483d355565d88cc1a"},
+ {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"},
+ {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"},
@@ -1733,12 +1776,0 @@ files = [
-[[package]]
-name = "tomlkit"
-version = "0.7.2"
-description = "Style preserving TOML library"
-category = "dev"
-optional = false
-python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
-files = [
- {file = "tomlkit-0.7.2-py2.py3-none-any.whl", hash = "sha256:173ad840fa5d2aac140528ca1933c29791b79a374a0861a80347f42ec9328117"},
- {file = "tomlkit-0.7.2.tar.gz", hash = "sha256:d7a454f319a7e9bd2e249f239168729327e4dd2d27b17dc68be264ad1ce36754"},
-]
-
@@ -1767,3 +1799,15 @@ telegram = ["requests"]
-name = "typed-ast"
-version = "1.4.3"
-description = "a fork of Python 2 and 3 ast modules with type comment support"
+name = "types-jsonschema"
+version = "4.17.0.4"
+description = "Typing stubs for jsonschema"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-jsonschema-4.17.0.4.tar.gz", hash = "sha256:f8e82ef9b618c7183ddbf604684b7a656119f388deea8fe8d4238a388bd919c2"},
+ {file = "types_jsonschema-4.17.0.4-py3-none-any.whl", hash = "sha256:2f5e7848c8f210a92973a92c08f661cbfe7d1b33ea4cbc5d4c34a49b4adc8cd7"},
+]
+
+[[package]]
+name = "types-psutil"
+version = "5.9.5.6"
+description = "Typing stubs for psutil"
@@ -1774,30 +1818,29 @@ files = [
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:2068531575a125b87a41802130fa7e29f26c09a2833fea68d9a40cf33902eba6"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:c907f561b1e83e93fad565bac5ba9c22d96a54e7ea0267c708bffe863cbe4075"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux2014_aarch64.whl", hash = "sha256:1b3ead4a96c9101bef08f9f7d1217c096f31667617b58de957f690c92378b528"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win32.whl", hash = "sha256:dde816ca9dac1d9c01dd504ea5967821606f02e510438120091b84e852367428"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win_amd64.whl", hash = "sha256:777a26c84bea6cd934422ac2e3b78863a37017618b6e5c08f92ef69853e765d3"},
- {file = "typed_ast-1.4.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:f8afcf15cc511ada719a88e013cec87c11aff7b91f019295eb4530f96fe5ef2f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:52b1eb8c83f178ab787f3a4283f68258525f8d70f778a2f6dd54d3b5e5fb4341"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:01ae5f73431d21eead5015997ab41afa53aa1fbe252f9da060be5dad2c730ace"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:c190f0899e9f9f8b6b7863debfb739abcb21a5c054f911ca3596d12b8a4c4c7f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win32.whl", hash = "sha256:398e44cd480f4d2b7ee8d98385ca104e35c81525dd98c519acff1b79bdaac363"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win_amd64.whl", hash = "sha256:bff6ad71c81b3bba8fa35f0f1921fb24ff4476235a6e94a26ada2e54370e6da7"},
- {file = "typed_ast-1.4.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0fb71b8c643187d7492c1f8352f2c15b4c4af3f6338f21681d3681b3dc31a266"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:760ad187b1041a154f0e4d0f6aae3e40fdb51d6de16e5c99aedadd9246450e9e"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:5feca99c17af94057417d744607b82dd0a664fd5e4ca98061480fd8b14b18d04"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:95431a26309a21874005845c21118c83991c63ea800dd44843e42a916aec5899"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win32.whl", hash = "sha256:aee0c1256be6c07bd3e1263ff920c325b59849dc95392a05f258bb9b259cf39c"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9ad2c92ec681e02baf81fdfa056fe0d818645efa9af1f1cd5fd6f1bd2bdfd805"},
- {file = "typed_ast-1.4.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b36b4f3920103a25e1d5d024d155c504080959582b928e91cb608a65c3a49e1a"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_i686.whl", hash = "sha256:067a74454df670dcaa4e59349a2e5c81e567d8d65458d480a5b3dfecec08c5ff"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:7538e495704e2ccda9b234b82423a4038f324f3a10c43bc088a1636180f11a41"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:af3d4a73793725138d6b334d9d247ce7e5f084d96284ed23f22ee626a7b88e39"},
- {file = "typed_ast-1.4.3-cp38-cp38-win32.whl", hash = "sha256:f2362f3cb0f3172c42938946dbc5b7843c2a28aec307c49100c8b38764eb6927"},
- {file = "typed_ast-1.4.3-cp38-cp38-win_amd64.whl", hash = "sha256:dd4a21253f42b8d2b48410cb31fe501d32f8b9fbeb1f55063ad102fe9c425e40"},
- {file = "typed_ast-1.4.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:f328adcfebed9f11301eaedfa48e15bdece9b519fb27e6a8c01aa52a17ec31b3"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2c726c276d09fc5c414693a2de063f521052d9ea7c240ce553316f70656c84d4"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:cae53c389825d3b46fb37538441f75d6aecc4174f615d048321b716df2757fb0"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:b9574c6f03f685070d859e75c7f9eeca02d6933273b5e69572e5ff9d5e3931c3"},
- {file = "typed_ast-1.4.3-cp39-cp39-win32.whl", hash = "sha256:209596a4ec71d990d71d5e0d312ac935d86930e6eecff6ccc7007fe54d703808"},
- {file = "typed_ast-1.4.3-cp39-cp39-win_amd64.whl", hash = "sha256:9c6d1a54552b5330bc657b7ef0eae25d00ba7ffe85d9ea8ae6540d2197a3788c"},
- {file = "typed_ast-1.4.3.tar.gz", hash = "sha256:fb1bbeac803adea29cedd70781399c99138358c26d05fcbd23c13016b7f5ec65"},
+ {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
+ {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+]
+
+[[package]]
+name = "types-requests"
+version = "2.28.11.12"
+description = "Typing stubs for requests"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
+ {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+]
+
+[package.dependencies]
+types-urllib3 = "<1.27"
+
+[[package]]
+name = "types-urllib3"
+version = "1.26.25.5"
+description = "Typing stubs for urllib3"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
+ {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
@@ -1808 +1851 @@ name = "typing-extensions"
-version = "4.4.0"
+version = "4.5.0"
@@ -1814,2 +1857,2 @@ files = [
- {file = "typing_extensions-4.4.0-py3-none-any.whl", hash = "sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e"},
- {file = "typing_extensions-4.4.0.tar.gz", hash = "sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa"},
+ {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"},
+ {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"},
@@ -1837 +1880 @@ name = "uvicorn"
-version = "0.14.0"
+version = "0.20.0"
@@ -1841 +1884 @@ optional = false
-python-versions = "*"
+python-versions = ">=3.7"
@@ -1843,2 +1886,2 @@ files = [
- {file = "uvicorn-0.14.0-py3-none-any.whl", hash = "sha256:2a76bb359171a504b3d1c853409af3adbfa5cef374a4a59e5881945a97a93eae"},
- {file = "uvicorn-0.14.0.tar.gz", hash = "sha256:45ad7dfaaa7d55cab4cd1e85e03f27e9d60bc067ddc59db52a2b0aeca8870292"},
+ {file = "uvicorn-0.20.0-py3-none-any.whl", hash = "sha256:c3ed1598a5668208723f2bb49336f4509424ad198d6ab2615b7783db58d919fd"},
+ {file = "uvicorn-0.20.0.tar.gz", hash = "sha256:a4e12017b940247f836bc90b72e725d7dfd0c8ed1c51eb365f5ba30d9f5127d8"},
@@ -1848,2 +1891 @@ files = [
-asgiref = ">=3.3.4"
-click = ">=7"
+click = ">=7.0"
@@ -1853 +1895 @@ h11 = ">=0.8"
-standard = ["PyYAML (>=5.1)", "colorama (>=0.4)", "httptools (>=0.2.0,<0.3.0)", "python-dotenv (>=0.13)", "uvloop (>=0.14.0,!=0.15.0,!=0.15.1)", "watchgod (>=0.6)", "websockets (>=9.1)"]
+standard = ["colorama (>=0.4)", "httptools (>=0.5.0)", "python-dotenv (>=0.13)", "pyyaml (>=5.1)", "uvloop (>=0.14.0,!=0.15.0,!=0.15.1)", "watchfiles (>=0.13)", "websockets (>=10.4)"]
@@ -1913 +1955 @@ name = "werkzeug"
-version = "2.2.2"
+version = "2.2.3"
@@ -1919,2 +1961,2 @@ files = [
- {file = "Werkzeug-2.2.2-py3-none-any.whl", hash = "sha256:f979ab81f58d7318e064e99c4506445d60135ac5cd2e177a2de0089bfd4c9bd5"},
- {file = "Werkzeug-2.2.2.tar.gz", hash = "sha256:7ea2d48322cc7c0f8b3a215ed73eabd7b5d75d0b50e31ab006286ccff9e00b8f"},
+ {file = "Werkzeug-2.2.3-py3-none-any.whl", hash = "sha256:56433961bc1f12533306c624f3be5e744389ac61d722175d543e1751285da612"},
+ {file = "Werkzeug-2.2.3.tar.gz", hash = "sha256:2e1ccc9417d4da358b9de6f174e3ac094391ea1d4fbef2d667865d819dfd0afe"},
@@ -1932 +1974 @@ python-versions = "3.9.15"
-content-hash = "073a785848bf1132d6407c6b57b962ef52e7ec8752c9bbf165d7d772ca9ab6bb"
+content-hash = "7b9462e0e596249181d714cb2a687ade42839f6ea4e972aa97ed86fb44640b66"
diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml
index b65adfdd..fdd88d31 100644
--- a/services/api/pyproject.toml
+++ b/services/api/pyproject.toml
@@ -10,2 +10 @@ environs = "^9.5.0"
-huggingface-hub = "^0.12.0"
-jsonschema = "^4.16.0"
+jsonschema = "^4.17.0"
@@ -14 +13,2 @@ python = "3.9.15"
-starlette = "^0.16.0"
+requests = "^2.28.2"
+starlette = "^0.25.0"
@@ -16,2 +16,2 @@ starlette-prometheus = "^0.9.0"
-uvicorn = "^0.14.0"
-watchdog = { extras = ["watchmedo"], version = "^2.1.3" }
+uvicorn = "^0.20.0"
+watchdog = { extras = ["watchmedo"], version = "^2.2.1" }
@@ -20,2 +20,2 @@ watchdog = { extras = ["watchmedo"], version = "^2.1.3" }
-bandit = "^1.7.0"
-black = "^22.1.0"
+bandit = "^1.7.4"
+black = "^22.12.0"
@@ -23,5 +23,5 @@ flake8 = "^3.9.2"
-isort = "^5.9.3"
-mypy = "0.812"
-pip-audit = "^2.4.6"
-poetryup = "^0.3.8"
-pytest = "^7.2.0"
+httpx = "^0.23.3"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pip-audit = "^2.4.14"
+pytest = "^7.2.1"
@@ -29,0 +30,3 @@ pytest-httpserver = "^1.0.6"
+types-psutil = "^5.9.5"
+types-requests = "^2.28.11"
+types-jsonschema = "^4.17.0.4"
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 107c6ab9..5e77f013 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -6 +6 @@ from typing import List
-import uvicorn # type: ignore
+import uvicorn
diff --git a/services/api/src/api/prometheus.py b/services/api/src/api/prometheus.py
index df7366e7..ba515612 100644
--- a/services/api/src/api/prometheus.py
+++ b/services/api/src/api/prometheus.py
@@ -4,0 +5 @@ import os
+from typing import Any
@@ -6 +7 @@ import os
-from prometheus_client import ( # type: ignore # https://github.com/prometheus/client_python/issues/491
+from prometheus_client import ( # type: ignore
@@ -12,3 +13,4 @@ from prometheus_client import ( # type: ignore # https://github.com/prometheus/
-from prometheus_client.multiprocess import ( # type: ignore # https://github.com/prometheus/client_python/issues/491
- MultiProcessCollector,
-)
+from prometheus_client.multiprocess import MultiProcessCollector # type: ignore
+
+# ^ type: ignore can be removed on next release:
+# https://github.com/prometheus/client_python/issues/491#issuecomment-1429287314
@@ -30,2 +32,4 @@ class Prometheus:
- def getLatestContent(self) -> str:
- return generate_latest(self.getRegistry()).decode("utf-8")
+ def getLatestContent(self) -> Any:
+ # ^ returns Any because we cannot be sure latest are UTF8Bytes
+ latest = generate_latest(self.getRegistry())
+ return latest.decode("utf-8")
diff --git a/services/api/src/api/routes/processing_step.py b/services/api/src/api/routes/processing_step.py
index 0875a58e..7468ba8a 100644
--- a/services/api/src/api/routes/processing_step.py
+++ b/services/api/src/api/routes/processing_step.py
@@ -42 +42 @@ def create_processing_step_endpoint(
- if not are_valid_parameters([dataset]):
+ if not are_valid_parameters([dataset]) or not dataset:
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index bd3c80d0..cd3450fe 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -76 +76 @@ def create_is_valid_endpoint(
- if not are_valid_parameters([dataset]):
+ if not are_valid_parameters([dataset]) or not dataset:
diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py
index cc11ee6c..c1931b2c 100644
--- a/services/api/src/api/routes/webhook.py
+++ b/services/api/src/api/routes/webhook.py
@@ -7 +7 @@ from typing import Any, List, Literal, Optional, TypedDict
-from jsonschema import ValidationError, validate # type: ignore
+from jsonschema import ValidationError, validate
@@ -53 +53,2 @@ def parse_payload(json: Any) -> MoonWebhookV2Payload:
- return json
+ return json # type: ignore
+ # ^ validate() ensures the content is correct, but does not give the type
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index d0afa6f2..40850f49 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -17 +17 @@ from api.config import AppConfig, UvicornConfig
-def monkeypatch_session():
+def monkeypatch_session() -> Iterator[MonkeyPatch]:
@@ -45 +45 @@ def processing_steps(app_config: AppConfig) -> List[ProcessingStep]:
-def first_dataset_processing_step(processing_steps: List[ProcessingStep]):
+def first_dataset_processing_step(processing_steps: List[ProcessingStep]) -> ProcessingStep:
@@ -50 +50 @@ def first_dataset_processing_step(processing_steps: List[ProcessingStep]):
-def first_config_processing_step(processing_steps: List[ProcessingStep]):
+def first_config_processing_step(processing_steps: List[ProcessingStep]) -> ProcessingStep:
@@ -55 +55 @@ def first_config_processing_step(processing_steps: List[ProcessingStep]):
-def first_split_processing_step(processing_steps: List[ProcessingStep]):
+def first_split_processing_step(processing_steps: List[ProcessingStep]) -> ProcessingStep:
@@ -74 +74 @@ def queue_mongo_resource(app_config: AppConfig) -> Iterator[QueueMongoResource]:
-def uvicorn_config(monkeypatch_session: MonkeyPatch):
+def uvicorn_config(monkeypatch_session: MonkeyPatch) -> UvicornConfig:
@@ -79 +79 @@ def uvicorn_config(monkeypatch_session: MonkeyPatch):
-def httpserver_listen_address(uvicorn_config: UvicornConfig):
+def httpserver_listen_address(uvicorn_config: UvicornConfig) -> tuple[str, int]:
@@ -84 +84 @@ def httpserver_listen_address(uvicorn_config: UvicornConfig):
-def hf_endpoint(app_config: AppConfig):
+def hf_endpoint(app_config: AppConfig) -> str:
@@ -89 +89 @@ def hf_endpoint(app_config: AppConfig):
-def hf_auth_path(app_config: AppConfig):
+def hf_auth_path(app_config: AppConfig) -> str:
diff --git a/services/api/tests/test_app_real.py b/services/api/tests/test_app_real.py
index 59f90255..787c2912 100644
--- a/services/api/tests/test_app_real.py
+++ b/services/api/tests/test_app_real.py
@@ -3,0 +4 @@
+from typing import Iterator
@@ -14 +15 @@ from api.config import AppConfig
-def real_monkeypatch():
+def real_monkeypatch() -> Iterator[MonkeyPatch]:
diff --git a/services/api/tests/test_authentication.py b/services/api/tests/test_authentication.py
index d221f242..64c5c554 100644
--- a/services/api/tests/test_authentication.py
+++ b/services/api/tests/test_authentication.py
@@ -9 +9,2 @@ from pytest_httpserver import HTTPServer
-from starlette.requests import Headers, Request
+from starlette.datastructures import Headers
+from starlette.requests import Request
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 789ac4de..58eb2dcf 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -17 +17 @@ name = "aiohttp"
-version = "3.8.3"
+version = "3.8.4"
@@ -23,87 +23,87 @@ files = [
- {file = "aiohttp-3.8.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ba71c9b4dcbb16212f334126cc3d8beb6af377f6703d9dc2d9fb3874fd667ee9"},
- {file = "aiohttp-3.8.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d24b8bb40d5c61ef2d9b6a8f4528c2f17f1c5d2d31fed62ec860f6006142e83e"},
- {file = "aiohttp-3.8.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f88df3a83cf9df566f171adba39d5bd52814ac0b94778d2448652fc77f9eb491"},
- {file = "aiohttp-3.8.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b97decbb3372d4b69e4d4c8117f44632551c692bb1361b356a02b97b69e18a62"},
- {file = "aiohttp-3.8.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:309aa21c1d54b8ef0723181d430347d7452daaff93e8e2363db8e75c72c2fb2d"},
- {file = "aiohttp-3.8.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ad5383a67514e8e76906a06741febd9126fc7c7ff0f599d6fcce3e82b80d026f"},
- {file = "aiohttp-3.8.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:20acae4f268317bb975671e375493dbdbc67cddb5f6c71eebdb85b34444ac46b"},
- {file = "aiohttp-3.8.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:05a3c31c6d7cd08c149e50dc7aa2568317f5844acd745621983380597f027a18"},
- {file = "aiohttp-3.8.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d6f76310355e9fae637c3162936e9504b4767d5c52ca268331e2756e54fd4ca5"},
- {file = "aiohttp-3.8.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:256deb4b29fe5e47893fa32e1de2d73c3afe7407738bd3c63829874661d4822d"},
- {file = "aiohttp-3.8.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:5c59fcd80b9049b49acd29bd3598cada4afc8d8d69bd4160cd613246912535d7"},
- {file = "aiohttp-3.8.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:059a91e88f2c00fe40aed9031b3606c3f311414f86a90d696dd982e7aec48142"},
- {file = "aiohttp-3.8.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:2feebbb6074cdbd1ac276dbd737b40e890a1361b3cc30b74ac2f5e24aab41f7b"},
- {file = "aiohttp-3.8.3-cp310-cp310-win32.whl", hash = "sha256:5bf651afd22d5f0c4be16cf39d0482ea494f5c88f03e75e5fef3a85177fecdeb"},
- {file = "aiohttp-3.8.3-cp310-cp310-win_amd64.whl", hash = "sha256:653acc3880459f82a65e27bd6526e47ddf19e643457d36a2250b85b41a564715"},
- {file = "aiohttp-3.8.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:86fc24e58ecb32aee09f864cb11bb91bc4c1086615001647dbfc4dc8c32f4008"},
- {file = "aiohttp-3.8.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:75e14eac916f024305db517e00a9252714fce0abcb10ad327fb6dcdc0d060f1d"},
- {file = "aiohttp-3.8.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d1fde0f44029e02d02d3993ad55ce93ead9bb9b15c6b7ccd580f90bd7e3de476"},
- {file = "aiohttp-3.8.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ab94426ddb1ecc6a0b601d832d5d9d421820989b8caa929114811369673235c"},
- {file = "aiohttp-3.8.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:89d2e02167fa95172c017732ed7725bc8523c598757f08d13c5acca308e1a061"},
- {file = "aiohttp-3.8.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:02f9a2c72fc95d59b881cf38a4b2be9381b9527f9d328771e90f72ac76f31ad8"},
- {file = "aiohttp-3.8.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9c7149272fb5834fc186328e2c1fa01dda3e1fa940ce18fded6d412e8f2cf76d"},
- {file = "aiohttp-3.8.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:512bd5ab136b8dc0ffe3fdf2dfb0c4b4f49c8577f6cae55dca862cd37a4564e2"},
- {file = "aiohttp-3.8.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:7018ecc5fe97027214556afbc7c502fbd718d0740e87eb1217b17efd05b3d276"},
- {file = "aiohttp-3.8.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:88c70ed9da9963d5496d38320160e8eb7e5f1886f9290475a881db12f351ab5d"},
- {file = "aiohttp-3.8.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:da22885266bbfb3f78218dc40205fed2671909fbd0720aedba39b4515c038091"},
- {file = "aiohttp-3.8.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:e65bc19919c910127c06759a63747ebe14f386cda573d95bcc62b427ca1afc73"},
- {file = "aiohttp-3.8.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:08c78317e950e0762c2983f4dd58dc5e6c9ff75c8a0efeae299d363d439c8e34"},
- {file = "aiohttp-3.8.3-cp311-cp311-win32.whl", hash = "sha256:45d88b016c849d74ebc6f2b6e8bc17cabf26e7e40c0661ddd8fae4c00f015697"},
- {file = "aiohttp-3.8.3-cp311-cp311-win_amd64.whl", hash = "sha256:96372fc29471646b9b106ee918c8eeb4cca423fcbf9a34daa1b93767a88a2290"},
- {file = "aiohttp-3.8.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:c971bf3786b5fad82ce5ad570dc6ee420f5b12527157929e830f51c55dc8af77"},
- {file = "aiohttp-3.8.3-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ff25f48fc8e623d95eca0670b8cc1469a83783c924a602e0fbd47363bb54aaca"},
- {file = "aiohttp-3.8.3-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e381581b37db1db7597b62a2e6b8b57c3deec95d93b6d6407c5b61ddc98aca6d"},
- {file = "aiohttp-3.8.3-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:db19d60d846283ee275d0416e2a23493f4e6b6028825b51290ac05afc87a6f97"},
- {file = "aiohttp-3.8.3-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:25892c92bee6d9449ffac82c2fe257f3a6f297792cdb18ad784737d61e7a9a85"},
- {file = "aiohttp-3.8.3-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:398701865e7a9565d49189f6c90868efaca21be65c725fc87fc305906be915da"},
- {file = "aiohttp-3.8.3-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4a4fbc769ea9b6bd97f4ad0b430a6807f92f0e5eb020f1e42ece59f3ecfc4585"},
- {file = "aiohttp-3.8.3-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:b29bfd650ed8e148f9c515474a6ef0ba1090b7a8faeee26b74a8ff3b33617502"},
- {file = "aiohttp-3.8.3-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:1e56b9cafcd6531bab5d9b2e890bb4937f4165109fe98e2b98ef0dcfcb06ee9d"},
- {file = "aiohttp-3.8.3-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:ec40170327d4a404b0d91855d41bfe1fe4b699222b2b93e3d833a27330a87a6d"},
- {file = "aiohttp-3.8.3-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:2df5f139233060578d8c2c975128fb231a89ca0a462b35d4b5fcf7c501ebdbe1"},
- {file = "aiohttp-3.8.3-cp36-cp36m-win32.whl", hash = "sha256:f973157ffeab5459eefe7b97a804987876dd0a55570b8fa56b4e1954bf11329b"},
- {file = "aiohttp-3.8.3-cp36-cp36m-win_amd64.whl", hash = "sha256:437399385f2abcd634865705bdc180c8314124b98299d54fe1d4c8990f2f9494"},
- {file = "aiohttp-3.8.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:09e28f572b21642128ef31f4e8372adb6888846f32fecb288c8b0457597ba61a"},
- {file = "aiohttp-3.8.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6f3553510abdbec67c043ca85727396ceed1272eef029b050677046d3387be8d"},
- {file = "aiohttp-3.8.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e168a7560b7c61342ae0412997b069753f27ac4862ec7867eff74f0fe4ea2ad9"},
- {file = "aiohttp-3.8.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:db4c979b0b3e0fa7e9e69ecd11b2b3174c6963cebadeecfb7ad24532ffcdd11a"},
- {file = "aiohttp-3.8.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e164e0a98e92d06da343d17d4e9c4da4654f4a4588a20d6c73548a29f176abe2"},
- {file = "aiohttp-3.8.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e8a78079d9a39ca9ca99a8b0ac2fdc0c4d25fc80c8a8a82e5c8211509c523363"},
- {file = "aiohttp-3.8.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:21b30885a63c3f4ff5b77a5d6caf008b037cb521a5f33eab445dc566f6d092cc"},
- {file = "aiohttp-3.8.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:4b0f30372cef3fdc262f33d06e7b411cd59058ce9174ef159ad938c4a34a89da"},
- {file = "aiohttp-3.8.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:8135fa153a20d82ffb64f70a1b5c2738684afa197839b34cc3e3c72fa88d302c"},
- {file = "aiohttp-3.8.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:ad61a9639792fd790523ba072c0555cd6be5a0baf03a49a5dd8cfcf20d56df48"},
- {file = "aiohttp-3.8.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:978b046ca728073070e9abc074b6299ebf3501e8dee5e26efacb13cec2b2dea0"},
- {file = "aiohttp-3.8.3-cp37-cp37m-win32.whl", hash = "sha256:0d2c6d8c6872df4a6ec37d2ede71eff62395b9e337b4e18efd2177de883a5033"},
- {file = "aiohttp-3.8.3-cp37-cp37m-win_amd64.whl", hash = "sha256:21d69797eb951f155026651f7e9362877334508d39c2fc37bd04ff55b2007091"},
- {file = "aiohttp-3.8.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:2ca9af5f8f5812d475c5259393f52d712f6d5f0d7fdad9acdb1107dd9e3cb7eb"},
- {file = "aiohttp-3.8.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1d90043c1882067f1bd26196d5d2db9aa6d268def3293ed5fb317e13c9413ea4"},
- {file = "aiohttp-3.8.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d737fc67b9a970f3234754974531dc9afeea11c70791dcb7db53b0cf81b79784"},
- {file = "aiohttp-3.8.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ebf909ea0a3fc9596e40d55d8000702a85e27fd578ff41a5500f68f20fd32e6c"},
- {file = "aiohttp-3.8.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5835f258ca9f7c455493a57ee707b76d2d9634d84d5d7f62e77be984ea80b849"},
- {file = "aiohttp-3.8.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:da37dcfbf4b7f45d80ee386a5f81122501ec75672f475da34784196690762f4b"},
- {file = "aiohttp-3.8.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87f44875f2804bc0511a69ce44a9595d5944837a62caecc8490bbdb0e18b1342"},
- {file = "aiohttp-3.8.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:527b3b87b24844ea7865284aabfab08eb0faf599b385b03c2aa91fc6edd6e4b6"},
- {file = "aiohttp-3.8.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:d5ba88df9aa5e2f806650fcbeedbe4f6e8736e92fc0e73b0400538fd25a4dd96"},
- {file = "aiohttp-3.8.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:e7b8813be97cab8cb52b1375f41f8e6804f6507fe4660152e8ca5c48f0436017"},
- {file = "aiohttp-3.8.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:2dea10edfa1a54098703cb7acaa665c07b4e7568472a47f4e64e6319d3821ccf"},
- {file = "aiohttp-3.8.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:713d22cd9643ba9025d33c4af43943c7a1eb8547729228de18d3e02e278472b6"},
- {file = "aiohttp-3.8.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:2d252771fc85e0cf8da0b823157962d70639e63cb9b578b1dec9868dd1f4f937"},
- {file = "aiohttp-3.8.3-cp38-cp38-win32.whl", hash = "sha256:66bd5f950344fb2b3dbdd421aaa4e84f4411a1a13fca3aeb2bcbe667f80c9f76"},
- {file = "aiohttp-3.8.3-cp38-cp38-win_amd64.whl", hash = "sha256:84b14f36e85295fe69c6b9789b51a0903b774046d5f7df538176516c3e422446"},
- {file = "aiohttp-3.8.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16c121ba0b1ec2b44b73e3a8a171c4f999b33929cd2397124a8c7fcfc8cd9e06"},
- {file = "aiohttp-3.8.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8d6aaa4e7155afaf994d7924eb290abbe81a6905b303d8cb61310a2aba1c68ba"},
- {file = "aiohttp-3.8.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:43046a319664a04b146f81b40e1545d4c8ac7b7dd04c47e40bf09f65f2437346"},
- {file = "aiohttp-3.8.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:599418aaaf88a6d02a8c515e656f6faf3d10618d3dd95866eb4436520096c84b"},
- {file = "aiohttp-3.8.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92a2964319d359f494f16011e23434f6f8ef0434acd3cf154a6b7bec511e2fb7"},
- {file = "aiohttp-3.8.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:73a4131962e6d91109bca6536416aa067cf6c4efb871975df734f8d2fd821b37"},
- {file = "aiohttp-3.8.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:598adde339d2cf7d67beaccda3f2ce7c57b3b412702f29c946708f69cf8222aa"},
- {file = "aiohttp-3.8.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:75880ed07be39beff1881d81e4a907cafb802f306efd6d2d15f2b3c69935f6fb"},
- {file = "aiohttp-3.8.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:a0239da9fbafd9ff82fd67c16704a7d1bccf0d107a300e790587ad05547681c8"},
- {file = "aiohttp-3.8.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:4e3a23ec214e95c9fe85a58470b660efe6534b83e6cbe38b3ed52b053d7cb6ad"},
- {file = "aiohttp-3.8.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:47841407cc89a4b80b0c52276f3cc8138bbbfba4b179ee3acbd7d77ae33f7ac4"},
- {file = "aiohttp-3.8.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:54d107c89a3ebcd13228278d68f1436d3f33f2dd2af5415e3feaeb1156e1a62c"},
- {file = "aiohttp-3.8.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c37c5cce780349d4d51739ae682dec63573847a2a8dcb44381b174c3d9c8d403"},
- {file = "aiohttp-3.8.3-cp39-cp39-win32.whl", hash = "sha256:f178d2aadf0166be4df834c4953da2d7eef24719e8aec9a65289483eeea9d618"},
- {file = "aiohttp-3.8.3-cp39-cp39-win_amd64.whl", hash = "sha256:88e5be56c231981428f4f506c68b6a46fa25c4123a2e86d156c58a8369d31ab7"},
- {file = "aiohttp-3.8.3.tar.gz", hash = "sha256:3828fb41b7203176b82fe5d699e0d845435f2374750a44b480ea6b930f6be269"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"},
+ {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"},
@@ -116 +116 @@ attrs = ">=17.3.0"
-charset-normalizer = ">=2.0,<3.0"
+charset-normalizer = ">=2.0,<4.0"
@@ -636 +636 @@ name = "charset-normalizer"
-version = "2.1.1"
+version = "3.0.1"
@@ -640 +640 @@ optional = false
-python-versions = ">=3.6.0"
+python-versions = "*"
@@ -642,2 +642,88 @@ files = [
- {file = "charset-normalizer-2.1.1.tar.gz", hash = "sha256:5a3d016c7c547f69d6f81fb0db9449ce888b418b5b9952cc5e6e66843e9dd845"},
- {file = "charset_normalizer-2.1.1-py3-none-any.whl", hash = "sha256:83e9a75d1911279afd89352c68b45348559d1fc0506b054b346651b5e7fee29f"},
+ {file = "charset-normalizer-3.0.1.tar.gz", hash = "sha256:ebea339af930f8ca5d7a699b921106c6e29c617fe9606fa7baa043c1cdae326f"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:88600c72ef7587fe1708fd242b385b6ed4b8904976d5da0893e31df8b3480cb6"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c75ffc45f25324e68ab238cb4b5c0a38cd1c3d7f1fb1f72b5541de469e2247db"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:db72b07027db150f468fbada4d85b3b2729a3db39178abf5c543b784c1254539"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:62595ab75873d50d57323a91dd03e6966eb79c41fa834b7a1661ed043b2d404d"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ff6f3db31555657f3163b15a6b7c6938d08df7adbfc9dd13d9d19edad678f1e8"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:772b87914ff1152b92a197ef4ea40efe27a378606c39446ded52c8f80f79702e"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:70990b9c51340e4044cfc394a81f614f3f90d41397104d226f21e66de668730d"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:292d5e8ba896bbfd6334b096e34bffb56161c81408d6d036a7dfa6929cff8783"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:2edb64ee7bf1ed524a1da60cdcd2e1f6e2b4f66ef7c077680739f1641f62f555"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:31a9ddf4718d10ae04d9b18801bd776693487cbb57d74cc3458a7673f6f34639"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:44ba614de5361b3e5278e1241fda3dc1838deed864b50a10d7ce92983797fa76"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:12db3b2c533c23ab812c2b25934f60383361f8a376ae272665f8e48b88e8e1c6"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c512accbd6ff0270939b9ac214b84fb5ada5f0409c44298361b2f5e13f9aed9e"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-win32.whl", hash = "sha256:502218f52498a36d6bf5ea77081844017bf7982cdbe521ad85e64cabee1b608b"},
+ {file = "charset_normalizer-3.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:601f36512f9e28f029d9481bdaf8e89e5148ac5d89cffd3b05cd533eeb423b59"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0298eafff88c99982a4cf66ba2efa1128e4ddaca0b05eec4c456bbc7db691d8d"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:a8d0fc946c784ff7f7c3742310cc8a57c5c6dc31631269876a88b809dbeff3d3"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:87701167f2a5c930b403e9756fab1d31d4d4da52856143b609e30a1ce7160f3c"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:14e76c0f23218b8f46c4d87018ca2e441535aed3632ca134b10239dfb6dadd6b"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0c0a590235ccd933d9892c627dec5bc7511ce6ad6c1011fdf5b11363022746c1"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8c7fe7afa480e3e82eed58e0ca89f751cd14d767638e2550c77a92a9e749c317"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:79909e27e8e4fcc9db4addea88aa63f6423ebb171db091fb4373e3312cb6d603"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8ac7b6a045b814cf0c47f3623d21ebd88b3e8cf216a14790b455ea7ff0135d18"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:72966d1b297c741541ca8cf1223ff262a6febe52481af742036a0b296e35fa5a"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f9d0c5c045a3ca9bedfc35dca8526798eb91a07aa7a2c0fee134c6c6f321cbd7"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:5995f0164fa7df59db4746112fec3f49c461dd6b31b841873443bdb077c13cfc"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:4a8fcf28c05c1f6d7e177a9a46a1c52798bfe2ad80681d275b10dcf317deaf0b"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:761e8904c07ad053d285670f36dd94e1b6ab7f16ce62b9805c475b7aa1cffde6"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-win32.whl", hash = "sha256:71140351489970dfe5e60fc621ada3e0f41104a5eddaca47a7acb3c1b851d6d3"},
+ {file = "charset_normalizer-3.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:9ab77acb98eba3fd2a85cd160851816bfce6871d944d885febf012713f06659c"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:84c3990934bae40ea69a82034912ffe5a62c60bbf6ec5bc9691419641d7d5c9a"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:74292fc76c905c0ef095fe11e188a32ebd03bc38f3f3e9bcb85e4e6db177b7ea"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c95a03c79bbe30eec3ec2b7f076074f4281526724c8685a42872974ef4d36b72"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f4c39b0e3eac288fedc2b43055cfc2ca7a60362d0e5e87a637beac5d801ef478"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:df2c707231459e8a4028eabcd3cfc827befd635b3ef72eada84ab13b52e1574d"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:93ad6d87ac18e2a90b0fe89df7c65263b9a99a0eb98f0a3d2e079f12a0735837"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:59e5686dd847347e55dffcc191a96622f016bc0ad89105e24c14e0d6305acbc6"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:cd6056167405314a4dc3c173943f11249fa0f1b204f8b51ed4bde1a9cd1834dc"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:083c8d17153ecb403e5e1eb76a7ef4babfc2c48d58899c98fcaa04833e7a2f9a"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:f5057856d21e7586765171eac8b9fc3f7d44ef39425f85dbcccb13b3ebea806c"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:7eb33a30d75562222b64f569c642ff3dc6689e09adda43a082208397f016c39a"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-win32.whl", hash = "sha256:95dea361dd73757c6f1c0a1480ac499952c16ac83f7f5f4f84f0658a01b8ef41"},
+ {file = "charset_normalizer-3.0.1-cp36-cp36m-win_amd64.whl", hash = "sha256:eaa379fcd227ca235d04152ca6704c7cb55564116f8bc52545ff357628e10602"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:3e45867f1f2ab0711d60c6c71746ac53537f1684baa699f4f668d4c6f6ce8e14"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cadaeaba78750d58d3cc6ac4d1fd867da6fc73c88156b7a3212a3cd4819d679d"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:911d8a40b2bef5b8bbae2e36a0b103f142ac53557ab421dc16ac4aafee6f53dc"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:503e65837c71b875ecdd733877d852adbc465bd82c768a067badd953bf1bc5a3"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a60332922359f920193b1d4826953c507a877b523b2395ad7bc716ddd386d866"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16a8663d6e281208d78806dbe14ee9903715361cf81f6d4309944e4d1e59ac5b"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:a16418ecf1329f71df119e8a65f3aa68004a3f9383821edcb20f0702934d8087"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:9d9153257a3f70d5f69edf2325357251ed20f772b12e593f3b3377b5f78e7ef8"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:02a51034802cbf38db3f89c66fb5d2ec57e6fe7ef2f4a44d070a593c3688667b"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:2e396d70bc4ef5325b72b593a72c8979999aa52fb8bcf03f701c1b03e1166918"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:11b53acf2411c3b09e6af37e4b9005cba376c872503c8f28218c7243582df45d"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-win32.whl", hash = "sha256:0bf2dae5291758b6f84cf923bfaa285632816007db0330002fa1de38bfcb7154"},
+ {file = "charset_normalizer-3.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:2c03cc56021a4bd59be889c2b9257dae13bf55041a3372d3295416f86b295fb5"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:024e606be3ed92216e2b6952ed859d86b4cfa52cd5bc5f050e7dc28f9b43ec42"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:4b0d02d7102dd0f997580b51edc4cebcf2ab6397a7edf89f1c73b586c614272c"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:358a7c4cb8ba9b46c453b1dd8d9e431452d5249072e4f56cfda3149f6ab1405e"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81d6741ab457d14fdedc215516665050f3822d3e56508921cc7239f8c8e66a58"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8b8af03d2e37866d023ad0ddea594edefc31e827fee64f8de5611a1dbc373174"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9cf4e8ad252f7c38dd1f676b46514f92dc0ebeb0db5552f5f403509705e24753"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e696f0dd336161fca9adbb846875d40752e6eba585843c768935ba5c9960722b"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c22d3fe05ce11d3671297dc8973267daa0f938b93ec716e12e0f6dee81591dc1"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:109487860ef6a328f3eec66f2bf78b0b72400280d8f8ea05f69c51644ba6521a"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:37f8febc8ec50c14f3ec9637505f28e58d4f66752207ea177c1d67df25da5aed"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:f97e83fa6c25693c7a35de154681fcc257c1c41b38beb0304b9c4d2d9e164479"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:a152f5f33d64a6be73f1d30c9cc82dfc73cec6477ec268e7c6e4c7d23c2d2291"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:39049da0ffb96c8cbb65cbf5c5f3ca3168990adf3551bd1dee10c48fce8ae820"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-win32.whl", hash = "sha256:4457ea6774b5611f4bed5eaa5df55f70abde42364d498c5134b7ef4c6958e20e"},
+ {file = "charset_normalizer-3.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:e62164b50f84e20601c1ff8eb55620d2ad25fb81b59e3cd776a1902527a788af"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:8eade758719add78ec36dc13201483f8e9b5d940329285edcd5f70c0a9edbd7f"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8499ca8f4502af841f68135133d8258f7b32a53a1d594aa98cc52013fff55678"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3fc1c4a2ffd64890aebdb3f97e1278b0cc72579a08ca4de8cd2c04799a3a22be"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:00d3ffdaafe92a5dc603cb9bd5111aaa36dfa187c8285c543be562e61b755f6b"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c2ac1b08635a8cd4e0cbeaf6f5e922085908d48eb05d44c5ae9eabab148512ca"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f6f45710b4459401609ebebdbcfb34515da4fc2aa886f95107f556ac69a9147e"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3ae1de54a77dc0d6d5fcf623290af4266412a7c4be0b1ff7444394f03f5c54e3"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3b590df687e3c5ee0deef9fc8c547d81986d9a1b56073d82de008744452d6541"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ab5de034a886f616a5668aa5d098af2b5385ed70142090e2a31bcbd0af0fdb3d"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:9cb3032517f1627cc012dbc80a8ec976ae76d93ea2b5feaa9d2a5b8882597579"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:608862a7bf6957f2333fc54ab4399e405baad0163dc9f8d99cb236816db169d4"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0f438ae3532723fb6ead77e7c604be7c8374094ef4ee2c5e03a3a17f1fca256c"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:356541bf4381fa35856dafa6a965916e54bed415ad8a24ee6de6e37deccf2786"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-win32.whl", hash = "sha256:39cf9ed17fe3b1bc81f33c9ceb6ce67683ee7526e65fde1447c772afc54a1bb8"},
+ {file = "charset_normalizer-3.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:0a11e971ed097d24c534c037d298ad32c6ce81a45736d31e0ff0ad37ab437d59"},
+ {file = "charset_normalizer-3.0.1-py3-none-any.whl", hash = "sha256:7e189e2e1d3ed2f4aebabd2d5b0f931e883676e51c7624826e0a4e5fe8a0bf24"},
@@ -646,3 +731,0 @@ files = [
-[package.extras]
-unicode-backport = ["unicodedata2"]
-
@@ -1247 +1330 @@ name = "gdown"
-version = "4.6.0"
+version = "4.6.3"
@@ -1253,2 +1336,2 @@ files = [
- {file = "gdown-4.6.0-py3-none-any.whl", hash = "sha256:e75c5aa8be8ea1cac642d4793f884339d887ab5e07aaa57fafa16c8a56a0cde5"},
- {file = "gdown-4.6.0.tar.gz", hash = "sha256:5ce3db0aeda54f46caacb2df86f31c3e3ecd17c355689e6456d85fb528ba9749"},
+ {file = "gdown-4.6.3-py3-none-any.whl", hash = "sha256:36bbb962b373f9634af1b9a10c73e231d3ad9b084659efd93bf3c62d57c25a97"},
+ {file = "gdown-4.6.3.tar.gz", hash = "sha256:ec6ea922ef4e3616a054073656d9aa2cdc717fbdbd0a94d2f77d52d323943307"},
@@ -1727 +1810 @@ reference = "master"
-resolved_reference = "bcd4af619a2fa45f5876d8855f7876cc09f663af"
+resolved_reference = "ba17d213a27bbc7263cc30d4f293967aa1021cff"
@@ -1784,3 +1867,3 @@ mongo-types = "0.15.1"
-mongoengine = "^0.24.1"
-orjson = "^3.6.4"
-psutil = "^5.9.2"
+mongoengine = "^0.24.2"
+orjson = "^3.8.6"
+psutil = "^5.9.4"
@@ -1787,0 +1871 @@ pymongo = {version = "^3.13.0", extras = ["srv"]}
+requests = "^2.28.2"
@@ -2339 +2423 @@ name = "mypy"
-version = "0.812"
+version = "1.0.0"
@@ -2343 +2427 @@ optional = false
-python-versions = ">=3.5"
+python-versions = ">=3.7"
@@ -2345,22 +2429,26 @@ files = [
- {file = "mypy-0.812-cp35-cp35m-macosx_10_9_x86_64.whl", hash = "sha256:a26f8ec704e5a7423c8824d425086705e381b4f1dfdef6e3a1edab7ba174ec49"},
- {file = "mypy-0.812-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:28fb5479c494b1bab244620685e2eb3c3f988d71fd5d64cc753195e8ed53df7c"},
- {file = "mypy-0.812-cp35-cp35m-manylinux2010_x86_64.whl", hash = "sha256:9743c91088d396c1a5a3c9978354b61b0382b4e3c440ce83cf77994a43e8c521"},
- {file = "mypy-0.812-cp35-cp35m-win_amd64.whl", hash = "sha256:d7da2e1d5f558c37d6e8c1246f1aec1e7349e4913d8fb3cb289a35de573fe2eb"},
- {file = "mypy-0.812-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:4eec37370483331d13514c3f55f446fc5248d6373e7029a29ecb7b7494851e7a"},
- {file = "mypy-0.812-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:d65cc1df038ef55a99e617431f0553cd77763869eebdf9042403e16089fe746c"},
- {file = "mypy-0.812-cp36-cp36m-manylinux2010_x86_64.whl", hash = "sha256:61a3d5b97955422964be6b3baf05ff2ce7f26f52c85dd88db11d5e03e146a3a6"},
- {file = "mypy-0.812-cp36-cp36m-win_amd64.whl", hash = "sha256:25adde9b862f8f9aac9d2d11971f226bd4c8fbaa89fb76bdadb267ef22d10064"},
- {file = "mypy-0.812-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:552a815579aa1e995f39fd05dde6cd378e191b063f031f2acfe73ce9fb7f9e56"},
- {file = "mypy-0.812-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:499c798053cdebcaa916eef8cd733e5584b5909f789de856b482cd7d069bdad8"},
- {file = "mypy-0.812-cp37-cp37m-manylinux2010_x86_64.whl", hash = "sha256:5873888fff1c7cf5b71efbe80e0e73153fe9212fafdf8e44adfe4c20ec9f82d7"},
- {file = "mypy-0.812-cp37-cp37m-win_amd64.whl", hash = "sha256:9f94aac67a2045ec719ffe6111df543bac7874cee01f41928f6969756e030564"},
- {file = "mypy-0.812-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d23e0ea196702d918b60c8288561e722bf437d82cb7ef2edcd98cfa38905d506"},
- {file = "mypy-0.812-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:674e822aa665b9fd75130c6c5f5ed9564a38c6cea6a6432ce47eafb68ee578c5"},
- {file = "mypy-0.812-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:abf7e0c3cf117c44d9285cc6128856106183938c68fd4944763003decdcfeb66"},
- {file = "mypy-0.812-cp38-cp38-win_amd64.whl", hash = "sha256:0d0a87c0e7e3a9becdfbe936c981d32e5ee0ccda3e0f07e1ef2c3d1a817cf73e"},
- {file = "mypy-0.812-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:7ce3175801d0ae5fdfa79b4f0cfed08807af4d075b402b7e294e6aa72af9aa2a"},
- {file = "mypy-0.812-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:b09669bcda124e83708f34a94606e01b614fa71931d356c1f1a5297ba11f110a"},
- {file = "mypy-0.812-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:33f159443db0829d16f0a8d83d94df3109bb6dd801975fe86bacb9bf71628e97"},
- {file = "mypy-0.812-cp39-cp39-win_amd64.whl", hash = "sha256:3f2aca7f68580dc2508289c729bd49ee929a436208d2b2b6aab15745a70a57df"},
- {file = "mypy-0.812-py3-none-any.whl", hash = "sha256:2f9b3407c58347a452fc0736861593e105139b905cca7d097e413453a1d650b4"},
- {file = "mypy-0.812.tar.gz", hash = "sha256:cd07039aa5df222037005b08fbbfd69b3ab0b0bd7a07d7906de75ae52c4e3119"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e0626db16705ab9f7fa6c249c017c887baf20738ce7f9129da162bb3075fc1af"},
+ {file = "mypy-1.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1ace23f6bb4aec4604b86c4843276e8fa548d667dbbd0cb83a3ae14b18b2db6c"},
+ {file = "mypy-1.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:87edfaf344c9401942883fad030909116aa77b0fa7e6e8e1c5407e14549afe9a"},
+ {file = "mypy-1.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0ab090d9240d6b4e99e1fa998c2d0aa5b29fc0fb06bd30e7ad6183c95fa07593"},
+ {file = "mypy-1.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:7cc2c01dfc5a3cbddfa6c13f530ef3b95292f926329929001d45e124342cd6b7"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14d776869a3e6c89c17eb943100f7868f677703c8a4e00b3803918f86aafbc52"},
+ {file = "mypy-1.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb2782a036d9eb6b5a6efcdda0986774bf798beef86a62da86cb73e2a10b423d"},
+ {file = "mypy-1.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cfca124f0ac6707747544c127880893ad72a656e136adc935c8600740b21ff5"},
+ {file = "mypy-1.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8845125d0b7c57838a10fd8925b0f5f709d0e08568ce587cc862aacce453e3dd"},
+ {file = "mypy-1.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b1b9e1ed40544ef486fa8ac022232ccc57109f379611633ede8e71630d07d2"},
+ {file = "mypy-1.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c7cf862aef988b5fbaa17764ad1d21b4831436701c7d2b653156a9497d92c83c"},
+ {file = "mypy-1.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5cd187d92b6939617f1168a4fe68f68add749902c010e66fe574c165c742ed88"},
+ {file = "mypy-1.0.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:4e5175026618c178dfba6188228b845b64131034ab3ba52acaffa8f6c361f805"},
+ {file = "mypy-1.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2f6ac8c87e046dc18c7d1d7f6653a66787a4555085b056fe2d599f1f1a2a2d21"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7306edca1c6f1b5fa0bc9aa645e6ac8393014fa82d0fa180d0ebc990ebe15964"},
+ {file = "mypy-1.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3cfad08f16a9c6611e6143485a93de0e1e13f48cfb90bcad7d5fde1c0cec3d36"},
+ {file = "mypy-1.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67cced7f15654710386e5c10b96608f1ee3d5c94ca1da5a2aad5889793a824c1"},
+ {file = "mypy-1.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:a86b794e8a56ada65c573183756eac8ac5b8d3d59daf9d5ebd72ecdbb7867a43"},
+ {file = "mypy-1.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:50979d5efff8d4135d9db293c6cb2c42260e70fb010cbc697b1311a4d7a39ddb"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3ae4c7a99e5153496243146a3baf33b9beff714464ca386b5f62daad601d87af"},
+ {file = "mypy-1.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5e398652d005a198a7f3c132426b33c6b85d98aa7dc852137a2a3be8890c4072"},
+ {file = "mypy-1.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be78077064d016bc1b639c2cbcc5be945b47b4261a4f4b7d8923f6c69c5c9457"},
+ {file = "mypy-1.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:92024447a339400ea00ac228369cd242e988dd775640755fa4ac0c126e49bb74"},
+ {file = "mypy-1.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:fe523fcbd52c05040c7bee370d66fee8373c5972171e4fbc323153433198592d"},
+ {file = "mypy-1.0.0-py3-none-any.whl", hash = "sha256:2efa963bdddb27cb4a0d42545cd137a8d2b883bd181bbc4525b568ef6eca258f"},
+ {file = "mypy-1.0.0.tar.gz", hash = "sha256:f34495079c8d9da05b183f9f7daec2878280c2ad7cc81da686ef0b484cea2ecf"},
@@ -2370,3 +2458,3 @@ files = [
-mypy-extensions = ">=0.4.3,<0.5.0"
-typed-ast = ">=1.4.0,<1.5.0"
-typing-extensions = ">=3.7.4"
+mypy-extensions = ">=0.4.3"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=3.10"
@@ -2375,0 +2464,3 @@ dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
@@ -2379,2 +2470,2 @@ name = "mypy-extensions"
-version = "0.4.4"
-description = "Experimental type system extensions for programs checked with the mypy typechecker."
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
@@ -2383 +2474 @@ optional = false
-python-versions = ">=2.7"
+python-versions = ">=3.5"
@@ -2385 +2476,2 @@ files = [
- {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
@@ -2540 +2632 @@ name = "openpyxl"
-version = "3.1.0"
+version = "3.1.1"
@@ -2546,2 +2638,2 @@ files = [
- {file = "openpyxl-3.1.0-py2.py3-none-any.whl", hash = "sha256:24d7d361025d186ba91eff58135d50855cf035a84371b891e58fb6eb5125660f"},
- {file = "openpyxl-3.1.0.tar.gz", hash = "sha256:eccedbe1cdd8b2494057e73959b496821141038dbb7eb9266ea59e3f34208231"},
+ {file = "openpyxl-3.1.1-py2.py3-none-any.whl", hash = "sha256:a0266e033e65f33ee697254b66116a5793c15fc92daf64711080000df4cfe0a8"},
+ {file = "openpyxl-3.1.1.tar.gz", hash = "sha256:f06d44e2c973781068bce5ecf860a09bcdb1c7f5ce1facd5e9aa82c92c93ae72"},
@@ -2701 +2793 @@ name = "pandas-stubs"
-version = "1.5.3.230203"
+version = "1.5.3.230214"
@@ -2707,2 +2799,2 @@ files = [
- {file = "pandas_stubs-1.5.3.230203-py3-none-any.whl", hash = "sha256:04da026f9933e1534ba480a2b06139a2dbd624c57ce0697e5873460d7815c67d"},
- {file = "pandas_stubs-1.5.3.230203.tar.gz", hash = "sha256:7a4eedb210c77e0911fe32a4178673ba68ef569c034117ea19ada79451db4af9"},
+ {file = "pandas_stubs-1.5.3.230214-py3-none-any.whl", hash = "sha256:be0f3056fef543c37248951a9868cba77e17ace3b815b0a23c72ed2ec13fbd9e"},
+ {file = "pandas_stubs-1.5.3.230214.tar.gz", hash = "sha256:5fc1fc9c470d5332f9025a53ecab4c36cb80b2c52a596d597d6c5e33b4078153"},
@@ -2945,15 +3036,0 @@ testing = ["pytest", "pytest-benchmark"]
-[[package]]
-name = "poetryup"
-version = "0.3.15"
-description = "Update dependencies and bump their version in the pyproject.toml file"
-category = "dev"
-optional = false
-python-versions = ">=3.6,<4.0"
-files = [
- {file = "poetryup-0.3.15-py3-none-any.whl", hash = "sha256:db068f55d10c0f89c76ea2b62c6bb81c0b0512454f7a83bdc0a13c146e5fb13e"},
- {file = "poetryup-0.3.15.tar.gz", hash = "sha256:efa4e7bb0cd005db4aff3cc678c8bfba9474ef42d5759c0168f2a55fc0f17bc3"},
-]
-
-[package.dependencies]
-tomlkit = ">=0.7.2,<0.8.0"
-
@@ -4242 +4319 @@ name = "setuptools"
-version = "67.2.0"
+version = "67.3.1"
@@ -4248,2 +4325,2 @@ files = [
- {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
- {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
+ {file = "setuptools-67.3.1-py3-none-any.whl", hash = "sha256:23c86b4e44432bfd8899384afc08872ec166a24f48a3f99f293b0a557e6a6b5d"},
+ {file = "setuptools-67.3.1.tar.gz", hash = "sha256:daec07fd848d80676694d6bf69c009d28910aeece68a38dbe88b7e1bb6dba12e"},
@@ -4295 +4372 @@ name = "soundfile"
-version = "0.11.0"
+version = "0.12.0"
@@ -4301,6 +4378,7 @@ files = [
- {file = "soundfile-0.11.0-py2.py3-none-any.whl", hash = "sha256:f4e4f832b1958403fb9726eeea54e0ebf1c7fc2599ff296a7ab1ac062f8048c9"},
- {file = "soundfile-0.11.0-py2.py3-none-macosx_10_9_arm64.macosx_11_0_arm64.whl", hash = "sha256:9e6a62eefad0a7f856cc8f5ede7f1a0c196b65d2901c00fffc74a3d7e81d89c8"},
- {file = "soundfile-0.11.0-py2.py3-none-macosx_10_9_x86_64.macosx_11_0_x86_64.whl", hash = "sha256:12f66fe9dcddedaa6c808bc3e104fc67fcee59dc64214bf7f43605e69836c497"},
- {file = "soundfile-0.11.0-py2.py3-none-win32.whl", hash = "sha256:08d9636815692f332e042990d449e79b888d288f0752226d8602e91523a0a29b"},
- {file = "soundfile-0.11.0-py2.py3-none-win_amd64.whl", hash = "sha256:a4ab6f66ad222d8e144dcb6abc73fbb867c11da2934b677f9b129778a6c65112"},
- {file = "soundfile-0.11.0.tar.gz", hash = "sha256:931738a1c93e8684c2d3e1d514ac63440ce827ec783ea0a2d3e4730e3dc58c18"},
+ {file = "soundfile-0.12.0-py2.py3-none-any.whl", hash = "sha256:9e7dc313f4f1e835f3b3bf35fd7c2908bba31bd6080727f075799b7a03f8d426"},
+ {file = "soundfile-0.12.0-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:ff687a6db726229adc93e7df93e6bd80ffb041a0c4d84fdf6659f14fa32854bd"},
+ {file = "soundfile-0.12.0-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:29623eae59f22d8c1b18d8cab90d39de1c834270081e3e50bdcafc95d6dd7bc9"},
+ {file = "soundfile-0.12.0-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:a90a27ce766dbc2fbdee91bf549d4e60ebbc2b94a2c2a7f12ca48b8bbe7b0123"},
+ {file = "soundfile-0.12.0-py2.py3-none-win32.whl", hash = "sha256:63f320b284781a346f1244f1bc8505567f1340d0ceb8841b4e3b598c9eec9d0c"},
+ {file = "soundfile-0.12.0-py2.py3-none-win_amd64.whl", hash = "sha256:5e34870f8e8aef60be350fd4c0aa7f294497b615e39e4bbbe5e87181af93aad5"},
+ {file = "soundfile-0.12.0.tar.gz", hash = "sha256:e50c42733ff5396e49a6a689722fa362387b2c403273bcc195994bf4a8e2df3f"},
@@ -4317 +4395 @@ name = "soupsieve"
-version = "2.3.2.post1"
+version = "2.4"
@@ -4321 +4399 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -4323,2 +4401,2 @@ files = [
- {file = "soupsieve-2.3.2.post1-py3-none-any.whl", hash = "sha256:3b2503d3c7084a42b1ebd08116e5f81aadfaea95863628c80a3b774a11b7c759"},
- {file = "soupsieve-2.3.2.post1.tar.gz", hash = "sha256:fc53893b3da2c33de295667a0e19f078c14bf86544af307354de5fcf12a3f30d"},
+ {file = "soupsieve-2.4-py3-none-any.whl", hash = "sha256:49e5368c2cda80ee7e84da9dbe3e110b70a4575f196efb74e51b94549d921955"},
+ {file = "soupsieve-2.4.tar.gz", hash = "sha256:e28dba9ca6c7c00173e34e4ba57448f0688bb681b7c5e8bf4971daafc093d69a"},
@@ -4329 +4407 @@ name = "stevedore"
-version = "4.1.1"
+version = "5.0.0"
@@ -4335,2 +4413,2 @@ files = [
- {file = "stevedore-4.1.1-py3-none-any.whl", hash = "sha256:aa6436565c069b2946fe4ebff07f5041e0c8bf18c7376dd29edf80cf7d524e4e"},
- {file = "stevedore-4.1.1.tar.gz", hash = "sha256:7f8aeb6e3f90f96832c301bff21a7eb5eefbe894c88c506483d355565d88cc1a"},
+ {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"},
+ {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"},
@@ -4653,12 +4730,0 @@ files = [
-[[package]]
-name = "tomlkit"
-version = "0.7.2"
-description = "Style preserving TOML library"
-category = "dev"
-optional = false
-python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
-files = [
- {file = "tomlkit-0.7.2-py2.py3-none-any.whl", hash = "sha256:173ad840fa5d2aac140528ca1933c29791b79a374a0861a80347f42ec9328117"},
- {file = "tomlkit-0.7.2.tar.gz", hash = "sha256:d7a454f319a7e9bd2e249f239168729327e4dd2d27b17dc68be264ad1ce36754"},
-]
-
@@ -4808,40 +4873,0 @@ url = "vendors/trec-car-tools/python3"
-[[package]]
-name = "typed-ast"
-version = "1.4.3"
-description = "a fork of Python 2 and 3 ast modules with type comment support"
-category = "dev"
-optional = false
-python-versions = "*"
-files = [
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:2068531575a125b87a41802130fa7e29f26c09a2833fea68d9a40cf33902eba6"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:c907f561b1e83e93fad565bac5ba9c22d96a54e7ea0267c708bffe863cbe4075"},
- {file = "typed_ast-1.4.3-cp35-cp35m-manylinux2014_aarch64.whl", hash = "sha256:1b3ead4a96c9101bef08f9f7d1217c096f31667617b58de957f690c92378b528"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win32.whl", hash = "sha256:dde816ca9dac1d9c01dd504ea5967821606f02e510438120091b84e852367428"},
- {file = "typed_ast-1.4.3-cp35-cp35m-win_amd64.whl", hash = "sha256:777a26c84bea6cd934422ac2e3b78863a37017618b6e5c08f92ef69853e765d3"},
- {file = "typed_ast-1.4.3-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:f8afcf15cc511ada719a88e013cec87c11aff7b91f019295eb4530f96fe5ef2f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:52b1eb8c83f178ab787f3a4283f68258525f8d70f778a2f6dd54d3b5e5fb4341"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:01ae5f73431d21eead5015997ab41afa53aa1fbe252f9da060be5dad2c730ace"},
- {file = "typed_ast-1.4.3-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:c190f0899e9f9f8b6b7863debfb739abcb21a5c054f911ca3596d12b8a4c4c7f"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win32.whl", hash = "sha256:398e44cd480f4d2b7ee8d98385ca104e35c81525dd98c519acff1b79bdaac363"},
- {file = "typed_ast-1.4.3-cp36-cp36m-win_amd64.whl", hash = "sha256:bff6ad71c81b3bba8fa35f0f1921fb24ff4476235a6e94a26ada2e54370e6da7"},
- {file = "typed_ast-1.4.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0fb71b8c643187d7492c1f8352f2c15b4c4af3f6338f21681d3681b3dc31a266"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:760ad187b1041a154f0e4d0f6aae3e40fdb51d6de16e5c99aedadd9246450e9e"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:5feca99c17af94057417d744607b82dd0a664fd5e4ca98061480fd8b14b18d04"},
- {file = "typed_ast-1.4.3-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:95431a26309a21874005845c21118c83991c63ea800dd44843e42a916aec5899"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win32.whl", hash = "sha256:aee0c1256be6c07bd3e1263ff920c325b59849dc95392a05f258bb9b259cf39c"},
- {file = "typed_ast-1.4.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9ad2c92ec681e02baf81fdfa056fe0d818645efa9af1f1cd5fd6f1bd2bdfd805"},
- {file = "typed_ast-1.4.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b36b4f3920103a25e1d5d024d155c504080959582b928e91cb608a65c3a49e1a"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_i686.whl", hash = "sha256:067a74454df670dcaa4e59349a2e5c81e567d8d65458d480a5b3dfecec08c5ff"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:7538e495704e2ccda9b234b82423a4038f324f3a10c43bc088a1636180f11a41"},
- {file = "typed_ast-1.4.3-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:af3d4a73793725138d6b334d9d247ce7e5f084d96284ed23f22ee626a7b88e39"},
- {file = "typed_ast-1.4.3-cp38-cp38-win32.whl", hash = "sha256:f2362f3cb0f3172c42938946dbc5b7843c2a28aec307c49100c8b38764eb6927"},
- {file = "typed_ast-1.4.3-cp38-cp38-win_amd64.whl", hash = "sha256:dd4a21253f42b8d2b48410cb31fe501d32f8b9fbeb1f55063ad102fe9c425e40"},
- {file = "typed_ast-1.4.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:f328adcfebed9f11301eaedfa48e15bdece9b519fb27e6a8c01aa52a17ec31b3"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2c726c276d09fc5c414693a2de063f521052d9ea7c240ce553316f70656c84d4"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:cae53c389825d3b46fb37538441f75d6aecc4174f615d048321b716df2757fb0"},
- {file = "typed_ast-1.4.3-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:b9574c6f03f685070d859e75c7f9eeca02d6933273b5e69572e5ff9d5e3931c3"},
- {file = "typed_ast-1.4.3-cp39-cp39-win32.whl", hash = "sha256:209596a4ec71d990d71d5e0d312ac935d86930e6eecff6ccc7007fe54d703808"},
- {file = "typed_ast-1.4.3-cp39-cp39-win_amd64.whl", hash = "sha256:9c6d1a54552b5330bc657b7ef0eae25d00ba7ffe85d9ea8ae6540d2197a3788c"},
- {file = "typed_ast-1.4.3.tar.gz", hash = "sha256:fb1bbeac803adea29cedd70781399c99138358c26d05fcbd23c13016b7f5ec65"},
-]
-
@@ -4871 +4897 @@ name = "types-psutil"
-version = "5.9.5.6"
+version = "5.9.5.7"
@@ -4877,2 +4903,2 @@ files = [
- {file = "types-psutil-5.9.5.6.tar.gz", hash = "sha256:65f93589711ca48859602c955c4247c834d96d4d33a9cbe4142d89593ef33b3c"},
- {file = "types_psutil-5.9.5.6-py3-none-any.whl", hash = "sha256:07acd57594ff254285250ef70be0fe4efe0b11a30065b6dce62a856235a8ca10"},
+ {file = "types-psutil-5.9.5.7.tar.gz", hash = "sha256:b38e376ba0d53b9a08d0eb77c3ccf621f9e0e0f90f9bdfb2ecc473d62c6828a3"},
+ {file = "types_psutil-5.9.5.7-py3-none-any.whl", hash = "sha256:de57e5610fa568354b7bc0df05b2392a05fd0c1bf7a804f32f988126a5767a00"},
@@ -4895 +4921 @@ name = "types-requests"
-version = "2.28.11.12"
+version = "2.28.11.13"
@@ -4901,2 +4927,2 @@ files = [
- {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
- {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
+ {file = "types-requests-2.28.11.13.tar.gz", hash = "sha256:3fd332842e8759ea5f7eb7789df8aa772ba155216ccf10ef4aa3b0e5b42e1b46"},
+ {file = "types_requests-2.28.11.13-py3-none-any.whl", hash = "sha256:94896f6f8e9f3db11e422c6e3e4abbc5d7ccace853eac74b23bdd65eeee3cdee"},
@@ -4910 +4936 @@ name = "types-urllib3"
-version = "1.26.25.5"
+version = "1.26.25.6"
@@ -4916,2 +4942,2 @@ files = [
- {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
- {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
+ {file = "types-urllib3-1.26.25.6.tar.gz", hash = "sha256:35586727cbd7751acccf2c0f34a88baffc092f435ab62458f10776466590f2d5"},
+ {file = "types_urllib3-1.26.25.6-py3-none-any.whl", hash = "sha256:a6c23c41bd03e542eaee5423a018f833077b51c4bf9ceb5aa544e12b812d5604"},
@@ -4922 +4948 @@ name = "typing-extensions"
-version = "4.4.0"
+version = "4.5.0"
@@ -4928,2 +4954,2 @@ files = [
- {file = "typing_extensions-4.4.0-py3-none-any.whl", hash = "sha256:16fa4864408f655d35ec496218b85f79b3437c829e93320c7c9215ccfd92489e"},
- {file = "typing_extensions-4.4.0.tar.gz", hash = "sha256:1511434bb92bf8dd198c12b1cc812e800d4181cfcb867674e0f8279cc93087aa"},
+ {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"},
+ {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"},
@@ -5038 +5064 @@ name = "werkzeug"
-version = "2.2.2"
+version = "2.2.3"
@@ -5044,2 +5070,2 @@ files = [
- {file = "Werkzeug-2.2.2-py3-none-any.whl", hash = "sha256:f979ab81f58d7318e064e99c4506445d60135ac5cd2e177a2de0089bfd4c9bd5"},
- {file = "Werkzeug-2.2.2.tar.gz", hash = "sha256:7ea2d48322cc7c0f8b3a215ed73eabd7b5d75d0b50e31ab006286ccff9e00b8f"},
+ {file = "Werkzeug-2.2.3-py3-none-any.whl", hash = "sha256:56433961bc1f12533306c624f3be5e744389ac61d722175d543e1751285da612"},
+ {file = "Werkzeug-2.2.3.tar.gz", hash = "sha256:2e1ccc9417d4da358b9de6f174e3ac094391ea1d4fbef2d667865d819dfd0afe"},
@@ -5436 +5462 @@ python-versions = "3.9.15"
-content-hash = "84b6ae8841c72d6f4955e9255622d2b826171732e65fb8104d8cb1ecd8556413"
+content-hash = "4b1fde55056862f335ccae9046704cced7bf1d9599e54a00e5b35c5dde787528"
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index b425f0a2..b885443c 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -9,3 +9,3 @@ license = "Apache-2.0"
-Pillow = "^9.3.0"
-PyICU = "^2.7.4"
-aiohttp = "^3.7.4.post0"
+Pillow = "^9.4.0"
+PyICU = "^2.10.2"
+aiohttp = "^3.8.4"
@@ -14 +14 @@ bs4 = "^0.0.1"
-conllu = "^4.4.1"
+conllu = "^4.5.2"
@@ -16 +16,2 @@ datasets = { extras = ["audio", "vision"], version = "~2.9.0" }
-gdown = "^4.2.0"
+environs = "^9.5.0"
+gdown = "^4.6.3"
@@ -22 +23 @@ lm-dataformat = "^0.0.20"
-lxml = "^4.9.1"
+lxml = "^4.9.2"
@@ -24,2 +25,2 @@ nlp = "^0.4.0"
-nltk = "^3.6.5"
-openpyxl = "^3.0.9"
+nltk = "^3.8.1"
+openpyxl = "^3.1.1"
@@ -27 +28 @@ pdf2image = "^1.16.2"
-py7zr = "^0.20.1"
+py7zr = "^0.20.4"
@@ -32,3 +33,3 @@ rarfile = "^4.0"
-scikit-learn = "^1.0"
-tensorflow-cpu = {version = "^2.9.1", markers = "sys_platform != 'darwin' or platform_machine != 'arm64'"}
-tensorflow-macos = {version = "^2.9.1", markers = "sys_platform == 'darwin' and platform_machine == 'arm64'"}
+scikit-learn = "^1.2.1"
+tensorflow-cpu = {version = "^2.11.0", markers = "sys_platform != 'darwin' or platform_machine != 'arm64'"}
+tensorflow-macos = {version = "^2.11.0", markers = "sys_platform == 'darwin' and platform_machine == 'arm64'"}
@@ -38 +39 @@ torchaudio = { url = "https://download.pytorch.org/whl/cpu/torchaudio-0.13.1%2Bc
-transformers = "^4.11.3"
+transformers = "^4.26.1"
@@ -40 +41 @@ trec-car-tools = { path = "vendors/trec-car-tools/python3" }
-typer = "^0.4.0"
+typer = "^0.4.2"
@@ -44,2 +45,2 @@ wget = "^3.2"
-bandit = "^1.7.0"
-black = "^22.1.0"
+bandit = "^1.7.4"
+black = "^22.12.0"
@@ -47,5 +48,5 @@ flake8 = "^3.9.2"
-isort = "^5.9.3"
-mypy = "0.812"
-pandas-stubs = "^1.5.2.221124"
-poetryup = "^0.3.8"
-pytest = "^7.2.0"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pandas-stubs = "^1.5.3"
+pip-audit = "^2.4.14"
+pytest = "^7.2.1"
@@ -53,2 +54 @@ pytest-cov = "^2.12.1"
-pip-audit = "^2.4.6"
-types-psutil = "^5.9.5.5"
+types-psutil = "^5.9.5"
diff --git a/services/worker/src/worker/asset.py b/services/worker/src/worker/asset.py
index b6e65ba2..c26a8b57 100644
--- a/services/worker/src/worker/asset.py
+++ b/services/worker/src/worker/asset.py
@@ -10 +10 @@ from libcommon.storage import StrPath
-from numpy import ndarray # type:ignore
+from numpy import ndarray
@@ -67 +67 @@ def create_audio_files(
- array: ndarray,
+ array: ndarray, # type: ignore
diff --git a/services/worker/src/worker/features.py b/services/worker/src/worker/features.py
index aafe6460..d74931c7 100644
--- a/services/worker/src/worker/features.py
+++ b/services/worker/src/worker/features.py
@@ -5 +5 @@ import json
-from typing import Any, List, Union
+from typing import Any, List, Optional, Union
@@ -22 +22 @@ from libcommon.storage import StrPath
-from numpy import ndarray # type:ignore
+from numpy import ndarray
@@ -28 +28 @@ from worker.asset import create_audio_files, create_image_file
-def append_hash_suffix(string, json_path: List[Union[str, int]] = None) -> str:
+def append_hash_suffix(string: str, json_path: Optional[List[Union[str, int]]] = None) -> str:
@@ -54 +54 @@ def image(
- json_path: List[Union[str, int]] = None,
+ json_path: Optional[List[Union[str, int]]] = None,
@@ -91 +91 @@ def audio(
- json_path: List[Union[str, int]] = None,
+ json_path: Optional[List[Union[str, int]]] = None,
@@ -129 +129 @@ def get_cell_value(
- json_path: List[Union[str, int]] = None,
+ json_path: Optional[List[Union[str, int]]] = None,
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index 16e27bbd..ac5d5693 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -205 +205 @@ class JobRunner(ABC):
- def __str__(self):
+ def __str__(self) -> str:
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index d793ccfa..ec516d5e 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -8,0 +9 @@ from libcommon.processing_graph import ProcessingGraph
+from libcommon.queue import JobInfo
@@ -12 +13 @@ from worker.config import AppConfig, FirstRowsConfig, ParquetAndDatasetInfoConfi
-from worker.job_runner import JobInfo, JobRunner
+from worker.job_runner import JobRunner
diff --git a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
index 70751adb..0b4f2803 100644
--- a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
+++ b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
@@ -13,0 +14 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import JobInfo
@@ -17 +18 @@ from worker.config import AppConfig, DatasetsBasedConfig
-from worker.job_runner import JobInfo, JobRunner
+from worker.job_runner import JobRunner
diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py
index 7b88b67c..bc498e71 100644
--- a/services/worker/src/worker/job_runners/first_rows.py
+++ b/services/worker/src/worker/job_runners/first_rows.py
@@ -11 +11,12 @@ from pathlib import Path
-from typing import Any, List, Literal, Mapping, Optional, TypedDict, Union
+from typing import (
+ Any,
+ Callable,
+ List,
+ Literal,
+ Mapping,
+ Optional,
+ TypedDict,
+ TypeVar,
+ Union,
+ cast,
+)
@@ -23,0 +35 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import JobInfo
@@ -30,6 +42 @@ from worker.features import get_cell_value
-from worker.job_runner import (
- ConfigNotFoundError,
- JobInfo,
- JobRunnerError,
- SplitNotFoundError,
-)
+from worker.job_runner import ConfigNotFoundError, JobRunnerError, SplitNotFoundError
@@ -130,5 +137,2 @@ class TooBigContentError(FirstRowsJobRunnerError):
-def retry():
- def decorator_retry(func):
- """retries with an increasing sleep before every attempt"""
- SLEEPS = [1, 7, 70, 7 * 60, 70 * 60]
- MAX_ATTEMPTS = len(SLEEPS)
+FuncT = TypeVar("FuncT", bound=Callable[..., Any])
+
@@ -136,16 +140,4 @@ def retry():
- @functools.wraps(func)
- def decorator(*args, **kwargs):
- attempt = 0
- last_err = None
- while attempt < MAX_ATTEMPTS:
- try:
- """always sleep before calling the function. It will prevent rate limiting in the first place"""
- duration = SLEEPS[attempt]
- logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.")
- time.sleep(duration)
- return func(*args, **kwargs)
- except ConnectionError as err:
- logging.info("Got a ConnectionError, possibly due to rate limiting. Let's retry.")
- last_err = err
- attempt += 1
- raise RuntimeError(f"Give up after {attempt} attempts with ConnectionError") from last_err
+def retry(func: FuncT) -> FuncT:
+ """retries with an increasing sleep before every attempt"""
+ SLEEPS = [1, 7, 70, 7 * 60, 70 * 60]
+ MAX_ATTEMPTS = len(SLEEPS)
@@ -153 +145,16 @@ def retry():
- return decorator
+ @functools.wraps(func)
+ def decorator(*args: Any, **kwargs: Any) -> Any:
+ attempt = 0
+ last_err = None
+ while attempt < MAX_ATTEMPTS:
+ try:
+ """always sleep before calling the function. It will prevent rate limiting in the first place"""
+ duration = SLEEPS[attempt]
+ logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.")
+ time.sleep(duration)
+ return func(*args, **kwargs)
+ except ConnectionError as err:
+ logging.info("Got a ConnectionError, possibly due to rate limiting. Let's retry.")
+ last_err = err
+ attempt += 1
+ raise RuntimeError(f"Give up after {attempt} attempts with ConnectionError") from last_err
@@ -155 +162 @@ def retry():
- return decorator_retry
+ return cast(FuncT, decorator)
@@ -181 +188 @@ class FirstRowsResponse(TypedDict):
-@retry()
+@retry
@@ -371 +378 @@ def to_features_list(features: Features) -> List[FeatureItem]:
- features_dict = features.to_dict()
+ features_dict = features.to_dict() # type: ignore
@@ -511 +518 @@ def compute_first_rows_response(
- iterable_dataset = iterable_dataset._resolve_features()
+ iterable_dataset = iterable_dataset._resolve_features() # type: ignore
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index 07a6a5ef..0d4d30a9 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -35 +35 @@ from datasets.utils.py_utils import asdict, map_nested
-from huggingface_hub.hf_api import (
+from huggingface_hub._commit_api import (
@@ -39,3 +38,0 @@ from huggingface_hub.hf_api import (
- DatasetInfo,
- HfApi,
- RepoFile,
@@ -43,2 +40,3 @@ from huggingface_hub.hf_api import (
-from huggingface_hub.utils import RepositoryNotFoundError, RevisionNotFoundError
-from libcommon.dataset import ask_access
+from huggingface_hub.hf_api import DatasetInfo, HfApi, RepoFile
+from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFoundError
+from libcommon.dataset import DatasetNotFoundError, ask_access
@@ -45,0 +44 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import JobInfo
@@ -49 +48 @@ from worker.config import AppConfig, ParquetAndDatasetInfoConfig
-from worker.job_runner import DatasetNotFoundError, JobInfo, JobRunnerError
+from worker.job_runner import JobRunnerError
@@ -508 +507 @@ class _MockStreamingDownloadManager(StreamingDownloadManager):
- def __init__(self, *args, **kwargs):
+ def __init__(self, *args: Any, **kwargs: Any) -> None:
@@ -510,4 +509,12 @@ class _MockStreamingDownloadManager(StreamingDownloadManager):
- self.ext_data_files = []
-
- def download(self, url_or_urls):
- url_or_urls = map_nested(self._download, url_or_urls, map_tuple=True, parallel_min_length=np.inf)
+ self.ext_data_files: List[str] = []
+
+ def download(self, url_or_urls: Any) -> Any:
+ url_or_urls = map_nested(
+ self._download,
+ url_or_urls,
+ map_tuple=True,
+ parallel_min_length=np.inf, # type: ignore
+ # ^ parallel_min_length has int type, but is currently used in datasets for a comparison only
+ # and it works with np.inf. No conversion is involved
+ # (would raise: OverflowError: cannot convert float infinity to integer)
+ )
@@ -516,3 +523,3 @@ class _MockStreamingDownloadManager(StreamingDownloadManager):
- def _download(self, urlpath):
- urlpath = str(urlpath)
- if is_relative_path(urlpath):
+ def _download(self, urlpath: Any) -> str:
+ urlpath_str = str(urlpath)
+ if is_relative_path(urlpath_str):
@@ -520,2 +527,2 @@ class _MockStreamingDownloadManager(StreamingDownloadManager):
- urlpath = url_or_path_join(self._base_path, urlpath)
- elif not urlpath.startswith(self._base_path):
+ urlpath_str = url_or_path_join(self._base_path, urlpath_str)
+ elif not urlpath_str.startswith(self._base_path):
@@ -523,2 +530,2 @@ class _MockStreamingDownloadManager(StreamingDownloadManager):
- self.ext_data_files.append(urlpath)
- return urlpath
+ self.ext_data_files.append(urlpath_str)
+ return urlpath_str
@@ -550 +557 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -559 +566 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -568 +575 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -577 +584 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -586 +593 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -616 +623 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -625 +632 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -634 +641 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -643 +650 @@ def raise_if_too_big_from_external_data_files(
- )
+ ) from error
@@ -728 +735 @@ def compute_parquet_and_dataset_info_response(
- If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
@@ -730 +737 @@ def compute_parquet_and_dataset_info_response(
- If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
@@ -732 +739 @@ def compute_parquet_and_dataset_info_response(
- If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
@@ -734 +741 @@ def compute_parquet_and_dataset_info_response(
- If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
@@ -736 +743 @@ def compute_parquet_and_dataset_info_response(
- If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
@@ -778 +785 @@ def compute_parquet_and_dataset_info_response(
- dataset_info[config] = asdict(builder.info)
+ dataset_info[config] = asdict(builder.info) # type: ignore
@@ -780,0 +788 @@ def compute_parquet_and_dataset_info_response(
+ # note that asdict() is not typed in the datasets library, hence type: ignore
diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py
index 1c82e4a7..e4cffbf6 100644
--- a/services/worker/src/worker/loop.py
+++ b/services/worker/src/worker/loop.py
@@ -55,2 +55,2 @@ class Loop:
- virtual_memory_used: int = virtual_memory().used # type: ignore
- virtual_memory_total: int = virtual_memory().total # type: ignore
+ virtual_memory_used = int(virtual_memory().used)
+ virtual_memory_total = int(virtual_memory().total)
diff --git a/services/worker/src/worker/resources.py b/services/worker/src/worker/resources.py
index f3c743d3..363b1ad8 100644
--- a/services/worker/src/worker/resources.py
+++ b/services/worker/src/worker/resources.py
@@ -25 +25 @@ class LibrariesResource(Resource):
- def allocate(self):
+ def allocate(self) -> None:
@@ -54 +54 @@ class LibrariesResource(Resource):
- def release(self):
+ def release(self) -> None:
diff --git a/services/worker/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py
index af99eb29..b2b11a0e 100644
--- a/services/worker/tests/fixtures/datasets.py
+++ b/services/worker/tests/fixtures/datasets.py
@@ -6 +6 @@ from pathlib import Path
-from typing import Any, Mapping
+from typing import Any, Mapping, Optional
@@ -33,6 +33,6 @@ def value(content: Any, dtype: Any) -> Dataset:
-def other(content: Any, feature_type: FeatureType = None) -> Dataset:
- return (
- Dataset.from_dict({"col": [content]})
- if feature_type is None
- else Dataset.from_dict({"col": [content]}, features=Features({"col": feature_type}))
- )
+def other(content: Any, feature_type: Optional[FeatureType] = None) -> Dataset:
+ if feature_type:
+ features = Features({"col": feature_type}) # type: ignore
+ return Dataset.from_dict({"col": [content]}, features=features)
+ else:
+ return Dataset.from_dict({"col": [content]})
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 0510d83f..89cc1a79 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -9 +9 @@ from pathlib import Path
-from typing import Any, Iterable, List, Literal, Mapping, Optional, Tuple, TypedDict
+from typing import Any, Iterator, List, Literal, Mapping, Optional, Tuple, TypedDict
@@ -13,7 +13,4 @@ import requests
-from datasets import Dataset, load_dataset_builder
-from huggingface_hub.hf_api import (
- REPO_TYPES,
- REPO_TYPES_URL_PREFIXES,
- HfApi,
- hf_raise_for_status,
-)
+from datasets import Dataset, DatasetBuilder, load_dataset_builder
+from huggingface_hub.constants import REPO_TYPES, REPO_TYPES_URL_PREFIXES
+from huggingface_hub.hf_api import HfApi
+from huggingface_hub.utils._errors import hf_raise_for_status
@@ -41,2 +38,2 @@ def update_repo_settings(
- name: str = None,
-) -> Mapping[str, bool]:
+ name: Optional[str] = None,
+) -> Any:
@@ -102 +99,6 @@ def create_hub_dataset_repo(
- *, prefix: str, file_paths: List[str] = None, dataset: Dataset = None, private=False, gated=False
+ *,
+ prefix: str,
+ file_paths: Optional[List[str]] = None,
+ dataset: Optional[Dataset] = None,
+ private: bool = False,
+ gated: bool = False,
@@ -134 +136 @@ def delete_hub_dataset_repo(repo_id: str) -> None:
-def hub_public_empty() -> Iterable[str]:
+def hub_public_empty() -> Iterator[str]:
@@ -141 +143 @@ def hub_public_empty() -> Iterable[str]:
-def hub_public_csv(csv_path: str) -> Iterable[str]:
+def hub_public_csv(csv_path: str) -> Iterator[str]:
@@ -148 +150 @@ def hub_public_csv(csv_path: str) -> Iterable[str]:
-def hub_private_csv(csv_path: str) -> Iterable[str]:
+def hub_private_csv(csv_path: str) -> Iterator[str]:
@@ -155 +157 @@ def hub_private_csv(csv_path: str) -> Iterable[str]:
-def hub_gated_csv(csv_path: str) -> Iterable[str]:
+def hub_gated_csv(csv_path: str) -> Iterator[str]:
@@ -162 +164 @@ def hub_gated_csv(csv_path: str) -> Iterable[str]:
-def hub_public_jsonl(jsonl_path: str) -> Iterable[str]:
+def hub_public_jsonl(jsonl_path: str) -> Iterator[str]:
@@ -169 +171 @@ def hub_public_jsonl(jsonl_path: str) -> Iterable[str]:
-def hub_gated_extra_fields_csv(csv_path: str, extra_fields_readme: str) -> Iterable[str]:
+def hub_gated_extra_fields_csv(csv_path: str, extra_fields_readme: str) -> Iterator[str]:
@@ -178 +180 @@ def hub_gated_extra_fields_csv(csv_path: str, extra_fields_readme: str) -> Itera
-def hub_public_audio(datasets: Mapping[str, Dataset]) -> Iterable[str]:
+def hub_public_audio(datasets: Mapping[str, Dataset]) -> Iterator[str]:
@@ -185 +187 @@ def hub_public_audio(datasets: Mapping[str, Dataset]) -> Iterable[str]:
-def hub_public_image(datasets: Mapping[str, Dataset]) -> Iterable[str]:
+def hub_public_image(datasets: Mapping[str, Dataset]) -> Iterator[str]:
@@ -192 +194 @@ def hub_public_image(datasets: Mapping[str, Dataset]) -> Iterable[str]:
-def hub_public_images_list(datasets: Mapping[str, Dataset]) -> Iterable[str]:
+def hub_public_images_list(datasets: Mapping[str, Dataset]) -> Iterator[str]:
@@ -199 +201 @@ def hub_public_images_list(datasets: Mapping[str, Dataset]) -> Iterable[str]:
-def hub_public_big(datasets: Mapping[str, Dataset]) -> Iterable[str]:
+def hub_public_big(datasets: Mapping[str, Dataset]) -> Iterator[str]:
@@ -206 +208 @@ def hub_public_big(datasets: Mapping[str, Dataset]) -> Iterable[str]:
-def hub_public_external_files(dataset_script_with_external_files_path) -> Iterable[str]:
+def hub_public_external_files(dataset_script_with_external_files_path: str) -> Iterator[str]:
@@ -213 +215 @@ def hub_public_external_files(dataset_script_with_external_files_path) -> Iterab
-def external_files_dataset_builder(hub_public_external_files):
+def external_files_dataset_builder(hub_public_external_files: str) -> DatasetBuilder:
@@ -229 +231 @@ HubDatasets = Mapping[str, HubDatasetTest]
-def create_config_names_response(dataset: str):
+def create_config_names_response(dataset: str) -> Any:
@@ -241 +243 @@ def create_config_names_response(dataset: str):
-def create_split_names_response(dataset: str):
+def create_split_names_response(dataset: str) -> Any:
@@ -254 +256 @@ def create_split_names_response(dataset: str):
-def create_splits_response(dataset: str):
+def create_splits_response(dataset: str) -> Any:
@@ -267 +269 @@ def create_splits_response(dataset: str):
-def create_first_rows_response(dataset: str, cols: Mapping[str, Any], rows: List[Any]):
+def create_first_rows_response(dataset: str, cols: Mapping[str, Any], rows: List[Any]) -> Any:
@@ -292 +294 @@ def create_first_rows_response(dataset: str, cols: Mapping[str, Any], rows: List
-def create_dataset_info_response_for_csv(dataset: str, config: str):
+def create_dataset_info_response_for_csv(dataset: str, config: str) -> Any:
@@ -315 +317 @@ def create_dataset_info_response_for_csv(dataset: str, config: str):
-def create_dataset_info_response_for_audio(dataset: str, config: str):
+def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any:
@@ -335 +337 @@ def create_dataset_info_response_for_audio(dataset: str, config: str):
-def create_parquet_and_dataset_info_response(dataset: str, data_type: Literal["csv", "audio"]):
+def create_parquet_and_dataset_info_response(dataset: str, data_type: Literal["csv", "audio"]) -> Any:
@@ -398 +400 @@ AUDIO_cols = {
-def get_AUDIO_rows(dataset: str):
+def get_AUDIO_rows(dataset: str) -> Any:
@@ -421 +423 @@ IMAGE_cols = {
-def get_IMAGE_rows(dataset: str):
+def get_IMAGE_rows(dataset: str) -> Any:
@@ -439 +441 @@ IMAGES_LIST_cols = {
-def get_IMAGES_LIST_rows(dataset: str):
+def get_IMAGES_LIST_rows(dataset: str) -> Any:
@@ -483,11 +485,11 @@ def hub_datasets(
- hub_public_empty,
- hub_public_csv,
- hub_private_csv,
- hub_gated_csv,
- hub_public_jsonl,
- hub_gated_extra_fields_csv,
- hub_public_audio,
- hub_public_image,
- hub_public_images_list,
- hub_public_big,
- hub_public_external_files,
+ hub_public_empty: str,
+ hub_public_csv: str,
+ hub_private_csv: str,
+ hub_gated_csv: str,
+ hub_public_jsonl: str,
+ hub_gated_extra_fields_csv: str,
+ hub_public_audio: str,
+ hub_public_image: str,
+ hub_public_images_list: str,
+ hub_public_big: str,
+ hub_public_external_files: str,
diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py
index 3f0e5d81..ba24c4c5 100644
--- a/services/worker/tests/job_runners/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__datasets_based_worker.py
@@ -8 +8 @@ from pathlib import Path
-from typing import Any, Mapping, Optional
+from typing import Any, Callable, Mapping, Optional
@@ -41,0 +42,3 @@ class DummyJobRunner(DatasetsBasedJobRunner):
+GetJobRunner = Callable[[str, Optional[str], Optional[str], AppConfig, bool], DummyJobRunner]
+
+
@@ -47 +50 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -53 +56 @@ def get_job_runner(
- force: bool = False,
+ force: bool,
@@ -81 +84 @@ def get_job_runner(
-def test_version(app_config: AppConfig, get_job_runner) -> None:
+def test_version(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -83 +86 @@ def test_version(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset, config, split, app_config)
+ job_runner = get_job_runner(dataset, config, split, app_config, False)
@@ -111 +114 @@ def test_get_cache_subdirectory(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -119 +122 @@ def test_get_cache_subdirectory(
- job_runner = get_job_runner(dataset, config, split, app_config, force=force)
+ job_runner = get_job_runner(dataset, config, split, app_config, force)
@@ -123 +126 @@ def test_get_cache_subdirectory(
-def test_set_and_unset_datasets_cache(app_config: AppConfig, get_job_runner) -> None:
+def test_set_and_unset_datasets_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -125 +128 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig, get_job_runner) ->
- job_runner = get_job_runner(dataset, config, split, app_config)
+ job_runner = get_job_runner(dataset, config, split, app_config, False)
@@ -134 +137 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig, get_job_runner) ->
-def test_set_and_unset_cache(app_config: AppConfig, get_job_runner) -> None:
+def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -136 +139 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset, config, split, app_config)
+ job_runner = get_job_runner(dataset, config, split, app_config, False)
@@ -146 +149 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner) -> None:
-def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str, config: str) -> None:
+def test_process(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public_csv: str, config: str) -> None:
@@ -151 +154 @@ def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str, con
- job_runner = get_job_runner(dataset, config, split, app_config)
+ job_runner = get_job_runner(dataset, config, split, app_config, False)
@@ -168 +171 @@ def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) ->
-def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, get_job_runner) -> None:
+def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -171,4 +174 @@ def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, g
- dataset,
- config,
- split,
- app_config=replace(app_config, worker=replace(app_config.worker, content_max_bytes=10)),
+ dataset, config, split, replace(app_config, worker=replace(app_config.worker, content_max_bytes=10)), False
diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py
index 14a6e883..a82a6912 100644
--- a/services/worker/tests/job_runners/test_config_names.py
+++ b/services/worker/tests/job_runners/test_config_names.py
@@ -5,0 +6 @@ from http import HTTPStatus
+from typing import Callable
@@ -19,0 +21,2 @@ from ..fixtures.hub import HubDatasets
+GetJobRunner = Callable[[str, AppConfig, bool], ConfigNamesJobRunner]
+
@@ -26 +29 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -58 +61 @@ def get_job_runner(
-def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
+def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner: GetJobRunner) -> None:
@@ -60 +63 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_run
- job_runner = get_job_runner(dataset, app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -65 +68 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_run
- job_runner = get_job_runner(dataset, app_config, force=True)
+ job_runner = get_job_runner(dataset, app_config, True)
@@ -69 +72 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_run
-def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
+def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner: GetJobRunner) -> None:
@@ -71 +74 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) ->
- job_runner = get_job_runner(dataset, app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -83 +86 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) ->
-def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -85 +88 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset, app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -108 +111 @@ def test_compute_splits_response_simple_csv(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -119,0 +123 @@ def test_compute_splits_response_simple_csv(
+ False,
diff --git a/services/worker/tests/job_runners/test_dataset_info.py b/services/worker/tests/job_runners/test_dataset_info.py
index a5010afb..5674de6a 100644
--- a/services/worker/tests/job_runners/test_dataset_info.py
+++ b/services/worker/tests/job_runners/test_dataset_info.py
@@ -5 +5 @@ from http import HTTPStatus
-from typing import Any
+from typing import Any, Callable
@@ -7,0 +8 @@ import pytest
+from libcommon.dataset import DatasetNotFoundError
@@ -16 +16,0 @@ from worker.job_runners.dataset_info import (
- DatasetNotFoundError,
@@ -27,0 +28,3 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
+GetJobRunner = Callable[[str, AppConfig, bool], DatasetInfoJobRunner]
+
+
@@ -32 +35 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -88 +91 @@ def test_compute(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -99 +102 @@ def test_compute(
- job_runner = get_job_runner(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -108 +111 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -110 +113 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py
index 27863b9c..66adf01d 100644
--- a/services/worker/tests/job_runners/test_first_rows.py
+++ b/services/worker/tests/job_runners/test_first_rows.py
@@ -5,0 +6 @@ from http import HTTPStatus
+from typing import Callable
@@ -21,0 +23,2 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
+GetJobRunner = Callable[[str, str, str, AppConfig, FirstRowsConfig, bool], FirstRowsJobRunner]
+
@@ -29 +32 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -67 +70 @@ def test_should_skip_job(
- app_config: AppConfig, get_job_runner, first_rows_config: FirstRowsConfig, hub_public_csv: str
+ app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig, hub_public_csv: str
@@ -70 +73 @@ def test_should_skip_job(
- job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config)
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False)
@@ -75 +78 @@ def test_should_skip_job(
- job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, force=True)
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, True)
@@ -80 +83 @@ def test_compute(
- app_config: AppConfig, get_job_runner, first_rows_config: FirstRowsConfig, hub_public_csv: str
+ app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig, hub_public_csv: str
@@ -83 +86 @@ def test_compute(
- job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config)
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False)
@@ -101 +104 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, get_job_runner, first_rows_config: FirstRowsConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig) -> None:
@@ -104 +107 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner, first_rows_config:
- job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config)
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False)
@@ -130 +133 @@ def test_number_rows(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -141,2 +144,2 @@ def test_number_rows(
- if hasattr(csv, "_patched_for_streaming") and csv._patched_for_streaming: # type: ignore
- csv._patched_for_streaming = False # type: ignore
+ if hasattr(csv, "_patched_for_streaming") and csv._patched_for_streaming:
+ csv._patched_for_streaming = False
@@ -152,0 +156 @@ def test_number_rows(
+ False,
@@ -191 +195 @@ def test_truncation(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -204,2 +208,2 @@ def test_truncation(
- app_config=replace(app_config, common=replace(app_config.common, hf_token=None)),
- first_rows_config=replace(
+ replace(app_config, common=replace(app_config.common, hf_token=None)),
+ replace(
@@ -212,0 +217 @@ def test_truncation(
+ False,
diff --git a/services/worker/tests/job_runners/test_parquet.py b/services/worker/tests/job_runners/test_parquet.py
index ac4e4a07..12dc805b 100644
--- a/services/worker/tests/job_runners/test_parquet.py
+++ b/services/worker/tests/job_runners/test_parquet.py
@@ -5 +5 @@ from http import HTTPStatus
-from typing import Any
+from typing import Any, Callable
@@ -7,0 +8 @@ import pytest
+from libcommon.dataset import DatasetNotFoundError
@@ -15 +15,0 @@ from worker.job_runners.parquet import (
- DatasetNotFoundError,
@@ -27,0 +28,3 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
+GetJobRunner = Callable[[str, AppConfig, bool], ParquetJobRunner]
+
+
@@ -32 +35 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -88 +91 @@ def test_compute(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -99 +102 @@ def test_compute(
- job_runner = get_job_runner(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -108 +111 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -110 +113 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index e2d16573..5706f2ef 100644
--- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Any, Iterator, List, Optional
+from typing import Any, Callable, Iterator, List, Optional
@@ -61,0 +62,3 @@ def parquet_and_dataset_info_config(
+GetJobRunner = Callable[[str, AppConfig, ParquetAndDatasetInfoConfig, bool], ParquetAndDatasetInfoJobRunner]
+
+
@@ -67 +70 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -120 +123 @@ def test_compute(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -125,3 +128 @@ def test_compute(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
@@ -140 +141 @@ def test_doesnotexist(
- app_config: AppConfig, get_job_runner, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig
+ app_config: AppConfig, get_job_runner: GetJobRunner, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig
@@ -143,3 +144 @@ def test_doesnotexist(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
@@ -229 +228 @@ def test_raise_if_too_big_from_datasets(
- "max_dataset_size,max_external_data_files,error",
+ "max_dataset_size,max_external_data_files,raises",
@@ -231,3 +230,2 @@ def test_raise_if_too_big_from_datasets(
- (None, None, None),
- (10, None, DatasetWithTooBigExternalFilesError),
- (None, 1, DatasetWithTooManyExternalFilesError),
+ (None, None, False),
+ (10, None, True),
@@ -236 +234 @@ def test_raise_if_too_big_from_datasets(
-def test_raise_if_too_big_from_external_files(
+def test_raise_if_too_big_external_files(
@@ -238 +236 @@ def test_raise_if_too_big_from_external_files(
- error,
+ raises: bool,
@@ -246,2 +244,36 @@ def test_raise_if_too_big_from_external_files(
- if error:
- with pytest.raises(error):
+ if raises:
+ with pytest.raises(DatasetWithTooBigExternalFilesError):
+ raise_if_too_big_from_external_data_files(
+ builder=external_files_dataset_builder,
+ hf_token=app_config.common.hf_token,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ )
+ else:
+ raise_if_too_big_from_external_data_files(
+ builder=external_files_dataset_builder,
+ hf_token=app_config.common.hf_token,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ )
+
+
[email protected](
+ "max_dataset_size,max_external_data_files,raises",
+ [
+ (None, None, False),
+ (None, 1, True),
+ ],
+)
+def test_raise_if_too_many_external_files(
+ external_files_dataset_builder: "datasets.builder.DatasetBuilder",
+ raises: bool,
+ max_dataset_size: Optional[int],
+ max_external_data_files: Optional[int],
+ app_config: AppConfig,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
+) -> None:
+ max_dataset_size = max_dataset_size or parquet_and_dataset_info_config.max_dataset_size
+ max_external_data_files = max_external_data_files or parquet_and_dataset_info_config.max_external_data_files
+ if raises:
+ with pytest.raises(DatasetWithTooManyExternalFilesError):
@@ -304 +336 @@ def test_not_supported_if_big(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -310,3 +342 @@ def test_not_supported_if_big(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
@@ -321 +351 @@ def test_supported_if_gated(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -327,3 +357 @@ def test_supported_if_gated(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
@@ -338 +366 @@ def test_not_supported_if_gated_with_extra_fields(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -344,3 +372 @@ def test_not_supported_if_gated_with_extra_fields(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
@@ -355 +381 @@ def test_blocked(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -361,3 +387 @@ def test_blocked(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
@@ -376 +400 @@ def test_compute_splits_response_simple_csv_ok(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -384,3 +408 @@ def test_compute_splits_response_simple_csv_ok(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
@@ -419 +441 @@ def test_compute_splits_response_simple_csv_error(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -427,3 +449 @@ def test_compute_splits_response_simple_csv_error(
- job_runner = get_job_runner(
- dataset=dataset, app_config=app_config, parquet_and_dataset_info_config=parquet_and_dataset_info_config
- )
+ job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False)
diff --git a/services/worker/tests/job_runners/test_sizes.py b/services/worker/tests/job_runners/test_sizes.py
index 858fe019..5e272fff 100644
--- a/services/worker/tests/job_runners/test_sizes.py
+++ b/services/worker/tests/job_runners/test_sizes.py
@@ -5 +5 @@ from http import HTTPStatus
-from typing import Any
+from typing import Any, Callable
@@ -7,0 +8 @@ import pytest
+from libcommon.dataset import DatasetNotFoundError
@@ -15 +15,0 @@ from worker.job_runners.sizes import (
- DatasetNotFoundError,
@@ -27,0 +28,3 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
+GetJobRunner = Callable[[str, AppConfig, bool], SizesJobRunner]
+
+
@@ -32 +35 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -243 +246 @@ def test_compute(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -254 +257 @@ def test_compute(
- job_runner = get_job_runner(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -263 +266 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -265 +268 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
diff --git a/services/worker/tests/job_runners/test_split_names.py b/services/worker/tests/job_runners/test_split_names.py
index 4ca27e23..4d45e534 100644
--- a/services/worker/tests/job_runners/test_split_names.py
+++ b/services/worker/tests/job_runners/test_split_names.py
@@ -5,0 +6 @@ from http import HTTPStatus
+from typing import Callable
@@ -19,0 +21,2 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
+GetJobRunner = Callable[[str, str, AppConfig, bool], SplitNamesJobRunner]
+
@@ -26 +29 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -59 +62 @@ def get_job_runner(
-def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str) -> None:
+def test_process(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public_csv: str) -> None:
@@ -61 +64 @@ def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str) ->
- job_runner = get_job_runner(dataset, config, app_config)
+ job_runner = get_job_runner(dataset, config, app_config, False)
@@ -73 +76 @@ def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str) ->
-def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -76 +79 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset, config, app_config)
+ job_runner = get_job_runner(dataset, config, app_config, False)
@@ -99 +102 @@ def test_compute_split_names_response(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -107 +109,0 @@ def test_compute_split_names_response(
- job_runner = get_job_runner(dataset, config, app_config)
@@ -112,0 +115 @@ def test_compute_split_names_response(
+ False,
diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py
index 404beabf..a3e2d93a 100644
--- a/services/worker/tests/job_runners/test_splits.py
+++ b/services/worker/tests/job_runners/test_splits.py
@@ -5,0 +6 @@ from http import HTTPStatus
+from typing import Callable
@@ -19,0 +21,2 @@ from ..fixtures.hub import HubDatasets
+GetJobRunner = Callable[[str, AppConfig, bool], SplitsJobRunner]
+
@@ -26 +29 @@ def get_job_runner(
-):
+) -> GetJobRunner:
@@ -58 +61 @@ def get_job_runner(
-def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
+def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner: GetJobRunner) -> None:
@@ -60 +63 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner)
- job_runner = get_job_runner(dataset, app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -65 +68 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner)
- job_runner = get_job_runner(dataset, app_config, force=True)
+ job_runner = get_job_runner(dataset, app_config, True)
@@ -69 +72 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner)
-def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
+def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner: GetJobRunner) -> None:
@@ -71 +74 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) ->
- job_runner = get_job_runner(dataset, app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -84 +87 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) ->
-def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -86 +89 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- job_runner = get_job_runner(dataset, app_config)
+ job_runner = get_job_runner(dataset, app_config, False)
@@ -109 +112 @@ def test_compute_splits_response_simple_csv(
- get_job_runner,
+ get_job_runner: GetJobRunner,
@@ -120,0 +124 @@ def test_compute_splits_response_simple_csv(
+ False,
diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py
index 0ed36874..7034927b 100644
--- a/services/worker/tests/test_job_runner_factory.py
+++ b/services/worker/tests/test_job_runner_factory.py
@@ -8 +8 @@ from libcommon.processing_graph import ProcessingGraph
-from libcommon.queue import Priority
+from libcommon.queue import JobInfo, Priority
@@ -12 +11,0 @@ from worker.config import AppConfig
-from worker.job_runner import JobInfo
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 6647f403..21ed88ad 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -4,0 +5 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import JobInfo
@@ -8 +9 @@ from worker.config import AppConfig, WorkerConfig
-from worker.job_runner import JobInfo, JobRunner
+from worker.job_runner import JobRunner
diff --git a/services/worker/tests/test_resources.py b/services/worker/tests/test_resources.py
index f7dd816c..d13e85c4 100644
--- a/services/worker/tests/test_resources.py
+++ b/services/worker/tests/test_resources.py
@@ -5,0 +6 @@ import pytest
+from pytest import TempPathFactory
@@ -14 +15,3 @@ from worker.resources import LibrariesResource
-def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define_numba_path: bool) -> None:
+def test_libraries(
+ tmp_path_factory: TempPathFactory, define_init_hf_datasets_cache: bool, define_numba_path: bool
+) -> None:
@@ -16,2 +19,2 @@ def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define
- init_hf_datasets_cache: str = (
- tmp_path_factory.mktemp("hf_datasets_cache") if define_init_hf_datasets_cache else None
+ init_hf_datasets_cache = (
+ str(tmp_path_factory.mktemp("hf_datasets_cache")) if define_init_hf_datasets_cache else None
@@ -19 +22 @@ def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define
- numba_path: str = tmp_path_factory.mktemp("numba_path") if define_numba_path else None
+ numba_path = str(tmp_path_factory.mktemp("numba_path")) if define_numba_path else None
@@ -22,3 +25 @@ def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define
- hf_endpoint=hf_endpoint,
- init_hf_datasets_cache=init_hf_datasets_cache,
- numba_path=numba_path,
+ hf_endpoint=hf_endpoint, init_hf_datasets_cache=init_hf_datasets_cache, numba_path=numba_path
@@ -31 +32 @@ def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define
- assert (resource.hf_datasets_cache == init_hf_datasets_cache) == define_init_hf_datasets_cache
+ assert (str(resource.hf_datasets_cache) == init_hf_datasets_cache) == define_init_hf_datasets_cache
@@ -38 +39 @@ def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define
-def test_libraries_context_manager(tmp_path_factory) -> None:
+def test_libraries_context_manager(tmp_path_factory: TempPathFactory) -> None:
@@ -40,2 +41,2 @@ def test_libraries_context_manager(tmp_path_factory) -> None:
- init_hf_datasets_cache: str = tmp_path_factory.mktemp("hf_datasets_cache")
- numba_path: str = tmp_path_factory.mktemp("numba_path")
+ init_hf_datasets_cache = str(tmp_path_factory.mktemp("hf_datasets_cache"))
+ numba_path = str(tmp_path_factory.mktemp("numba_path"))
@@ -44,3 +45 @@ def test_libraries_context_manager(tmp_path_factory) -> None:
- hf_endpoint=hf_endpoint,
- init_hf_datasets_cache=init_hf_datasets_cache,
- numba_path=numba_path,
+ hf_endpoint=hf_endpoint, init_hf_datasets_cache=init_hf_datasets_cache, numba_path=numba_path
|
|
4f66f176117711580db0d82daabcb5d4a4eabc97
|
Quentin Lhoest
| 2023-02-15T14:10:18 |
Ignore big datasets from external files (#809)
|
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index fcfec3ca..b0ae7cda 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -99,0 +100 @@ PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE = 100_000_000
+PARQUET_AND_DATASET_INFO_MAX_EXTERNAL_DATA_FILES = 10_000
@@ -111,0 +113 @@ class ParquetAndDatasetInfoConfig:
+ max_external_data_files: int = PARQUET_AND_DATASET_INFO_MAX_EXTERNAL_DATA_FILES
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index efc4f42b..07a6a5ef 100644
--- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -7,0 +8 @@ import re
+from functools import partial
@@ -8,0 +10 @@ from http import HTTPStatus
+from multiprocessing.pool import ThreadPool
@@ -15 +17,9 @@ import datasets.config
-from datasets import get_dataset_config_names, get_dataset_infos, load_dataset_builder
+import numpy as np
+import requests
+from datasets import (
+ DownloadConfig,
+ get_dataset_config_names,
+ get_dataset_infos,
+ load_dataset_builder,
+)
+from datasets.builder import DatasetBuilder
@@ -17 +27,8 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError
-from datasets.utils.py_utils import asdict
+from datasets.download import StreamingDownloadManager
+from datasets.utils.file_utils import (
+ get_authentication_headers_for_url,
+ http_head,
+ is_relative_path,
+ url_or_path_join,
+)
+from datasets.utils.py_utils import asdict, map_nested
@@ -41,0 +59,7 @@ ParquetAndDatasetInfoJobRunnerErrorCode = Literal[
+ "UnsupportedExternalFilesError",
+ "DatasetWithTooManyExternalFilesError",
+ "DatasetWithTooBigExternalFilesError",
+ "ExternalFilesSizeRequestHTTPError",
+ "ExternalFilesSizeRequestConnectionError",
+ "ExternalFilesSizeRequestTimeoutError",
+ "ExternalFilesSizeRequestError",
@@ -425,0 +450,196 @@ class EmptyFeaturesError(Exception):
+class DatasetWithTooManyExternalFilesError(ParquetAndDatasetInfoJobRunnerError):
+ """Raised when the dataset size (sum of config sizes given by the datasets library) is too big."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyExternalFilesError", cause, True)
+
+
+class DatasetWithTooBigExternalFilesError(ParquetAndDatasetInfoJobRunnerError):
+ """Raised when the dataset size (sum of config sizes given by the datasets library) is too big."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooBigExternalFilesError", cause, True)
+
+
+class UnsupportedExternalFilesError(ParquetAndDatasetInfoJobRunnerError):
+ """Raised when we failed to get the size of the external files."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "UnsupportedExternalFilesError", cause, True)
+
+
+class ExternalFilesSizeRequestHTTPError(ParquetAndDatasetInfoJobRunnerError):
+ """Raised when we failed to get the size of the external files."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestHTTPError", cause, True)
+
+
+class ExternalFilesSizeRequestConnectionError(ParquetAndDatasetInfoJobRunnerError):
+ """Raised when we failed to get the size of the external files."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestConnectionError", cause, True)
+
+
+class ExternalFilesSizeRequestTimeoutError(ParquetAndDatasetInfoJobRunnerError):
+ """Raised when we failed to get the size of the external files."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestTimeoutError", cause, True)
+
+
+class ExternalFilesSizeRequestError(ParquetAndDatasetInfoJobRunnerError):
+ """Raised when we failed to get the size of the external files."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestError", cause, True)
+
+
+def _request_size(url: str, hf_token: Optional[str] = None) -> Optional[int]:
+ headers = get_authentication_headers_for_url(url, use_auth_token=hf_token)
+ response = http_head(url, headers=headers, max_retries=3)
+ response.raise_for_status()
+ size = response.headers.get("Content-Length") if response.ok else None
+ return int(size) if size is not None else size
+
+
+class _MockStreamingDownloadManager(StreamingDownloadManager):
+ def __init__(self, *args, **kwargs):
+ super().__init__(*args, **kwargs)
+ self.ext_data_files = []
+
+ def download(self, url_or_urls):
+ url_or_urls = map_nested(self._download, url_or_urls, map_tuple=True, parallel_min_length=np.inf)
+ return url_or_urls
+
+ def _download(self, urlpath):
+ urlpath = str(urlpath)
+ if is_relative_path(urlpath):
+ # append the relative path to the base_path
+ urlpath = url_or_path_join(self._base_path, urlpath)
+ elif not urlpath.startswith(self._base_path):
+ # it's an external file
+ self.ext_data_files.append(urlpath)
+ return urlpath
+
+
+def raise_if_too_big_from_external_data_files(
+ builder: DatasetBuilder, max_dataset_size: int, max_external_data_files: int, hf_token: Optional[str]
+) -> None:
+ # Packaged dataset modules only load data files that are inside the dataset repository.
+ # No need to check them since they're already caught by `raise_if_too_big_from_hub`
+ if type(builder).__module__.startswith("datasets."):
+ return
+ # For datasets with a loading script however, we need to check the downloaded files
+ mock_dl_manager = _MockStreamingDownloadManager(
+ base_path=builder.base_path, download_config=DownloadConfig(use_auth_token=hf_token)
+ )
+ try:
+ builder._split_generators(mock_dl_manager) # type: ignore
+ except (requests.exceptions.RequestException, NotImplementedError) as error:
+ if isinstance(error, NotImplementedError):
+ # we can ignore the errors from functions not implemented in streaming mode like `.extract()` on TAR files
+ if "is not implemented in streaming mode." not in str(error):
+ raise UnsupportedExternalFilesError(
+ (
+ "Couldn't get the list of external files in `_split_generators` because it doesn't support"
+ f" streaming:\n{error}"
+ ),
+ error,
+ )
+ elif isinstance(error, requests.exceptions.HTTPError):
+ raise ExternalFilesSizeRequestHTTPError(
+ (
+ "Couldn't get the list of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+ elif isinstance(error, requests.exceptions.ConnectionError):
+ raise ExternalFilesSizeRequestConnectionError(
+ (
+ "Couldn't get the list of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+ elif isinstance(error, requests.exceptions.Timeout):
+ raise ExternalFilesSizeRequestTimeoutError(
+ (
+ "Couldn't get the list of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+ else:
+ raise ExternalFilesSizeRequestError(
+ (
+ "Couldn't get the list of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+ ext_data_files = mock_dl_manager.ext_data_files
+ if len(ext_data_files) > max_external_data_files:
+ raise DatasetWithTooManyExternalFilesError(
+ f"The conversion to parquet is limited to datasets with less than {max_external_data_files} files. "
+ f"However it uses {len(ext_data_files)} data files."
+ )
+ elif ext_data_files:
+ try:
+ with ThreadPool(16) as pool:
+ total_size = 0
+ get_size = partial(_request_size, hf_token=hf_token)
+ for i, size in enumerate(pool.imap_unordered(get_size, ext_data_files)):
+ if size is not None:
+ total_size += size
+ if total_size > max_dataset_size:
+ raise DatasetWithTooBigExternalFilesError(
+ f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes."
+ f" However {i + 1} data files of {len(ext_data_files)} are already bigger than"
+ f" {total_size} bytes."
+ )
+ except requests.exceptions.RequestException as error:
+ if isinstance(error, requests.exceptions.HTTPError):
+ raise ExternalFilesSizeRequestHTTPError(
+ (
+ "Couldn't get the size of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+ elif isinstance(error, requests.exceptions.ConnectionError):
+ raise ExternalFilesSizeRequestConnectionError(
+ (
+ "Couldn't get the size of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+ elif isinstance(error, requests.exceptions.Timeout):
+ raise ExternalFilesSizeRequestTimeoutError(
+ (
+ "Couldn't get the size of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+ else:
+ raise ExternalFilesSizeRequestError(
+ (
+ "Couldn't get the size of external files in `_split_generators` because a request"
+ f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead"
+ " (e.g. inside a data/ folder)."
+ ),
+ error,
+ )
+
+
@@ -437,0 +658 @@ def compute_parquet_and_dataset_info_response(
+ max_external_data_files: int,
@@ -469,0 +691,2 @@ def compute_parquet_and_dataset_info_response(
+ max_external_data_files (`int`):
+ The maximum number of external data files of a dataset. This is for datasets with loading scripts only.
@@ -499,0 +723,14 @@ def compute_parquet_and_dataset_info_response(
+ - [`~job_runners.parquet_and_dataset_info.DatasetWithTooManyExternalFilesError`]
+ If the dataset has too many external files to be converted to parquet
+ - [`~job_runners.parquet_and_dataset_info.DatasetWithTooBigExternalFilesError`]
+ If the dataset is too big external files be converted to parquet
+ - [`~job_runners.parquet_and_dataset_info.UnsupportedExternalFilesError`]
+ If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestHTTPError`]
+ If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestConnectionError`]
+ If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestTimeoutError`]
+ If we failed to get the external files sizes to make sure we can convert the dataet to parquet
+ - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestError`]
+ If we failed to get the external files sizes to make sure we can convert the dataet to parquet
@@ -533,0 +771,6 @@ def compute_parquet_and_dataset_info_response(
+ raise_if_too_big_from_external_data_files(
+ builder=builder,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ hf_token=hf_token,
+ )
@@ -612 +855 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner):
- return "1.0.0"
+ return "1.1.0"
@@ -642,0 +886 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner):
+ max_external_data_files=self.parquet_and_dataset_info_config.max_external_data_files,
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index e8686eca..279ae7d6 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -52,0 +53 @@ def set_env_vars(datasets_cache_directory: Path, modules_cache_directory: Path)
+ mp.setenv("PARQUET_AND_DATASET_INFO_MAX_EXTERNAL_DATA_FILES", "10")
diff --git a/services/worker/tests/fixtures/files.py b/services/worker/tests/fixtures/files.py
index edab9e8f..50a3b7af 100644
--- a/services/worker/tests/fixtures/files.py
+++ b/services/worker/tests/fixtures/files.py
@@ -66,0 +67,49 @@ def extra_fields_readme(tmp_path_factory: pytest.TempPathFactory) -> str:
+
+
+DATASET_SCRIPT_WITH_EXTERNAL_FILES_CONTENT = """
+import datasets
+
+_URLS = {
+ "train": [
+ "https://huggingface.co/datasets/lhoestq/test/resolve/main/some_text.txt",
+ "https://huggingface.co/datasets/lhoestq/test/resolve/main/another_text.txt",
+ ]
+}
+
+
+class Test(datasets.GeneratorBasedBuilder):
+
+
+ def _info(self):
+ return datasets.DatasetInfo(
+ features=datasets.Features(
+ {
+ "text": datasets.Value("string"),
+ }
+ ),
+ homepage="https://huggingface.co/datasets/lhoestq/test",
+ )
+
+ def _split_generators(self, dl_manager):
+ downloaded_files = dl_manager.download_and_extract(_URLS)
+
+ return [
+ datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": downloaded_files["train"]}),
+ ]
+
+ def _generate_examples(self, filepaths):
+ _id = 0
+ for filepath in filepaths:
+ with open(filepath, encoding="utf-8") as f:
+ for line in f:
+ yield _id, {"text": line.rstrip()}
+ _id += 1
+"""
+
+
[email protected](scope="session")
+def dataset_script_with_external_files_path(tmp_path_factory: pytest.TempPathFactory) -> str:
+ path = str(tmp_path_factory.mktemp("data") / "{dataset_name}.py")
+ with open(path, "w", newline="") as f:
+ f.write(DATASET_SCRIPT_WITH_EXTERNAL_FILES_CONTENT)
+ return path
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 244ab0fd..0510d83f 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -13 +13 @@ import requests
-from datasets import Dataset
+from datasets import Dataset, load_dataset_builder
@@ -104 +104,2 @@ def create_hub_dataset_repo(
- repo_id = f"{CI_USER}/{prefix}-{int(time.time() * 10e3)}"
+ dataset_name = f"{prefix}-{int(time.time() * 10e3)}"
+ repo_id = f"{CI_USER}/{dataset_name}"
@@ -116 +117 @@ def create_hub_dataset_repo(
- path_in_repo=Path(file_path).name,
+ path_in_repo=Path(file_path).name.replace("{dataset_name}", dataset_name),
@@ -203,0 +205,12 @@ def hub_public_big(datasets: Mapping[str, Dataset]) -> Iterable[str]:
[email protected](scope="session")
+def hub_public_external_files(dataset_script_with_external_files_path) -> Iterable[str]:
+ repo_id = create_hub_dataset_repo(prefix="external_files", file_paths=[dataset_script_with_external_files_path])
+ yield repo_id
+ delete_hub_dataset_repo(repo_id=repo_id)
+
+
[email protected]
+def external_files_dataset_builder(hub_public_external_files):
+ return load_dataset_builder(hub_public_external_files)
+
+
@@ -451,0 +465,15 @@ BIG_rows = ["a" * 1_234 for _ in range(4_567)]
+TEXT_cols = {
+ "text": {"_type": "Value", "dtype": "string"},
+}
+
+TEXT_rows = [
+ {"text": text}
+ for text in [
+ "foo",
+ "bar",
+ "foobar",
+ "- Hello there !",
+ "- General Kenobi !",
+ ]
+]
+
@@ -464,0 +493 @@ def hub_datasets(
+ hub_public_external_files,
@@ -570,0 +600,8 @@ def hub_datasets(
+ "external_files": {
+ "name": hub_public_external_files,
+ "config_names_response": create_config_names_response(hub_public_external_files),
+ "split_names_response": create_split_names_response(hub_public_external_files),
+ "splits_response": create_splits_response(hub_public_external_files),
+ "first_rows_response": create_first_rows_response(hub_public_external_files, TEXT_cols, TEXT_rows),
+ "parquet_and_dataset_info_response": None,
+ },
diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index c9b735d7..e2d16573 100644
--- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Any, Iterator, List
+from typing import Any, Iterator, List, Optional
@@ -7,0 +8 @@ from typing import Any, Iterator, List
+import datasets.builder
@@ -21,0 +23,2 @@ from worker.job_runners.parquet_and_dataset_info import (
+ DatasetWithTooBigExternalFilesError,
+ DatasetWithTooManyExternalFilesError,
@@ -27,0 +31 @@ from worker.job_runners.parquet_and_dataset_info import (
+ raise_if_too_big_from_external_data_files,
@@ -223,0 +228,35 @@ def test_raise_if_too_big_from_datasets(
[email protected](
+ "max_dataset_size,max_external_data_files,error",
+ [
+ (None, None, None),
+ (10, None, DatasetWithTooBigExternalFilesError),
+ (None, 1, DatasetWithTooManyExternalFilesError),
+ ],
+)
+def test_raise_if_too_big_from_external_files(
+ external_files_dataset_builder: "datasets.builder.DatasetBuilder",
+ error,
+ max_dataset_size: Optional[int],
+ max_external_data_files: Optional[int],
+ app_config: AppConfig,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
+) -> None:
+ max_dataset_size = max_dataset_size or parquet_and_dataset_info_config.max_dataset_size
+ max_external_data_files = max_external_data_files or parquet_and_dataset_info_config.max_external_data_files
+ if error:
+ with pytest.raises(error):
+ raise_if_too_big_from_external_data_files(
+ builder=external_files_dataset_builder,
+ hf_token=app_config.common.hf_token,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ )
+ else:
+ raise_if_too_big_from_external_data_files(
+ builder=external_files_dataset_builder,
+ hf_token=app_config.common.hf_token,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ )
+
+
|
|
c7869e7417aea924c69f927437d163b305bf57aa
|
dependabot[bot]
|
49699333+dependabot[bot]@users.noreply.github.com
| 2023-02-15T14:00:51 |
chore(deps): bump starlette from 0.23.1 to 0.25.0 in /front/admin_ui (#821)
|
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index f090ec88..94eba3e0 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -458 +458 @@ name = "fastapi"
-version = "0.90.1"
+version = "0.92.0"
@@ -464,2 +464,2 @@ files = [
- {file = "fastapi-0.90.1-py3-none-any.whl", hash = "sha256:d4e4bd820204eeaa19879bf129bbce73db09bdc209d5d83a064b40b2b00557b0"},
- {file = "fastapi-0.90.1.tar.gz", hash = "sha256:d17e85deb3a350b731467e7bf035e158faffa381310a1b7356421e916fcc64f2"},
+ {file = "fastapi-0.92.0-py3-none-any.whl", hash = "sha256:ae7b97c778e2f2ec3fb3cb4fb14162129411d99907fb71920f6d69a524340ebf"},
+ {file = "fastapi-0.92.0.tar.gz", hash = "sha256:023a0f5bd2c8b2609014d3bba1e14a1d7df96c6abea0a73070621c9862b9a4de"},
@@ -470 +470 @@ pydantic = ">=1.6.2,<1.7 || >1.7,<1.7.1 || >1.7.1,<1.7.2 || >1.7.2,<1.7.3 || >1.
-starlette = ">=0.22.0,<0.24.0"
+starlette = ">=0.25.0,<0.26.0"
@@ -1270,0 +1271,7 @@ files = [
+ {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
+ {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
+ {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
+ {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
+ {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
+ {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
+ {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -1657 +1664 @@ name = "starlette"
-version = "0.23.1"
+version = "0.25.0"
@@ -1663,2 +1670,2 @@ files = [
- {file = "starlette-0.23.1-py3-none-any.whl", hash = "sha256:ec69736c90be8dbfc6ec6800ba6feb79c8c44f9b1706c0b2bb27f936bcf362cc"},
- {file = "starlette-0.23.1.tar.gz", hash = "sha256:8510e5b3d670326326c5c1d4cb657cc66832193fe5d5b7015a51c7b1e1b1bf42"},
+ {file = "starlette-0.25.0-py3-none-any.whl", hash = "sha256:774f1df1983fd594b9b6fb3ded39c2aa1979d10ac45caac0f4255cbe2acb8628"},
+ {file = "starlette-0.25.0.tar.gz", hash = "sha256:854c71e73736c429c2bdb07801f2c76c9cba497e7c3cf4988fde5e95fe4cdb3c"},
@@ -1939 +1946 @@ lock-version = "2.0"
-python-versions = "~3.9.15"
+python-versions = "~3.9"
|
f8eb4fb3f5267a34237f18d739390da5037bbb5a
|
Sylvain Lesage
| 2023-02-15T08:40:47 |
feat: 🎸 add a new admin endpoint: /dataset-status (#815)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 2f16c5c2..10c6abb3 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -586,0 +587,12 @@ class Queue:
+ def get_dataset_pending_jobs_for_type(self, dataset: str, job_type: str) -> List[JobDict]:
+ """Get the pending jobs of a dataset for a given job type.
+
+ Returns: an array of the pending jobs for the dataset and the given job type
+ """
+ return [
+ d.to_dict()
+ for d in Job.objects(
+ type=job_type, dataset=dataset, status__in=[Status.WAITING.value, Status.STARTED.value]
+ )
+ ]
+
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index b9fb1b0f..906bfdde 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -179,0 +180,15 @@ def get_response_without_content(
+def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> List[CacheEntryWithoutContent]:
+ responses = CachedResponse.objects(kind=kind, dataset=dataset).only(
+ "http_status", "error_code", "worker_version", "dataset_git_revision"
+ )
+ return [
+ {
+ "http_status": response.http_status,
+ "error_code": response.error_code,
+ "worker_version": response.worker_version,
+ "dataset_git_revision": response.dataset_git_revision,
+ }
+ for response in responses
+ ]
+
+
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 6c0244a7..291373e4 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -229,0 +230,32 @@ def test_count_by_status() -> None:
+def test_get_dataset_pending_jobs_for_type() -> None:
+ queue = Queue(max_jobs_per_namespace=100)
+ test_type = "test_type"
+ test_another_type = "test_another_type"
+ test_dataset = "test_dataset"
+ test_another_dataset = "test_another_dataset"
+ test_configs_waiting = ["test_config_waiting_1", "test_config_waiting_2"]
+ test_configs_started = ["test_config_started_1", "test_config_started_2"]
+ test_configs_finished = ["test_config_finished_1", "test_config_finished_2"]
+ for config in test_configs_finished:
+ for dataset in [test_dataset, test_another_dataset]:
+ for job_type in [test_type, test_another_type]:
+ queue.upsert_job(job_type=job_type, dataset=dataset, config=config, split=None)
+ job_info = queue.start_job()
+ queue.finish_job(job_info["job_id"], finished_status=Status.SUCCESS)
+ for config in test_configs_started:
+ for dataset in [test_dataset, test_another_dataset]:
+ for job_type in [test_type, test_another_type]:
+ queue.upsert_job(job_type=job_type, dataset=dataset, config=config, split=None)
+ job_info = queue.start_job()
+ for config in test_configs_waiting:
+ for dataset in [test_dataset, test_another_dataset]:
+ for job_type in [test_type, test_another_type]:
+ queue.upsert_job(job_type=job_type, dataset=dataset, config=config, split=None)
+ result = queue.get_dataset_pending_jobs_for_type(dataset=test_dataset, job_type=test_type)
+ assert len(result) == len(test_configs_waiting) + len(test_configs_started)
+ for r in result:
+ assert r["dataset"] == test_dataset
+ assert r["type"] == test_type
+ assert r["status"] in [Status.WAITING.value, Status.STARTED.value]
+
+
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index 2db35861..3a1c59c2 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -23,0 +24 @@ from libcommon.simple_cache import (
+ get_dataset_responses_without_content_for_kind,
@@ -454,0 +456,11 @@ def test_get_cache_reports() -> None:
+ upsert_response(
+ kind=kind_2,
+ dataset=dataset_c,
+ config=config_c,
+ split=split_c,
+ content=content_c,
+ details=details_c,
+ http_status=http_status_c,
+ error_code=error_code_c,
+ )
+
@@ -560,0 +573,13 @@ def test_get_cache_reports() -> None:
+ result_a = get_dataset_responses_without_content_for_kind(kind=kind, dataset=dataset_a)
+ assert len(result_a) == 1
+ assert result_a[0]["http_status"] == HTTPStatus.OK.value
+ assert result_a[0]["error_code"] is None
+
+ assert not get_dataset_responses_without_content_for_kind(kind=kind_2, dataset=dataset_a)
+
+ result_c = get_dataset_responses_without_content_for_kind(kind=kind_2, dataset=dataset_c)
+ assert len(result_c) == 2
+ for result in result_c:
+ assert result["http_status"] == http_status_c.value
+ assert result["error_code"] == error_code_c
+
diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py
index 513d2445..d28addc7 100644
--- a/services/admin/src/admin/app.py
+++ b/services/admin/src/admin/app.py
@@ -23,0 +24 @@ from admin.routes.cancel_jobs import create_cancel_jobs_endpoint
+from admin.routes.dataset_status import create_dataset_status_endpoint
@@ -73,0 +75,9 @@ def create_app() -> Starlette:
+ Route(
+ "/dataset-status",
+ endpoint=create_dataset_status_endpoint(
+ processing_steps=processing_steps,
+ max_age=app_config.admin.max_age,
+ external_auth_url=app_config.admin.external_auth_url,
+ organization=app_config.admin.hf_organization,
+ ),
+ ),
diff --git a/services/admin/src/admin/routes/dataset_status.py b/services/admin/src/admin/routes/dataset_status.py
new file mode 100644
index 00000000..602cb17a
--- /dev/null
+++ b/services/admin/src/admin/routes/dataset_status.py
@@ -0,0 +1,60 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+from typing import List, Optional
+
+from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import Queue
+from libcommon.simple_cache import get_dataset_responses_without_content_for_kind
+from starlette.requests import Request
+from starlette.responses import Response
+
+from admin.authentication import auth_check
+from admin.utils import (
+ AdminCustomError,
+ Endpoint,
+ MissingRequiredParameterError,
+ UnexpectedError,
+ are_valid_parameters,
+ get_json_admin_error_response,
+ get_json_ok_response,
+)
+
+
+def create_dataset_status_endpoint(
+ processing_steps: List[ProcessingStep],
+ max_age: int,
+ external_auth_url: Optional[str] = None,
+ organization: Optional[str] = None,
+) -> Endpoint:
+ async def dataset_status_endpoint(request: Request) -> Response:
+ try:
+ dataset = request.query_params.get("dataset")
+ if not are_valid_parameters([dataset]):
+ raise MissingRequiredParameterError("Parameter 'dataset' is required")
+ logging.info(f"/dataset-status, dataset={dataset}")
+
+ # if auth_check fails, it will raise an exception that will be caught below
+ auth_check(external_auth_url=external_auth_url, request=request, organization=organization)
+ queue = Queue()
+ return get_json_ok_response(
+ {
+ processing_step.endpoint: {
+ "cached_responses": get_dataset_responses_without_content_for_kind(
+ kind=processing_step.cache_kind, dataset=dataset
+ ),
+ "jobs": queue.get_dataset_pending_jobs_for_type(
+ dataset=dataset, job_type=processing_step.job_type
+ ),
+ }
+ for processing_step in processing_steps
+ },
+ max_age=max_age,
+ )
+ except AdminCustomError as e:
+ return get_json_admin_error_response(e, max_age=max_age)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age)
+
+ return dataset_status_endpoint
diff --git a/services/admin/tests/test_app.py b/services/admin/tests/test_app.py
index ca5c7e23..afee6b0a 100644
--- a/services/admin/tests/test_app.py
+++ b/services/admin/tests/test_app.py
@@ -72,0 +73,11 @@ def test_pending_jobs(client: TestClient, processing_steps: List[ProcessingStep]
+def test_dataset_status(client: TestClient, processing_steps: List[ProcessingStep]) -> None:
+ response = client.get("/dataset-status")
+ assert response.status_code == 422
+ response = client.get("/dataset-status", params={"dataset": "test-dataset"})
+ assert response.status_code == 200
+ json = response.json()
+ for processing_step in processing_steps:
+ assert not json[processing_step.job_type]["cached_responses"]
+ assert not json[processing_step.job_type]["jobs"]
+
+
|
|
276e2734f59e4f359c19a1c89c50f8243b3c24fc
|
Sylvain Lesage
| 2023-02-14T16:12:23 |
refactor: 💡 factorize the workers templates (#814)
|
diff --git a/.github/workflows/chart.yml b/.github/workflows/chart.yml
index 0876bcce..0a2ff9d0 100644
--- a/.github/workflows/chart.yml
+++ b/.github/workflows/chart.yml
@@ -23 +23 @@ jobs:
- - name: Lint chart
+ - name: Lint chart with default values
@@ -25,0 +26,6 @@ jobs:
+ - name: Lint chart with dev values
+ run: helm lint --values env/dev.yaml
+ working-directory: chart
+ - name: Lint chart with prod values
+ run: helm lint --values env/prod.yaml
+ working-directory: chart
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index eb0ccede..218b9243 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.1.1
+version: 1.2.0
diff --git a/chart/Makefile b/chart/Makefile
index 236f370f..555a89f3 100644
--- a/chart/Makefile
+++ b/chart/Makefile
@@ -45,0 +46,2 @@ quality:
+ helm lint --values env/dev.yaml
+ helm lint --values env/prod.yaml
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index 4cc899be..e7ef62d0 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -73,0 +74,3 @@ common:
+parquetAndDatasetInfo:
+ maxDatasetSize: "500_000_000"
+
@@ -157,95 +160,43 @@ api:
-# --- workers ---
-
-configNames:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
-
-splitNames:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
-
-splits:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
-
-firstRows:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
- columnsMaxNumber: 1_000
-
-parquetAndDatasetInfo:
-
- # the maximum size of the supported datasets. Bigger datasets, or datasets that cannot provide the size, are ignored.
- maxDatasetSize: "500_000_000" # support up to 500 MB
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
-
-parquet:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
-
-datasetInfo:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
-
-sizes:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
-
-genericWorker:
- replicas: 1
- resources:
- requests:
- cpu: 100m
- memory: "512Mi"
- limits:
- cpu: 1
- memory: "4Gi"
+workers:
+ -
+ deployName: "generic"
+ maxJobsPerNamespace: 1
+ workerOnlyJobTypes: []
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 100m
+ memory: "512Mi"
+ limits:
+ cpu: 1
+ memory: "4Gi"
+ tolerations: []
+ -
+ deployName: "first-rows"
+ maxJobsPerNamespace: 1
+ workerOnlyJobTypes: ["/first-rows"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 100m
+ memory: "512Mi"
+ limits:
+ cpu: 1
+ memory: "4Gi"
+ tolerations: []
+ -
+ deployName: "parquet-and-dataset-info"
+ maxJobsPerNamespace: 1
+ workerOnlyJobTypes: []
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 100m
+ memory: "512Mi"
+ limits:
+ cpu: 1
+ memory: "4Gi"
+ tolerations: []
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index d3aa45e2..75d1382a 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -63,0 +64,8 @@ common:
+firstRows:
+ maxBytes: "200_000"
+
+parquetAndDatasetInfo:
+ maxDatasetSize: "5_000_000_000"
+ blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,bigcode/the-stack,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,bigcode/the-stack-dedup,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
+
+
@@ -167,23 +175,33 @@ api:
-# --- workers ---
-
-configNames:
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
- maxJobsPerNamespace: 3
-
- nodeSelector:
- role-datasets-server: "true"
- replicas: 8
- resources:
- requests:
- cpu: 1
- memory: "1Gi"
- limits:
- cpu: 2
- memory: "30Gi"
-
-splitNames:
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
+workers:
+ -
+ deployName: "generic"
+ maxJobsPerNamespace: 4
+ workerOnlyJobTypes: []
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 20
+ resources:
+ requests:
+ cpu: 1
+ memory: "8Gi"
+ limits:
+ cpu: 2
+ memory: "30Gi"
+ tolerations: []
+ -
+ deployName: "config-names"
+ maxJobsPerNamespace: 4
+ workerOnlyJobTypes: ["/config-names"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 8
+ resources:
+ requests:
+ cpu: 1
+ memory: "1Gi"
+ limits:
+ cpu: 2
+ memory: "30Gi"
+ tolerations: []
+ -
+ deployName: "split-names"
@@ -191,16 +209,14 @@ splitNames:
-
- nodeSelector:
- role-datasets-server: "true"
- replicas: 12
- resources:
- requests:
- cpu: 1
- memory: "8Gi"
- limits:
- cpu: 2
- memory: "30Gi"
-
-splits:
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
+ workerOnlyJobTypes: ["/split-names"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 12
+ resources:
+ requests:
+ cpu: 1
+ memory: "8Gi"
+ limits:
+ cpu: 2
+ memory: "30Gi"
+ tolerations: []
+ -
+ deployName: "splits"
@@ -208,19 +224,14 @@ splits:
-
- nodeSelector:
- role-datasets-server: "true"
- replicas: 8
- resources:
- requests:
- cpu: 1
- memory: "8Gi"
- limits:
- cpu: 2
- memory: "30Gi"
-
-firstRows:
- # Max size of the /first-rows endpoint response in bytes
- maxBytes: "200_000"
-
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
+ workerOnlyJobTypes: ["/splits"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 8
+ resources:
+ requests:
+ cpu: 1
+ memory: "8Gi"
+ limits:
+ cpu: 2
+ memory: "30Gi"
+ tolerations: []
+ -
+ deployName: "first-rows"
@@ -228,25 +239,14 @@ firstRows:
-
- # Max number of columns in the /first-rows endpoint response
- columnsMaxNumber: 1_000
-
- nodeSelector:
- role-datasets-server: "true"
- replicas: 80
- resources:
- requests:
- cpu: 1
- memory: "8Gi"
- limits:
- cpu: 2
- memory: "30Gi"
-
-parquetAndDatasetInfo:
- # comma-separated list of the blocked datasets. Defaults to empty.
- blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,bigcode/the-stack,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,bigcode/the-stack-dedup,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
- # comma-separated list of the supported datasets. If empty, all the datasets are processed. Defaults to empty.
- supportedDatasets: ""
- # the maximum size of the supported datasets. Bigger datasets, or datasets that cannot provide the size, are ignored.
- maxDatasetSize: "5_000_000_000" # support up to 5 GB
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
+ workerOnlyJobTypes: ["/first-rows"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 80
+ resources:
+ requests:
+ cpu: 1
+ memory: "8Gi"
+ limits:
+ cpu: 2
+ memory: "30Gi"
+ tolerations: []
+ -
+ deployName: "parquet-and-dataset-info"
@@ -254,15 +254,14 @@ parquetAndDatasetInfo:
- nodeSelector:
- role-datasets-server: "true"
- replicas: 20
- resources:
- requests:
- cpu: 1
- memory: "8Gi"
- limits:
- cpu: 2
- memory: "30Gi"
-
-parquet:
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
+ workerOnlyJobTypes: ["/parquet-and-dataset-info"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 20
+ resources:
+ requests:
+ cpu: 1
+ memory: "8Gi"
+ limits:
+ cpu: 2
+ memory: "30Gi"
+ tolerations: []
+ -
+ deployName: "parquet"
@@ -270,15 +269,14 @@ parquet:
- nodeSelector:
- role-datasets-server: "true"
- replicas: 2
- resources:
- requests:
- cpu: 1
- memory: "100Mi"
- limits:
- cpu: 2
- memory: "1Gi"
-
-datasetInfo:
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
+ workerOnlyJobTypes: ["/parquet"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 2
+ resources:
+ requests:
+ cpu: 1
+ memory: "100Mi"
+ limits:
+ cpu: 2
+ memory: "1Gi"
+ tolerations: []
+ -
+ deployName: "dataset-info"
@@ -286,15 +284,14 @@ datasetInfo:
- nodeSelector:
- role-datasets-server: "true"
- replicas: 2
- resources:
- requests:
- cpu: 1
- memory: "100Mi"
- limits:
- cpu: 2
- memory: "1Gi"
-
-sizes:
- # override the common queue parameters
- queue:
- # Maximum number of jobs running at the same time for the same namespace
+ workerOnlyJobTypes: ["/dataset-info"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 2
+ resources:
+ requests:
+ cpu: 1
+ memory: "100Mi"
+ limits:
+ cpu: 2
+ memory: "1Gi"
+ tolerations: []
+ -
+ deployName: "sizes"
@@ -302,25 +299,12 @@ sizes:
- nodeSelector:
- role-datasets-server: "true"
- replicas: 2
- resources:
- requests:
- cpu: 1
- memory: "100Mi"
- limits:
- cpu: 2
- memory: "1Gi"
-
-genericWorker:
- queue:
- # Maximum number of jobs running at the same time for the same namespace
- maxJobsPerNamespace: 4
- nodeSelector:
- role-datasets-server: "true"
- replicas: 20
- resources:
- requests:
- cpu: 1
- memory: "8Gi"
- limits:
- cpu: 2
- memory: "30Gi"
+ workerOnlyJobTypes: ["/sizes"]
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 2
+ resources:
+ requests:
+ cpu: 1
+ memory: "100Mi"
+ limits:
+ cpu: 2
+ memory: "1Gi"
+ tolerations: []
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
index daa35eb4..22a93428 100644
--- a/chart/templates/_envWorker.tpl
+++ b/chart/templates/_envWorker.tpl
@@ -7 +6,0 @@
- # WORKER_ENDPOINT is not defined here, it's hard-coded in the template
@@ -18 +17,36 @@
- # ^ note: for datasets_based workers, the datasets cache is automatically added, so no need to add it here
+# specific to the /first-rows job runner
+- name: FIRST_ROWS_MAX_BYTES
+ value: {{ .Values.firstRows.maxBytes | quote }}
+- name: FIRST_ROWS_MAX_NUMBER
+ value: {{ .Values.firstRows.maxNumber | quote }}
+- name: FIRST_ROWS_MIN_CELL_BYTES
+ value: {{ .Values.firstRows.minCellBytes | quote }}
+- name: FIRST_ROWS_MIN_NUMBER
+ value: {{ .Values.firstRows.minNumber| quote }}
+- name: FIRST_ROWS_COLUMNS_MAX_NUMBER
+ value: {{ .Values.firstRows.columnsMaxNumber| quote }}
+# specific to the /parquet-and-dataset-info job runner
+- name: PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS
+ value: {{ .Values.parquetAndDatasetInfo.blockedDatasets | quote }}
+- name: PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE
+ value: {{ .Values.parquetAndDatasetInfo.commitMessage | quote }}
+- name: PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN
+ {{- if .Values.secrets.userHfToken.fromSecret }}
+ valueFrom:
+ secretKeyRef:
+ name: {{ .Values.secrets.userHfToken.secretName | quote }}
+ key: HF_TOKEN
+ optional: false
+ {{- else }}
+ value: {{ .Values.secrets.userHfToken.value }}
+ {{- end }}
+- name: PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE
+ value: {{ .Values.parquetAndDatasetInfo.maxDatasetSize | quote }}
+- name: PARQUET_AND_DATASET_INFO_SOURCE_REVISION
+ value: {{ .Values.parquetAndDatasetInfo.sourceRevision | quote }}
+- name: PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS
+ value: {{ .Values.parquetAndDatasetInfo.supportedDatasets | quote }}
+- name: PARQUET_AND_DATASET_INFO_TARGET_REVISION
+ value: {{ .Values.parquetAndDatasetInfo.targetRevision | quote }}
+- name: PARQUET_AND_DATASET_INFO_URL_TEMPLATE
+ value: {{ .Values.parquetAndDatasetInfo.urlTemplate | quote }}
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index 3f899e4b..15e997f2 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -129 +129 @@ app.kubernetes.io/component: "{{ include "name" . }}-api"
-{{- define "labels.configNames" -}}
+{{- define "labels.worker" -}}
@@ -131,41 +131 @@ app.kubernetes.io/component: "{{ include "name" . }}-api"
-app.kubernetes.io/component: "{{ include "name" . }}-worker-config-names"
-{{- end -}}
-
-{{- define "labels.splitNames" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-split-names"
-{{- end -}}
-
-{{- define "labels.splits" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-splits"
-{{- end -}}
-
-{{- define "labels.firstRows" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-first-rows"
-{{- end -}}
-
-{{- define "labels.parquetAndDatasetInfo" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-parquet-and-dataset-info"
-{{- end -}}
-
-{{- define "labels.parquet" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-parquet"
-{{- end -}}
-
-{{- define "labels.datasetInfo" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-dataset-info"
-{{- end -}}
-
-{{- define "labels.sizes" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-sizes"
-{{- end -}}
-
-{{- define "labels.genericWorker" -}}
-{{ include "datasetServer.labels" . }}
-app.kubernetes.io/component: "{{ include "name" . }}-worker-generic"
+app.kubernetes.io/component: "{{ include "name" . }}-worker"
diff --git a/chart/templates/worker/sizes/_container.tpl b/chart/templates/worker/_container.tpl
similarity index 64%
rename from chart/templates/worker/sizes/_container.tpl
rename to chart/templates/worker/_container.tpl
index 87260406..0df0ade7 100644
--- a/chart/templates/worker/sizes/_container.tpl
+++ b/chart/templates/worker/_container.tpl
@@ -4,2 +4,2 @@
-{{- define "containerWorkerSizes" -}}
-- name: "{{ include "name" . }}-worker-sizes"
+{{- define "containerWorker" -}}
+- name: "{{ include "name" . }}-worker"
@@ -9,3 +8,0 @@
- - name: WORKER_ONLY_JOB_TYPES
- value: "/sizes"
- # ^ hard-coded
@@ -19 +16 @@
- value: {{ printf "%s/sizes/datasets" .Values.cacheDirectory | quote }}
+ value: {{ printf "%s/%s/datasets" .Values.cacheDirectory .workerValues.deployName | quote }}
@@ -21,3 +18,3 @@
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.sizes.queue.maxJobsPerNamespace | quote }}
+ value: {{ .workerValues.maxJobsPerNamespace | quote }}
+ - name: WORKER_ONLY_JOB_TYPES
+ value: {{ .workerValues.workerOnlyJobTypes | quote }}
@@ -29 +26 @@
- resources: {{ toYaml .Values.sizes.resources | nindent 4 }}
+ resources: {{ toYaml .workerValues.resources | nindent 4 }}
diff --git a/chart/templates/worker/sizes/deployment.yaml b/chart/templates/worker/_deployment.yaml
similarity index 55%
rename from chart/templates/worker/sizes/deployment.yaml
rename to chart/templates/worker/_deployment.yaml
index d1417f1d..b5443536 100644
--- a/chart/templates/worker/sizes/deployment.yaml
+++ b/chart/templates/worker/_deployment.yaml
@@ -3,0 +4,2 @@
+{{- define "deploymentWorker" -}}
+---
@@ -7,2 +9,2 @@ metadata:
- labels: {{ include "labels.sizes" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-sizes"
+ labels: {{ include "labels.worker" . | nindent 4 }}
+ name: "{{ include "name" . }}-worker-{{ .workerValues.deployName }}"
@@ -12 +14 @@ spec:
- replicas: {{ .Values.sizes.replicas }}
+ replicas: {{ .workerValues.replicas }}
@@ -15 +17 @@ spec:
- matchLabels: {{ include "labels.sizes" . | nindent 6 }}
+ matchLabels: {{ include "labels.worker" . | nindent 6 }}
@@ -20 +22 @@ spec:
- labels: {{ include "labels.sizes" . | nindent 8 }}
+ labels: {{ include "labels.worker" . | nindent 8 }}
@@ -26,3 +28,3 @@ spec:
- containers: {{ include "containerWorkerSizes" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.sizes.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.sizes.tolerations | nindent 8 }}
+ containers: {{ include "containerWorker" . | nindent 8 }}
+ nodeSelector: {{ toYaml .workerValues.nodeSelector | nindent 8 }}
+ tolerations: {{ toYaml .workerValues.tolerations | nindent 8 }}
@@ -30,0 +33 @@ spec:
+{{- end -}}
diff --git a/chart/templates/worker/config-names/_container.tpl b/chart/templates/worker/config-names/_container.tpl
deleted file mode 100644
index 6a6fb063..00000000
--- a/chart/templates/worker/config-names/_container.tpl
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerConfigNames" -}}
-- name: "{{ include "name" . }}-worker-config-names"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- - name: WORKER_ONLY_JOB_TYPES
- value: "/config-names"
- # ^ hard-coded
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/config-names/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.configNames.queue.maxJobsPerNamespace | quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.configNames.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/config-names/deployment.yaml b/chart/templates/worker/config-names/deployment.yaml
deleted file mode 100644
index 72c63c12..00000000
--- a/chart/templates/worker/config-names/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.configNames" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-config-names"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.configNames.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.configNames" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.configNames" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerConfigNames" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.configNames.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.configNames.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/dataset-info/_container.tpl b/chart/templates/worker/dataset-info/_container.tpl
deleted file mode 100644
index f8eac5ac..00000000
--- a/chart/templates/worker/dataset-info/_container.tpl
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerDatasetInfo" -}}
-- name: "{{ include "name" . }}-worker-dataset-info"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- - name: WORKER_ONLY_JOB_TYPES
- value: "/dataset-info"
- # ^ hard-coded
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/datasets-info/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.datasetInfo.queue.maxJobsPerNamespace | quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.datasetInfo.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/dataset-info/deployment.yaml b/chart/templates/worker/dataset-info/deployment.yaml
deleted file mode 100644
index 753f1143..00000000
--- a/chart/templates/worker/dataset-info/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.datasetInfo" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-dataset-info"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.datasetInfo.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.datasetInfo" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.datasetInfo" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerDatasetInfo" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.datasetInfo.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.datasetInfo.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/deployment.yaml b/chart/templates/worker/deployment.yaml
new file mode 100644
index 00000000..56eb3a38
--- /dev/null
+++ b/chart/templates/worker/deployment.yaml
@@ -0,0 +1,6 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+{{- range $index, $workerValues := .Values.workers }}
+{{ include "deploymentWorker" (merge (dict "workerValues" $workerValues) $ ) }}
+{{- end }}
diff --git a/chart/templates/worker/first-rows/_container.tpl b/chart/templates/worker/first-rows/_container.tpl
deleted file mode 100644
index 41d44fc8..00000000
--- a/chart/templates/worker/first-rows/_container.tpl
+++ /dev/null
@@ -1,40 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerFirstRows" -}}
-- name: "{{ include "name" . }}-worker-first-rows"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- - name: WORKER_ONLY_JOB_TYPES
- value: "/first-rows"
- # ^ hard-coded
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/first-rows/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.firstRows.queue.maxJobsPerNamespace | quote }}
- - name: FIRST_ROWS_MAX_BYTES
- value: {{ .Values.firstRows.maxBytes | quote }}
- - name: FIRST_ROWS_MAX_NUMBER
- value: {{ .Values.firstRows.maxNumber | quote }}
- - name: FIRST_ROWS_MIN_CELL_BYTES
- value: {{ .Values.firstRows.minCellBytes | quote }}
- - name: FIRST_ROWS_MIN_NUMBER
- value: {{ .Values.firstRows.minNumber| quote }}
- - name: FIRST_ROWS_COLUMNS_MAX_NUMBER
- value: {{ .Values.firstRows.columnsMaxNumber| quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.firstRows.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/first-rows/deployment.yaml b/chart/templates/worker/first-rows/deployment.yaml
deleted file mode 100644
index ba4a72f9..00000000
--- a/chart/templates/worker/first-rows/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.firstRows" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-first-rows"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.firstRows.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.firstRows" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.firstRows" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerFirstRows" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.firstRows.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.firstRows.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/generic/_container.tpl b/chart/templates/worker/generic/_container.tpl
deleted file mode 100644
index 15ec95ae..00000000
--- a/chart/templates/worker/generic/_container.tpl
+++ /dev/null
@@ -1,61 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerGeneric" -}}
-- name: "{{ include "name" . }}-worker-generic"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/generic/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.genericWorker.queue.maxJobsPerNamespace | quote }}
- - name: FIRST_ROWS_MAX_BYTES
- value: {{ .Values.firstRows.maxBytes | quote }}
- - name: FIRST_ROWS_MAX_NUMBER
- value: {{ .Values.firstRows.maxNumber | quote }}
- - name: FIRST_ROWS_MIN_CELL_BYTES
- value: {{ .Values.firstRows.minCellBytes | quote }}
- - name: FIRST_ROWS_MIN_NUMBER
- value: {{ .Values.firstRows.minNumber| quote }}
- - name: FIRST_ROWS_COLUMNS_MAX_NUMBER
- value: {{ .Values.firstRows.columnsMaxNumber| quote }}
- - name: PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS
- value: {{ .Values.parquetAndDatasetInfo.blockedDatasets | quote }}
- - name: PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE
- value: {{ .Values.parquetAndDatasetInfo.commitMessage | quote }}
- - name: PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN
- {{- if .Values.secrets.userHfToken.fromSecret }}
- valueFrom:
- secretKeyRef:
- name: {{ .Values.secrets.userHfToken.secretName | quote }}
- key: HF_TOKEN
- optional: false
- {{- else }}
- value: {{ .Values.secrets.userHfToken.value }}
- {{- end }}
- - name: PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE
- value: {{ .Values.parquetAndDatasetInfo.maxDatasetSize | quote }}
- - name: PARQUET_AND_DATASET_INFO_SOURCE_REVISION
- value: {{ .Values.parquetAndDatasetInfo.sourceRevision | quote }}
- - name: PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS
- value: {{ .Values.parquetAndDatasetInfo.supportedDatasets | quote }}
- - name: PARQUET_AND_DATASET_INFO_TARGET_REVISION
- value: {{ .Values.parquetAndDatasetInfo.targetRevision | quote }}
- - name: PARQUET_AND_DATASET_INFO_URL_TEMPLATE
- value: {{ .Values.parquetAndDatasetInfo.urlTemplate | quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.genericWorker.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/generic/deployment.yaml b/chart/templates/worker/generic/deployment.yaml
deleted file mode 100644
index cab65da2..00000000
--- a/chart/templates/worker/generic/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.genericWorker" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-generic"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.genericWorker.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.genericWorker" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.genericWorker" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerGeneric" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.genericWorker.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.genericWorker.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/parquet-and-dataset-info/_container.tpl b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
deleted file mode 100644
index 92e40ea7..00000000
--- a/chart/templates/worker/parquet-and-dataset-info/_container.tpl
+++ /dev/null
@@ -1,54 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerParquetAndDatasetInfo" -}}
-- name: "{{ include "name" . }}-worker-parquet-and-dataset-info"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- - name: WORKER_ONLY_JOB_TYPES
- value: "/parquet-and-dataset-info"
- # ^ hard-coded
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/parquet-and-dataset-info/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.parquetAndDatasetInfo.queue.maxJobsPerNamespace | quote }}
- - name: PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS
- value: {{ .Values.parquetAndDatasetInfo.blockedDatasets | quote }}
- - name: PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE
- value: {{ .Values.parquetAndDatasetInfo.commitMessage | quote }}
- - name: PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN
- {{- if .Values.secrets.userHfToken.fromSecret }}
- valueFrom:
- secretKeyRef:
- name: {{ .Values.secrets.userHfToken.secretName | quote }}
- key: HF_TOKEN
- optional: false
- {{- else }}
- value: {{ .Values.secrets.userHfToken.value }}
- {{- end }}
- - name: PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE
- value: {{ .Values.parquetAndDatasetInfo.maxDatasetSize | quote }}
- - name: PARQUET_AND_DATASET_INFO_SOURCE_REVISION
- value: {{ .Values.parquetAndDatasetInfo.sourceRevision | quote }}
- - name: PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS
- value: {{ .Values.parquetAndDatasetInfo.supportedDatasets | quote }}
- - name: PARQUET_AND_DATASET_INFO_TARGET_REVISION
- value: {{ .Values.parquetAndDatasetInfo.targetRevision | quote }}
- - name: PARQUET_AND_DATASET_INFO_URL_TEMPLATE
- value: {{ .Values.parquetAndDatasetInfo.urlTemplate | quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.parquetAndDatasetInfo.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/parquet-and-dataset-info/deployment.yaml b/chart/templates/worker/parquet-and-dataset-info/deployment.yaml
deleted file mode 100644
index 671a21d5..00000000
--- a/chart/templates/worker/parquet-and-dataset-info/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.parquetAndDatasetInfo" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-parquet-and-dataset-info"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.parquetAndDatasetInfo.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.parquetAndDatasetInfo" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.parquetAndDatasetInfo" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerParquetAndDatasetInfo" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.parquetAndDatasetInfo.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.parquetAndDatasetInfo.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/parquet/_container.tpl b/chart/templates/worker/parquet/_container.tpl
deleted file mode 100644
index 16cb5e22..00000000
--- a/chart/templates/worker/parquet/_container.tpl
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerParquet" -}}
-- name: "{{ include "name" . }}-worker-parquet"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- - name: WORKER_ONLY_JOB_TYPES
- value: "/parquet"
- # ^ hard-coded
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/parquet/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.parquet.queue.maxJobsPerNamespace | quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.parquet.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/parquet/deployment.yaml b/chart/templates/worker/parquet/deployment.yaml
deleted file mode 100644
index 50e3bab9..00000000
--- a/chart/templates/worker/parquet/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.parquet" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-parquet"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.parquet.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.parquet" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.parquet" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerParquet" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.parquet.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.parquet.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/split-names/_container.tpl b/chart/templates/worker/split-names/_container.tpl
deleted file mode 100644
index f1f9a353..00000000
--- a/chart/templates/worker/split-names/_container.tpl
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerSplitNames" -}}
-- name: "{{ include "name" . }}-worker-split-names"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- - name: WORKER_ONLY_JOB_TYPES
- value: "/split-names"
- # ^ hard-coded
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/split-names/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.splitNames.queue.maxJobsPerNamespace | quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.splitNames.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/split-names/deployment.yaml b/chart/templates/worker/split-names/deployment.yaml
deleted file mode 100644
index df03783b..00000000
--- a/chart/templates/worker/split-names/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.splitNames" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-split-names"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.splitNames.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.splitNames" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.splitNames" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerSplitNames" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.splitNames.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.splitNames.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/splits/_container.tpl b/chart/templates/worker/splits/_container.tpl
deleted file mode 100644
index 0d95479a..00000000
--- a/chart/templates/worker/splits/_container.tpl
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "containerWorkerSplits" -}}
-- name: "{{ include "name" . }}-worker-splits"
- image: {{ include "services.worker.image" . }}
- imagePullPolicy: {{ .Values.images.pullPolicy }}
- env:
- - name: WORKER_ONLY_JOB_TYPES
- value: "/splits"
- # ^ hard-coded
- {{ include "envAssets" . | nindent 2 }}
- {{ include "envCache" . | nindent 2 }}
- {{ include "envQueue" . | nindent 2 }}
- {{ include "envCommon" . | nindent 2 }}
- {{ include "envWorker" . | nindent 2 }}
- {{ include "envDatasetsBased" . | nindent 2 }}
- - name: DATASETS_BASED_HF_DATASETS_CACHE
- value: {{ printf "%s/splits/datasets" .Values.cacheDirectory | quote }}
- - name: QUEUE_MAX_JOBS_PER_NAMESPACE
- # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
- # overridden
- value: {{ .Values.splits.queue.maxJobsPerNamespace | quote }}
- volumeMounts:
- {{ include "volumeMountAssetsRW" . | nindent 2 }}
- {{ include "volumeMountCache" . | nindent 2 }}
- securityContext:
- allowPrivilegeEscalation: false
- resources: {{ toYaml .Values.splits.resources | nindent 4 }}
-{{- end -}}
diff --git a/chart/templates/worker/splits/deployment.yaml b/chart/templates/worker/splits/deployment.yaml
deleted file mode 100644
index cd1872a9..00000000
--- a/chart/templates/worker/splits/deployment.yaml
+++ /dev/null
@@ -1,30 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
- labels: {{ include "labels.splits" . | nindent 4 }}
- name: "{{ include "name" . }}-worker-splits"
- namespace: {{ .Release.Namespace }}
-spec:
- progressDeadlineSeconds: 600
- replicas: {{ .Values.splits.replicas }}
- revisionHistoryLimit: 10
- selector:
- matchLabels: {{ include "labels.splits" . | nindent 6 }}
- strategy:
- type: Recreate
- template:
- metadata:
- labels: {{ include "labels.splits" . | nindent 8 }}
- spec:
- {{- include "image.imagePullSecrets" . | nindent 6 }}
- initContainers:
- {{ include "initContainerAssets" . | nindent 8 }}
- {{ include "initContainerCache" . | nindent 8 }}
- containers: {{ include "containerWorkerSplits" . | nindent 8 }}
- nodeSelector: {{ toYaml .Values.splits.nodeSelector | nindent 8 }}
- tolerations: {{ toYaml .Values.splits.tolerations | nindent 8 }}
- volumes: {{ include "volumeData" . | nindent 8 }}
- securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 02588117..4de55d6d 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -108,0 +109,28 @@ worker:
+firstRows:
+ # Max size of the /first-rows endpoint response in bytes
+ maxBytes: "1_000_000"
+ # Max number of rows in the /first-rows endpoint response
+ maxNumber: 100
+ # Min size of a cell in the /first-rows endpoint response in bytes
+ minCellBytes: 100
+ # Min number of rows in the /first-rows endpoint response
+ minNumber: 10
+ # Max number of columns in the /first-rows endpoint response
+ columnsMaxNumber: 1_000
+
+parquetAndDatasetInfo:
+ # comma-separated list of the blocked datasets. Defaults to empty.
+ blockedDatasets: ""
+ # the git commit message when the parquet files are uploaded to the Hub. Defaults to `Update parquet files`.
+ commitMessage: "Update parquet files"
+ # the maximum size of the supported datasets. Bigger datasets, or datasets that cannot provide the size, are ignored.
+ maxDatasetSize: "100_000_000"
+ # the git revision of the dataset to use to prepare the parquet files. Defaults to `main`.
+ sourceRevision: "main"
+ # comma-separated list of the supported datasets. If empty, all the datasets are processed. Defaults to empty.
+ supportedDatasets: ""
+ # the git revision of the dataset where to store the parquet files. Make sure the hf_token (see the "Common" section) allows to write there. Defaults to `refs/convert/parquet`.
+ targetRevision: "refs/convert/parquet"
+ # the URL template to build the parquet file URLs. Defaults to `/datasets/%s/resolve/%s/%s`.
+ urlTemplate: "/datasets/%s/resolve/%s/%s"
+
@@ -239,5 +267,4 @@ api:
-# --- workers ---
-
-configNames:
- # override the common queue parameters
- queue:
+workers:
+ -
+ # name of the deployment
+ deployName: "generic"
@@ -246,13 +273,13 @@ configNames:
-
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 1
- limits:
- cpu: 1
- tolerations: []
-
-splitNames:
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: []
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "config-names"
@@ -261,13 +288,13 @@ splitNames:
-
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 1
- limits:
- cpu: 1
- tolerations: []
-
-splits:
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/config-names"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 1
+ limits:
+ cpu: 1
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "split-names"
@@ -276,21 +303,13 @@ splits:
-
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 0
- limits:
- cpu: 0
- tolerations: []
-
-firstRows:
- # Max size of the /first-rows endpoint response in bytes
- maxBytes: "1_000_000"
- # Max number of rows in the /first-rows endpoint response
- maxNumber: 100
- # Min size of a cell in the /first-rows endpoint response in bytes
- minCellBytes: 100
- # Min number of rows in the /first-rows endpoint response
- minNumber: 10
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/split-names"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 1
+ limits:
+ cpu: 1
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "splits"
@@ -299,30 +318,13 @@ firstRows:
- # Max number of columns in the /first-rows endpoint response
- columnsMaxNumber: 1_000
-
-
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 0
- limits:
- cpu: 0
- tolerations: []
-
-parquetAndDatasetInfo:
- # comma-separated list of the blocked datasets. Defaults to empty.
- blockedDatasets: ""
- # the git commit message when the parquet files are uploaded to the Hub. Defaults to `Update parquet files`.
- commitMessage: "Update parquet files"
- # the maximum size of the supported datasets. Bigger datasets, or datasets that cannot provide the size, are ignored.
- maxDatasetSize: "100_000_000"
- # the git revision of the dataset to use to prepare the parquet files. Defaults to `main`.
- sourceRevision: "main"
- # comma-separated list of the supported datasets. If empty, all the datasets are processed. Defaults to empty.
- supportedDatasets: ""
- # the git revision of the dataset where to store the parquet files. Make sure the hf_token (see the "Common" section) allows to write there. Defaults to `refs/convert/parquet`.
- targetRevision: "refs/convert/parquet"
- # the URL template to build the parquet file URLs. Defaults to `/datasets/%s/resolve/%s/%s`.
- urlTemplate: "/datasets/%s/resolve/%s/%s"
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/splits"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "first-rows"
@@ -331,12 +333,13 @@ parquetAndDatasetInfo:
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 0
- limits:
- cpu: 0
- tolerations: []
-
-parquet:
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/first-rows"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "parquet-and-dataset-info"
@@ -345,12 +348,13 @@ parquet:
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 0
- limits:
- cpu: 0
- tolerations: []
-
-datasetInfo:
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/parquet-and-dataset-info"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "parquet"
@@ -359,12 +363,13 @@ datasetInfo:
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 0
- limits:
- cpu: 0
- tolerations: []
-
-sizes:
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/parquet"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "dataset-info"
@@ -373,12 +378,13 @@ sizes:
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 0
- limits:
- cpu: 0
- tolerations: []
-
-genericWorker:
- # override the common queue parameters
- queue:
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/dataset-info"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
+ -
+ # name of the deployment
+ deployName: "sizes"
@@ -387,8 +393,10 @@ genericWorker:
- nodeSelector: {}
- replicas: 1
- resources:
- requests:
- cpu: 0
- limits:
- cpu: 0
- tolerations: []
+ # job types that this worker can process
+ workerOnlyJobTypes: ["/sizes"]
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
|
|
ded9a8ca8a1ec7be2d271078397c5abb393ffd83
|
Sylvain Lesage
| 2023-02-14T09:52:09 |
fix: 🐛 add missing config (#813)
|
diff --git a/chart/templates/worker/config-names/_container.tpl b/chart/templates/worker/config-names/_container.tpl
index 1f2a97fe..6a6fb063 100644
--- a/chart/templates/worker/config-names/_container.tpl
+++ b/chart/templates/worker/config-names/_container.tpl
@@ -11,0 +12 @@
+ {{ include "envAssets" . | nindent 2 }}
diff --git a/chart/templates/worker/dataset-info/_container.tpl b/chart/templates/worker/dataset-info/_container.tpl
index ec68373d..f8eac5ac 100644
--- a/chart/templates/worker/dataset-info/_container.tpl
+++ b/chart/templates/worker/dataset-info/_container.tpl
@@ -11,0 +12 @@
+ {{ include "envAssets" . | nindent 2 }}
diff --git a/chart/templates/worker/parquet-and-dataset-info/_container.tpl b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
index d492d9d8..92e40ea7 100644
--- a/chart/templates/worker/parquet-and-dataset-info/_container.tpl
+++ b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
@@ -11,0 +12 @@
+ {{ include "envAssets" . | nindent 2 }}
diff --git a/chart/templates/worker/parquet/_container.tpl b/chart/templates/worker/parquet/_container.tpl
index 74baed1c..16cb5e22 100644
--- a/chart/templates/worker/parquet/_container.tpl
+++ b/chart/templates/worker/parquet/_container.tpl
@@ -11,0 +12 @@
+ {{ include "envAssets" . | nindent 2 }}
diff --git a/chart/templates/worker/sizes/_container.tpl b/chart/templates/worker/sizes/_container.tpl
index 21b92b7c..87260406 100644
--- a/chart/templates/worker/sizes/_container.tpl
+++ b/chart/templates/worker/sizes/_container.tpl
@@ -11,0 +12 @@
+ {{ include "envAssets" . | nindent 2 }}
diff --git a/chart/templates/worker/split-names/_container.tpl b/chart/templates/worker/split-names/_container.tpl
index 07b29ab6..f1f9a353 100644
--- a/chart/templates/worker/split-names/_container.tpl
+++ b/chart/templates/worker/split-names/_container.tpl
@@ -11,0 +12 @@
+ {{ include "envAssets" . | nindent 2 }}
diff --git a/chart/templates/worker/splits/_container.tpl b/chart/templates/worker/splits/_container.tpl
index e8cb31c5..0d95479a 100644
--- a/chart/templates/worker/splits/_container.tpl
+++ b/chart/templates/worker/splits/_container.tpl
@@ -11,0 +12 @@
+ {{ include "envAssets" . | nindent 2 }}
|
|
03eb5738e63385a95020a4dd21a76f6607771591
|
Sylvain Lesage
| 2023-02-14T09:22:02 |
fix: 🐛 add missing volumes (#812)
|
diff --git a/chart/templates/worker/dataset-info/deployment.yaml b/chart/templates/worker/dataset-info/deployment.yaml
index 62130353..753f1143 100644
--- a/chart/templates/worker/dataset-info/deployment.yaml
+++ b/chart/templates/worker/dataset-info/deployment.yaml
@@ -28,0 +29 @@ spec:
+ volumes: {{ include "volumeData" . | nindent 8 }}
diff --git a/chart/templates/worker/parquet/deployment.yaml b/chart/templates/worker/parquet/deployment.yaml
index bd9fba65..50e3bab9 100644
--- a/chart/templates/worker/parquet/deployment.yaml
+++ b/chart/templates/worker/parquet/deployment.yaml
@@ -28,0 +29 @@ spec:
+ volumes: {{ include "volumeData" . | nindent 8 }}
diff --git a/chart/templates/worker/sizes/deployment.yaml b/chart/templates/worker/sizes/deployment.yaml
index d5b0d473..d1417f1d 100644
--- a/chart/templates/worker/sizes/deployment.yaml
+++ b/chart/templates/worker/sizes/deployment.yaml
@@ -28,0 +29 @@ spec:
+ volumes: {{ include "volumeData" . | nindent 8 }}
|
|
70b90f49527006b1c0637e44263b2dcc79bae0c1
|
Sylvain Lesage
| 2023-02-14T07:54:56 |
fix: 🐛 ensure all the workers have the same access to the disk (#811)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 9b04759e..eb0ccede 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.1.0
+version: 1.1.1
diff --git a/chart/templates/worker/dataset-info/_container.tpl b/chart/templates/worker/dataset-info/_container.tpl
index 02653524..ec68373d 100644
--- a/chart/templates/worker/dataset-info/_container.tpl
+++ b/chart/templates/worker/dataset-info/_container.tpl
@@ -15,0 +16,3 @@
+ {{ include "envDatasetsBased" . | nindent 2 }}
+ - name: DATASETS_BASED_HF_DATASETS_CACHE
+ value: {{ printf "%s/datasets-info/datasets" .Values.cacheDirectory | quote }}
@@ -21,0 +25 @@
+ {{ include "volumeMountCache" . | nindent 2 }}
diff --git a/chart/templates/worker/dataset-info/deployment.yaml b/chart/templates/worker/dataset-info/deployment.yaml
index f17170e1..62130353 100644
--- a/chart/templates/worker/dataset-info/deployment.yaml
+++ b/chart/templates/worker/dataset-info/deployment.yaml
@@ -24,0 +25 @@ spec:
+ {{ include "initContainerCache" . | nindent 8 }}
diff --git a/chart/templates/worker/parquet/_container.tpl b/chart/templates/worker/parquet/_container.tpl
index 06be53fb..74baed1c 100644
--- a/chart/templates/worker/parquet/_container.tpl
+++ b/chart/templates/worker/parquet/_container.tpl
@@ -15,0 +16,3 @@
+ {{ include "envDatasetsBased" . | nindent 2 }}
+ - name: DATASETS_BASED_HF_DATASETS_CACHE
+ value: {{ printf "%s/parquet/datasets" .Values.cacheDirectory | quote }}
@@ -21,0 +25 @@
+ {{ include "volumeMountCache" . | nindent 2 }}
diff --git a/chart/templates/worker/parquet/deployment.yaml b/chart/templates/worker/parquet/deployment.yaml
index 126e8fc6..bd9fba65 100644
--- a/chart/templates/worker/parquet/deployment.yaml
+++ b/chart/templates/worker/parquet/deployment.yaml
@@ -24,0 +25 @@ spec:
+ {{ include "initContainerCache" . | nindent 8 }}
diff --git a/chart/templates/worker/sizes/_container.tpl b/chart/templates/worker/sizes/_container.tpl
index 7146122f..21b92b7c 100644
--- a/chart/templates/worker/sizes/_container.tpl
+++ b/chart/templates/worker/sizes/_container.tpl
@@ -15,0 +16,3 @@
+ {{ include "envDatasetsBased" . | nindent 2 }}
+ - name: DATASETS_BASED_HF_DATASETS_CACHE
+ value: {{ printf "%s/sizes/datasets" .Values.cacheDirectory | quote }}
@@ -21,0 +25 @@
+ {{ include "volumeMountCache" . | nindent 2 }}
diff --git a/chart/templates/worker/sizes/deployment.yaml b/chart/templates/worker/sizes/deployment.yaml
index 6dbd20b7..d5b0d473 100644
--- a/chart/templates/worker/sizes/deployment.yaml
+++ b/chart/templates/worker/sizes/deployment.yaml
@@ -24,0 +25 @@ spec:
+ {{ include "initContainerCache" . | nindent 8 }}
|
|
760538a4d82049208ee12467339d35abbf28809a
|
Sylvain Lesage
| 2023-02-13T18:16:07 |
Update chart (#808)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 2b4e2f76..9b04759e 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21,3 +21 @@ type: application
-#
-# We don't increment it, as it will not be distributed.
-version: 1.0.0
+version: 1.1.0
@@ -31 +29 @@ version: 1.0.0
-appVersion: "0.20.2"
+appVersion: "0.21.0"
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index 46a976ca..4cc899be 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -29 +29 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -35 +35 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -40 +40 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -45 +45 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 72ab7c2d..d3aa45e2 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -19 +19 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -25 +25 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -30 +30 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -35 +35 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
diff --git a/chart/values.yaml b/chart/values.yaml
index 23ecce17..02588117 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -26 +26 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -32 +32 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -37 +37 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
@@ -42 +42 @@ images:
- tag: sha-27ad2f7
+ tag: sha-fb3399a
|
|
1621a71f4b677e215529ce4a622f9a7894d408aa
|
Sylvain Lesage
| 2023-02-13T18:15:34 |
chore: 🤖 add VERSION file (#807)
|
diff --git a/VERSION b/VERSION
new file mode 100644
index 00000000..88541566
--- /dev/null
+++ b/VERSION
@@ -0,0 +1 @@
+0.21.0
|
|
fb3399aa3a96abe0f46b1c6a9ee8ac9f3358f394
|
Sylvain Lesage
| 2023-02-13T15:29:23 |
Generic worker (#802)
|
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index 6f1a775d..46a976ca 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -241,0 +242,10 @@ sizes:
+
+genericWorker:
+ replicas: 1
+ resources:
+ requests:
+ cpu: 100m
+ memory: "512Mi"
+ limits:
+ cpu: 1
+ memory: "4Gi"
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 8b324fd2..72ab7c2d 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -234 +234 @@ firstRows:
- replicas: 90
+ replicas: 80
@@ -256 +256 @@ parquetAndDatasetInfo:
- replicas: 30
+ replicas: 20
@@ -311,0 +312,15 @@ sizes:
+
+genericWorker:
+ queue:
+ # Maximum number of jobs running at the same time for the same namespace
+ maxJobsPerNamespace: 4
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 20
+ resources:
+ requests:
+ cpu: 1
+ memory: "8Gi"
+ limits:
+ cpu: 2
+ memory: "30Gi"
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index 3ad176a4..3f899e4b 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -168,0 +169,5 @@ app.kubernetes.io/component: "{{ include "name" . }}-worker-sizes"
+{{- define "labels.genericWorker" -}}
+{{ include "datasetServer.labels" . }}
+app.kubernetes.io/component: "{{ include "name" . }}-worker-generic"
+{{- end -}}
+
diff --git a/chart/templates/worker/config-names/_container.tpl b/chart/templates/worker/config-names/_container.tpl
index def7afb6..1f2a97fe 100644
--- a/chart/templates/worker/config-names/_container.tpl
+++ b/chart/templates/worker/config-names/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/templates/worker/dataset-info/_container.tpl b/chart/templates/worker/dataset-info/_container.tpl
index 36283497..02653524 100644
--- a/chart/templates/worker/dataset-info/_container.tpl
+++ b/chart/templates/worker/dataset-info/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/templates/worker/first-rows/_container.tpl b/chart/templates/worker/first-rows/_container.tpl
index 884b136f..41d44fc8 100644
--- a/chart/templates/worker/first-rows/_container.tpl
+++ b/chart/templates/worker/first-rows/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/templates/worker/generic/_container.tpl b/chart/templates/worker/generic/_container.tpl
new file mode 100644
index 00000000..15ec95ae
--- /dev/null
+++ b/chart/templates/worker/generic/_container.tpl
@@ -0,0 +1,61 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+{{- define "containerWorkerGeneric" -}}
+- name: "{{ include "name" . }}-worker-generic"
+ image: {{ include "services.worker.image" . }}
+ imagePullPolicy: {{ .Values.images.pullPolicy }}
+ env:
+ {{ include "envAssets" . | nindent 2 }}
+ {{ include "envCache" . | nindent 2 }}
+ {{ include "envQueue" . | nindent 2 }}
+ {{ include "envCommon" . | nindent 2 }}
+ {{ include "envWorker" . | nindent 2 }}
+ {{ include "envDatasetsBased" . | nindent 2 }}
+ - name: DATASETS_BASED_HF_DATASETS_CACHE
+ value: {{ printf "%s/generic/datasets" .Values.cacheDirectory | quote }}
+ - name: QUEUE_MAX_JOBS_PER_NAMESPACE
+ # value: {{ .Values.queue.maxJobsPerNamespace | quote }}
+ # overridden
+ value: {{ .Values.genericWorker.queue.maxJobsPerNamespace | quote }}
+ - name: FIRST_ROWS_MAX_BYTES
+ value: {{ .Values.firstRows.maxBytes | quote }}
+ - name: FIRST_ROWS_MAX_NUMBER
+ value: {{ .Values.firstRows.maxNumber | quote }}
+ - name: FIRST_ROWS_MIN_CELL_BYTES
+ value: {{ .Values.firstRows.minCellBytes | quote }}
+ - name: FIRST_ROWS_MIN_NUMBER
+ value: {{ .Values.firstRows.minNumber| quote }}
+ - name: FIRST_ROWS_COLUMNS_MAX_NUMBER
+ value: {{ .Values.firstRows.columnsMaxNumber| quote }}
+ - name: PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS
+ value: {{ .Values.parquetAndDatasetInfo.blockedDatasets | quote }}
+ - name: PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE
+ value: {{ .Values.parquetAndDatasetInfo.commitMessage | quote }}
+ - name: PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN
+ {{- if .Values.secrets.userHfToken.fromSecret }}
+ valueFrom:
+ secretKeyRef:
+ name: {{ .Values.secrets.userHfToken.secretName | quote }}
+ key: HF_TOKEN
+ optional: false
+ {{- else }}
+ value: {{ .Values.secrets.userHfToken.value }}
+ {{- end }}
+ - name: PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE
+ value: {{ .Values.parquetAndDatasetInfo.maxDatasetSize | quote }}
+ - name: PARQUET_AND_DATASET_INFO_SOURCE_REVISION
+ value: {{ .Values.parquetAndDatasetInfo.sourceRevision | quote }}
+ - name: PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS
+ value: {{ .Values.parquetAndDatasetInfo.supportedDatasets | quote }}
+ - name: PARQUET_AND_DATASET_INFO_TARGET_REVISION
+ value: {{ .Values.parquetAndDatasetInfo.targetRevision | quote }}
+ - name: PARQUET_AND_DATASET_INFO_URL_TEMPLATE
+ value: {{ .Values.parquetAndDatasetInfo.urlTemplate | quote }}
+ volumeMounts:
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
+ {{ include "volumeMountCache" . | nindent 2 }}
+ securityContext:
+ allowPrivilegeEscalation: false
+ resources: {{ toYaml .Values.genericWorker.resources | nindent 4 }}
+{{- end -}}
diff --git a/chart/templates/worker/generic/deployment.yaml b/chart/templates/worker/generic/deployment.yaml
new file mode 100644
index 00000000..cab65da2
--- /dev/null
+++ b/chart/templates/worker/generic/deployment.yaml
@@ -0,0 +1,30 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+apiVersion: apps/v1
+kind: Deployment
+metadata:
+ labels: {{ include "labels.genericWorker" . | nindent 4 }}
+ name: "{{ include "name" . }}-worker-generic"
+ namespace: {{ .Release.Namespace }}
+spec:
+ progressDeadlineSeconds: 600
+ replicas: {{ .Values.genericWorker.replicas }}
+ revisionHistoryLimit: 10
+ selector:
+ matchLabels: {{ include "labels.genericWorker" . | nindent 6 }}
+ strategy:
+ type: Recreate
+ template:
+ metadata:
+ labels: {{ include "labels.genericWorker" . | nindent 8 }}
+ spec:
+ {{- include "image.imagePullSecrets" . | nindent 6 }}
+ initContainers:
+ {{ include "initContainerAssets" . | nindent 8 }}
+ {{ include "initContainerCache" . | nindent 8 }}
+ containers: {{ include "containerWorkerGeneric" . | nindent 8 }}
+ nodeSelector: {{ toYaml .Values.genericWorker.nodeSelector | nindent 8 }}
+ tolerations: {{ toYaml .Values.genericWorker.tolerations | nindent 8 }}
+ volumes: {{ include "volumeData" . | nindent 8 }}
+ securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/worker/parquet-and-dataset-info/_container.tpl b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
index ef54ff92..d492d9d8 100644
--- a/chart/templates/worker/parquet-and-dataset-info/_container.tpl
+++ b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/templates/worker/parquet/_container.tpl b/chart/templates/worker/parquet/_container.tpl
index 33613ab5..06be53fb 100644
--- a/chart/templates/worker/parquet/_container.tpl
+++ b/chart/templates/worker/parquet/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/templates/worker/sizes/_container.tpl b/chart/templates/worker/sizes/_container.tpl
index e42ab783..7146122f 100644
--- a/chart/templates/worker/sizes/_container.tpl
+++ b/chart/templates/worker/sizes/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/templates/worker/split-names/_container.tpl b/chart/templates/worker/split-names/_container.tpl
index 694349fb..07b29ab6 100644
--- a/chart/templates/worker/split-names/_container.tpl
+++ b/chart/templates/worker/split-names/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/templates/worker/splits/_container.tpl b/chart/templates/worker/splits/_container.tpl
index c5211441..e8cb31c5 100644
--- a/chart/templates/worker/splits/_container.tpl
+++ b/chart/templates/worker/splits/_container.tpl
@@ -9 +9 @@
- - name: WORKER_ENDPOINT
+ - name: WORKER_ONLY_JOB_TYPES
diff --git a/chart/values.yaml b/chart/values.yaml
index b451368c..23ecce17 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -380,0 +381,14 @@ sizes:
+
+genericWorker:
+ # override the common queue parameters
+ queue:
+ # Maximum number of jobs running at the same time for the same namespace
+ maxJobsPerNamespace: 1
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ tolerations: []
diff --git a/e2e/tests/test_15_is_valid.py b/e2e/tests/test_15_is_valid.py
index a1c42a67..fe0fbcb6 100644
--- a/e2e/tests/test_15_is_valid.py
+++ b/e2e/tests/test_15_is_valid.py
@@ -16 +16 @@ def test_is_valid_after_datasets_processed(hf_dataset_repos_csv_data: DatasetRep
- assert response.json()["valid"] is True, response.text
+ assert response.json()["valid"], response.text
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py
index 50abc75a..f7340a60 100644
--- a/libs/libcommon/src/libcommon/dataset.py
+++ b/libs/libcommon/src/libcommon/dataset.py
@@ -151 +151 @@ def get_dataset_info_for_supported_datasets(
- if dataset_info.private is True:
+ if dataset_info.private:
diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py
index 07c0674a..43a2a6c4 100644
--- a/libs/libcommon/src/libcommon/operations.py
+++ b/libs/libcommon/src/libcommon/operations.py
@@ -50,0 +51 @@ def update_dataset(
+ queue = Queue()
@@ -53 +54 @@ def update_dataset(
- Queue(type=init_processing_step.job_type).upsert_job(dataset=dataset, force=force, priority=priority)
+ queue.upsert_job(job_type=init_processing_step.job_type, dataset=dataset, force=force, priority=priority)
@@ -137,0 +139 @@ def check_in_process(
+ queue = Queue()
@@ -139 +141,2 @@ def check_in_process(
- Queue(type=step.job_type).is_job_in_process(dataset=dataset, config=config, split=split) for step in all_steps
+ queue.is_job_in_process(job_type=step.job_type, dataset=dataset, config=config, split=split)
+ for step in all_steps
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index f86e08bf..2f16c5c2 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -172 +172 @@ class Queue:
- """A queue manages jobs of a given type.
+ """A queue manages jobs.
@@ -188 +187,0 @@ class Queue:
- type (`str`, required): Type of the job. It identifies the queue.
@@ -195,2 +194 @@ class Queue:
- def __init__(self, type: str, max_jobs_per_namespace: Optional[int] = None):
- self.type = type
+ def __init__(self, max_jobs_per_namespace: Optional[int] = None):
@@ -202,0 +201 @@ class Queue:
+ job_type: str,
@@ -213,0 +213 @@ class Queue:
+ job_type (`str`): The type of the job
@@ -223 +223 @@ class Queue:
- type=self.type,
+ type=job_type,
@@ -227 +227 @@ class Queue:
- unicity_id=f"Job[{self.type}][{dataset}][{config}][{split}]",
+ unicity_id=f"Job[{job_type}][{dataset}][{config}][{split}]",
@@ -236,0 +237 @@ class Queue:
+ job_type: str,
@@ -250,0 +252 @@ class Queue:
+ job_type (`str`): The type of the job
@@ -259 +261 @@ class Queue:
- existing = Job.objects(type=self.type, dataset=dataset, config=config, split=split, status=Status.WAITING)
+ existing = Job.objects(type=job_type, dataset=dataset, config=config, split=split, status=Status.WAITING)
@@ -265 +267,3 @@ class Queue:
- return self._add_job(dataset=dataset, config=config, split=split, force=force, priority=priority)
+ return self._add_job(
+ job_type=job_type, dataset=dataset, config=config, split=split, force=force, priority=priority
+ )
@@ -267 +271,3 @@ class Queue:
- def _get_next_waiting_job_for_priority(self, priority: Priority) -> Job:
+ def _get_next_waiting_job_for_priority(
+ self, priority: Priority, only_job_types: Optional[list[str]] = None
+ ) -> Job:
@@ -275,0 +282,4 @@ class Queue:
+ Args:
+ priority (`Priority`): The priority of the job.
+ only_job_types: if not None, only jobs of the given types are considered.
+
@@ -281 +291,7 @@ class Queue:
- started_jobs = Job.objects(type=self.type, status=Status.STARTED)
+ logging.debug("Getting next waiting job for priority %s, among types %s", priority, only_job_types or "all")
+ started_jobs = (
+ Job.objects(type__in=only_job_types, status=Status.STARTED)
+ if only_job_types
+ else Job.objects(status=Status.STARTED)
+ )
+ logging.debug(f"Number of started jobs: {started_jobs.count()}")
@@ -282,0 +299 @@ class Queue:
+ logging.debug(f"Started job namespaces: {started_job_namespaces}")
@@ -285,5 +302,13 @@ class Queue:
- Job.objects(
- type=self.type,
- status=Status.WAITING,
- namespace__nin=set(started_job_namespaces),
- priority=priority,
+ (
+ Job.objects(
+ type__in=only_job_types,
+ status=Status.WAITING,
+ namespace__nin=set(started_job_namespaces),
+ priority=priority,
+ )
+ if only_job_types
+ else Job.objects(
+ status=Status.WAITING,
+ namespace__nin=set(started_job_namespaces),
+ priority=priority,
+ )
@@ -292 +317 @@ class Queue:
- .only("dataset", "config", "split", "force")
+ .only("type", "dataset", "config", "split", "force")
@@ -298,0 +324 @@ class Queue:
+ logging.debug("No waiting job for namespace without started job")
@@ -312,0 +339,4 @@ class Queue:
+ logging.debug(
+ f"Descending frequency namespace counts, with less than {self.max_jobs_per_namespace} started jobs:"
+ f" {descending_frequency_namespace_counts}"
+ )
@@ -318,0 +349 @@ class Queue:
+ logging.debug(f"Least common namespaces group: {least_common_namespaces_group}")
@@ -320,6 +351,15 @@ class Queue:
- Job.objects(
- type=self.type,
- status=Status.WAITING,
- namespace__in=least_common_namespaces_group,
- unicity_id__nin=started_unicity_ids,
- priority=priority,
+ (
+ Job.objects(
+ type__in=only_job_types,
+ status=Status.WAITING,
+ namespace__in=least_common_namespaces_group,
+ unicity_id__nin=started_unicity_ids,
+ priority=priority,
+ )
+ if only_job_types
+ else Job.objects(
+ status=Status.WAITING,
+ namespace__in=least_common_namespaces_group,
+ unicity_id__nin=started_unicity_ids,
+ priority=priority,
+ )
@@ -328 +368 @@ class Queue:
- .only("dataset", "config", "split", "force")
+ .only("type", "dataset", "config", "split", "force")
@@ -339 +379 @@ class Queue:
- def get_next_waiting_job(self) -> Job:
+ def get_next_waiting_job(self, only_job_types: Optional[list[str]] = None) -> Job:
@@ -348,0 +389,3 @@ class Queue:
+ Args:
+ only_job_types: if not None, only jobs of the given types are considered.
+
@@ -356 +399 @@ class Queue:
- return self._get_next_waiting_job_for_priority(priority)
+ return self._get_next_waiting_job_for_priority(priority=priority, only_job_types=only_job_types)
@@ -361 +404 @@ class Queue:
- def start_job(self) -> JobInfo:
+ def start_job(self, only_job_types: Optional[list[str]] = None) -> JobInfo:
@@ -365,0 +409,3 @@ class Queue:
+ Args:
+ only_job_types: if not None, only jobs of the given types are considered.
+
@@ -372 +418,3 @@ class Queue:
- next_waiting_job = self.get_next_waiting_job()
+ logging.debug("looking for a job to start, among the following types: %s", only_job_types or "all")
+ next_waiting_job = self.get_next_waiting_job(only_job_types=only_job_types)
+ logging.debug(f"job found: {next_waiting_job}")
@@ -374,0 +423,4 @@ class Queue:
+ if only_job_types and next_waiting_job.type not in only_job_types:
+ raise RuntimeError(
+ f"The job type {next_waiting_job.type} is not in the list of allowed job types {only_job_types}"
+ )
@@ -377 +429 @@ class Queue:
- "type": self.type,
+ "type": next_waiting_job.type,
@@ -384,0 +437,14 @@ class Queue:
+ def get_job_type(self, job_id: str) -> str:
+ """Get the job type for a given job id.
+
+ Args:
+ job_id (`str`, required): id of the job
+
+ Returns: the job type
+
+ Raises:
+ DoesNotExist: if the job does not exist
+ """
+ job = Job.objects(pk=job_id).get()
+ return job.type
+
@@ -411 +477,3 @@ class Queue:
- def is_job_in_process(self, dataset: str, config: Optional[str] = None, split: Optional[str] = None) -> bool:
+ def is_job_in_process(
+ self, job_type: str, dataset: str, config: Optional[str] = None, split: Optional[str] = None
+ ) -> bool:
@@ -414,0 +483 @@ class Queue:
+ job_type (`str`, required): job type
@@ -424 +493 @@ class Queue:
- type=self.type,
+ type=job_type,
@@ -433,3 +502,3 @@ class Queue:
- def cancel_started_jobs(self) -> None:
- """Cancel all started jobs."""
- for job in Job.objects(type=self.type, status=Status.STARTED.value):
+ def cancel_started_jobs(self, job_type: str) -> None:
+ """Cancel all started jobs for a given type."""
+ for job in Job.objects(type=job_type, status=Status.STARTED.value):
@@ -437 +506 @@ class Queue:
- self.upsert_job(dataset=job.dataset, config=job.config, split=job.split)
+ self.upsert_job(job_type=job.type, dataset=job.dataset, config=job.config, split=job.split)
@@ -440,2 +509,2 @@ class Queue:
- def count_jobs(self, status: Status) -> int:
- """Count the number of jobs with a given status.
+ def count_jobs(self, status: Status, job_type: str) -> int:
+ """Count the number of jobs with a given status and the given type.
@@ -444,0 +514 @@ class Queue:
+ job_type (`str`, required): job type
@@ -446 +516 @@ class Queue:
- Returns: the number of jobs with the given status
+ Returns: the number of jobs with the given status and the given type.
@@ -448 +518 @@ class Queue:
- return Job.objects(type=self.type, status=status.value).count()
+ return Job.objects(type=job_type, status=status.value).count()
@@ -450,2 +520,2 @@ class Queue:
- def get_jobs_count_by_status(self) -> CountByStatus:
- """Count the number of jobs by status.
+ def get_jobs_count_by_status(self, job_type: str) -> CountByStatus:
+ """Count the number of jobs by status for a given job type.
@@ -461,6 +531,6 @@ class Queue:
- "waiting": self.count_jobs(status=Status.WAITING),
- "started": self.count_jobs(status=Status.STARTED),
- "success": self.count_jobs(status=Status.SUCCESS),
- "error": self.count_jobs(status=Status.ERROR),
- "cancelled": self.count_jobs(status=Status.CANCELLED),
- "skipped": self.count_jobs(status=Status.SKIPPED),
+ "waiting": self.count_jobs(status=Status.WAITING, job_type=job_type),
+ "started": self.count_jobs(status=Status.STARTED, job_type=job_type),
+ "success": self.count_jobs(status=Status.SUCCESS, job_type=job_type),
+ "error": self.count_jobs(status=Status.ERROR, job_type=job_type),
+ "cancelled": self.count_jobs(status=Status.CANCELLED, job_type=job_type),
+ "skipped": self.count_jobs(status=Status.SKIPPED, job_type=job_type),
@@ -469,2 +539,2 @@ class Queue:
- def get_dump_with_status(self, status: Status) -> List[JobDict]:
- """Get the dump of the jobs with a given status.
+ def get_dump_with_status(self, status: Status, job_type: str) -> List[JobDict]:
+ """Get the dump of the jobs with a given status and a given type.
@@ -473,0 +544 @@ class Queue:
+ job_type (`str`, required): job type
@@ -475 +546 @@ class Queue:
- Returns: a list of jobs with the given status
+ Returns: a list of jobs with the given status and the given type
@@ -477 +548 @@ class Queue:
- return [d.to_dict() for d in Job.objects(type=self.type, status=status.value)]
+ return [d.to_dict() for d in Job.objects(type=job_type, status=status.value)]
@@ -479,2 +550,2 @@ class Queue:
- def get_dump_by_pending_status(self) -> DumpByPendingStatus:
- """Get the dump of the jobs by pending status.
+ def get_dump_by_pending_status(self, job_type: str) -> DumpByPendingStatus:
+ """Get the dump of the jobs by pending status for a given job type.
@@ -485,2 +556,2 @@ class Queue:
- "waiting": self.get_dump_with_status(status=Status.WAITING),
- "started": self.get_dump_with_status(status=Status.STARTED),
+ "waiting": self.get_dump_with_status(job_type=job_type, status=Status.WAITING),
+ "started": self.get_dump_with_status(job_type=job_type, status=Status.STARTED),
@@ -489,2 +560,2 @@ class Queue:
- def get_total_duration_per_dataset(self) -> Dict[str, int]:
- """Get the total duration for the last 30 days of the finished jobs for every dataset
+ def get_total_duration_per_dataset(self, job_type: str) -> Dict[str, int]:
+ """Get the total duration for the last 30 days of the finished jobs of a given type for every dataset
@@ -499 +570 @@ class Queue:
- type=self.type,
+ type=job_type,
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 6f444bd1..6c0244a7 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -25 +25,3 @@ def queue_mongo_resource(queue_mongo_host: str) -> Iterator[QueueMongoResource]:
- with QueueMongoResource(database=database, host=host) as queue_mongo_resource:
+ with QueueMongoResource(database=database, host=host, server_selection_timeout_ms=3_000) as queue_mongo_resource:
+ if not queue_mongo_resource.is_available():
+ raise RuntimeError("Mongo resource is not available")
@@ -34 +36 @@ def test__add_job() -> None:
- queue = Queue(test_type)
+ queue = Queue()
@@ -36 +38 @@ def test__add_job() -> None:
- queue._add_job(dataset=test_dataset, force=True)
+ queue._add_job(job_type=test_type, dataset=test_dataset, force=True)
@@ -38,2 +40,2 @@ def test__add_job() -> None:
- queue._add_job(dataset=test_dataset)
- assert queue.is_job_in_process(dataset=test_dataset) is True
+ queue._add_job(job_type=test_type, dataset=test_dataset)
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -46,2 +48,2 @@ def test__add_job() -> None:
- assert job_info["force"] is True
- assert queue.is_job_in_process(dataset=test_dataset) is True
+ assert job_info["force"]
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -50 +52 @@ def test__add_job() -> None:
- queue._add_job(dataset=test_dataset, force=True)
+ queue._add_job(job_type=test_type, dataset=test_dataset, force=True)
@@ -57 +59 @@ def test__add_job() -> None:
- assert queue.is_job_in_process(dataset=test_dataset) is True
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -60 +62 @@ def test__add_job() -> None:
- assert job_info["force"] is False
+ assert not job_info["force"]
@@ -64 +66 @@ def test__add_job() -> None:
- assert job_info["force"] is True
+ assert job_info["force"]
@@ -71 +73 @@ def test__add_job() -> None:
- assert queue.is_job_in_process(dataset=test_dataset) is False
+ assert not queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -81 +83 @@ def test_upsert_job() -> None:
- queue = Queue(test_type)
+ queue = Queue()
@@ -83 +85 @@ def test_upsert_job() -> None:
- queue.upsert_job(dataset=test_dataset, force=True)
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, force=True)
@@ -85,2 +87,2 @@ def test_upsert_job() -> None:
- queue.upsert_job(dataset=test_dataset)
- assert queue.is_job_in_process(dataset=test_dataset) is True
+ queue.upsert_job(job_type=test_type, dataset=test_dataset)
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -93,2 +95,2 @@ def test_upsert_job() -> None:
- assert job_info["force"] is True # the new job inherits from waiting forced jobs
- assert queue.is_job_in_process(dataset=test_dataset) is True
+ assert job_info["force"] # the new job inherits from waiting forced jobs
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -96 +98 @@ def test_upsert_job() -> None:
- queue.upsert_job(dataset=test_dataset, force=False)
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, force=False)
@@ -103 +105 @@ def test_upsert_job() -> None:
- assert queue.is_job_in_process(dataset=test_dataset) is True
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -106 +108 @@ def test_upsert_job() -> None:
- assert job_info["force"] is False # the new jobs does not inherit from started forced jobs
+ assert not job_info["force"] # the new jobs does not inherit from started forced jobs
@@ -109 +111 @@ def test_upsert_job() -> None:
- assert queue.is_job_in_process(dataset=test_dataset) is False
+ assert not queue.is_job_in_process(job_type=test_type, dataset=test_dataset)
@@ -123,11 +125,13 @@ def test_priority_logic() -> None:
- queue = Queue(test_type)
- queue.upsert_job(dataset="dataset1", config="config", split="split1")
- queue.upsert_job(dataset="dataset1/dataset", config="config", split="split1")
- queue.upsert_job(dataset="dataset1", config="config", split="split2")
- queue.upsert_job(dataset="dataset2", config="config", split="split1", priority=Priority.LOW)
- queue.upsert_job(dataset="dataset2/dataset", config="config", split="split1", priority=Priority.LOW)
- queue.upsert_job(dataset="dataset2", config="config", split="split2")
- queue.upsert_job(dataset="dataset3", config="config", split="split1")
- queue.upsert_job(dataset="dataset3", config="config", split="split1", priority=Priority.LOW)
- queue.upsert_job(dataset="dataset1", config="config", split="split1")
- queue.upsert_job(dataset="dataset2", config="config", split="split1", priority=Priority.LOW)
+ queue = Queue()
+ queue.upsert_job(job_type=test_type, dataset="dataset1", config="config", split="split1")
+ queue.upsert_job(job_type=test_type, dataset="dataset1/dataset", config="config", split="split1")
+ queue.upsert_job(job_type=test_type, dataset="dataset1", config="config", split="split2")
+ queue.upsert_job(job_type=test_type, dataset="dataset2", config="config", split="split1", priority=Priority.LOW)
+ queue.upsert_job(
+ job_type=test_type, dataset="dataset2/dataset", config="config", split="split1", priority=Priority.LOW
+ )
+ queue.upsert_job(job_type=test_type, dataset="dataset2", config="config", split="split2")
+ queue.upsert_job(job_type=test_type, dataset="dataset3", config="config", split="split1")
+ queue.upsert_job(job_type=test_type, dataset="dataset3", config="config", split="split1", priority=Priority.LOW)
+ queue.upsert_job(job_type=test_type, dataset="dataset1", config="config", split="split1")
+ queue.upsert_job(job_type=test_type, dataset="dataset2", config="config", split="split1", priority=Priority.LOW)
@@ -155,5 +159,5 @@ def test_max_jobs_per_namespace(max_jobs_per_namespace: Optional[int]) -> None:
- queue = Queue(test_type, max_jobs_per_namespace=max_jobs_per_namespace)
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split1")
- assert queue.is_job_in_process(dataset=test_dataset, config=test_config, split="split1") is True
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split2")
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split3")
+ queue = Queue(max_jobs_per_namespace=max_jobs_per_namespace)
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split1")
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, config=test_config, split="split1")
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split2")
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split3")
@@ -164 +168 @@ def test_max_jobs_per_namespace(max_jobs_per_namespace: Optional[int]) -> None:
- assert queue.is_job_in_process(dataset=test_dataset, config=test_config, split="split1") is True
+ assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, config=test_config, split="split1")
@@ -178 +182,23 @@ def test_max_jobs_per_namespace(max_jobs_per_namespace: Optional[int]) -> None:
- assert queue.is_job_in_process(dataset=test_dataset, config=test_config, split="split1") is False
+ assert not queue.is_job_in_process(job_type=test_type, dataset=test_dataset, config=test_config, split="split1")
+
+
[email protected](
+ "job_type,only_job_types",
+ [
+ ("test_type", None),
+ ("test_type", ["test_type"]),
+ ("test_type", ["other_type", "test_type"]),
+ ("test_type", ["other_type"]),
+ ],
+)
+def test_only_job_types(job_type: str, only_job_types: Optional[list[str]]) -> None:
+ test_dataset = "test_dataset"
+ queue = Queue(max_jobs_per_namespace=100)
+ queue.upsert_job(job_type=job_type, dataset=test_dataset, config=None, split=None)
+ assert queue.is_job_in_process(job_type=job_type, dataset=test_dataset, config=None, split=None)
+ if only_job_types and job_type not in only_job_types:
+ with pytest.raises(EmptyQueueError):
+ queue.start_job(only_job_types=only_job_types)
+ else:
+ job_info = queue.start_job(only_job_types=only_job_types)
+ assert job_info["dataset"] == test_dataset
@@ -185,2 +211 @@ def test_count_by_status() -> None:
- queue = Queue(test_type)
- queue_other = Queue(test_other_type)
+ queue = Queue()
@@ -191,2 +216,2 @@ def test_count_by_status() -> None:
- assert queue.get_jobs_count_by_status() == expected_empty
- assert queue_other.get_jobs_count_by_status() == expected_empty
+ assert queue.get_jobs_count_by_status(job_type=test_type) == expected_empty
+ assert queue.get_jobs_count_by_status(job_type=test_other_type) == expected_empty
@@ -194 +219 @@ def test_count_by_status() -> None:
- queue.upsert_job(dataset=test_dataset)
+ queue.upsert_job(job_type=test_type, dataset=test_dataset)
@@ -196,2 +221,2 @@ def test_count_by_status() -> None:
- assert queue.get_jobs_count_by_status() == expected_one_waiting
- assert queue_other.get_jobs_count_by_status() == expected_empty
+ assert queue.get_jobs_count_by_status(job_type=test_type) == expected_one_waiting
+ assert queue.get_jobs_count_by_status(job_type=test_other_type) == expected_empty
@@ -199 +224 @@ def test_count_by_status() -> None:
- queue_other.upsert_job(dataset=test_dataset)
+ queue.upsert_job(job_type=test_other_type, dataset=test_dataset)
@@ -201,2 +226,2 @@ def test_count_by_status() -> None:
- assert queue.get_jobs_count_by_status() == expected_one_waiting
- assert queue_other.get_jobs_count_by_status() == expected_one_waiting
+ assert queue.get_jobs_count_by_status(job_type=test_type) == expected_one_waiting
+ assert queue.get_jobs_count_by_status(job_type=test_other_type) == expected_one_waiting
@@ -209,6 +234,6 @@ def test_get_total_duration_per_dataset() -> None:
- queue = Queue(test_type)
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split1")
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split2")
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split3")
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split4")
- queue.upsert_job(dataset=test_dataset, config=test_config, split="split5")
+ queue = Queue()
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split1")
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split2")
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split3")
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split4")
+ queue.upsert_job(job_type=test_type, dataset=test_dataset, config=test_config, split="split5")
@@ -226 +251 @@ def test_get_total_duration_per_dataset() -> None:
- queue.cancel_started_jobs()
+ queue.cancel_started_jobs(job_type=test_type)
@@ -228 +253 @@ def test_get_total_duration_per_dataset() -> None:
- assert queue.get_total_duration_per_dataset()[test_dataset] >= duration * 3
+ assert queue.get_total_duration_per_dataset(job_type=test_type)[test_dataset] >= duration * 3
diff --git a/libs/libcommon/tests/test_storage.py b/libs/libcommon/tests/test_storage.py
index a0307134..b3a8aa55 100644
--- a/libs/libcommon/tests/test_storage.py
+++ b/libs/libcommon/tests/test_storage.py
@@ -111,4 +111,4 @@ def test_remove_dir(tmp_path_factory: pytest.TempPathFactory, exists: bool, is_s
- assert tmp_path.exists() is False
- assert tmp_path.is_dir() is False
- assert tmp_file.exists() is False
- assert tmp_file.is_file() is False
+ assert not tmp_path.exists()
+ assert not tmp_path.is_dir()
+ assert not tmp_file.exists()
+ assert not tmp_file.is_file()
diff --git a/services/admin/src/admin/prometheus.py b/services/admin/src/admin/prometheus.py
index fc5221a4..b668efa4 100644
--- a/services/admin/src/admin/prometheus.py
+++ b/services/admin/src/admin/prometheus.py
@@ -63,0 +64 @@ class Prometheus:
+ queue = Queue()
@@ -65 +66 @@ class Prometheus:
- for status, total in Queue(type=processing_step.job_type).get_jobs_count_by_status().items():
+ for status, total in queue.get_jobs_count_by_status(job_type=processing_step.job_type).items():
diff --git a/services/admin/src/admin/routes/cancel_jobs.py b/services/admin/src/admin/routes/cancel_jobs.py
index 059a9ca7..22292cf3 100644
--- a/services/admin/src/admin/routes/cancel_jobs.py
+++ b/services/admin/src/admin/routes/cancel_jobs.py
@@ -33 +33 @@ def create_cancel_jobs_endpoint(
- Queue(type=processing_step.job_type).cancel_started_jobs()
+ Queue().cancel_started_jobs(job_type=processing_step.job_type)
diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py
index 8f053829..40b97558 100644
--- a/services/admin/src/admin/routes/force_refresh.py
+++ b/services/admin/src/admin/routes/force_refresh.py
@@ -57 +57,3 @@ def create_force_refresh_endpoint(
- Queue(type=processing_step.job_type).upsert_job(dataset=dataset, config=config, split=split, force=True)
+ Queue().upsert_job(
+ job_type=processing_step.job_type, dataset=dataset, config=config, split=split, force=True
+ )
diff --git a/services/admin/src/admin/routes/jobs_duration.py b/services/admin/src/admin/routes/jobs_duration.py
index 3dc5b6d4..053cacb9 100644
--- a/services/admin/src/admin/routes/jobs_duration.py
+++ b/services/admin/src/admin/routes/jobs_duration.py
@@ -32,0 +33 @@ def create_jobs_duration_per_dataset_endpoint(
+ queue = Queue()
@@ -34 +35 @@ def create_jobs_duration_per_dataset_endpoint(
- Queue(type=processing_step.job_type).get_total_duration_per_dataset(),
+ queue.get_total_duration_per_dataset(job_type=processing_step.job_type),
diff --git a/services/admin/src/admin/routes/pending_jobs.py b/services/admin/src/admin/routes/pending_jobs.py
index 451e66d9..1f2bbf3d 100644
--- a/services/admin/src/admin/routes/pending_jobs.py
+++ b/services/admin/src/admin/routes/pending_jobs.py
@@ -32,0 +33 @@ def create_pending_jobs_endpoint(
+ queue = Queue()
@@ -35 +36 @@ def create_pending_jobs_endpoint(
- processing_step.endpoint: Queue(type=processing_step.job_type).get_dump_by_pending_status()
+ processing_step.endpoint: queue.get_dump_by_pending_status(job_type=processing_step.job_type)
diff --git a/services/admin/tests/test_authentication.py b/services/admin/tests/test_authentication.py
index 866e6893..3c285bb1 100644
--- a/services/admin/tests/test_authentication.py
+++ b/services/admin/tests/test_authentication.py
@@ -17 +17 @@ def test_no_auth_check() -> None:
- assert auth_check() is True
+ assert auth_check()
@@ -42 +42 @@ def test_external_auth_responses_without_request(status: int, error: Optional[Ty
- assert auth_check(external_auth_url=url, organization="org1") is True
+ assert auth_check(external_auth_url=url, organization="org1")
@@ -58 +58 @@ def test_org(org: str, status: int, error: Optional[Type[Exception]]) -> None:
- assert auth_check(external_auth_url=url, organization=org) is True
+ assert auth_check(external_auth_url=url, organization=org)
@@ -93,7 +93,4 @@ def test_valid_responses_with_request() -> None:
- assert (
- auth_check(
- external_auth_url=url,
- request=create_request(headers={}),
- organization=organization,
- )
- is True
+ assert auth_check(
+ external_auth_url=url,
+ request=create_request(headers={}),
+ organization=organization,
diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py
index 07e7361b..6b93d9e0 100644
--- a/services/api/tests/routes/test_valid.py
+++ b/services/api/tests/routes/test_valid.py
@@ -112,3 +112,3 @@ def test_errors() -> None:
- assert is_valid(dataset=dataset_a, processing_steps_for_valid=processing_steps_for_valid) is True
- assert is_valid(dataset=dataset_b, processing_steps_for_valid=processing_steps_for_valid) is True
- assert is_valid(dataset=dataset_c, processing_steps_for_valid=processing_steps_for_valid) is False
+ assert is_valid(dataset=dataset_a, processing_steps_for_valid=processing_steps_for_valid)
+ assert is_valid(dataset=dataset_b, processing_steps_for_valid=processing_steps_for_valid)
+ assert not is_valid(dataset=dataset_c, processing_steps_for_valid=processing_steps_for_valid)
diff --git a/services/api/tests/test_authentication.py b/services/api/tests/test_authentication.py
index 3762c81e..d221f242 100644
--- a/services/api/tests/test_authentication.py
+++ b/services/api/tests/test_authentication.py
@@ -18 +18 @@ def test_no_auth_check() -> None:
- assert auth_check("dataset") is True
+ assert auth_check("dataset")
diff --git a/services/worker/README.md b/services/worker/README.md
index 35df2da6..e2b979c9 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -14 +13,0 @@ Set environment variables to configure the worker.
-- `WORKER_ENDPOINT`: the endpoint on which the worker will work (pre-compute and cache the response). The same worker is used for different endpoints to reuse shared code and dependencies. But at runtime, the worker is assigned only one endpoint. Allowed values: `/splits`, `/first_rows`, `/parquet-and-dataset-info`, etc. Defaults to `/splits`.
@@ -17,0 +17 @@ Set environment variables to configure the worker.
+- `WORKER_ONLY_JOB_TYPES`: comma-separated list of the job types to process, e.g. "/splits,/first-rows". If empty, the worker processes all the jobs. Defaults to empty.
@@ -49,2 +48,0 @@ If the Hub is not https://huggingface.co (i.e., if you set the `COMMON_HF_ENDPOI
-Only needed when the `WORKER_ENDPOINT` is set to `/first-rows`.
-
@@ -63,2 +60,0 @@ Also, set the assets-related configuration for the first-rows worker. See [../..
-Only needed when the `WORKER_ENDPOINT` is set to `/parquet-and-dataset-info`.
-
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index 545f7901..fcfec3ca 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -31 +30,0 @@ class WorkerConfig:
- endpoint: str = WORKER_ENDPOINT
@@ -34,0 +34 @@ class WorkerConfig:
+ only_job_types: list[str] = field(default_factory=get_empty_str_list)
@@ -44 +43,0 @@ class WorkerConfig:
- endpoint=env.str(name="ENDPOINT", default=WORKER_ENDPOINT),
@@ -48,0 +48 @@ class WorkerConfig:
+ only_job_types=env.list(name="ONLY_JOB_TYPES", default=get_empty_str_list()),
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index 942e7dbe..16e27bbd 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -443,0 +444 @@ class JobRunner(ABC):
+ queue = Queue()
@@ -446 +447,2 @@ class JobRunner(ABC):
- Queue(type=processing_step.job_type).upsert_job(
+ queue.upsert_job(
+ job_type=processing_step.job_type,
diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py
index 32b11a56..1c82e4a7 100644
--- a/services/worker/src/worker/loop.py
+++ b/services/worker/src/worker/loop.py
@@ -15,0 +16,4 @@ from worker.job_runner_factory import BaseJobRunnerFactory
+class UnknownJobTypeError(Exception):
+ pass
+
+
@@ -25,4 +28,0 @@ class Loop:
- library_cache_paths (`set[str]`):
- The paths of the library caches. Used to check if the disk is full.
- queue (`Queue`):
- The job queue.
@@ -31,0 +32,2 @@ class Loop:
+ library_cache_paths (`set[str]`):
+ The paths of the library caches. Used to check if the disk is full.
@@ -33,0 +36,3 @@ class Loop:
+ max_jobs_per_namespace (`int`):
+ The maximum number of jobs that can be processed per namespace. If a namespace has more jobs, the loop will
+ wait until some jobs are finished.
@@ -36,2 +40,0 @@ class Loop:
- library_cache_paths: set[str]
- queue: Queue
@@ -38,0 +42 @@ class Loop:
+ library_cache_paths: set[str]
@@ -39,0 +44 @@ class Loop:
+ max_jobs_per_namespace: int
@@ -43,0 +49 @@ class Loop:
+ self.queue = Queue(max_jobs_per_namespace=self.max_jobs_per_namespace)
@@ -46,15 +51,0 @@ class Loop:
- def log(self, level: int, msg: str) -> None:
- logging.log(level=level, msg=f"[{self.queue.type}] {msg}")
-
- def debug(self, msg: str) -> None:
- self.log(level=logging.DEBUG, msg=msg)
-
- def info(self, msg: str) -> None:
- self.log(level=logging.INFO, msg=msg)
-
- def critical(self, msg: str) -> None:
- self.log(level=logging.CRITICAL, msg=msg)
-
- def exception(self, msg: str) -> None:
- self.log(level=logging.ERROR, msg=msg)
-
@@ -69 +60 @@ class Loop:
- self.info(
+ logging.info(
@@ -81 +72 @@ class Loop:
- self.info(f"cpu load is too high: {load_pct:.0f}% - max is {self.worker_config.max_load_pct}%")
+ logging.info(f"cpu load is too high: {load_pct:.0f}% - max is {self.worker_config.max_load_pct}%")
@@ -104 +95 @@ class Loop:
- self.debug(f"sleep during {duration:.2f} seconds")
+ logging.debug(f"sleep during {duration:.2f} seconds")
@@ -108 +99 @@ class Loop:
- self.info("Worker started")
+ logging.info("Worker started")
@@ -116,2 +107,2 @@ class Loop:
- except BaseException as e:
- self.critical(f"quit due to an uncaught error while processing the job: {e}")
+ except BaseException:
+ logging.exception("quit due to an uncaught error while processing the job")
@@ -121 +112 @@ class Loop:
- self.debug("try to process a job")
+ logging.debug("try to process a job")
@@ -124,2 +115,8 @@ class Loop:
- job_runner = self.job_runner_factory.create_job_runner(self.queue.start_job())
- self.debug(f"job assigned: {job_runner}")
+ job_info = self.queue.start_job(only_job_types=self.worker_config.only_job_types)
+ if self.worker_config.only_job_types and job_info["type"] not in self.worker_config.only_job_types:
+ raise UnknownJobTypeError(
+ f"Job of type {job_info['type']} is not supported (only"
+ f" ${', '.join(self.worker_config.only_job_types)}). The queue should not have provided this"
+ " job. It is in an inconsistent state. Please report this issue to the datasets team."
+ )
+ logging.debug(f"job assigned: {job_info}")
@@ -127 +124 @@ class Loop:
- self.debug("no job in the queue")
+ logging.debug("no job in the queue")
@@ -129,0 +127 @@ class Loop:
+ job_runner = self.job_runner_factory.create_job_runner(job_info)
@@ -132 +130 @@ class Loop:
- self.debug(f"job finished with {finished_status.value}: {job_runner}")
+ logging.debug(f"job finished with {finished_status.value}: {job_runner}")
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index b1ae25d6..a7c1816d 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -6 +5,0 @@ from libcommon.processing_graph import ProcessingGraph
-from libcommon.queue import Queue
@@ -23 +21,0 @@ if __name__ == "__main__":
- processing_step = processing_graph.get_step(app_config.worker.endpoint)
@@ -49 +46,0 @@ if __name__ == "__main__":
- queue = Queue(type=processing_step.job_type, max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace)
@@ -51 +47,0 @@ if __name__ == "__main__":
- queue=queue,
@@ -53,0 +50 @@ if __name__ == "__main__":
+ max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace,
diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py
index 108c67d2..3f0e5d81 100644
--- a/services/worker/tests/job_runners/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__datasets_based_worker.py
@@ -177 +177 @@ def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, g
- assert worker.process() is False
+ assert not worker.process()
diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py
index ed9bfbac..14a6e883 100644
--- a/services/worker/tests/job_runners/test_config_names.py
+++ b/services/worker/tests/job_runners/test_config_names.py
@@ -61 +61 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_run
- assert job_runner.should_skip_job() is False
+ assert not job_runner.should_skip_job()
@@ -64 +64 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_run
- assert job_runner.should_skip_job() is True
+ assert job_runner.should_skip_job()
@@ -66 +66 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_run
- assert job_runner.should_skip_job() is False
+ assert not job_runner.should_skip_job()
@@ -72 +72 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) ->
- assert job_runner.process() is True
+ assert job_runner.process()
@@ -86 +86 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -130 +130 @@ def test_compute_splits_response_simple_csv(
- assert exc_info.value.disclose_cause is False
+ assert not exc_info.value.disclose_cause
@@ -133 +133 @@ def test_compute_splits_response_simple_csv(
- assert exc_info.value.disclose_cause is True
+ assert exc_info.value.disclose_cause
diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py
index e6779406..27863b9c 100644
--- a/services/worker/tests/job_runners/test_first_rows.py
+++ b/services/worker/tests/job_runners/test_first_rows.py
@@ -71 +71 @@ def test_should_skip_job(
- assert job_runner.should_skip_job() is False
+ assert not job_runner.should_skip_job()
@@ -74 +74 @@ def test_should_skip_job(
- assert job_runner.should_skip_job() is True
+ assert job_runner.should_skip_job()
@@ -76 +76 @@ def test_should_skip_job(
- assert job_runner.should_skip_job() is False
+ assert not job_runner.should_skip_job()
@@ -84 +84 @@ def test_compute(
- assert job_runner.process() is True
+ assert job_runner.process()
@@ -105 +105 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner, first_rows_config:
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -162 +162 @@ def test_number_rows(
- assert exc_info.value.disclose_cause is False
+ assert not exc_info.value.disclose_cause
@@ -165 +165 @@ def test_number_rows(
- assert exc_info.value.disclose_cause is True
+ assert exc_info.value.disclose_cause
diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index ff5f7fb3..c9b735d7 100644
--- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -124 +124 @@ def test_compute(
- assert job_runner.process() is True
+ assert job_runner.process()
@@ -142 +142 @@ def test_doesnotexist(
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -274 +274 @@ def test_not_supported_if_big(
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -291 +291 @@ def test_supported_if_gated(
- assert job_runner.process() is True
+ assert job_runner.process()
@@ -308 +308 @@ def test_not_supported_if_gated_with_extra_fields(
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -325 +325 @@ def test_blocked(
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -395 +395 @@ def test_compute_splits_response_simple_csv_error(
- assert exc_info.value.disclose_cause is False
+ assert not exc_info.value.disclose_cause
@@ -398 +398 @@ def test_compute_splits_response_simple_csv_error(
- assert exc_info.value.disclose_cause is True
+ assert exc_info.value.disclose_cause
diff --git a/services/worker/tests/job_runners/test_split_names.py b/services/worker/tests/job_runners/test_split_names.py
index bf9fbe4d..4ca27e23 100644
--- a/services/worker/tests/job_runners/test_split_names.py
+++ b/services/worker/tests/job_runners/test_split_names.py
@@ -62 +62 @@ def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str) ->
- assert job_runner.process() is True
+ assert job_runner.process()
@@ -77 +77 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -123 +123 @@ def test_compute_split_names_response(
- assert exc_info.value.disclose_cause is False
+ assert not exc_info.value.disclose_cause
@@ -126 +126 @@ def test_compute_split_names_response(
- assert exc_info.value.disclose_cause is True
+ assert exc_info.value.disclose_cause
diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py
index effca782..404beabf 100644
--- a/services/worker/tests/job_runners/test_splits.py
+++ b/services/worker/tests/job_runners/test_splits.py
@@ -61 +61 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner)
- assert job_runner.should_skip_job() is False
+ assert not job_runner.should_skip_job()
@@ -64 +64 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner)
- assert job_runner.should_skip_job() is True
+ assert job_runner.should_skip_job()
@@ -66 +66 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner)
- assert job_runner.should_skip_job() is False
+ assert not job_runner.should_skip_job()
@@ -72 +72 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) ->
- assert job_runner.process() is True
+ assert job_runner.process()
@@ -87 +87 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
- assert job_runner.process() is False
+ assert not job_runner.process()
@@ -131 +131 @@ def test_compute_splits_response_simple_csv(
- assert exc_info.value.disclose_cause is False
+ assert not exc_info.value.disclose_cause
@@ -134 +134 @@ def test_compute_splits_response_simple_csv(
- assert exc_info.value.disclose_cause is True
+ assert exc_info.value.disclose_cause
diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py
index 686acc72..d79af11d 100644
--- a/services/worker/tests/test_job_runner.py
+++ b/services/worker/tests/test_job_runner.py
@@ -293 +293 @@ def test_create_children_jobs() -> None:
- assert job_runner.should_skip_job() is False
+ assert not job_runner.should_skip_job()
@@ -296 +296 @@ def test_create_children_jobs() -> None:
- assert job_runner.should_skip_job() is True
+ assert job_runner.should_skip_job()
@@ -298 +298,2 @@ def test_create_children_jobs() -> None:
- child_dataset_jobs = Queue(type="/child-dataset").get_dump_with_status(status=Status.WAITING)
+ queue = Queue()
+ child_dataset_jobs = queue.get_dump_with_status(job_type="/child-dataset", status=Status.WAITING)
@@ -304 +305 @@ def test_create_children_jobs() -> None:
- child_config_jobs = Queue(type="/child-config").get_dump_with_status(status=Status.WAITING)
+ child_config_jobs = queue.get_dump_with_status(job_type="/child-config", status=Status.WAITING)
@@ -310 +311 @@ def test_create_children_jobs() -> None:
- child_split_jobs = Queue(type="/child-split").get_dump_with_status(status=Status.WAITING)
+ child_split_jobs = queue.get_dump_with_status(job_type="/child-split", status=Status.WAITING)
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 59058ee4..6647f403 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -5 +4,0 @@ from libcommon.processing_graph import ProcessingStep
-from libcommon.queue import Queue
@@ -55 +53,0 @@ def test_process_next_job(
- queue = Queue(type=test_processing_step.endpoint, max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace)
@@ -57,2 +54,0 @@ def test_process_next_job(
- library_cache_paths=libraries_resource.storage_paths,
- queue=queue,
@@ -59,0 +56 @@ def test_process_next_job(
+ library_cache_paths=libraries_resource.storage_paths,
@@ -60,0 +58 @@ def test_process_next_job(
+ max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace,
@@ -62 +60 @@ def test_process_next_job(
- assert loop.process_next_job() is False
+ assert not loop.process_next_job()
@@ -66,4 +64,8 @@ def test_process_next_job(
- loop.queue.upsert_job(dataset=dataset, config=config, split=split)
- loop.queue.is_job_in_process(dataset=dataset, config=config, split=split) is True
- assert loop.process_next_job() is True
- loop.queue.is_job_in_process(dataset=dataset, config=config, split=split) is False
+ loop.queue.upsert_job(job_type=test_processing_step.job_type, dataset=dataset, config=config, split=split)
+ assert loop.queue.is_job_in_process(
+ job_type=test_processing_step.job_type, dataset=dataset, config=config, split=split
+ )
+ assert loop.process_next_job()
+ assert not loop.queue.is_job_in_process(
+ job_type=test_processing_step.job_type, dataset=dataset, config=config, split=split
+ )
diff --git a/services/worker/tests/test_resources.py b/services/worker/tests/test_resources.py
index 7fa7872d..f7dd816c 100644
--- a/services/worker/tests/test_resources.py
+++ b/services/worker/tests/test_resources.py
@@ -30 +30 @@ def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define
- assert datasets.config.HF_UPDATE_DOWNLOAD_COUNTS is False
+ assert not datasets.config.HF_UPDATE_DOWNLOAD_COUNTS
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index f3a81dcc..4a3e818b 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -86 +86 @@ services:
- WORKER_ENDPOINT: "/config-names" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/config-names" # hard-coded
@@ -103 +103 @@ services:
- WORKER_ENDPOINT: "/split-names" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/split-names" # hard-coded
@@ -120 +120 @@ services:
- WORKER_ENDPOINT: "/splits" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/splits" # hard-coded
@@ -139 +139 @@ services:
- WORKER_ENDPOINT: "/first-rows" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/first-rows" # hard-coded
@@ -163 +163 @@ services:
- WORKER_ENDPOINT: "/parquet-and-dataset-info" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/parquet-and-dataset-info" # hard-coded
@@ -185 +185 @@ services:
- WORKER_ENDPOINT: "/parquet" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/parquet" # hard-coded
@@ -199 +199 @@ services:
- WORKER_ENDPOINT: "/dataset-info" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/dataset-info" # hard-coded
@@ -213 +213,32 @@ services:
- WORKER_ENDPOINT: "/sizes" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/sizes" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ worker-anything:
+ build:
+ context: ..
+ dockerfile: services/worker/Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ ASSETS_BASE_URL: "http://localhost:${PORT_REVERSE_PROXY-8000}/assets" # hard-coded to work with the reverse-proxy
+ ASSETS_STORAGE_DIRECTORY: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ FIRST_ROWS_MAX_BYTES: ${FIRST_ROWS_MAX_BYTES-1_000_000}
+ FIRST_ROWS_MAX_NUMBER: ${FIRST_ROWS_MAX_NUMBER-100}
+ FIRST_ROWS_MIN_CELL_BYTES: ${FIRST_ROWS_MIN_CELL_BYTES-100}
+ FIRST_ROWS_MIN_NUMBER: ${FIRST_ROWS_MIN_NUMBER-10}
+ FIRST_ROWS_COLUMNS_MAX_NUMBER: ${FIRST_ROWS_COLUMNS_MAX_NUMBER-1_000}
+ PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS: ${PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS-}
+ PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE: ${PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE-Update parquet files}
+ PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-}
+ PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE-100_000_000}
+ PARQUET_AND_DATASET_INFO_SOURCE_REVISION: ${PARQUET_AND_DATASET_INFO_SOURCE_REVISION-main}
+ PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS-}
+ PARQUET_AND_DATASET_INFO_TARGET_REVISION: ${PARQUET_AND_DATASET_INFO_TARGET_REVISION-refs/convert/parquet}
+ PARQUET_AND_DATASET_INFO_URL_TEMPLATE: ${PARQUET_AND_DATASET_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
+ WORKER_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ # ^ note: the datasets cache is automatically added, so no need to add it here
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index d8cb0633..e713b72f 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -86 +86 @@ services:
- WORKER_ENDPOINT: "/config-names" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/config-names" # hard-coded
@@ -103 +103 @@ services:
- WORKER_ENDPOINT: "/split-names" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/split-names" # hard-coded
@@ -120 +120 @@ services:
- WORKER_ENDPOINT: "/splits" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/splits" # hard-coded
@@ -139 +139 @@ services:
- WORKER_ENDPOINT: "/first-rows" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/first-rows" # hard-coded
@@ -163 +163 @@ services:
- WORKER_ENDPOINT: "/parquet-and-dataset-info" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/parquet-and-dataset-info" # hard-coded
@@ -185 +185 @@ services:
- WORKER_ENDPOINT: "/parquet" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/parquet" # hard-coded
@@ -199 +199 @@ services:
- WORKER_ENDPOINT: "/dataset-info" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/dataset-info" # hard-coded
@@ -213 +213,32 @@ services:
- WORKER_ENDPOINT: "/sizes" # hard-coded
+ WORKER_ONLY_JOB_TYPES: "/sizes" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ worker-anything:
+ build:
+ context: ..
+ dockerfile: services/worker/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
+ environment:
+ ASSETS_BASE_URL: "http://localhost:${PORT_REVERSE_PROXY-8000}/assets" # hard-coded to work with the reverse-proxy
+ ASSETS_STORAGE_DIRECTORY: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ FIRST_ROWS_MAX_BYTES: ${FIRST_ROWS_MAX_BYTES-1_000_000}
+ FIRST_ROWS_MAX_NUMBER: ${FIRST_ROWS_MAX_NUMBER-100}
+ FIRST_ROWS_MIN_CELL_BYTES: ${FIRST_ROWS_MIN_CELL_BYTES-100}
+ FIRST_ROWS_MIN_NUMBER: ${FIRST_ROWS_MIN_NUMBER-10}
+ FIRST_ROWS_COLUMNS_MAX_NUMBER: ${FIRST_ROWS_COLUMNS_MAX_NUMBER-1_000}
+ PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS: ${PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS-}
+ PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE: ${PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE-Update parquet files}
+ PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-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}
+ PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE-100_000_000}
+ PARQUET_AND_DATASET_INFO_SOURCE_REVISION: ${PARQUET_AND_DATASET_INFO_SOURCE_REVISION-main}
+ PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS-}
+ PARQUET_AND_DATASET_INFO_TARGET_REVISION: ${PARQUET_AND_DATASET_INFO_TARGET_REVISION-refs/convert/parquet}
+ PARQUET_AND_DATASET_INFO_URL_TEMPLATE: ${PARQUET_AND_DATASET_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
+ WORKER_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ # ^ note: the datasets cache is automatically added, so no need to add it here
+ extends:
+ file: docker-compose-dev-base.yml
+ service: datasets-worker
diff --git a/tools/docker-compose-mongo.yml b/tools/docker-compose-mongo.yml
index bd6634ac..9fa1263e 100644
--- a/tools/docker-compose-mongo.yml
+++ b/tools/docker-compose-mongo.yml
@@ -4,3 +4 @@ services:
- image: mongo
- volumes:
- - mongo-test:/data/db:rw
+ image: mongo:latest
@@ -9,2 +6,0 @@ services:
-volumes:
- mongo-test:
|
|
69ce0faf0c3ea69990d707bbfde664a9f71df785
|
Sylvain Lesage
| 2023-02-13T14:40:09 |
Rename obsolete mentions to datasets_based (#805)
|
diff --git a/DEVELOPER_GUIDE.md b/DEVELOPER_GUIDE.md
index d80c95b8..d446e6d0 100644
--- a/DEVELOPER_GUIDE.md
+++ b/DEVELOPER_GUIDE.md
@@ -158 +158 @@ GITHUB_TOKEN=xxx
-To install the [datasets based worker](./services/worker) on Mac OS, you can follow the next steps.
+To install the [worker](./services/worker) on Mac OS, you can follow the next steps.
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index 7fb0392a..6f1a775d 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -242,4 +241,0 @@ sizes:
-
-# --- datasets_based ---
-datasetsBased:
- contentMaxBytes: "10_000_000"
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 13bdb598..8b324fd2 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -312,4 +311,0 @@ sizes:
-
-# --- datasets_based ---
-datasetsBased:
- contentMaxBytes: "10_000_000"
diff --git a/chart/templates/_envDatasetsBased.tpl b/chart/templates/_envDatasetsBased.tpl
index ee8fea75..05d1ed29 100644
--- a/chart/templates/_envDatasetsBased.tpl
+++ b/chart/templates/_envDatasetsBased.tpl
@@ -11,2 +10,0 @@
-- name: CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes}}
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
new file mode 100644
index 00000000..daa35eb4
--- /dev/null
+++ b/chart/templates/_envWorker.tpl
@@ -0,0 +1,19 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+{{- define "envWorker" -}}
+- name: WORKER_CONTENT_MAX_BYTES
+ value: {{ .Values.worker.contentMaxBytes | quote}}
+ # WORKER_ENDPOINT is not defined here, it's hard-coded in the template
+- name: WORKER_MAX_DISK_USAGE_PCT
+ value: {{ .Values.worker.maxDiskUsagePct | quote }}
+- name: WORKER_MAX_LOAD_PCT
+ value: {{ .Values.worker.maxLoadPct | quote }}
+- name: WORKER_MAX_MEMORY_PCT
+ value: {{ .Values.worker.maxMemoryPct | quote }}
+- name: WORKER_SLEEP_SECONDS
+ value: {{ .Values.worker.sleepSeconds | quote }}
+- name: WORKER_STORAGE_PATHS
+ value: {{ .Values.assets.storageDirectory | quote }}
+ # ^ note: for datasets_based workers, the datasets cache is automatically added, so no need to add it here
+{{- end -}}
diff --git a/chart/templates/_envWorkerLoop.tpl b/chart/templates/_envWorkerLoop.tpl
deleted file mode 100644
index e4d922bc..00000000
--- a/chart/templates/_envWorkerLoop.tpl
+++ /dev/null
@@ -1,13 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-{{- define "envWorkerLoop" -}}
-- name: WORKER_LOOP_MAX_DISK_USAGE_PCT
- value: {{ .Values.workerLoop.maxDiskUsagePct | quote }}
-- name: WORKER_LOOP_MAX_LOAD_PCT
- value: {{ .Values.workerLoop.maxLoadPct | quote }}
-- name: WORKER_LOOP_MAX_MEMORY_PCT
- value: {{ .Values.workerLoop.maxMemoryPct | quote }}
-- name: WORKER_LOOP_SLEEP_SECONDS
- value: {{ .Values.workerLoop.sleepSeconds | quote }}
-{{- end -}}
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index 84a19bf8..3ad176a4 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -85 +85 @@ imagePullSecrets:
-{{- define "workers.datasetsBased.image" -}}
+{{- define "services.worker.image" -}}
@@ -266 +266 @@ http://{{ $hubName }}
-{{- end -}}
\ No newline at end of file
+{{- end -}}
diff --git a/chart/templates/worker/config-names/_container.tpl b/chart/templates/worker/config-names/_container.tpl
index 3a87636b..def7afb6 100644
--- a/chart/templates/worker/config-names/_container.tpl
+++ b/chart/templates/worker/config-names/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -15 +15 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
+ {{ include "envWorker" . | nindent 2 }}
@@ -19,2 +18,0 @@
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/dataset-info/_container.tpl b/chart/templates/worker/dataset-info/_container.tpl
index 03bb31c0..36283497 100644
--- a/chart/templates/worker/dataset-info/_container.tpl
+++ b/chart/templates/worker/dataset-info/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -15,3 +15 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
+ {{ include "envWorker" . | nindent 2 }}
diff --git a/chart/templates/worker/first-rows/_container.tpl b/chart/templates/worker/first-rows/_container.tpl
index 83e2fd72..884b136f 100644
--- a/chart/templates/worker/first-rows/_container.tpl
+++ b/chart/templates/worker/first-rows/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -16,4 +16 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
- - name: WORKER_LOOP_STORAGE_PATHS
- value: {{ .Values.assets.storageDirectory | quote }}
- # ^ note: the datasets cache is automatically added, so no need to add it here
+ {{ include "envWorker" . | nindent 2 }}
@@ -23,2 +19,0 @@
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/parquet-and-dataset-info/_container.tpl b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
index 2a0f4b85..ef54ff92 100644
--- a/chart/templates/worker/parquet-and-dataset-info/_container.tpl
+++ b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -15 +15 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
+ {{ include "envWorker" . | nindent 2 }}
@@ -19,2 +18,0 @@
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/parquet/_container.tpl b/chart/templates/worker/parquet/_container.tpl
index ec04cbea..33613ab5 100644
--- a/chart/templates/worker/parquet/_container.tpl
+++ b/chart/templates/worker/parquet/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -15,3 +15 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
+ {{ include "envWorker" . | nindent 2 }}
diff --git a/chart/templates/worker/sizes/_container.tpl b/chart/templates/worker/sizes/_container.tpl
index a575818d..e42ab783 100644
--- a/chart/templates/worker/sizes/_container.tpl
+++ b/chart/templates/worker/sizes/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -15,3 +15 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
+ {{ include "envWorker" . | nindent 2 }}
diff --git a/chart/templates/worker/split-names/_container.tpl b/chart/templates/worker/split-names/_container.tpl
index 53f0d88c..694349fb 100644
--- a/chart/templates/worker/split-names/_container.tpl
+++ b/chart/templates/worker/split-names/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -15 +15 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
+ {{ include "envWorker" . | nindent 2 }}
@@ -19,2 +18,0 @@
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/splits/_container.tpl b/chart/templates/worker/splits/_container.tpl
index d862a20a..c5211441 100644
--- a/chart/templates/worker/splits/_container.tpl
+++ b/chart/templates/worker/splits/_container.tpl
@@ -6 +6 @@
- image: {{ include "workers.datasetsBased.image" . }}
+ image: {{ include "services.worker.image" . }}
@@ -9 +9 @@
- - name: DATASETS_BASED_ENDPOINT
+ - name: WORKER_ENDPOINT
@@ -15 +15 @@
- {{ include "envWorkerLoop" . | nindent 2 }}
+ {{ include "envWorker" . | nindent 2 }}
@@ -19,2 +18,0 @@
- - name: DATASETS_BASED_CONTENT_MAX_BYTES
- value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/values.yaml b/chart/values.yaml
index 3b1aff69..b451368c 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -97 +97,3 @@ queue:
-workerLoop:
+worker:
+ # maximum size in bytes of the response content computed by a worker
+ contentMaxBytes: "10_000_000"
@@ -379,4 +380,0 @@ sizes:
-
-# --- datasets_based ---
-datasetsBased:
- contentMaxBytes: "10_000_000"
diff --git a/services/worker/README.md b/services/worker/README.md
index 04211f8c..35df2da6 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -8,0 +9,12 @@ Use environment variables to configure the worker. The prefix of each environmen
+## Worker configuration
+
+Set environment variables to configure the worker.
+
+- `WORKER_CONTENT_MAX_BYTES`: the maximum size in bytes of the response content computed by a worker (to prevent returning big responses in the REST API). Defaults to `10_000_000`.
+- `WORKER_ENDPOINT`: the endpoint on which the worker will work (pre-compute and cache the response). The same worker is used for different endpoints to reuse shared code and dependencies. But at runtime, the worker is assigned only one endpoint. Allowed values: `/splits`, `/first_rows`, `/parquet-and-dataset-info`, etc. Defaults to `/splits`.
+- `WORKER_MAX_DISK_USAGE_PCT`: maximum disk usage of every storage disk in the list (in percentage) to allow a job to start. Set to 0 to disable the test. Defaults to 90.
+- `WORKER_MAX_LOAD_PCT`: maximum load of the machine (in percentage: the max between the 1m load and the 5m load divided by the number of CPUs \*100) allowed to start a job. Set to 0 to disable the test. Defaults to 70.
+- `WORKER_MAX_MEMORY_PCT`: maximum memory (RAM + SWAP) usage of the machine (in percentage) allowed to start a job. Set to 0 to disable the test. Defaults to 80.
+- `WORKER_SLEEP_SECONDS`: wait duration in seconds at each loop iteration before checking if resources are available and processing a job if any is available. Note that the loop doesn't wait just after finishing a job: the next job is immediately processed. Defaults to `15`.
+- `WORKER_STORAGE_PATHS`: comma-separated list of paths to check for disk usage. Defaults to empty.
+
@@ -13 +24,0 @@ Set environment variables to configure the datasets-based worker (`DATASETS_BASE
-- `DATASETS_BASED_ENDPOINT`: the endpoint on which the worker will work (pre-compute and cache the response). The same worker is used for different endpoints to reuse shared code and dependencies. But at runtime, the worker is assigned only one endpoint. Allowed values: `/splits`, `/first_rows`, and ` /parquet-and-dataset-info`. Defaults to `/splits`.
@@ -15 +25,0 @@ Set environment variables to configure the datasets-based worker (`DATASETS_BASE
-- `DATASETS_BASED_CONTENT_MAX_BYTES`: the maximum size in bytes of the response content computed by a worker (to prevent returning big responses in the REST API). Defaults to `10_000_000`.
@@ -21 +31 @@ Also, set the modules cache configuration for the datasets-based worker. See [..
-Note that both directories will be appended to `WORKER_LOOP_STORAGE_PATHS` (see [../../libs/libcommon/README.md](../../libs/libcommon/README.md)) to hold the workers when the disk is full.
+Note that both directories will be appended to `WORKER_STORAGE_PATHS` (see [../../libs/libcommon/README.md](../../libs/libcommon/README.md)) to hold the workers when the disk is full.
@@ -29 +39 @@ Numba requires setting the `NUMBA_CACHE_DIR` environment variable to a writable
-Note that this directory will be appended to `WORKER_LOOP_STORAGE_PATHS` (see [../../libs/libcommon/README.md](../../libs/libcommon/README.md)) to hold the workers when the disk is full.
+Note that this directory will be appended to `WORKER_STORAGE_PATHS` (see [../../libs/libcommon/README.md](../../libs/libcommon/README.md)) to hold the workers when the disk is full.
@@ -39 +49 @@ If the Hub is not https://huggingface.co (i.e., if you set the `COMMON_HF_ENDPOI
-Only needed when the `DATASETS_BASED_ENDPOINT` is set to `/first-rows`.
+Only needed when the `WORKER_ENDPOINT` is set to `/first-rows`.
@@ -53 +63 @@ Also, set the assets-related configuration for the first-rows worker. See [../..
-Only needed when the `DATASETS_BASED_ENDPOINT` is set to `/parquet-and-dataset-info`.
+Only needed when the `WORKER_ENDPOINT` is set to `/parquet-and-dataset-info`.
@@ -73,10 +82,0 @@ See [../../libs/libcommon/README.md](../../libs/libcommon/README.md) for more in
-
-## Worker loop configuration
-
-Set environment variables to configure the worker loop that processes the queue.
-
-- `WORKER_LOOP_MAX_DISK_USAGE_PCT`: maximum disk usage of every storage disk in the list (in percentage) to allow a job to start. Set to 0 to disable the test. Defaults to 90.
-- `WORKER_LOOP_MAX_LOAD_PCT`: maximum load of the machine (in percentage: the max between the 1m load and the 5m load divided by the number of CPUs \*100) allowed to start a job. Set to 0 to disable the test. Defaults to 70.
-- `WORKER_LOOP_MAX_MEMORY_PCT`: maximum memory (RAM + SWAP) usage of the machine (in percentage) allowed to start a job. Set to 0 to disable the test. Defaults to 80.
-- `WORKER_LOOP_SLEEP_SECONDS`: wait duration in seconds at each loop iteration before checking if resources are available and processing a job if any is available. Note that the loop doesn't wait just after finishing a job: the next job is immediately processed. Defaults to `15`.
-- `WORKER_LOOP_STORAGE_PATHS`: comma-separated list of paths to check for disk usage. Defaults to empty.
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index 1cacd1c3..545f7901 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -16,4 +16,6 @@ from libcommon.config import (
-WORKER_LOOP_MAX_DISK_USAGE_PCT = 90
-WORKER_LOOP_MAX_LOAD_PCT = 70
-WORKER_LOOP_MAX_MEMORY_PCT = 80
-WORKER_LOOP_SLEEP_SECONDS = 15
+WORKER_CONTENT_MAX_BYTES = 10_000_000
+WORKER_ENDPOINT = "/config-names"
+WORKER_MAX_DISK_USAGE_PCT = 90
+WORKER_MAX_LOAD_PCT = 70
+WORKER_MAX_MEMORY_PCT = 80
+WORKER_SLEEP_SECONDS = 15
@@ -27,5 +29,7 @@ def get_empty_str_list() -> List[str]:
-class LoopConfig:
- max_disk_usage_pct: int = WORKER_LOOP_MAX_DISK_USAGE_PCT
- max_load_pct: int = WORKER_LOOP_MAX_LOAD_PCT
- max_memory_pct: int = WORKER_LOOP_MAX_MEMORY_PCT
- sleep_seconds: int = WORKER_LOOP_SLEEP_SECONDS
+class WorkerConfig:
+ content_max_bytes: int = WORKER_CONTENT_MAX_BYTES
+ endpoint: str = WORKER_ENDPOINT
+ max_disk_usage_pct: int = WORKER_MAX_DISK_USAGE_PCT
+ max_load_pct: int = WORKER_MAX_LOAD_PCT
+ max_memory_pct: int = WORKER_MAX_MEMORY_PCT
+ sleep_seconds: int = WORKER_SLEEP_SECONDS
@@ -35 +39 @@ class LoopConfig:
- def from_env(cls) -> "LoopConfig":
+ def from_env(cls) -> "WorkerConfig":
@@ -37 +41 @@ class LoopConfig:
- with env.prefixed("WORKER_LOOP_"):
+ with env.prefixed("WORKER_"):
@@ -39,4 +43,6 @@ class LoopConfig:
- max_disk_usage_pct=env.int(name="MAX_DISK_USAGE_PCT", default=WORKER_LOOP_MAX_DISK_USAGE_PCT),
- max_load_pct=env.int(name="MAX_LOAD_PCT", default=WORKER_LOOP_MAX_LOAD_PCT),
- max_memory_pct=env.int(name="MAX_MEMORY_PCT", default=WORKER_LOOP_MAX_MEMORY_PCT),
- sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_LOOP_SLEEP_SECONDS),
+ content_max_bytes=env.int(name="CONTENT_MAX_BYTES", default=WORKER_CONTENT_MAX_BYTES),
+ endpoint=env.str(name="ENDPOINT", default=WORKER_ENDPOINT),
+ max_disk_usage_pct=env.int(name="MAX_DISK_USAGE_PCT", default=WORKER_MAX_DISK_USAGE_PCT),
+ max_load_pct=env.int(name="MAX_LOAD_PCT", default=WORKER_MAX_LOAD_PCT),
+ max_memory_pct=env.int(name="MAX_MEMORY_PCT", default=WORKER_MAX_MEMORY_PCT),
+ sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_SLEEP_SECONDS),
@@ -47 +52,0 @@ class LoopConfig:
-DATASETS_BASED_ENDPOINT = "/config-names"
@@ -49 +53,0 @@ DATASETS_BASED_HF_DATASETS_CACHE = None
-DATASETS_BASED_CONTENT_MAX_BYTES = 10_000_000
@@ -54 +57,0 @@ class DatasetsBasedConfig:
- endpoint: str = DATASETS_BASED_ENDPOINT
@@ -56 +58,0 @@ class DatasetsBasedConfig:
- content_max_bytes: int = DATASETS_BASED_CONTENT_MAX_BYTES
@@ -63 +64,0 @@ class DatasetsBasedConfig:
- endpoint=env.str(name="ENDPOINT", default=DATASETS_BASED_ENDPOINT),
@@ -65 +65,0 @@ class DatasetsBasedConfig:
- content_max_bytes=env.int(name="CONTENT_MAX_BYTES", default=DATASETS_BASED_CONTENT_MAX_BYTES),
@@ -157 +157 @@ class AppConfig:
- loop: LoopConfig = field(default_factory=LoopConfig)
+ worker: WorkerConfig = field(default_factory=WorkerConfig)
@@ -169 +169 @@ class AppConfig:
- loop=LoopConfig.from_env(),
+ worker=WorkerConfig.from_env(),
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index dd00b5a7..942e7dbe 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -26 +26 @@ from packaging import version
-from worker.config import DatasetsBasedConfig
+from worker.config import WorkerConfig
@@ -160 +160 @@ class JobRunner(ABC):
- datasets_based_config: DatasetsBasedConfig
+ worker_config: WorkerConfig
@@ -178 +178 @@ class JobRunner(ABC):
- datasets_based_config: DatasetsBasedConfig,
+ worker_config: WorkerConfig,
@@ -189 +189 @@ class JobRunner(ABC):
- self.datasets_based_config = datasets_based_config
+ self.worker_config = worker_config
@@ -326 +326 @@ class JobRunner(ABC):
- if len(orjson_dumps(content)) > self.datasets_based_config.content_max_bytes:
+ if len(orjson_dumps(content)) > self.worker_config.content_max_bytes:
@@ -329 +329 @@ class JobRunner(ABC):
- f" ({self.datasets_based_config.content_max_bytes})."
+ f" ({self.worker_config.content_max_bytes})."
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 89cf577c..d793ccfa 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -99 +99 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- datasets_based_config=self.app_config.datasets_based,
+ worker_config=self.app_config.worker,
@@ -106 +106 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- datasets_based_config=self.app_config.datasets_based,
+ worker_config=self.app_config.worker,
@@ -113 +113 @@ class JobRunnerFactory(BaseJobRunnerFactory):
- datasets_based_config=self.app_config.datasets_based,
+ worker_config=self.app_config.worker,
diff --git a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
index 82119cbb..70751adb 100644
--- a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
+++ b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
@@ -37 +37 @@ class DatasetsBasedJobRunner(JobRunner):
- datasets_based_config=app_config.datasets_based,
+ worker_config=app_config.worker,
diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py
index 7860ab36..32b11a56 100644
--- a/services/worker/src/worker/loop.py
+++ b/services/worker/src/worker/loop.py
@@ -12 +12 @@ from psutil import cpu_count, disk_usage, getloadavg, swap_memory, virtual_memor
-from worker.config import LoopConfig
+from worker.config import WorkerConfig
@@ -32,2 +32,2 @@ class Loop:
- loop_config (`LoopConfig`):
- Loop configuration.
+ worker_config (`WorkerConfig`):
+ Worker configuration.
@@ -39 +39 @@ class Loop:
- loop_config: LoopConfig
+ worker_config: WorkerConfig
@@ -44 +44 @@ class Loop:
- self.storage_paths = set(self.loop_config.storage_paths).union(self.library_cache_paths)
+ self.storage_paths = set(self.worker_config.storage_paths).union(self.library_cache_paths)
@@ -62 +62 @@ class Loop:
- if self.loop_config.max_memory_pct <= 0:
+ if self.worker_config.max_memory_pct <= 0:
@@ -67 +67 @@ class Loop:
- ok = percent < self.loop_config.max_memory_pct
+ ok = percent < self.worker_config.max_memory_pct
@@ -70 +70 @@ class Loop:
- f"memory usage (RAM + SWAP) is too high: {percent:.0f}% - max is {self.loop_config.max_memory_pct}%"
+ f"memory usage (RAM + SWAP) is too high: {percent:.0f}% - max is {self.worker_config.max_memory_pct}%"
@@ -75 +75 @@ class Loop:
- if self.loop_config.max_load_pct <= 0:
+ if self.worker_config.max_load_pct <= 0:
@@ -79 +79 @@ class Loop:
- ok = load_pct < self.loop_config.max_load_pct
+ ok = load_pct < self.worker_config.max_load_pct
@@ -81 +81 @@ class Loop:
- self.info(f"cpu load is too high: {load_pct:.0f}% - max is {self.loop_config.max_load_pct}%")
+ self.info(f"cpu load is too high: {load_pct:.0f}% - max is {self.worker_config.max_load_pct}%")
@@ -85 +85 @@ class Loop:
- if self.loop_config.max_disk_usage_pct <= 0:
+ if self.worker_config.max_disk_usage_pct <= 0:
@@ -90 +90 @@ class Loop:
- if usage.percent >= self.loop_config.max_disk_usage_pct:
+ if usage.percent >= self.worker_config.max_disk_usage_pct:
@@ -103 +103 @@ class Loop:
- duration = self.loop_config.sleep_seconds * jitter
+ duration = self.worker_config.sleep_seconds * jitter
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index 6de77fae..b1ae25d6 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -23 +23 @@ if __name__ == "__main__":
- processing_step = processing_graph.get_step(app_config.datasets_based.endpoint)
+ processing_step = processing_graph.get_step(app_config.worker.endpoint)
@@ -54 +54 @@ if __name__ == "__main__":
- loop_config=app_config.loop,
+ worker_config=app_config.worker,
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index 54048d0b..e8686eca 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -56 +56 @@ def set_env_vars(datasets_cache_directory: Path, modules_cache_directory: Path)
- mp.setenv("DATASETS_BASED_CONTENT_MAX_BYTES", "10_000_000")
+ mp.setenv("WORKER_CONTENT_MAX_BYTES", "10_000_000")
diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py
index 275e48bc..108c67d2 100644
--- a/services/worker/tests/job_runners/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__datasets_based_worker.py
@@ -174 +174 @@ def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, g
- app_config=replace(app_config, datasets_based=replace(app_config.datasets_based, content_max_bytes=10)),
+ app_config=replace(app_config, worker=replace(app_config.worker, content_max_bytes=10)),
diff --git a/services/worker/tests/job_runners/test_dataset_info.py b/services/worker/tests/job_runners/test_dataset_info.py
index a0dfd692..a5010afb 100644
--- a/services/worker/tests/job_runners/test_dataset_info.py
+++ b/services/worker/tests/job_runners/test_dataset_info.py
@@ -49 +49 @@ def get_job_runner(
- datasets_based_config=app_config.datasets_based,
+ worker_config=app_config.worker,
diff --git a/services/worker/tests/job_runners/test_parquet.py b/services/worker/tests/job_runners/test_parquet.py
index 16b7a25c..ac4e4a07 100644
--- a/services/worker/tests/job_runners/test_parquet.py
+++ b/services/worker/tests/job_runners/test_parquet.py
@@ -49 +49 @@ def get_job_runner(
- datasets_based_config=app_config.datasets_based,
+ worker_config=app_config.worker,
diff --git a/services/worker/tests/job_runners/test_sizes.py b/services/worker/tests/job_runners/test_sizes.py
index 83326b97..858fe019 100644
--- a/services/worker/tests/job_runners/test_sizes.py
+++ b/services/worker/tests/job_runners/test_sizes.py
@@ -49 +49 @@ def get_job_runner(
- datasets_based_config=app_config.datasets_based,
+ worker_config=app_config.worker,
diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py
index fa0ca682..686acc72 100644
--- a/services/worker/tests/test_job_runner.py
+++ b/services/worker/tests/test_job_runner.py
@@ -12 +12 @@ from libcommon.simple_cache import SplitFullName, upsert_response
-from worker.config import DatasetsBasedConfig
+from worker.config import WorkerConfig
@@ -81 +81 @@ def test_compare_major_version(
- datasets_based_config=DatasetsBasedConfig(),
+ worker_config=WorkerConfig(),
@@ -199 +199 @@ def test_should_skip_job(
- datasets_based_config=DatasetsBasedConfig(),
+ worker_config=WorkerConfig(),
@@ -240 +240 @@ def test_check_type(
- datasets_based_config=DatasetsBasedConfig(),
+ worker_config=WorkerConfig(),
@@ -265 +265 @@ def test_check_type(
- datasets_based_config=DatasetsBasedConfig(),
+ worker_config=WorkerConfig(),
@@ -291 +291 @@ def test_create_children_jobs() -> None:
- datasets_based_config=DatasetsBasedConfig(),
+ worker_config=WorkerConfig(),
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 39012540..59058ee4 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -8 +8 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from worker.config import AppConfig, DatasetsBasedConfig, LoopConfig
+from worker.config import AppConfig, WorkerConfig
@@ -35 +35 @@ class DummyJobRunnerFactory(BaseJobRunnerFactory):
- self.datasets_based_config = DatasetsBasedConfig()
+ self.worker_config = WorkerConfig()
@@ -42 +42 @@ class DummyJobRunnerFactory(BaseJobRunnerFactory):
- datasets_based_config=self.datasets_based_config,
+ worker_config=self.worker_config,
@@ -60 +60 @@ def test_process_next_job(
- loop_config=LoopConfig(),
+ worker_config=WorkerConfig(),
diff --git a/tools/docker-compose-base.yml b/tools/docker-compose-base.yml
index cb87973b..7da735f2 100644
--- a/tools/docker-compose-base.yml
+++ b/tools/docker-compose-base.yml
@@ -19,4 +19,5 @@ services:
- WORKER_LOOP_MAX_DISK_USAGE_PCT: ${WORKER_LOOP_MAX_DISK_USAGE_PCT-90}
- WORKER_LOOP_MAX_LOAD_PCT: ${WORKER_LOOP_MAX_LOAD_PCT-70}
- WORKER_LOOP_MAX_MEMORY_PCT: ${WORKER_LOOP_MAX_MEMORY_PCT-80}
- WORKER_LOOP_SLEEP_SECONDS: ${WORKER_LOOP_SLEEP_SECONDS-15}
+ WORKER_CONTENT_MAX_BYTES: ${WORKER_CONTENT_MAX_BYTES-10_000_000}
+ WORKER_MAX_DISK_USAGE_PCT: ${WORKER_MAX_DISK_USAGE_PCT-90}
+ WORKER_MAX_LOAD_PCT: ${WORKER_MAX_LOAD_PCT-70}
+ WORKER_MAX_MEMORY_PCT: ${WORKER_MAX_MEMORY_PCT-80}
+ WORKER_SLEEP_SECONDS: ${WORKER_SLEEP_SECONDS-15}
@@ -29 +29,0 @@ services:
- DATASETS_BASED_CONTENT_MAX_BYTES: ${DATASETS_BASED_CONTENT_MAX_BYTES-10_000_000}
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index 676cc299..f3a81dcc 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -77 +76,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -87 +86 @@ services:
- DATASETS_BASED_ENDPOINT: "/config-names" # hard-coded
+ WORKER_ENDPOINT: "/config-names" # hard-coded
@@ -95 +93,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -105 +103 @@ services:
- DATASETS_BASED_ENDPOINT: "/split-names" # hard-coded
+ WORKER_ENDPOINT: "/split-names" # hard-coded
@@ -113 +110,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -123 +120 @@ services:
- DATASETS_BASED_ENDPOINT: "/splits" # hard-coded
+ WORKER_ENDPOINT: "/splits" # hard-coded
@@ -131 +127,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -143 +139 @@ services:
- DATASETS_BASED_ENDPOINT: "/first-rows" # hard-coded
+ WORKER_ENDPOINT: "/first-rows" # hard-coded
@@ -149 +145 @@ services:
- WORKER_LOOP_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ WORKER_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
@@ -158 +153,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -168 +163 @@ services:
- DATASETS_BASED_ENDPOINT: "/parquet-and-dataset-info" # hard-coded
+ WORKER_ENDPOINT: "/parquet-and-dataset-info" # hard-coded
@@ -184 +178,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -191 +185 @@ services:
- DATASETS_BASED_ENDPOINT: "/parquet" # hard-coded
+ WORKER_ENDPOINT: "/parquet" # hard-coded
@@ -199 +192,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -206 +199 @@ services:
- DATASETS_BASED_ENDPOINT: "/dataset-info" # hard-coded
+ WORKER_ENDPOINT: "/dataset-info" # hard-coded
@@ -214 +206,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -221 +213 @@ services:
- DATASETS_BASED_ENDPOINT: "/sizes" # hard-coded
+ WORKER_ENDPOINT: "/sizes" # hard-coded
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
index 4b488ce9..617997bb 100644
--- a/tools/docker-compose-dev-base.yml
+++ b/tools/docker-compose-dev-base.yml
@@ -19,4 +19,5 @@ services:
- WORKER_LOOP_MAX_DISK_USAGE_PCT: ${WORKER_LOOP_MAX_DISK_USAGE_PCT-90}
- WORKER_LOOP_MAX_LOAD_PCT: ${WORKER_LOOP_MAX_LOAD_PCT-70}
- WORKER_LOOP_MAX_MEMORY_PCT: ${WORKER_LOOP_MAX_MEMORY_PCT-80}
- WORKER_LOOP_SLEEP_SECONDS: ${WORKER_LOOP_SLEEP_SECONDS-15}
+ WORKER_CONTENT_MAX_BYTES: ${WORKER_CONTENT_MAX_BYTES-10_000_000}
+ WORKER_MAX_DISK_USAGE_PCT: ${WORKER_MAX_DISK_USAGE_PCT-90}
+ WORKER_MAX_LOAD_PCT: ${WORKER_MAX_LOAD_PCT-70}
+ WORKER_MAX_MEMORY_PCT: ${WORKER_MAX_MEMORY_PCT-80}
+ WORKER_SLEEP_SECONDS: ${WORKER_SLEEP_SECONDS-15}
@@ -29 +29,0 @@ services:
- DATASETS_BASED_CONTENT_MAX_BYTES: ${DATASETS_BASED_CONTENT_MAX_BYTES-10_000_000}
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 208b0a4d..d8cb0633 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -77 +76,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -87 +86 @@ services:
- DATASETS_BASED_ENDPOINT: "/config-names" # hard-coded
+ WORKER_ENDPOINT: "/config-names" # hard-coded
@@ -95 +93,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -105 +103 @@ services:
- DATASETS_BASED_ENDPOINT: "/split-names" # hard-coded
+ WORKER_ENDPOINT: "/split-names" # hard-coded
@@ -113 +110,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -123 +120 @@ services:
- DATASETS_BASED_ENDPOINT: "/splits" # hard-coded
+ WORKER_ENDPOINT: "/splits" # hard-coded
@@ -131 +127,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -143 +139 @@ services:
- DATASETS_BASED_ENDPOINT: "/first-rows" # hard-coded
+ WORKER_ENDPOINT: "/first-rows" # hard-coded
@@ -149 +145 @@ services:
- WORKER_LOOP_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ WORKER_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
@@ -158 +153,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -168 +163 @@ services:
- DATASETS_BASED_ENDPOINT: "/parquet-and-dataset-info" # hard-coded
+ WORKER_ENDPOINT: "/parquet-and-dataset-info" # hard-coded
@@ -184 +178,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -191 +185 @@ services:
- DATASETS_BASED_ENDPOINT: "/parquet" # hard-coded
+ WORKER_ENDPOINT: "/parquet" # hard-coded
@@ -199 +192,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -206 +199 @@ services:
- DATASETS_BASED_ENDPOINT: "/dataset-info" # hard-coded
+ WORKER_ENDPOINT: "/dataset-info" # hard-coded
@@ -214 +206,0 @@ services:
- # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
@@ -221 +213 @@ services:
- DATASETS_BASED_ENDPOINT: "/sizes" # hard-coded
+ WORKER_ENDPOINT: "/sizes" # hard-coded
|
|
f43a0d23c2d84e22c18547ec002c2d6160e9d9e8
|
Sylvain Lesage
| 2023-02-13T08:30:34 |
Move workers/datasets_based to services/worker (#800)
|
diff --git a/.github/workflows/build_push_docker_hub.yml b/.github/workflows/build_push_docker_hub.yml
index 19e91ad7..dd1ac06e 100644
--- a/.github/workflows/build_push_docker_hub.yml
+++ b/.github/workflows/build_push_docker_hub.yml
@@ -23,2 +23,2 @@ jobs:
- - directory: workers
- project: datasets_based
+ - directory: services
+ project: worker
diff --git a/.github/workflows/e2e.yml b/.github/workflows/e2e.yml
index 8ba34a94..81a6ad18 100644
--- a/.github/workflows/e2e.yml
+++ b/.github/workflows/e2e.yml
@@ -14 +13,0 @@ on:
- - 'workers/**'
@@ -26 +24,0 @@ on:
- - 'workers/**'
diff --git a/.github/workflows/w-datasets_based.yml b/.github/workflows/s-worker.yml
similarity index 72%
rename from .github/workflows/w-datasets_based.yml
rename to .github/workflows/s-worker.yml
index 2082de4a..20f562e9 100644
--- a/.github/workflows/w-datasets_based.yml
+++ b/.github/workflows/s-worker.yml
@@ -4 +4 @@
-name: workers/datasets_based
+name: services/worker
@@ -12,2 +12,2 @@ on:
- - 'workers/datasets_based/**'
- - '.github/workflows/w-datasets_based.yml'
+ - 'services/worker/**'
+ - '.github/workflows/s-worker.yml'
@@ -21,2 +21,2 @@ on:
- - 'workers/datasets_based/**'
- - '.github/workflows/w-datasets_based.yml'
+ - 'services/worker/**'
+ - '.github/workflows/s-worker.yml'
@@ -31 +31 @@ jobs:
- working-directory: workers/datasets_based
+ working-directory: services/worker
@@ -36 +36 @@ jobs:
- working-directory: workers/datasets_based
+ working-directory: services/worker
diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace
index b65964e1..ad6493aa 100644
--- a/.vscode/monorepo.code-workspace
+++ b/.vscode/monorepo.code-workspace
@@ -28,2 +28,2 @@
- "name": "services/reverse-proxy",
- "path": "../services/reverse-proxy"
+ "name": "services/worker",
+ "path": "../services/worker"
@@ -32,2 +32,2 @@
- "name": "workers/datasets_based",
- "path": "../workers/datasets_based"
+ "name": "services/reverse-proxy",
+ "path": "../services/reverse-proxy"
@@ -41,2 +41 @@
- "services": true,
- "workers": true
+ "services": true
diff --git a/DEVELOPER_GUIDE.md b/DEVELOPER_GUIDE.md
index dc0be457..d80c95b8 100644
--- a/DEVELOPER_GUIDE.md
+++ b/DEVELOPER_GUIDE.md
@@ -31 +31 @@ You can just restart the worker's docker container and it will apply your change
-To install a single job (in [jobs](./jobs)), library (in [libs](./libs)), service (in [services](./services)) or worker (in [workers](./workers)), go to their respective directory, and install Python 3.9 (consider [pyenv](https://github.com/pyenv/pyenv)) and [poetry](https://python-poetry.org/docs/master/#installation) (don't forget to add `poetry` to the `PATH` environment variable).
+To install a single job (in [jobs](./jobs)), library (in [libs](./libs)) or service (in [services](./services)), go to their respective directory, and install Python 3.9 (consider [pyenv](https://github.com/pyenv/pyenv)) and [poetry](https://python-poetry.org/docs/master/#installation) (don't forget to add `poetry` to the `PATH` environment variable).
@@ -54 +54 @@ If you use VSCode, it might be useful to use the ["monorepo" workspace](./.vscod
-The repository is structured as a monorepo, with Python libraries and applications in [jobs](./jobs)), [libs](./libs), [services](./services) and [workers](./workers):
+The repository is structured as a monorepo, with Python libraries and applications in [jobs](./jobs)), [libs](./libs) and [services](./services):
@@ -58,2 +58 @@ The repository is structured as a monorepo, with Python libraries and applicatio
-- [services](./services) contains the applications: the public API, the admin API (which is separated from the public API and might be published under its own domain at some point) and the reverse proxy.
-- [workers](./workers) contains the workers that process the queue asynchronously: they get a "job" (caution: not the Helm jobs, but the jobs stored in the queue), process the expected response for the associated endpoint, and store the response in the cache.
+- [services](./services) contains the applications: the public API, the admin API (which is separated from the public API and might be published under its own domain at some point), the reverse proxy, and the worker that processes the queue asynchronously: it gets a "job" (caution: the jobs stored in the queue, not the Helm jobs), processes the expected response for the associated endpoint, and stores the response in the cache.
@@ -67 +66 @@ The application is distributed in several components.
-The precomputed responses are stored in a Mongo database called "cache". They are computed by [workers](./workers) which take their jobs from a job queue stored in a Mongo database called "queue", and store the results (error or valid response) into the "cache" (see [libcommon](./libs/libcommon)).
+The precomputed responses are stored in a Mongo database called "cache". They are computed by [workers](./services/worker) which take their jobs from a job queue stored in a Mongo database called "queue", and store the results (error or valid response) into the "cache" (see [libcommon](./libs/libcommon)).
@@ -159 +158 @@ GITHUB_TOKEN=xxx
-To install the [datasets based worker](./workers/datasets_based) on Mac OS, you can follow the next steps.
+To install the [datasets based worker](./services/worker) on Mac OS, you can follow the next steps.
@@ -222 +221 @@ Check that the expected local version of Python is used:
-$ cd workers/datasets_based
+$ cd services/worker
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index 7ff95ac6..7fb0392a 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -41,2 +41 @@ images:
- workers:
- datasetsBased:
+ worker:
@@ -45 +44 @@ images:
- repository: datasets-server-workers-datasets_based
+ repository: datasets-server-services-worker
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 337db015..13bdb598 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -31,2 +31 @@ images:
- workers:
- datasetsBased:
+ worker:
@@ -35 +34 @@ images:
- repository: datasets-server-workers-datasets_based
+ repository: datasets-server-services-worker
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index 3efacf03..84a19bf8 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -86 +86 @@ imagePullSecrets:
-{{ include "datasetsServer.images.image" (dict "imageRoot" .Values.images.workers.datasetsBased "global" .Values.global.huggingface) }}
+{{ include "datasetsServer.images.image" (dict "imageRoot" .Values.images.services.worker "global" .Values.global.huggingface) }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 30a1914d..3b1aff69 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -38,2 +38 @@ images:
- workers:
- datasetsBased:
+ worker:
@@ -42 +41 @@ images:
- repository: datasets-server-workers-datasets_based
+ repository: datasets-server-services-worker
diff --git a/docs/source/server.mdx b/docs/source/server.mdx
index 8a1d8a9c..3430b77e 100644
--- a/docs/source/server.mdx
+++ b/docs/source/server.mdx
@@ -28 +28 @@ Workers are responsible for executing the jobs in the queue. They complete the a
-Take a look at the [workers configuration](https://github.com/huggingface/datasets-server/tree/main/workers/datasets_based#configuration) for a complete list of the environment variables if you're interested in learning more.
+Take a look at the [workers configuration](https://github.com/huggingface/datasets-server/tree/main/services/worker#configuration) for a complete list of the environment variables if you're interested in learning more.
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index 12a65d67..f98f4f4e 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -547 +547 @@ version = "0.6.8"
-description = "Library for utils, common to all the services and workers"
+description = "Library for utils common to all the services"
diff --git a/jobs/mongodb_migration/src/mongodb_migration/main.py b/jobs/mongodb_migration/src/mongodb_migration/main.py
index 0f15ca0c..a5f41e6a 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/main.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/main.py
@@ -38 +38 @@ def run_job() -> None:
- if queue_resource.is_available() is False:
+ if not queue_resource.is_available():
@@ -43 +43 @@ def run_job() -> None:
- if migrations_database_resource.is_available() is False:
+ if not migrations_database_resource.is_available():
diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml
index df1074c0..d38c50ab 100644
--- a/libs/libcommon/pyproject.toml
+++ b/libs/libcommon/pyproject.toml
@@ -3 +3 @@ authors = ["Sylvain Lesage <[email protected]>"]
-description = "Library for utils, common to all the services and workers"
+description = "Library for utils common to all the services"
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 285bc086..6f444bd1 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -228 +228,2 @@ def test_get_total_duration_per_dataset() -> None:
- assert queue.get_total_duration_per_dataset() == {test_dataset: duration * 3}
+ assert queue.get_total_duration_per_dataset()[test_dataset] >= duration * 3
+ # ^ it should be equal, not >=, but if the runner is slow, it might take a bit more time
diff --git a/libs/libcommon/tests/test_resources.py b/libs/libcommon/tests/test_resources.py
index e93cf39b..48450db8 100644
--- a/libs/libcommon/tests/test_resources.py
+++ b/libs/libcommon/tests/test_resources.py
@@ -24 +24 @@ def test_database_resource(queue_mongo_host: str) -> None:
- server_selection_timeout_ms = 1_000
+ server_selection_timeout_ms = 5_000
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index 47bc9eb4..86171519 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -595 +595 @@ version = "0.6.8"
-description = "Library for utils, common to all the services and workers"
+description = "Library for utils common to all the services"
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index dd6111df..f9855ee0 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -615 +615 @@ version = "0.6.8"
-description = "Library for utils, common to all the services and workers"
+description = "Library for utils common to all the services"
diff --git a/workers/datasets_based/.flake8 b/services/worker/.flake8
similarity index 100%
rename from workers/datasets_based/.flake8
rename to services/worker/.flake8
diff --git a/workers/datasets_based/.python-version b/services/worker/.python-version
similarity index 100%
rename from workers/datasets_based/.python-version
rename to services/worker/.python-version
diff --git a/workers/datasets_based/Dockerfile b/services/worker/Dockerfile
similarity index 68%
rename from workers/datasets_based/Dockerfile
rename to services/worker/Dockerfile
index 80ac0507..93882054 100644
--- a/workers/datasets_based/Dockerfile
+++ b/services/worker/Dockerfile
@@ -28,3 +28,3 @@ WORKDIR /src
-COPY workers/datasets_based/vendors ./workers/datasets_based/vendors/
-COPY workers/datasets_based/poetry.lock ./workers/datasets_based/poetry.lock
-COPY workers/datasets_based/pyproject.toml ./workers/datasets_based/pyproject.toml
+COPY services/worker/vendors ./services/worker/vendors/
+COPY services/worker/poetry.lock ./services/worker/poetry.lock
+COPY services/worker/pyproject.toml ./services/worker/pyproject.toml
@@ -32 +32 @@ COPY libs/libcommon ./libs/libcommon
-WORKDIR /src/workers/datasets_based/
+WORKDIR /src/services/worker/
@@ -34 +34 @@ RUN poetry install --no-cache
-COPY workers/datasets_based/src ./src
+COPY services/worker/src ./src
@@ -37 +37 @@ RUN poetry install --no-cache
-ENTRYPOINT ["poetry", "run", "python", "src/datasets_based/main.py"]
+ENTRYPOINT ["poetry", "run", "python", "src/worker/main.py"]
diff --git a/workers/datasets_based/Makefile b/services/worker/Makefile
similarity index 81%
rename from workers/datasets_based/Makefile
rename to services/worker/Makefile
index 8c09fd2f..45cb01a7 100644
--- a/workers/datasets_based/Makefile
+++ b/services/worker/Makefile
@@ -2 +2 @@
-export COMPOSE_PROJECT_NAME := datasets_based
+export COMPOSE_PROJECT_NAME := worker
@@ -15 +15 @@ run:
- poetry run python src/datasets_based/main.py
+ poetry run python src/worker/main.py
diff --git a/workers/datasets_based/README.md b/services/worker/README.md
similarity index 100%
rename from workers/datasets_based/README.md
rename to services/worker/README.md
diff --git a/workers/datasets_based/dev.Dockerfile b/services/worker/dev.Dockerfile
similarity index 76%
rename from workers/datasets_based/dev.Dockerfile
rename to services/worker/dev.Dockerfile
index 7b9d54bf..4fd090fd 100644
--- a/workers/datasets_based/dev.Dockerfile
+++ b/services/worker/dev.Dockerfile
@@ -28,3 +28,3 @@ WORKDIR /src
-COPY workers/datasets_based/vendors ./workers/datasets_based/vendors/
-COPY workers/datasets_based/poetry.lock ./workers/datasets_based/poetry.lock
-COPY workers/datasets_based/pyproject.toml ./workers/datasets_based/pyproject.toml
+COPY services/worker/vendors ./services/worker/vendors/
+COPY services/worker/poetry.lock ./services/worker/poetry.lock
+COPY services/worker/pyproject.toml ./services/worker/pyproject.toml
@@ -32 +32 @@ COPY libs/libcommon ./libs/libcommon
-WORKDIR /src/workers/datasets_based/
+WORKDIR /src/services/worker/
@@ -37 +37 @@ RUN poetry install --no-cache
-# Removed: COPY workers/datasets_based/src ./src
+# Removed: COPY services/worker/src ./src
@@ -41 +41 @@ RUN poetry install --no-cache
-ENTRYPOINT ["/bin/sh", "-c" , "poetry install && poetry run python src/datasets_based/main.py"]
+ENTRYPOINT ["/bin/sh", "-c" , "poetry install && poetry run python src/worker/main.py"]
diff --git a/workers/datasets_based/poetry.lock b/services/worker/poetry.lock
similarity index 99%
rename from workers/datasets_based/poetry.lock
rename to services/worker/poetry.lock
index 2f85f69a..789ac4de 100644
--- a/workers/datasets_based/poetry.lock
+++ b/services/worker/poetry.lock
@@ -1772 +1772 @@ version = "0.6.8"
-description = "Library for utils, common to all the services and workers"
+description = "Library for utils common to all the services"
@@ -3064 +3064 @@ name = "py7zr"
-version = "0.20.2"
+version = "0.20.4"
@@ -3070,2 +3070,2 @@ files = [
- {file = "py7zr-0.20.2-py3-none-any.whl", hash = "sha256:f6615a5bb07a9252034f23e518affcf4f62725c3632f23b7cfee86aef8bae779"},
- {file = "py7zr-0.20.2.tar.gz", hash = "sha256:791ef912a295b61b91c5fe0c23adeddb80bf13500308062c082b8fec6c8c9653"},
+ {file = "py7zr-0.20.4-py3-none-any.whl", hash = "sha256:94d0c24217f6582741813ee94490a4ca82bd5f9bf35e4f8610cb588cf7445764"},
+ {file = "py7zr-0.20.4.tar.gz", hash = "sha256:1d01f98ea1e1f5c49940358691b2076f9a5848056426541e783de33834f59e21"},
@@ -3079 +3079 @@ multivolumefile = ">=0.2.3"
-psutil = "*"
+psutil = {version = "*", markers = "sys_platform != \"cygwin\""}
diff --git a/workers/datasets_based/poetry.toml b/services/worker/poetry.toml
similarity index 100%
rename from workers/datasets_based/poetry.toml
rename to services/worker/poetry.toml
diff --git a/workers/datasets_based/pyproject.toml b/services/worker/pyproject.toml
similarity index 94%
rename from workers/datasets_based/pyproject.toml
rename to services/worker/pyproject.toml
index 946ae1e0..b425f0a2 100644
--- a/workers/datasets_based/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -3,2 +3,2 @@ authors = ["Sylvain Lesage <[email protected]>"]
-description = "Worker for processing steps that need the datasets library"
-name = "datasets_based"
+description = "Worker that processes jobs and stores the responses in the cache"
+name = "worker"
@@ -68 +68 @@ markers = [
-source = ["datasets_based"]
+source = ["worker"]
diff --git a/workers/datasets_based/src/datasets_based/__init__.py b/services/worker/src/worker/__init__.py
similarity index 100%
rename from workers/datasets_based/src/datasets_based/__init__.py
rename to services/worker/src/worker/__init__.py
diff --git a/workers/datasets_based/src/datasets_based/asset.py b/services/worker/src/worker/asset.py
similarity index 100%
rename from workers/datasets_based/src/datasets_based/asset.py
rename to services/worker/src/worker/asset.py
diff --git a/workers/datasets_based/src/datasets_based/config.py b/services/worker/src/worker/config.py
similarity index 97%
rename from workers/datasets_based/src/datasets_based/config.py
rename to services/worker/src/worker/config.py
index fc3dd81d..1cacd1c3 100644
--- a/workers/datasets_based/src/datasets_based/config.py
+++ b/services/worker/src/worker/config.py
@@ -27 +27 @@ def get_empty_str_list() -> List[str]:
-class WorkerLoopConfig:
+class LoopConfig:
@@ -35 +35 @@ class WorkerLoopConfig:
- def from_env(cls) -> "WorkerLoopConfig":
+ def from_env(cls) -> "LoopConfig":
@@ -157 +157 @@ class AppConfig:
- worker_loop: WorkerLoopConfig = field(default_factory=WorkerLoopConfig)
+ loop: LoopConfig = field(default_factory=LoopConfig)
@@ -169 +169 @@ class AppConfig:
- worker_loop=WorkerLoopConfig.from_env(),
+ loop=LoopConfig.from_env(),
diff --git a/workers/datasets_based/src/datasets_based/features.py b/services/worker/src/worker/features.py
similarity index 99%
rename from workers/datasets_based/src/datasets_based/features.py
rename to services/worker/src/worker/features.py
index 1710b1e2..aafe6460 100644
--- a/workers/datasets_based/src/datasets_based/features.py
+++ b/services/worker/src/worker/features.py
@@ -25 +25 @@ from PIL import Image as PILImage # type: ignore
-from datasets_based.asset import create_audio_files, create_image_file
+from worker.asset import create_audio_files, create_image_file
diff --git a/workers/datasets_based/src/datasets_based/worker.py b/services/worker/src/worker/job_runner.py
similarity index 90%
rename from workers/datasets_based/src/datasets_based/worker.py
rename to services/worker/src/worker/job_runner.py
index 3268c9b0..dd00b5a7 100644
--- a/workers/datasets_based/src/datasets_based/worker.py
+++ b/services/worker/src/worker/job_runner.py
@@ -26 +26 @@ from packaging import version
-from datasets_based.config import DatasetsBasedConfig
+from worker.config import DatasetsBasedConfig
@@ -28 +28 @@ from datasets_based.config import DatasetsBasedConfig
-GeneralWorkerErrorCode = Literal[
+GeneralJobRunnerErrorCode = Literal[
@@ -40,2 +40,2 @@ ERROR_CODES_TO_RETRY: list[str] = ["ClientConnectionError"]
-class WorkerError(CustomError):
- """Base class for worker exceptions."""
+class JobRunnerError(CustomError):
+ """Base class for job runner exceptions."""
@@ -56,2 +56,2 @@ class WorkerError(CustomError):
-class GeneralWorkerError(WorkerError):
- """Base class for worker exceptions."""
+class GeneralJobRunnerError(JobRunnerError):
+ """General class for job runner exceptions."""
@@ -63 +63 @@ class GeneralWorkerError(WorkerError):
- code: GeneralWorkerErrorCode,
+ code: GeneralJobRunnerErrorCode,
@@ -72 +72 @@ class GeneralWorkerError(WorkerError):
-class ConfigNotFoundError(GeneralWorkerError):
+class ConfigNotFoundError(GeneralJobRunnerError):
@@ -85 +85 @@ class ConfigNotFoundError(GeneralWorkerError):
-class SplitNotFoundError(GeneralWorkerError):
+class SplitNotFoundError(GeneralJobRunnerError):
@@ -98 +98 @@ class SplitNotFoundError(GeneralWorkerError):
-class NoGitRevisionError(GeneralWorkerError):
+class NoGitRevisionError(GeneralJobRunnerError):
@@ -111 +111 @@ class NoGitRevisionError(GeneralWorkerError):
-class TooBigContentError(GeneralWorkerError):
+class TooBigContentError(GeneralJobRunnerError):
@@ -124,2 +124,2 @@ class TooBigContentError(GeneralWorkerError):
-class UnexpectedError(GeneralWorkerError):
- """Raised when the worker raised an unexpected error."""
+class UnexpectedError(GeneralJobRunnerError):
+ """Raised when the job runner raised an unexpected error."""
@@ -138 +138 @@ class UnexpectedError(GeneralWorkerError):
-class Worker(ABC):
+class JobRunner(ABC):
@@ -140 +140 @@ class Worker(ABC):
- Base class for workers. A worker is a class that processes a job, for a specific processing step.
+ Base class for job runners. A job runner is a class that processes a job, for a specific processing step.
@@ -194,2 +194,2 @@ class Worker(ABC):
- worker_job_type = self.get_job_type()
- if self.processing_step.job_type != worker_job_type:
+ job_type = self.get_job_type()
+ if self.processing_step.job_type != job_type:
@@ -197,2 +197,2 @@ class Worker(ABC):
- f"The processing step's job type is {self.processing_step.job_type}, but the worker only processes"
- f" {worker_job_type}"
+ f"The processing step's job type is {self.processing_step.job_type}, but the job runner only processes"
+ f" {job_type}"
@@ -200 +200 @@ class Worker(ABC):
- if self.job_type != worker_job_type:
+ if self.job_type != job_type:
@@ -202 +202 @@ class Worker(ABC):
- f"The submitted job type is {self.job_type}, but the worker only processes {worker_job_type}"
+ f"The submitted job type is {self.job_type}, but the job runner only processes {job_type}"
@@ -243 +243 @@ class Worker(ABC):
- Compare the major version of worker's self version and the other version's.
+ Compare the major version of job runner's self version and the other version's.
@@ -250 +250,2 @@ class Worker(ABC):
- 0 if they are equal. Negative if worker's major version is lower than other_version, positive otherwise.
+ 0 if they are equal. Negative if job runner's major version is lower than other_version, positive
+ otherwise.
@@ -252 +253 @@ class Worker(ABC):
- :obj:`ValueError`: if worker's version or other_version is not a valid semantic version.
+ :obj:`ValueError`: if job runner's version or other_version is not a valid semantic version.
@@ -275 +276 @@ class Worker(ABC):
- - and the cached entry has been created with the same major version of the worker
+ - and the cached entry has been created with the same major version of the job runner
@@ -297 +298,4 @@ class Worker(ABC):
- # no worker version in the cache, or the worker has been updated - we process the job to update the cache
+ # no job runner version in the cache, or the job runner has been updated - we process the job to update
+ # the cache
+ # note: the collection field is named "worker_version" for historical reasons, it might be renamed
+ # "job_runner_version" in the future.
diff --git a/workers/datasets_based/src/datasets_based/worker_factory.py b/services/worker/src/worker/job_runner_factory.py
similarity index 56%
rename from workers/datasets_based/src/datasets_based/worker_factory.py
rename to services/worker/src/worker/job_runner_factory.py
index 87323203..89cf577c 100644
--- a/workers/datasets_based/src/datasets_based/worker_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -11,14 +11,10 @@ from libcommon.storage import StrPath
-from datasets_based.config import (
- AppConfig,
- FirstRowsConfig,
- ParquetAndDatasetInfoConfig,
-)
-from datasets_based.worker import JobInfo, Worker
-from datasets_based.workers.config_names import ConfigNamesWorker
-from datasets_based.workers.dataset_info import DatasetInfoWorker
-from datasets_based.workers.first_rows import FirstRowsWorker
-from datasets_based.workers.parquet import ParquetWorker
-from datasets_based.workers.parquet_and_dataset_info import ParquetAndDatasetInfoWorker
-from datasets_based.workers.sizes import SizesWorker
-from datasets_based.workers.split_names import SplitNamesWorker
-from datasets_based.workers.splits import SplitsWorker
+from worker.config import AppConfig, FirstRowsConfig, ParquetAndDatasetInfoConfig
+from worker.job_runner import JobInfo, JobRunner
+from worker.job_runners.config_names import ConfigNamesJobRunner
+from worker.job_runners.dataset_info import DatasetInfoJobRunner
+from worker.job_runners.first_rows import FirstRowsJobRunner
+from worker.job_runners.parquet import ParquetJobRunner
+from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoJobRunner
+from worker.job_runners.sizes import SizesJobRunner
+from worker.job_runners.split_names import SplitNamesJobRunner
+from worker.job_runners.splits import SplitsJobRunner
@@ -27 +23 @@ from datasets_based.workers.splits import SplitsWorker
-class BaseWorkerFactory(ABC):
+class BaseJobRunnerFactory(ABC):
@@ -29 +25 @@ class BaseWorkerFactory(ABC):
- Base class for worker factories. A worker factory is a class that creates a worker.
+ Base class for job runner factories. A job runner factory is a class that creates a job runner.
@@ -31,0 +28,2 @@ class BaseWorkerFactory(ABC):
+
+ Note that this class is only implemented once in the code, but we need it for the tests.
@@ -34,2 +32,2 @@ class BaseWorkerFactory(ABC):
- def create_worker(self, job_info: JobInfo) -> Worker:
- return self._create_worker(job_info=job_info)
+ def create_job_runner(self, job_info: JobInfo) -> JobRunner:
+ return self._create_job_runner(job_info=job_info)
@@ -38 +36 @@ class BaseWorkerFactory(ABC):
- def _create_worker(self, job_info: JobInfo) -> Worker:
+ def _create_job_runner(self, job_info: JobInfo) -> JobRunner:
@@ -43 +41 @@ class BaseWorkerFactory(ABC):
-class WorkerFactory(BaseWorkerFactory):
+class JobRunnerFactory(BaseJobRunnerFactory):
@@ -49 +47 @@ class WorkerFactory(BaseWorkerFactory):
- def _create_worker(self, job_info: JobInfo) -> Worker:
+ def _create_job_runner(self, job_info: JobInfo) -> JobRunner:
@@ -58,2 +56,2 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == ConfigNamesWorker.get_job_type():
- return ConfigNamesWorker(
+ if job_type == ConfigNamesJobRunner.get_job_type():
+ return ConfigNamesJobRunner(
@@ -65,2 +63,2 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == SplitNamesWorker.get_job_type():
- return SplitNamesWorker(
+ if job_type == SplitNamesJobRunner.get_job_type():
+ return SplitNamesJobRunner(
@@ -72,2 +70,2 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == SplitsWorker.get_job_type():
- return SplitsWorker(
+ if job_type == SplitsJobRunner.get_job_type():
+ return SplitsJobRunner(
@@ -79 +77 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == FirstRowsWorker.get_job_type():
+ if job_type == FirstRowsJobRunner.get_job_type():
@@ -81 +79 @@ class WorkerFactory(BaseWorkerFactory):
- return FirstRowsWorker(
+ return FirstRowsJobRunner(
@@ -89,2 +87,2 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == ParquetAndDatasetInfoWorker.get_job_type():
- return ParquetAndDatasetInfoWorker(
+ if job_type == ParquetAndDatasetInfoJobRunner.get_job_type():
+ return ParquetAndDatasetInfoJobRunner(
@@ -97,2 +95,2 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == ParquetWorker.get_job_type():
- return ParquetWorker(
+ if job_type == ParquetJobRunner.get_job_type():
+ return ParquetJobRunner(
@@ -104,2 +102,2 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == DatasetInfoWorker.get_job_type():
- return DatasetInfoWorker(
+ if job_type == DatasetInfoJobRunner.get_job_type():
+ return DatasetInfoJobRunner(
@@ -111,2 +109,2 @@ class WorkerFactory(BaseWorkerFactory):
- if job_type == SizesWorker.get_job_type():
- return SizesWorker(
+ if job_type == SizesJobRunner.get_job_type():
+ return SizesJobRunner(
@@ -119,8 +117,8 @@ class WorkerFactory(BaseWorkerFactory):
- ConfigNamesWorker.get_job_type(),
- SplitNamesWorker.get_job_type(),
- SplitsWorker.get_job_type(),
- FirstRowsWorker.get_job_type(),
- ParquetAndDatasetInfoWorker.get_job_type(),
- ParquetWorker.get_job_type(),
- DatasetInfoWorker.get_job_type(),
- SizesWorker.get_job_type(),
+ ConfigNamesJobRunner.get_job_type(),
+ SplitNamesJobRunner.get_job_type(),
+ SplitsJobRunner.get_job_type(),
+ FirstRowsJobRunner.get_job_type(),
+ ParquetAndDatasetInfoJobRunner.get_job_type(),
+ ParquetJobRunner.get_job_type(),
+ DatasetInfoJobRunner.get_job_type(),
+ SizesJobRunner.get_job_type(),
diff --git a/workers/datasets_based/src/datasets_based/workers/__init__.py b/services/worker/src/worker/job_runners/__init__.py
similarity index 100%
rename from workers/datasets_based/src/datasets_based/workers/__init__.py
rename to services/worker/src/worker/job_runners/__init__.py
diff --git a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
similarity index 91%
rename from workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
rename to services/worker/src/worker/job_runners/_datasets_based_job_runner.py
index 300abb8f..82119cbb 100644
--- a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
+++ b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
@@ -16,2 +16,2 @@ from libcommon.storage import init_dir, remove_dir
-from datasets_based.config import AppConfig, DatasetsBasedConfig
-from datasets_based.worker import JobInfo, Worker
+from worker.config import AppConfig, DatasetsBasedConfig
+from worker.job_runner import JobInfo, JobRunner
@@ -20,2 +20,2 @@ from datasets_based.worker import JobInfo, Worker
-class DatasetsBasedWorker(Worker):
- """Base class for workers that use datasets."""
+class DatasetsBasedJobRunner(JobRunner):
+ """Base class for job runners that use datasets."""
@@ -27 +27 @@ class DatasetsBasedWorker(Worker):
- # the worker should have only one running job at the same time, then it should
+ # the job runner should have only one running job at the same time, then it should
diff --git a/workers/datasets_based/src/datasets_based/workers/config_names.py b/services/worker/src/worker/job_runners/config_names.py
similarity index 84%
rename from workers/datasets_based/src/datasets_based/workers/config_names.py
rename to services/worker/src/worker/job_runners/config_names.py
index 8e02ff24..323dfe26 100644
--- a/workers/datasets_based/src/datasets_based/workers/config_names.py
+++ b/services/worker/src/worker/job_runners/config_names.py
@@ -12,2 +12,2 @@ from libcommon.simple_cache import SplitFullName
-from datasets_based.worker import WorkerError
-from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
+from worker.job_runner import JobRunnerError
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
@@ -15 +15 @@ from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
-ConfigNamesWorkerErrorCode = Literal["EmptyDatasetError", "ConfigNamesError"]
+ConfigNamesJobRunnerErrorCode = Literal["EmptyDatasetError", "ConfigNamesError"]
@@ -18,2 +18,2 @@ ConfigNamesWorkerErrorCode = Literal["EmptyDatasetError", "ConfigNamesError"]
-class ConfigNamesWorkerError(WorkerError):
- """Base class for worker exceptions."""
+class ConfigNamesJobRunnerError(JobRunnerError):
+ """Base class for job runner exceptions."""
@@ -25 +25 @@ class ConfigNamesWorkerError(WorkerError):
- code: ConfigNamesWorkerErrorCode,
+ code: ConfigNamesJobRunnerErrorCode,
@@ -34 +34 @@ class ConfigNamesWorkerError(WorkerError):
-class EmptyDatasetError(ConfigNamesWorkerError):
+class EmptyDatasetError(ConfigNamesJobRunnerError):
@@ -41 +41 @@ class EmptyDatasetError(ConfigNamesWorkerError):
-class ConfigNamesError(ConfigNamesWorkerError):
+class ConfigNamesError(ConfigNamesJobRunnerError):
@@ -77 +77 @@ def compute_config_names_response(
- - [`~workers.config_names.EmptyDatasetError`]
+ - [`~job_runners.config_names.EmptyDatasetError`]
@@ -79 +79 @@ def compute_config_names_response(
- - [`~workers.config_names.ConfigNamesError`]
+ - [`~job_runners.config_names.ConfigNamesError`]
@@ -98 +98 @@ def compute_config_names_response(
-class ConfigNamesWorker(DatasetsBasedWorker):
+class ConfigNamesJobRunner(DatasetsBasedJobRunner):
diff --git a/workers/datasets_based/src/datasets_based/workers/dataset_info.py b/services/worker/src/worker/job_runners/dataset_info.py
similarity index 87%
rename from workers/datasets_based/src/datasets_based/workers/dataset_info.py
rename to services/worker/src/worker/job_runners/dataset_info.py
index 39dcb5fa..b0bf602f 100644
--- a/workers/datasets_based/src/datasets_based/workers/dataset_info.py
+++ b/services/worker/src/worker/job_runners/dataset_info.py
@@ -11 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.worker import Worker, WorkerError
+from worker.job_runner import JobRunner, JobRunnerError
@@ -13 +13 @@ from datasets_based.worker import Worker, WorkerError
-DatasetInfoWorkerErrorCode = Literal[
+DatasetInfoJobRunnerErrorCode = Literal[
@@ -23 +23 @@ class DatasetInfoResponse(TypedDict):
-class DatasetInfoWorkerError(WorkerError):
+class DatasetInfoJobRunnerError(JobRunnerError):
@@ -30 +30 @@ class DatasetInfoWorkerError(WorkerError):
- code: DatasetInfoWorkerErrorCode,
+ code: DatasetInfoJobRunnerErrorCode,
@@ -39 +39 @@ class DatasetInfoWorkerError(WorkerError):
-class PreviousStepStatusError(DatasetInfoWorkerError):
+class PreviousStepStatusError(DatasetInfoJobRunnerError):
@@ -46 +46 @@ class PreviousStepStatusError(DatasetInfoWorkerError):
-class PreviousStepFormatError(DatasetInfoWorkerError):
+class PreviousStepFormatError(DatasetInfoJobRunnerError):
@@ -64 +64 @@ def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse:
- - [`~workers.dataset_info.PreviousStepStatusError`]
+ - [`~job_runners.dataset_info.PreviousStepStatusError`]
@@ -66 +66 @@ def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse:
- - [`~workers.dataset_info.PreviousStepFormatError`]
+ - [`~job_runners.dataset_info.PreviousStepFormatError`]
@@ -88 +88 @@ def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse:
-class DatasetInfoWorker(Worker):
+class DatasetInfoJobRunner(JobRunner):
diff --git a/workers/datasets_based/src/datasets_based/workers/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py
similarity index 94%
rename from workers/datasets_based/src/datasets_based/workers/first_rows.py
rename to services/worker/src/worker/job_runners/first_rows.py
index 3a1eba1b..7b88b67c 100644
--- a/workers/datasets_based/src/datasets_based/workers/first_rows.py
+++ b/services/worker/src/worker/job_runners/first_rows.py
@@ -28,3 +28,3 @@ from libcommon.utils import orjson_dumps
-from datasets_based.config import AppConfig, FirstRowsConfig
-from datasets_based.features import get_cell_value
-from datasets_based.worker import (
+from worker.config import AppConfig, FirstRowsConfig
+from worker.features import get_cell_value
+from worker.job_runner import (
@@ -32,0 +33 @@ from datasets_based.worker import (
+ JobRunnerError,
@@ -34 +34,0 @@ from datasets_based.worker import (
- WorkerError,
@@ -36 +36 @@ from datasets_based.worker import (
-from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
@@ -38 +38 @@ from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
-FirstRowsWorkerErrorCode = Literal[
+FirstRowsJobRunnerErrorCode = Literal[
@@ -51 +51 @@ FirstRowsWorkerErrorCode = Literal[
-class FirstRowsWorkerError(WorkerError):
+class FirstRowsJobRunnerError(JobRunnerError):
@@ -58 +58 @@ class FirstRowsWorkerError(WorkerError):
- code: FirstRowsWorkerErrorCode,
+ code: FirstRowsJobRunnerErrorCode,
@@ -67 +67 @@ class FirstRowsWorkerError(WorkerError):
-class SplitsNamesError(FirstRowsWorkerError):
+class SplitsNamesError(FirstRowsJobRunnerError):
@@ -74 +74 @@ class SplitsNamesError(FirstRowsWorkerError):
-class EmptyDatasetError(FirstRowsWorkerError):
+class EmptyDatasetError(FirstRowsJobRunnerError):
@@ -81 +81 @@ class EmptyDatasetError(FirstRowsWorkerError):
-class InfoError(FirstRowsWorkerError):
+class InfoError(FirstRowsJobRunnerError):
@@ -88 +88 @@ class InfoError(FirstRowsWorkerError):
-class FeaturesError(FirstRowsWorkerError):
+class FeaturesError(FirstRowsJobRunnerError):
@@ -95 +95 @@ class FeaturesError(FirstRowsWorkerError):
-class StreamingRowsError(FirstRowsWorkerError):
+class StreamingRowsError(FirstRowsJobRunnerError):
@@ -102 +102 @@ class StreamingRowsError(FirstRowsWorkerError):
-class NormalRowsError(FirstRowsWorkerError):
+class NormalRowsError(FirstRowsJobRunnerError):
@@ -109 +109 @@ class NormalRowsError(FirstRowsWorkerError):
-class RowsPostProcessingError(FirstRowsWorkerError):
+class RowsPostProcessingError(FirstRowsJobRunnerError):
@@ -116 +116 @@ class RowsPostProcessingError(FirstRowsWorkerError):
-class TooManyColumnsError(FirstRowsWorkerError):
+class TooManyColumnsError(FirstRowsJobRunnerError):
@@ -123 +123 @@ class TooManyColumnsError(FirstRowsWorkerError):
-class TooBigContentError(FirstRowsWorkerError):
+class TooBigContentError(FirstRowsJobRunnerError):
@@ -448 +448 @@ def compute_first_rows_response(
- - [`libcommon.worker.ConfigNotFoundError`]
+ - [`~job_runner.ConfigNotFoundError`]
@@ -450 +450 @@ def compute_first_rows_response(
- - [`libcommon.worker.SplitNotFoundError`]
+ - [`~job_runner.SplitNotFoundError`]
@@ -452 +452 @@ def compute_first_rows_response(
- - [`~workers.first_rows.InfoError`]
+ - [`~job_runners.first_rows.InfoError`]
@@ -454 +454 @@ def compute_first_rows_response(
- - [`~workers.first_rows.FeaturesError`]
+ - [`~job_runners.first_rows.FeaturesError`]
@@ -456 +456 @@ def compute_first_rows_response(
- - [`~workers.first_rows.StreamingRowsError`]
+ - [`~job_runners.first_rows.StreamingRowsError`]
@@ -458 +458 @@ def compute_first_rows_response(
- - [`~workers.first_rows.NormalRowsError`]
+ - [`~job_runners.first_rows.NormalRowsError`]
@@ -460 +460 @@ def compute_first_rows_response(
- - [`~workers.first_rows.RowsPostProcessingError`]
+ - [`~job_runners.first_rows.RowsPostProcessingError`]
@@ -462 +462 @@ def compute_first_rows_response(
- - [`~workers.first_rows.TooManyColumnsError`]
+ - [`~job_runners.first_rows.TooManyColumnsError`]
@@ -464 +464 @@ def compute_first_rows_response(
- - [`~workers.first_rows.TooBigContentError`]
+ - [`~job_runners.first_rows.TooBigContentError`]
@@ -610 +610 @@ def compute_first_rows_response(
-class FirstRowsWorker(DatasetsBasedWorker):
+class FirstRowsJobRunner(DatasetsBasedJobRunner):
diff --git a/workers/datasets_based/src/datasets_based/workers/parquet.py b/services/worker/src/worker/job_runners/parquet.py
similarity index 83%
rename from workers/datasets_based/src/datasets_based/workers/parquet.py
rename to services/worker/src/worker/job_runners/parquet.py
index e156db21..a29ec018 100644
--- a/workers/datasets_based/src/datasets_based/workers/parquet.py
+++ b/services/worker/src/worker/job_runners/parquet.py
@@ -11,2 +11,2 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.worker import Worker, WorkerError
-from datasets_based.workers.parquet_and_dataset_info import ParquetFileItem
+from worker.job_runner import JobRunner, JobRunnerError
+from worker.job_runners.parquet_and_dataset_info import ParquetFileItem
@@ -14 +14 @@ from datasets_based.workers.parquet_and_dataset_info import ParquetFileItem
-ParquetWorkerErrorCode = Literal[
+ParquetJobRunnerErrorCode = Literal[
@@ -24 +24 @@ class ParquetResponse(TypedDict):
-class ParquetWorkerError(WorkerError):
+class ParquetJobRunnerError(JobRunnerError):
@@ -31 +31 @@ class ParquetWorkerError(WorkerError):
- code: ParquetWorkerErrorCode,
+ code: ParquetJobRunnerErrorCode,
@@ -40 +40 @@ class ParquetWorkerError(WorkerError):
-class PreviousStepStatusError(ParquetWorkerError):
+class PreviousStepStatusError(ParquetJobRunnerError):
@@ -47 +47 @@ class PreviousStepStatusError(ParquetWorkerError):
-class PreviousStepFormatError(ParquetWorkerError):
+class PreviousStepFormatError(ParquetJobRunnerError):
@@ -65 +65 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- - [`~workers.parquet.PreviousStepStatusError`]
+ - [`~job_runners.parquet.PreviousStepStatusError`]
@@ -67 +67 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- - [`~workers.parquet.PreviousStepFormatError`]
+ - [`~job_runners.parquet.PreviousStepFormatError`]
@@ -73,2 +73,2 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
- # TODO: we should move this dependency to the Worker class: defining which are the inputs, and just getting their
- # value here
+ # TODO: we should move this dependency to the JobRunner class: defining which are the inputs, and just getting
+ # their value here
@@ -91 +91 @@ def compute_parquet_response(dataset: str) -> ParquetResponse:
-class ParquetWorker(Worker):
+class ParquetJobRunner(JobRunner):
diff --git a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
similarity index 92%
rename from workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
rename to services/worker/src/worker/job_runners/parquet_and_dataset_info.py
index 74c37021..efc4f42b 100644
--- a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
+++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py
@@ -31,3 +31,3 @@ from libcommon.simple_cache import SplitFullName
-from datasets_based.config import AppConfig, ParquetAndDatasetInfoConfig
-from datasets_based.worker import DatasetNotFoundError, JobInfo, WorkerError
-from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
+from worker.config import AppConfig, ParquetAndDatasetInfoConfig
+from worker.job_runner import DatasetNotFoundError, JobInfo, JobRunnerError
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
@@ -35 +35 @@ from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
-ParquetAndDatasetInfoWorkerErrorCode = Literal[
+ParquetAndDatasetInfoJobRunnerErrorCode = Literal[
@@ -45 +45 @@ ParquetAndDatasetInfoWorkerErrorCode = Literal[
-class ParquetAndDatasetInfoWorkerError(WorkerError):
+class ParquetAndDatasetInfoJobRunnerError(JobRunnerError):
@@ -52 +52 @@ class ParquetAndDatasetInfoWorkerError(WorkerError):
- code: ParquetAndDatasetInfoWorkerErrorCode,
+ code: ParquetAndDatasetInfoJobRunnerErrorCode,
@@ -61 +61 @@ class ParquetAndDatasetInfoWorkerError(WorkerError):
-class DatasetRevisionNotFoundError(ParquetAndDatasetInfoWorkerError):
+class DatasetRevisionNotFoundError(ParquetAndDatasetInfoJobRunnerError):
@@ -68 +68 @@ class DatasetRevisionNotFoundError(ParquetAndDatasetInfoWorkerError):
-class ConfigNamesError(ParquetAndDatasetInfoWorkerError):
+class ConfigNamesError(ParquetAndDatasetInfoJobRunnerError):
@@ -75 +75 @@ class ConfigNamesError(ParquetAndDatasetInfoWorkerError):
-class EmptyDatasetError(ParquetAndDatasetInfoWorkerError):
+class EmptyDatasetError(ParquetAndDatasetInfoJobRunnerError):
@@ -82 +82 @@ class EmptyDatasetError(ParquetAndDatasetInfoWorkerError):
-class DatasetInBlockListError(ParquetAndDatasetInfoWorkerError):
+class DatasetInBlockListError(ParquetAndDatasetInfoJobRunnerError):
@@ -89 +89 @@ class DatasetInBlockListError(ParquetAndDatasetInfoWorkerError):
-class DatasetTooBigFromHubError(ParquetAndDatasetInfoWorkerError):
+class DatasetTooBigFromHubError(ParquetAndDatasetInfoJobRunnerError):
@@ -96 +96 @@ class DatasetTooBigFromHubError(ParquetAndDatasetInfoWorkerError):
-class DatasetTooBigFromDatasetsError(ParquetAndDatasetInfoWorkerError):
+class DatasetTooBigFromDatasetsError(ParquetAndDatasetInfoJobRunnerError):
@@ -196 +196 @@ def raise_if_blocked(
- - [`~parquet.worker.DatasetInBlockListError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`]
@@ -231 +231 @@ def get_dataset_info_or_raise(
- - [`~libcommon.worker.DatasetNotFoundError`]
+ - [`~.job_runner.DatasetNotFoundError`]
@@ -234 +234 @@ def get_dataset_info_or_raise(
- - [`~parquet.worker.DatasetRevisionNotFoundError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetRevisionNotFoundError`]
@@ -265 +265 @@ def raise_if_too_big_from_hub(
- - [`~parquet.worker.DatasetTooBigFromHubError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromHubError`]
@@ -306 +306 @@ def raise_if_too_big_from_datasets(
- - [`~parquet.worker.DatasetTooBigFromDatasetsError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromDatasetsError`]
@@ -369 +369 @@ def raise_if_not_supported(
- - [`~parquet.worker.DatasetInBlockListError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`]
@@ -379 +379 @@ def raise_if_not_supported(
- - [`~parquet.worker.DatasetRevisionNotFoundError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetRevisionNotFoundError`]
@@ -381 +381 @@ def raise_if_not_supported(
- - [`~parquet.worker.DatasetTooBigFromHubError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromHubError`]
@@ -385 +385 @@ def raise_if_not_supported(
- - [`~parquet.worker.DatasetTooBigFromDatasetsError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromDatasetsError`]
@@ -475 +475 @@ def compute_parquet_and_dataset_info_response(
- - [`~workers.parquet_and_dataset_info.DatasetInBlockListError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`]
@@ -486 +486 @@ def compute_parquet_and_dataset_info_response(
- - [`~workers.parquet_and_dataset_info.DatasetRevisionNotFoundError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetRevisionNotFoundError`]
@@ -488 +488 @@ def compute_parquet_and_dataset_info_response(
- - [`~workers.parquet_and_dataset_info.DatasetTooBigFromHubError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromHubError`]
@@ -492 +492 @@ def compute_parquet_and_dataset_info_response(
- - [`~workers.parquet_and_dataset_info.DatasetTooBigFromDatasetsError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromDatasetsError`]
@@ -494 +494 @@ def compute_parquet_and_dataset_info_response(
- - [`~workers.parquet_and_dataset_info.EmptyDatasetError`]
+ - [`~job_runners.parquet_and_dataset_info.EmptyDatasetError`]
@@ -496 +496 @@ def compute_parquet_and_dataset_info_response(
- - [`~workers.parquet_and_dataset_info.ConfigNamesError`]
+ - [`~job_runners.parquet_and_dataset_info.ConfigNamesError`]
@@ -498 +498 @@ def compute_parquet_and_dataset_info_response(
- - [`~workers.parquet_and_dataset_info.DatasetInBlockListError`]
+ - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`]
@@ -533 +532,0 @@ def compute_parquet_and_dataset_info_response(
- # TODO: run the loop in parallel, in different workers? with dagster?
@@ -604 +603 @@ def compute_parquet_and_dataset_info_response(
-class ParquetAndDatasetInfoWorker(DatasetsBasedWorker):
+class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner):
diff --git a/workers/datasets_based/src/datasets_based/workers/sizes.py b/services/worker/src/worker/job_runners/sizes.py
similarity index 93%
rename from workers/datasets_based/src/datasets_based/workers/sizes.py
rename to services/worker/src/worker/job_runners/sizes.py
index ff450908..fa48abf7 100644
--- a/workers/datasets_based/src/datasets_based/workers/sizes.py
+++ b/services/worker/src/worker/job_runners/sizes.py
@@ -11 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.worker import Worker, WorkerError
+from worker.job_runner import JobRunner, JobRunnerError
@@ -13 +13 @@ from datasets_based.worker import Worker, WorkerError
-SizesWorkerErrorCode = Literal[
+SizesJobRunnerErrorCode = Literal[
@@ -57 +57 @@ class SizesResponse(TypedDict):
-class SizesWorkerError(WorkerError):
+class SizesJobRunnerError(JobRunnerError):
@@ -64 +64 @@ class SizesWorkerError(WorkerError):
- code: SizesWorkerErrorCode,
+ code: SizesJobRunnerErrorCode,
@@ -73 +73 @@ class SizesWorkerError(WorkerError):
-class PreviousStepStatusError(SizesWorkerError):
+class PreviousStepStatusError(SizesJobRunnerError):
@@ -80 +80 @@ class PreviousStepStatusError(SizesWorkerError):
-class PreviousStepFormatError(SizesWorkerError):
+class PreviousStepFormatError(SizesJobRunnerError):
@@ -98 +98 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- - [`~workers.sizes.PreviousStepStatusError`]
+ - [`~job_runners.sizes.PreviousStepStatusError`]
@@ -100 +100 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
- - [`~workers.sizes.PreviousStepFormatError`]
+ - [`~job_runners.sizes.PreviousStepFormatError`]
@@ -172 +172 @@ def compute_sizes_response(dataset: str) -> SizesResponse:
-class SizesWorker(Worker):
+class SizesJobRunner(JobRunner):
diff --git a/workers/datasets_based/src/datasets_based/workers/split_names.py b/services/worker/src/worker/job_runners/split_names.py
similarity index 88%
rename from workers/datasets_based/src/datasets_based/workers/split_names.py
rename to services/worker/src/worker/job_runners/split_names.py
index f196b98c..c485f640 100644
--- a/workers/datasets_based/src/datasets_based/workers/split_names.py
+++ b/services/worker/src/worker/job_runners/split_names.py
@@ -12,2 +12,2 @@ from libcommon.simple_cache import SplitFullName
-from datasets_based.worker import WorkerError
-from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
+from worker.job_runner import JobRunnerError
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
@@ -15 +15 @@ from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
-SplitNamesWorkerErrorCode = Literal[
+SplitNamesJobRunnerErrorCode = Literal[
@@ -21,2 +21,2 @@ SplitNamesWorkerErrorCode = Literal[
-class SplitNamesWorkerError(WorkerError):
- """Base class for worker exceptions."""
+class SplitNamesJobRunnerError(JobRunnerError):
+ """Base class for split names job runner exceptions."""
@@ -28 +28 @@ class SplitNamesWorkerError(WorkerError):
- code: SplitNamesWorkerErrorCode,
+ code: SplitNamesJobRunnerErrorCode,
@@ -37 +37 @@ class SplitNamesWorkerError(WorkerError):
-class SplitNamesError(SplitNamesWorkerError):
+class SplitNamesError(SplitNamesJobRunnerError):
@@ -44 +44 @@ class SplitNamesError(SplitNamesWorkerError):
-class EmptyDatasetError(SplitNamesWorkerError):
+class EmptyDatasetError(SplitNamesJobRunnerError):
@@ -91 +91 @@ def compute_split_names_response(
- - [`~workers.split_names.EmptyDatasetError`]
+ - [`~job_runners.split_names.EmptyDatasetError`]
@@ -93 +93 @@ def compute_split_names_response(
- - [`~workers.split_names.SplitsNamesError`]
+ - [`~job_runners.split_names.SplitsNamesError`]
@@ -112 +112 @@ def compute_split_names_response(
-class SplitNamesWorker(DatasetsBasedWorker):
+class SplitNamesJobRunner(DatasetsBasedJobRunner):
diff --git a/workers/datasets_based/src/datasets_based/workers/splits.py b/services/worker/src/worker/job_runners/splits.py
similarity index 90%
rename from workers/datasets_based/src/datasets_based/workers/splits.py
rename to services/worker/src/worker/job_runners/splits.py
index fc9a9d71..fa8a9487 100644
--- a/workers/datasets_based/src/datasets_based/workers/splits.py
+++ b/services/worker/src/worker/job_runners/splits.py
@@ -12,2 +12,2 @@ from libcommon.simple_cache import SplitFullName
-from datasets_based.worker import WorkerError
-from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
+from worker.job_runner import JobRunnerError
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
@@ -15 +15 @@ from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
-SplitsWorkerErrorCode = Literal[
+SplitsJobRunnerErrorCode = Literal[
@@ -21,2 +21,2 @@ SplitsWorkerErrorCode = Literal[
-class SplitsWorkerError(WorkerError):
- """Base class for worker exceptions."""
+class SplitsJobRunnerError(JobRunnerError):
+ """Base class for splits job runner exceptions."""
@@ -28 +28 @@ class SplitsWorkerError(WorkerError):
- code: SplitsWorkerErrorCode,
+ code: SplitsJobRunnerErrorCode,
@@ -37 +37 @@ class SplitsWorkerError(WorkerError):
-class SplitsNamesError(SplitsWorkerError):
+class SplitsNamesError(SplitsJobRunnerError):
@@ -44 +44 @@ class SplitsNamesError(SplitsWorkerError):
-class EmptyDatasetError(SplitsWorkerError):
+class EmptyDatasetError(SplitsJobRunnerError):
@@ -109 +109 @@ def compute_splits_response(
- - [`~workers.splits.EmptyDatasetError`]
+ - [`~job_runners.splits.EmptyDatasetError`]
@@ -111 +111 @@ def compute_splits_response(
- - [`~workers.splits.SplitsNamesError`]
+ - [`~job_runners.splits.SplitsNamesError`]
@@ -127 +127 @@ def compute_splits_response(
-class SplitsWorker(DatasetsBasedWorker):
+class SplitsJobRunner(DatasetsBasedJobRunner):
diff --git a/workers/datasets_based/src/datasets_based/worker_loop.py b/services/worker/src/worker/loop.py
similarity index 67%
rename from workers/datasets_based/src/datasets_based/worker_loop.py
rename to services/worker/src/worker/loop.py
index 290075b1..7860ab36 100644
--- a/workers/datasets_based/src/datasets_based/worker_loop.py
+++ b/services/worker/src/worker/loop.py
@@ -12,2 +12,2 @@ from psutil import cpu_count, disk_usage, getloadavg, swap_memory, virtual_memor
-from datasets_based.config import WorkerLoopConfig
-from datasets_based.worker_factory import BaseWorkerFactory
+from worker.config import LoopConfig
+from worker.job_runner_factory import BaseJobRunnerFactory
@@ -17 +17 @@ from datasets_based.worker_factory import BaseWorkerFactory
-class WorkerLoop:
+class Loop:
@@ -19 +19 @@ class WorkerLoop:
- A worker loop gets jobs from a queue and processes them.
+ A loop gets jobs from a queue and processes them.
@@ -21 +21 @@ class WorkerLoop:
- Once initialized, the worker loop can be started with the `loop` method and will run until an uncaught exception
+ Once initialized, the loop can be started with the `run` method and will run until an uncaught exception
@@ -29,4 +29,5 @@ class WorkerLoop:
- worker_factory (`WorkerFactory`):
- The worker factory that will create a worker for each job. Must be able to process the jobs of the queue.
- worker_loop_config (`WorkerLoopConfig`):
- Worker loop configuration.
+ job_runner_factory (`JobRunnerFactory`):
+ The job runner factory that will create a job runner for each job. Must be able to process the jobs of the
+ queue.
+ loop_config (`LoopConfig`):
+ Loop configuration.
@@ -37,2 +38,2 @@ class WorkerLoop:
- worker_factory: BaseWorkerFactory
- worker_loop_config: WorkerLoopConfig
+ job_runner_factory: BaseJobRunnerFactory
+ loop_config: LoopConfig
@@ -43 +44 @@ class WorkerLoop:
- self.storage_paths = set(self.worker_loop_config.storage_paths).union(self.library_cache_paths)
+ self.storage_paths = set(self.loop_config.storage_paths).union(self.library_cache_paths)
@@ -61 +62 @@ class WorkerLoop:
- if self.worker_loop_config.max_memory_pct <= 0:
+ if self.loop_config.max_memory_pct <= 0:
@@ -66 +67 @@ class WorkerLoop:
- ok = percent < self.worker_loop_config.max_memory_pct
+ ok = percent < self.loop_config.max_memory_pct
@@ -69,2 +70 @@ class WorkerLoop:
- f"memory usage (RAM + SWAP) is too high: {percent:.0f}% - max is"
- f" {self.worker_loop_config.max_memory_pct}%"
+ f"memory usage (RAM + SWAP) is too high: {percent:.0f}% - max is {self.loop_config.max_memory_pct}%"
@@ -75 +75 @@ class WorkerLoop:
- if self.worker_loop_config.max_load_pct <= 0:
+ if self.loop_config.max_load_pct <= 0:
@@ -79 +79 @@ class WorkerLoop:
- ok = load_pct < self.worker_loop_config.max_load_pct
+ ok = load_pct < self.loop_config.max_load_pct
@@ -81 +81 @@ class WorkerLoop:
- self.info(f"cpu load is too high: {load_pct:.0f}% - max is {self.worker_loop_config.max_load_pct}%")
+ self.info(f"cpu load is too high: {load_pct:.0f}% - max is {self.loop_config.max_load_pct}%")
@@ -85 +85 @@ class WorkerLoop:
- if self.worker_loop_config.max_disk_usage_pct <= 0:
+ if self.loop_config.max_disk_usage_pct <= 0:
@@ -90 +90 @@ class WorkerLoop:
- if usage.percent >= self.worker_loop_config.max_disk_usage_pct:
+ if usage.percent >= self.loop_config.max_disk_usage_pct:
@@ -103 +103 @@ class WorkerLoop:
- duration = self.worker_loop_config.sleep_seconds * jitter
+ duration = self.loop_config.sleep_seconds * jitter
@@ -107 +107 @@ class WorkerLoop:
- def loop(self) -> None:
+ def run(self) -> None:
@@ -124,2 +124,2 @@ class WorkerLoop:
- worker = self.worker_factory.create_worker(self.queue.start_job())
- self.debug(f"job assigned: {worker}")
+ job_runner = self.job_runner_factory.create_job_runner(self.queue.start_job())
+ self.debug(f"job assigned: {job_runner}")
@@ -130,3 +130,3 @@ class WorkerLoop:
- finished_status = worker.run()
- self.queue.finish_job(job_id=worker.job_id, finished_status=finished_status)
- self.debug(f"job finished with {finished_status.value}: {worker}")
+ finished_status = job_runner.run()
+ self.queue.finish_job(job_id=job_runner.job_id, finished_status=finished_status)
+ self.debug(f"job finished with {finished_status.value}: {job_runner}")
diff --git a/workers/datasets_based/src/datasets_based/main.py b/services/worker/src/worker/main.py
similarity index 83%
rename from workers/datasets_based/src/datasets_based/main.py
rename to services/worker/src/worker/main.py
index 08f64331..6de77fae 100644
--- a/workers/datasets_based/src/datasets_based/main.py
+++ b/services/worker/src/worker/main.py
@@ -10,4 +10,4 @@ from libcommon.storage import init_assets_dir
-from datasets_based.config import AppConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.worker_factory import WorkerFactory
-from datasets_based.worker_loop import WorkerLoop
+from worker.config import AppConfig
+from worker.job_runner_factory import JobRunnerFactory
+from worker.loop import Loop
+from worker.resources import LibrariesResource
@@ -43 +43 @@ if __name__ == "__main__":
- worker_factory = WorkerFactory(
+ job_runner_factory = JobRunnerFactory(
@@ -50 +50 @@ if __name__ == "__main__":
- worker_loop = WorkerLoop(
+ loop = Loop(
@@ -53,2 +53,2 @@ if __name__ == "__main__":
- worker_factory=worker_factory,
- worker_loop_config=app_config.worker_loop,
+ job_runner_factory=job_runner_factory,
+ loop_config=app_config.loop,
@@ -56 +56 @@ if __name__ == "__main__":
- worker_loop.loop()
+ loop.run()
diff --git a/workers/datasets_based/src/datasets_based/py.typed b/services/worker/src/worker/py.typed
similarity index 100%
rename from workers/datasets_based/src/datasets_based/py.typed
rename to services/worker/src/worker/py.typed
diff --git a/workers/datasets_based/src/datasets_based/resources.py b/services/worker/src/worker/resources.py
similarity index 100%
rename from workers/datasets_based/src/datasets_based/resources.py
rename to services/worker/src/worker/resources.py
diff --git a/workers/datasets_based/tests/__init__.py b/services/worker/tests/__init__.py
similarity index 100%
rename from workers/datasets_based/tests/__init__.py
rename to services/worker/tests/__init__.py
diff --git a/workers/datasets_based/tests/conftest.py b/services/worker/tests/conftest.py
similarity index 97%
rename from workers/datasets_based/tests/conftest.py
rename to services/worker/tests/conftest.py
index 2292da13..54048d0b 100644
--- a/workers/datasets_based/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -14,2 +14,2 @@ from pytest import MonkeyPatch, fixture
-from datasets_based.config import AppConfig, FirstRowsConfig
-from datasets_based.resources import LibrariesResource
+from worker.config import AppConfig, FirstRowsConfig
+from worker.resources import LibrariesResource
diff --git a/workers/datasets_based/tests/constants.py b/services/worker/tests/constants.py
similarity index 100%
rename from workers/datasets_based/tests/constants.py
rename to services/worker/tests/constants.py
diff --git a/workers/datasets_based/tests/fixtures/__init__.py b/services/worker/tests/fixtures/__init__.py
similarity index 100%
rename from workers/datasets_based/tests/fixtures/__init__.py
rename to services/worker/tests/fixtures/__init__.py
diff --git a/workers/datasets_based/tests/fixtures/data/test_image_rgb.jpg b/services/worker/tests/fixtures/data/test_image_rgb.jpg
similarity index 100%
rename from workers/datasets_based/tests/fixtures/data/test_image_rgb.jpg
rename to services/worker/tests/fixtures/data/test_image_rgb.jpg
diff --git a/workers/datasets_based/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py
similarity index 100%
rename from workers/datasets_based/tests/fixtures/datasets.py
rename to services/worker/tests/fixtures/datasets.py
diff --git a/workers/datasets_based/tests/fixtures/files.py b/services/worker/tests/fixtures/files.py
similarity index 100%
rename from workers/datasets_based/tests/fixtures/files.py
rename to services/worker/tests/fixtures/files.py
diff --git a/workers/datasets_based/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
similarity index 100%
rename from workers/datasets_based/tests/fixtures/hub.py
rename to services/worker/tests/fixtures/hub.py
diff --git a/workers/datasets_based/tests/workers/__init__.py b/services/worker/tests/job_runners/__init__.py
similarity index 100%
rename from workers/datasets_based/tests/workers/__init__.py
rename to services/worker/tests/job_runners/__init__.py
diff --git a/workers/datasets_based/tests/workers/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py
similarity index 75%
rename from workers/datasets_based/tests/workers/test__datasets_based_worker.py
rename to services/worker/tests/job_runners/test__datasets_based_worker.py
index 5c527daa..275e48bc 100644
--- a/workers/datasets_based/tests/workers/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__datasets_based_worker.py
@@ -17,3 +17,3 @@ from libcommon.simple_cache import get_response
-from datasets_based.config import AppConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
+from worker.config import AppConfig
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.resources import LibrariesResource
@@ -24 +24 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
-class DummyWorker(DatasetsBasedWorker):
+class DummyJobRunner(DatasetsBasedJobRunner):
@@ -28 +28 @@ class DummyWorker(DatasetsBasedWorker):
- # ^ borrowing the endpoint, so that the processing step exists and the worker can be initialized
+ # ^ borrowing the endpoint, so that the processing step exists and the job runner can be initialized
@@ -43 +43 @@ class DummyWorker(DatasetsBasedWorker):
-def get_worker(
+def get_job_runner(
@@ -48 +48 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -54,2 +54,2 @@ def get_worker(
- ) -> DummyWorker:
- return DummyWorker(
+ ) -> DummyJobRunner:
+ return DummyJobRunner(
@@ -57 +57 @@ def get_worker(
- "type": DummyWorker.get_job_type(),
+ "type": DummyJobRunner.get_job_type(),
@@ -67 +67 @@ def get_worker(
- endpoint=DummyWorker.get_job_type(),
+ endpoint=DummyJobRunner.get_job_type(),
@@ -78 +78 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -81 +81 @@ def get_worker(
-def test_version(app_config: AppConfig, get_worker) -> None:
+def test_version(app_config: AppConfig, get_job_runner) -> None:
@@ -83,4 +83,4 @@ def test_version(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset, config, split, app_config)
- assert len(worker.get_version().split(".")) == 3
- assert worker.compare_major_version(other_version="0.0.0") > 0
- assert worker.compare_major_version(other_version="1000.0.0") < 0
+ job_runner = get_job_runner(dataset, config, split, app_config)
+ assert len(job_runner.get_version().split(".")) == 3
+ assert job_runner.compare_major_version(other_version="0.0.0") > 0
+ assert job_runner.compare_major_version(other_version="1000.0.0") < 0
@@ -111 +111 @@ def test_get_cache_subdirectory(
- get_worker,
+ get_job_runner,
@@ -119,2 +119,2 @@ def test_get_cache_subdirectory(
- worker = get_worker(dataset, config, split, app_config, force=force)
- assert worker.get_cache_subdirectory(date=date) == expected
+ job_runner = get_job_runner(dataset, config, split, app_config, force=force)
+ assert job_runner.get_cache_subdirectory(date=date) == expected
@@ -123 +123 @@ def test_get_cache_subdirectory(
-def test_set_and_unset_datasets_cache(app_config: AppConfig, get_worker) -> None:
+def test_set_and_unset_datasets_cache(app_config: AppConfig, get_job_runner) -> None:
@@ -125,2 +125,2 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig, get_worker) -> None
- worker = get_worker(dataset, config, split, app_config)
- base_path = worker.base_datasets_cache
+ job_runner = get_job_runner(dataset, config, split, app_config)
+ base_path = job_runner.base_datasets_cache
@@ -128 +128 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig, get_worker) -> None
- worker.set_datasets_cache(dummy_path)
+ job_runner.set_datasets_cache(dummy_path)
@@ -130 +130 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig, get_worker) -> None
- worker.unset_datasets_cache()
+ job_runner.unset_datasets_cache()
@@ -134 +134 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig, get_worker) -> None
-def test_set_and_unset_cache(app_config: AppConfig, get_worker) -> None:
+def test_set_and_unset_cache(app_config: AppConfig, get_job_runner) -> None:
@@ -136,3 +136,3 @@ def test_set_and_unset_cache(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset, config, split, app_config)
- datasets_base_path = worker.base_datasets_cache
- worker.set_cache()
+ job_runner = get_job_runner(dataset, config, split, app_config)
+ datasets_base_path = job_runner.base_datasets_cache
+ job_runner.set_cache()
@@ -141 +141 @@ def test_set_and_unset_cache(app_config: AppConfig, get_worker) -> None:
- worker.unset_cache()
+ job_runner.unset_cache()
@@ -146 +146 @@ def test_set_and_unset_cache(app_config: AppConfig, get_worker) -> None:
-def test_process(app_config: AppConfig, get_worker, hub_public_csv: str, config: str) -> None:
+def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str, config: str) -> None:
@@ -151,2 +151,2 @@ def test_process(app_config: AppConfig, get_worker, hub_public_csv: str, config:
- worker = get_worker(dataset, config, split, app_config)
- datasets_base_path = worker.base_datasets_cache
+ job_runner = get_job_runner(dataset, config, split, app_config)
+ datasets_base_path = job_runner.base_datasets_cache
@@ -155 +155 @@ def test_process(app_config: AppConfig, get_worker, hub_public_csv: str, config:
- result = worker.process()
+ result = job_runner.process()
@@ -168 +168 @@ def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) ->
-def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, get_worker) -> None:
+def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, get_job_runner) -> None:
@@ -170 +170 @@ def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, g
- worker = get_worker(
+ worker = get_job_runner(
diff --git a/workers/datasets_based/tests/workers/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py
similarity index 73%
rename from workers/datasets_based/tests/workers/test_config_names.py
rename to services/worker/tests/job_runners/test_config_names.py
index 6da4f23c..ed9bfbac 100644
--- a/workers/datasets_based/tests/workers/test_config_names.py
+++ b/services/worker/tests/job_runners/test_config_names.py
@@ -14,3 +14,3 @@ from libcommon.simple_cache import DoesNotExist, get_response
-from datasets_based.config import AppConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.workers.config_names import ConfigNamesWorker
+from worker.config import AppConfig
+from worker.job_runners.config_names import ConfigNamesJobRunner
+from worker.resources import LibrariesResource
@@ -22 +22 @@ from ..fixtures.hub import HubDatasets
-def get_worker(
+def get_job_runner(
@@ -27 +27 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -31,2 +31,2 @@ def get_worker(
- ) -> ConfigNamesWorker:
- return ConfigNamesWorker(
+ ) -> ConfigNamesJobRunner:
+ return ConfigNamesJobRunner(
@@ -34 +34 @@ def get_worker(
- "type": ConfigNamesWorker.get_job_type(),
+ "type": ConfigNamesJobRunner.get_job_type(),
@@ -44 +44 @@ def get_worker(
- endpoint=ConfigNamesWorker.get_job_type(),
+ endpoint=ConfigNamesJobRunner.get_job_type(),
@@ -55 +55 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -58 +58 @@ def get_worker(
-def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
+def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
@@ -60,2 +60,2 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker)
- worker = get_worker(dataset, app_config)
- assert worker.should_skip_job() is False
+ job_runner = get_job_runner(dataset, app_config)
+ assert job_runner.should_skip_job() is False
@@ -63,4 +63,4 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker)
- worker.process()
- assert worker.should_skip_job() is True
- worker = get_worker(dataset, app_config, force=True)
- assert worker.should_skip_job() is False
+ job_runner.process()
+ assert job_runner.should_skip_job() is True
+ job_runner = get_job_runner(dataset, app_config, force=True)
+ assert job_runner.should_skip_job() is False
@@ -69 +69 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker)
-def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
+def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
@@ -71,3 +71,3 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None
- worker = get_worker(dataset, app_config)
- assert worker.process() is True
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=hub_public_csv)
+ job_runner = get_job_runner(dataset, app_config)
+ assert job_runner.process() is True
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=hub_public_csv)
@@ -76 +76 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None
- assert cached_response["worker_version"] == worker.get_version()
+ assert cached_response["worker_version"] == job_runner.get_version()
@@ -83 +83 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None
-def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
@@ -85,2 +85,2 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset, app_config)
- assert worker.process() is False
+ job_runner = get_job_runner(dataset, app_config)
+ assert job_runner.process() is False
@@ -88 +88 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -108 +108 @@ def test_compute_splits_response_simple_csv(
- get_worker,
+ get_job_runner,
@@ -117 +117 @@ def test_compute_splits_response_simple_csv(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -122 +122 @@ def test_compute_splits_response_simple_csv(
- result = worker.compute()
+ result = job_runner.compute()
@@ -127 +127 @@ def test_compute_splits_response_simple_csv(
- worker.compute()
+ job_runner.compute()
diff --git a/workers/datasets_based/tests/workers/test_dataset_info.py b/services/worker/tests/job_runners/test_dataset_info.py
similarity index 78%
rename from workers/datasets_based/tests/workers/test_dataset_info.py
rename to services/worker/tests/job_runners/test_dataset_info.py
index 01f1c24a..a0dfd692 100644
--- a/workers/datasets_based/tests/workers/test_dataset_info.py
+++ b/services/worker/tests/job_runners/test_dataset_info.py
@@ -13,3 +13,3 @@ from libcommon.simple_cache import upsert_response
-from datasets_based.config import AppConfig
-from datasets_based.workers.dataset_info import (
- DatasetInfoWorker,
+from worker.config import AppConfig
+from worker.job_runners.dataset_info import (
+ DatasetInfoJobRunner,
@@ -29 +29 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-def get_worker(
+def get_job_runner(
@@ -33 +33 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -37,2 +37,2 @@ def get_worker(
- ) -> DatasetInfoWorker:
- return DatasetInfoWorker(
+ ) -> DatasetInfoJobRunner:
+ return DatasetInfoJobRunner(
@@ -40 +40 @@ def get_worker(
- "type": DatasetInfoWorker.get_job_type(),
+ "type": DatasetInfoJobRunner.get_job_type(),
@@ -51 +51 @@ def get_worker(
- endpoint=DatasetInfoWorker.get_job_type(),
+ endpoint=DatasetInfoJobRunner.get_job_type(),
@@ -61 +61 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -88 +88 @@ def test_compute(
- get_worker,
+ get_job_runner,
@@ -99 +99 @@ def test_compute(
- worker = get_worker(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset=dataset, app_config=app_config)
@@ -102 +102 @@ def test_compute(
- worker.compute()
+ job_runner.compute()
@@ -105 +105 @@ def test_compute(
- assert worker.compute() == expected_content
+ assert job_runner.compute() == expected_content
@@ -108 +108 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
@@ -110 +110 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset=dataset, app_config=app_config)
@@ -112 +112 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker.compute()
+ job_runner.compute()
diff --git a/workers/datasets_based/tests/workers/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py
similarity index 77%
rename from workers/datasets_based/tests/workers/test_first_rows.py
rename to services/worker/tests/job_runners/test_first_rows.py
index ab7ce7a0..e6779406 100644
--- a/workers/datasets_based/tests/workers/test_first_rows.py
+++ b/services/worker/tests/job_runners/test_first_rows.py
@@ -16,3 +16,3 @@ from libcommon.storage import StrPath
-from datasets_based.config import AppConfig, FirstRowsConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.workers.first_rows import FirstRowsWorker, get_json_size
+from worker.config import AppConfig, FirstRowsConfig
+from worker.job_runners.first_rows import FirstRowsJobRunner, get_json_size
+from worker.resources import LibrariesResource
@@ -24 +24 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
-def get_worker(
+def get_job_runner(
@@ -30 +30 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -37,2 +37,2 @@ def get_worker(
- ) -> FirstRowsWorker:
- return FirstRowsWorker(
+ ) -> FirstRowsJobRunner:
+ return FirstRowsJobRunner(
@@ -40 +40 @@ def get_worker(
- "type": FirstRowsWorker.get_job_type(),
+ "type": FirstRowsJobRunner.get_job_type(),
@@ -50 +50 @@ def get_worker(
- endpoint=FirstRowsWorker.get_job_type(),
+ endpoint=FirstRowsJobRunner.get_job_type(),
@@ -63 +63 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -67 +67 @@ def test_should_skip_job(
- app_config: AppConfig, get_worker, first_rows_config: FirstRowsConfig, hub_public_csv: str
+ app_config: AppConfig, get_job_runner, first_rows_config: FirstRowsConfig, hub_public_csv: str
@@ -70,2 +70,2 @@ def test_should_skip_job(
- worker = get_worker(dataset, config, split, app_config, first_rows_config)
- assert worker.should_skip_job() is False
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config)
+ assert job_runner.should_skip_job() is False
@@ -73,4 +73,4 @@ def test_should_skip_job(
- worker.process()
- assert worker.should_skip_job() is True
- worker = get_worker(dataset, config, split, app_config, first_rows_config, force=True)
- assert worker.should_skip_job() is False
+ job_runner.process()
+ assert job_runner.should_skip_job() is True
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, force=True)
+ assert job_runner.should_skip_job() is False
@@ -79 +79,3 @@ def test_should_skip_job(
-def test_compute(app_config: AppConfig, get_worker, first_rows_config: FirstRowsConfig, hub_public_csv: str) -> None:
+def test_compute(
+ app_config: AppConfig, get_job_runner, first_rows_config: FirstRowsConfig, hub_public_csv: str
+) -> None:
@@ -81,3 +83,5 @@ def test_compute(app_config: AppConfig, get_worker, first_rows_config: FirstRows
- worker = get_worker(dataset, config, split, app_config, first_rows_config)
- assert worker.process() is True
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=dataset, config=config, split=split)
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config)
+ assert job_runner.process() is True
+ cached_response = get_response(
+ kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config, split=split
+ )
@@ -86 +90 @@ def test_compute(app_config: AppConfig, get_worker, first_rows_config: FirstRows
- assert cached_response["worker_version"] == worker.get_version()
+ assert cached_response["worker_version"] == job_runner.get_version()
@@ -97 +101 @@ def test_compute(app_config: AppConfig, get_worker, first_rows_config: FirstRows
-def test_doesnotexist(app_config: AppConfig, get_worker, first_rows_config: FirstRowsConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner, first_rows_config: FirstRowsConfig) -> None:
@@ -100,2 +104,2 @@ def test_doesnotexist(app_config: AppConfig, get_worker, first_rows_config: Firs
- worker = get_worker(dataset, config, split, app_config, first_rows_config)
- assert worker.process() is False
+ job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config)
+ assert job_runner.process() is False
@@ -103 +107 @@ def test_doesnotexist(app_config: AppConfig, get_worker, first_rows_config: Firs
- get_response(kind=worker.processing_step.cache_kind, dataset=dataset, config=config, split=split)
+ get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config, split=split)
@@ -126 +130 @@ def test_number_rows(
- get_worker,
+ get_job_runner,
@@ -143 +147 @@ def test_number_rows(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -151 +155 @@ def test_number_rows(
- result = worker.compute()
+ result = job_runner.compute()
@@ -155 +159 @@ def test_number_rows(
- worker.compute()
+ job_runner.compute()
@@ -187 +191 @@ def test_truncation(
- get_worker,
+ get_job_runner,
@@ -196 +200 @@ def test_truncation(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -213 +217 @@ def test_truncation(
- worker.compute()
+ job_runner.compute()
@@ -216 +220 @@ def test_truncation(
- response = worker.compute()
+ response = job_runner.compute()
diff --git a/workers/datasets_based/tests/workers/test_parquet.py b/services/worker/tests/job_runners/test_parquet.py
similarity index 79%
rename from workers/datasets_based/tests/workers/test_parquet.py
rename to services/worker/tests/job_runners/test_parquet.py
index c022f651..16b7a25c 100644
--- a/workers/datasets_based/tests/workers/test_parquet.py
+++ b/services/worker/tests/job_runners/test_parquet.py
@@ -13,2 +13,2 @@ from libcommon.simple_cache import upsert_response
-from datasets_based.config import AppConfig
-from datasets_based.workers.parquet import (
+from worker.config import AppConfig
+from worker.job_runners.parquet import (
@@ -16 +16 @@ from datasets_based.workers.parquet import (
- ParquetWorker,
+ ParquetJobRunner,
@@ -29 +29 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-def get_worker(
+def get_job_runner(
@@ -33 +33 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -37,2 +37,2 @@ def get_worker(
- ) -> ParquetWorker:
- return ParquetWorker(
+ ) -> ParquetJobRunner:
+ return ParquetJobRunner(
@@ -40 +40 @@ def get_worker(
- "type": ParquetWorker.get_job_type(),
+ "type": ParquetJobRunner.get_job_type(),
@@ -51 +51 @@ def get_worker(
- endpoint=ParquetWorker.get_job_type(),
+ endpoint=ParquetJobRunner.get_job_type(),
@@ -61 +61 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -88 +88 @@ def test_compute(
- get_worker,
+ get_job_runner,
@@ -99 +99 @@ def test_compute(
- worker = get_worker(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset=dataset, app_config=app_config)
@@ -102 +102 @@ def test_compute(
- worker.compute()
+ job_runner.compute()
@@ -105 +105 @@ def test_compute(
- assert worker.compute() == expected_content
+ assert job_runner.compute() == expected_content
@@ -108 +108 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
@@ -110 +110 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset=dataset, app_config=app_config)
@@ -112 +112 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker.compute()
+ job_runner.compute()
diff --git a/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
similarity index 86%
rename from workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
rename to services/worker/tests/job_runners/test_parquet_and_dataset_info.py
index 78438dc7..ff5f7fb3 100644
--- a/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
+++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py
@@ -17,3 +17,2 @@ from libcommon.simple_cache import DoesNotExist, get_response
-from datasets_based.config import AppConfig, ParquetAndDatasetInfoConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.workers.parquet_and_dataset_info import (
+from worker.config import AppConfig, ParquetAndDatasetInfoConfig
+from worker.job_runners.parquet_and_dataset_info import (
@@ -23 +22 @@ from datasets_based.workers.parquet_and_dataset_info import (
- ParquetAndDatasetInfoWorker,
+ ParquetAndDatasetInfoJobRunner,
@@ -30,0 +30 @@ from datasets_based.workers.parquet_and_dataset_info import (
+from worker.resources import LibrariesResource
@@ -59 +59 @@ def parquet_and_dataset_info_config(
-def get_worker(
+def get_job_runner(
@@ -64 +64 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -69,2 +69,2 @@ def get_worker(
- ) -> ParquetAndDatasetInfoWorker:
- return ParquetAndDatasetInfoWorker(
+ ) -> ParquetAndDatasetInfoJobRunner:
+ return ParquetAndDatasetInfoJobRunner(
@@ -72 +72 @@ def get_worker(
- "type": ParquetAndDatasetInfoWorker.get_job_type(),
+ "type": ParquetAndDatasetInfoJobRunner.get_job_type(),
@@ -82 +82 @@ def get_worker(
- endpoint=ParquetAndDatasetInfoWorker.get_job_type(),
+ endpoint=ParquetAndDatasetInfoJobRunner.get_job_type(),
@@ -94 +94 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -116 +116 @@ def test_compute(
- get_worker,
+ get_job_runner,
@@ -121 +121 @@ def test_compute(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -124,2 +124,2 @@ def test_compute(
- assert worker.process() is True
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ assert job_runner.process() is True
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -128 +128 @@ def test_compute(
- assert cached_response["worker_version"] == worker.get_version()
+ assert cached_response["worker_version"] == job_runner.get_version()
@@ -136 +136 @@ def test_doesnotexist(
- app_config: AppConfig, get_worker, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig
+ app_config: AppConfig, get_job_runner, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig
@@ -139 +139 @@ def test_doesnotexist(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -142 +142 @@ def test_doesnotexist(
- assert worker.process() is False
+ assert job_runner.process() is False
@@ -144 +144 @@ def test_doesnotexist(
- get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -265 +265 @@ def test_not_supported_if_big(
- get_worker,
+ get_job_runner,
@@ -271 +271 @@ def test_not_supported_if_big(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -274,2 +274,2 @@ def test_not_supported_if_big(
- assert worker.process() is False
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ assert job_runner.process() is False
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -281 +281,4 @@ def test_supported_if_gated(
- app_config: AppConfig, get_worker, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, hub_gated_csv: str
+ app_config: AppConfig,
+ get_job_runner,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
+ hub_gated_csv: str,
@@ -285 +288 @@ def test_supported_if_gated(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -288,2 +291,2 @@ def test_supported_if_gated(
- assert worker.process() is True
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ assert job_runner.process() is True
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -296 +299 @@ def test_not_supported_if_gated_with_extra_fields(
- get_worker,
+ get_job_runner,
@@ -302 +305 @@ def test_not_supported_if_gated_with_extra_fields(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -305,2 +308,2 @@ def test_not_supported_if_gated_with_extra_fields(
- assert worker.process() is False
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ assert job_runner.process() is False
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -313 +316 @@ def test_blocked(
- get_worker,
+ get_job_runner,
@@ -319 +322 @@ def test_blocked(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -322,2 +325,2 @@ def test_blocked(
- assert worker.process() is False
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ assert job_runner.process() is False
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -334 +337 @@ def test_compute_splits_response_simple_csv_ok(
- get_worker,
+ get_job_runner,
@@ -342 +345 @@ def test_compute_splits_response_simple_csv_ok(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -345 +348 @@ def test_compute_splits_response_simple_csv_ok(
- result = worker.compute()
+ result = job_runner.compute()
@@ -377 +380 @@ def test_compute_splits_response_simple_csv_error(
- get_worker,
+ get_job_runner,
@@ -385 +388 @@ def test_compute_splits_response_simple_csv_error(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -389 +392 @@ def test_compute_splits_response_simple_csv_error(
- worker.compute()
+ job_runner.compute()
diff --git a/workers/datasets_based/tests/workers/test_sizes.py b/services/worker/tests/job_runners/test_sizes.py
similarity index 93%
rename from workers/datasets_based/tests/workers/test_sizes.py
rename to services/worker/tests/job_runners/test_sizes.py
index 4d4893ff..83326b97 100644
--- a/workers/datasets_based/tests/workers/test_sizes.py
+++ b/services/worker/tests/job_runners/test_sizes.py
@@ -13,2 +13,2 @@ from libcommon.simple_cache import upsert_response
-from datasets_based.config import AppConfig
-from datasets_based.workers.sizes import (
+from worker.config import AppConfig
+from worker.job_runners.sizes import (
@@ -18 +18 @@ from datasets_based.workers.sizes import (
- SizesWorker,
+ SizesJobRunner,
@@ -29 +29 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-def get_worker(
+def get_job_runner(
@@ -33 +33 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -37,2 +37,2 @@ def get_worker(
- ) -> SizesWorker:
- return SizesWorker(
+ ) -> SizesJobRunner:
+ return SizesJobRunner(
@@ -40 +40 @@ def get_worker(
- "type": SizesWorker.get_job_type(),
+ "type": SizesJobRunner.get_job_type(),
@@ -51 +51 @@ def get_worker(
- endpoint=SizesWorker.get_job_type(),
+ endpoint=SizesJobRunner.get_job_type(),
@@ -61 +61 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -243 +243 @@ def test_compute(
- get_worker,
+ get_job_runner,
@@ -254 +254 @@ def test_compute(
- worker = get_worker(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset=dataset, app_config=app_config)
@@ -257 +257 @@ def test_compute(
- worker.compute()
+ job_runner.compute()
@@ -260 +260 @@ def test_compute(
- assert worker.compute() == expected_content
+ assert job_runner.compute() == expected_content
@@ -263 +263 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
@@ -265 +265 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset=dataset, app_config=app_config)
+ job_runner = get_job_runner(dataset=dataset, app_config=app_config)
@@ -267 +267 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker.compute()
+ job_runner.compute()
diff --git a/workers/datasets_based/tests/workers/test_split_names.py b/services/worker/tests/job_runners/test_split_names.py
similarity index 75%
rename from workers/datasets_based/tests/workers/test_split_names.py
rename to services/worker/tests/job_runners/test_split_names.py
index 876a2a63..bf9fbe4d 100644
--- a/workers/datasets_based/tests/workers/test_split_names.py
+++ b/services/worker/tests/job_runners/test_split_names.py
@@ -14,3 +14,3 @@ from libcommon.simple_cache import DoesNotExist, get_response
-from datasets_based.config import AppConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.workers.split_names import SplitNamesWorker
+from worker.config import AppConfig
+from worker.job_runners.split_names import SplitNamesJobRunner
+from worker.resources import LibrariesResource
@@ -22 +22 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
-def get_worker(
+def get_job_runner(
@@ -27 +27 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -32,2 +32,2 @@ def get_worker(
- ) -> SplitNamesWorker:
- return SplitNamesWorker(
+ ) -> SplitNamesJobRunner:
+ return SplitNamesJobRunner(
@@ -35 +35 @@ def get_worker(
- "type": SplitNamesWorker.get_job_type(),
+ "type": SplitNamesJobRunner.get_job_type(),
@@ -45 +45 @@ def get_worker(
- endpoint=SplitNamesWorker.get_job_type(),
+ endpoint=SplitNamesJobRunner.get_job_type(),
@@ -56 +56 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -59 +59 @@ def get_worker(
-def test_process(app_config: AppConfig, get_worker, hub_public_csv: str) -> None:
+def test_process(app_config: AppConfig, get_job_runner, hub_public_csv: str) -> None:
@@ -61,3 +61,3 @@ def test_process(app_config: AppConfig, get_worker, hub_public_csv: str) -> None
- worker = get_worker(dataset, config, app_config)
- assert worker.process() is True
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=hub_public_csv, config=config)
+ job_runner = get_job_runner(dataset, config, app_config)
+ assert job_runner.process() is True
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=hub_public_csv, config=config)
@@ -66 +66 @@ def test_process(app_config: AppConfig, get_worker, hub_public_csv: str) -> None
- assert cached_response["worker_version"] == worker.get_version()
+ assert cached_response["worker_version"] == job_runner.get_version()
@@ -73 +73 @@ def test_process(app_config: AppConfig, get_worker, hub_public_csv: str) -> None
-def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
@@ -76,2 +76,2 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset, config, app_config)
- assert worker.process() is False
+ job_runner = get_job_runner(dataset, config, app_config)
+ assert job_runner.process() is False
@@ -79 +79 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- get_response(kind=worker.processing_step.cache_kind, dataset=dataset, config=config)
+ get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config)
@@ -99 +99 @@ def test_compute_split_names_response(
- get_worker,
+ get_job_runner,
@@ -107 +107 @@ def test_compute_split_names_response(
- worker = get_worker(dataset, config, app_config)
+ job_runner = get_job_runner(dataset, config, app_config)
@@ -109 +109 @@ def test_compute_split_names_response(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -115 +115 @@ def test_compute_split_names_response(
- result = worker.compute()
+ result = job_runner.compute()
@@ -120 +120 @@ def test_compute_split_names_response(
- worker.compute()
+ job_runner.compute()
diff --git a/workers/datasets_based/tests/workers/test_splits.py b/services/worker/tests/job_runners/test_splits.py
similarity index 74%
rename from workers/datasets_based/tests/workers/test_splits.py
rename to services/worker/tests/job_runners/test_splits.py
index 7e85c31a..effca782 100644
--- a/workers/datasets_based/tests/workers/test_splits.py
+++ b/services/worker/tests/job_runners/test_splits.py
@@ -14,3 +14,3 @@ from libcommon.simple_cache import DoesNotExist, get_response
-from datasets_based.config import AppConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.workers.splits import SplitsWorker
+from worker.config import AppConfig
+from worker.job_runners.splits import SplitsJobRunner
+from worker.resources import LibrariesResource
@@ -22 +22 @@ from ..fixtures.hub import HubDatasets
-def get_worker(
+def get_job_runner(
@@ -27 +27 @@ def get_worker(
- def _get_worker(
+ def _get_job_runner(
@@ -31,2 +31,2 @@ def get_worker(
- ) -> SplitsWorker:
- return SplitsWorker(
+ ) -> SplitsJobRunner:
+ return SplitsJobRunner(
@@ -34 +34 @@ def get_worker(
- "type": SplitsWorker.get_job_type(),
+ "type": SplitsJobRunner.get_job_type(),
@@ -44 +44 @@ def get_worker(
- endpoint=SplitsWorker.get_job_type(),
+ endpoint=SplitsJobRunner.get_job_type(),
@@ -55 +55 @@ def get_worker(
- return _get_worker
+ return _get_job_runner
@@ -58 +58 @@ def get_worker(
-def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
+def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
@@ -60,2 +60,2 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker) -> N
- worker = get_worker(dataset, app_config)
- assert worker.should_skip_job() is False
+ job_runner = get_job_runner(dataset, app_config)
+ assert job_runner.should_skip_job() is False
@@ -63,4 +63,4 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker) -> N
- worker.process()
- assert worker.should_skip_job() is True
- worker = get_worker(dataset, app_config, force=True)
- assert worker.should_skip_job() is False
+ job_runner.process()
+ assert job_runner.should_skip_job() is True
+ job_runner = get_job_runner(dataset, app_config, force=True)
+ assert job_runner.should_skip_job() is False
@@ -69 +69 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker) -> N
-def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
+def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner) -> None:
@@ -71,3 +71,3 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None
- worker = get_worker(dataset, app_config)
- assert worker.process() is True
- cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=hub_public_csv)
+ job_runner = get_job_runner(dataset, app_config)
+ assert job_runner.process() is True
+ cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=hub_public_csv)
@@ -76 +76 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None
- assert cached_response["worker_version"] == worker.get_version()
+ assert cached_response["worker_version"] == job_runner.get_version()
@@ -84 +84 @@ def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None
-def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
+def test_doesnotexist(app_config: AppConfig, get_job_runner) -> None:
@@ -86,2 +86,2 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- worker = get_worker(dataset, app_config)
- assert worker.process() is False
+ job_runner = get_job_runner(dataset, app_config)
+ assert job_runner.process() is False
@@ -89 +89 @@ def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
- get_response(kind=worker.processing_step.cache_kind, dataset=dataset)
+ get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset)
@@ -109 +109 @@ def test_compute_splits_response_simple_csv(
- get_worker,
+ get_job_runner,
@@ -118 +118 @@ def test_compute_splits_response_simple_csv(
- worker = get_worker(
+ job_runner = get_job_runner(
@@ -123 +123 @@ def test_compute_splits_response_simple_csv(
- result = worker.compute()
+ result = job_runner.compute()
@@ -128 +128 @@ def test_compute_splits_response_simple_csv(
- worker.compute()
+ job_runner.compute()
diff --git a/workers/datasets_based/tests/test_features.py b/services/worker/tests/test_features.py
similarity index 99%
rename from workers/datasets_based/tests/test_features.py
rename to services/worker/tests/test_features.py
index ca2b8057..ef6c076e 100644
--- a/workers/datasets_based/tests/test_features.py
+++ b/services/worker/tests/test_features.py
@@ -13,2 +13,2 @@ from libcommon.storage import StrPath
-from datasets_based.config import AppConfig
-from datasets_based.features import get_cell_value
+from worker.config import AppConfig
+from worker.features import get_cell_value
diff --git a/workers/datasets_based/tests/test_worker.py b/services/worker/tests/test_job_runner.py
similarity index 84%
rename from workers/datasets_based/tests/test_worker.py
rename to services/worker/tests/test_job_runner.py
index 1c5e3211..fa0ca682 100644
--- a/workers/datasets_based/tests/test_worker.py
+++ b/services/worker/tests/test_job_runner.py
@@ -12,2 +12,2 @@ from libcommon.simple_cache import SplitFullName, upsert_response
-from datasets_based.config import DatasetsBasedConfig
-from datasets_based.worker import ERROR_CODES_TO_RETRY, Worker
+from worker.config import DatasetsBasedConfig
+from worker.job_runner import ERROR_CODES_TO_RETRY, JobRunner
@@ -25 +25 @@ def prepare_and_clean_mongo(
-class DummyWorker(Worker):
+class DummyJobRunner(JobRunner):
@@ -28 +28 @@ class DummyWorker(Worker):
- return DummyWorker._get_dataset_git_revision()
+ return DummyJobRunner._get_dataset_git_revision()
@@ -69 +69 @@ def test_compare_major_version(
- worker = DummyWorker(
+ job_runner = DummyJobRunner(
@@ -85 +85 @@ def test_compare_major_version(
- worker.compare_major_version(other_version)
+ job_runner.compare_major_version(other_version)
@@ -87 +87 @@ def test_compare_major_version(
- assert worker.compare_major_version(other_version) == expected
+ assert job_runner.compare_major_version(other_version) == expected
@@ -105,2 +105,2 @@ class CacheEntry:
- worker_version=DummyWorker.get_version(),
- dataset_git_revision=DummyWorker._get_dataset_git_revision(),
+ worker_version=DummyJobRunner.get_version(),
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
@@ -114,2 +114,2 @@ class CacheEntry:
- worker_version=DummyWorker.get_version(),
- dataset_git_revision=DummyWorker._get_dataset_git_revision(),
+ worker_version=DummyJobRunner.get_version(),
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
@@ -123,2 +123,2 @@ class CacheEntry:
- worker_version=DummyWorker.get_version(),
- dataset_git_revision=DummyWorker._get_dataset_git_revision(),
+ worker_version=DummyJobRunner.get_version(),
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
@@ -137,2 +137,2 @@ class CacheEntry:
- worker_version=DummyWorker.get_version(),
- dataset_git_revision=DummyWorker._get_dataset_git_revision(),
+ worker_version=DummyJobRunner.get_version(),
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
@@ -146,2 +146,2 @@ class CacheEntry:
- worker_version=None, # no worker version
- dataset_git_revision=DummyWorker._get_dataset_git_revision(),
+ worker_version=None, # no version
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
@@ -155,2 +155,2 @@ class CacheEntry:
- worker_version="0.0.1", # a different worker version
- dataset_git_revision=DummyWorker._get_dataset_git_revision(),
+ worker_version="0.0.1", # a different version
+ dataset_git_revision=DummyJobRunner._get_dataset_git_revision(),
@@ -164 +164 @@ class CacheEntry:
- worker_version=DummyWorker.get_version(),
+ worker_version=DummyJobRunner.get_version(),
@@ -173 +173 @@ class CacheEntry:
- worker_version=DummyWorker.get_version(),
+ worker_version=DummyJobRunner.get_version(),
@@ -187 +187 @@ def test_should_skip_job(
- worker = DummyWorker(
+ job_runner = DummyJobRunner(
@@ -214 +214 @@ def test_should_skip_job(
- assert worker.should_skip_job() is expected_skip
+ assert job_runner.should_skip_job() is expected_skip
@@ -228 +228 @@ def test_check_type(
- DummyWorker(
+ DummyJobRunner(
@@ -253 +253 @@ def test_check_type(
- DummyWorker(
+ DummyJobRunner(
@@ -279 +279 @@ def test_create_children_jobs() -> None:
- worker = DummyWorker(
+ job_runner = DummyJobRunner(
@@ -293 +293 @@ def test_create_children_jobs() -> None:
- assert worker.should_skip_job() is False
+ assert job_runner.should_skip_job() is False
@@ -295,2 +295,2 @@ def test_create_children_jobs() -> None:
- worker.run()
- assert worker.should_skip_job() is True
+ job_runner.run()
+ assert job_runner.should_skip_job() is True
diff --git a/workers/datasets_based/tests/test_worker_factory.py b/services/worker/tests/test_job_runner_factory.py
similarity index 55%
rename from workers/datasets_based/tests/test_worker_factory.py
rename to services/worker/tests/test_job_runner_factory.py
index 279d45d6..0ed36874 100644
--- a/workers/datasets_based/tests/test_worker_factory.py
+++ b/services/worker/tests/test_job_runner_factory.py
@@ -11,4 +11,4 @@ from libcommon.storage import StrPath
-from datasets_based.config import AppConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.worker import JobInfo
-from datasets_based.worker_factory import WorkerFactory
+from worker.config import AppConfig
+from worker.job_runner import JobInfo
+from worker.job_runner_factory import JobRunnerFactory
+from worker.resources import LibrariesResource
@@ -23 +23 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph:
- "job_type,expected_worker",
+ "job_type,expected_job_runner",
@@ -25,7 +25,7 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph:
- ("/config-names", "ConfigNamesWorker"),
- ("/splits", "SplitsWorker"),
- ("/first-rows", "FirstRowsWorker"),
- ("/parquet-and-dataset-info", "ParquetAndDatasetInfoWorker"),
- ("/parquet", "ParquetWorker"),
- ("/dataset-info", "DatasetInfoWorker"),
- ("/sizes", "SizesWorker"),
+ ("/config-names", "ConfigNamesJobRunner"),
+ ("/splits", "SplitsJobRunner"),
+ ("/first-rows", "FirstRowsJobRunner"),
+ ("/parquet-and-dataset-info", "ParquetAndDatasetInfoJobRunner"),
+ ("/parquet", "ParquetJobRunner"),
+ ("/dataset-info", "DatasetInfoJobRunner"),
+ ("/sizes", "SizesJobRunner"),
@@ -35 +35 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph:
-def test_create_worker(
+def test_create_job_runner(
@@ -41 +41 @@ def test_create_worker(
- expected_worker: Optional[str],
+ expected_job_runner: Optional[str],
@@ -43 +43 @@ def test_create_worker(
- worker_factory = WorkerFactory(
+ factory = JobRunnerFactory(
@@ -58 +58 @@ def test_create_worker(
- if expected_worker is None:
+ if expected_job_runner is None:
@@ -60 +60 @@ def test_create_worker(
- worker_factory.create_worker(job_info=job_info)
+ factory.create_job_runner(job_info=job_info)
@@ -62,2 +62,2 @@ def test_create_worker(
- worker = worker_factory.create_worker(job_info=job_info)
- assert worker.__class__.__name__ == expected_worker
+ job_runner = factory.create_job_runner(job_info=job_info)
+ assert job_runner.__class__.__name__ == expected_job_runner
diff --git a/workers/datasets_based/tests/test_worker_loop.py b/services/worker/tests/test_loop.py
similarity index 60%
rename from workers/datasets_based/tests/test_worker_loop.py
rename to services/worker/tests/test_loop.py
index e843c976..39012540 100644
--- a/workers/datasets_based/tests/test_worker_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -8,5 +8,5 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from datasets_based.config import AppConfig, DatasetsBasedConfig, WorkerLoopConfig
-from datasets_based.resources import LibrariesResource
-from datasets_based.worker import JobInfo, Worker
-from datasets_based.worker_factory import BaseWorkerFactory
-from datasets_based.worker_loop import WorkerLoop
+from worker.config import AppConfig, DatasetsBasedConfig, LoopConfig
+from worker.job_runner import JobInfo, JobRunner
+from worker.job_runner_factory import BaseJobRunnerFactory
+from worker.loop import Loop
+from worker.resources import LibrariesResource
@@ -15 +15 @@ from datasets_based.worker_loop import WorkerLoop
-class DummyWorker(Worker):
+class DummyJobRunner(JobRunner):
@@ -32 +32 @@ class DummyWorker(Worker):
-class DummyWorkerFactory(BaseWorkerFactory):
+class DummyJobRunnerFactory(BaseJobRunnerFactory):
@@ -38,2 +38,2 @@ class DummyWorkerFactory(BaseWorkerFactory):
- def _create_worker(self, job_info: JobInfo) -> Worker:
- return DummyWorker(
+ def _create_job_runner(self, job_info: JobInfo) -> JobRunner:
+ return DummyJobRunner(
@@ -54 +54 @@ def test_process_next_job(
- worker_factory = DummyWorkerFactory(processing_step=test_processing_step)
+ factory = DummyJobRunnerFactory(processing_step=test_processing_step)
@@ -56 +56 @@ def test_process_next_job(
- worker_loop = WorkerLoop(
+ loop = Loop(
@@ -59,2 +59,2 @@ def test_process_next_job(
- worker_factory=worker_factory,
- worker_loop_config=WorkerLoopConfig(),
+ job_runner_factory=factory,
+ loop_config=LoopConfig(),
@@ -62 +62 @@ def test_process_next_job(
- assert worker_loop.process_next_job() is False
+ assert loop.process_next_job() is False
@@ -66,4 +66,4 @@ def test_process_next_job(
- worker_loop.queue.upsert_job(dataset=dataset, config=config, split=split)
- worker_loop.queue.is_job_in_process(dataset=dataset, config=config, split=split) is True
- assert worker_loop.process_next_job() is True
- worker_loop.queue.is_job_in_process(dataset=dataset, config=config, split=split) is False
+ loop.queue.upsert_job(dataset=dataset, config=config, split=split)
+ loop.queue.is_job_in_process(dataset=dataset, config=config, split=split) is True
+ assert loop.process_next_job() is True
+ loop.queue.is_job_in_process(dataset=dataset, config=config, split=split) is False
diff --git a/workers/datasets_based/tests/test_resources.py b/services/worker/tests/test_resources.py
similarity index 97%
rename from workers/datasets_based/tests/test_resources.py
rename to services/worker/tests/test_resources.py
index b83582d0..7fa7872d 100644
--- a/workers/datasets_based/tests/test_resources.py
+++ b/services/worker/tests/test_resources.py
@@ -7 +7 @@ import pytest
-from datasets_based.resources import LibrariesResource
+from worker.resources import LibrariesResource
diff --git a/workers/datasets_based/vendors/trec-car-tools/.gitignore b/services/worker/vendors/trec-car-tools/.gitignore
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/.gitignore
rename to services/worker/vendors/trec-car-tools/.gitignore
diff --git a/workers/datasets_based/vendors/trec-car-tools/.travis.yml b/services/worker/vendors/trec-car-tools/.travis.yml
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/.travis.yml
rename to services/worker/vendors/trec-car-tools/.travis.yml
diff --git a/workers/datasets_based/vendors/trec-car-tools/.travis/test.sh b/services/worker/vendors/trec-car-tools/.travis/test.sh
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/.travis/test.sh
rename to services/worker/vendors/trec-car-tools/.travis/test.sh
diff --git a/workers/datasets_based/vendors/trec-car-tools/LICENSE b/services/worker/vendors/trec-car-tools/LICENSE
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/LICENSE
rename to services/worker/vendors/trec-car-tools/LICENSE
diff --git a/workers/datasets_based/vendors/trec-car-tools/README.mkd b/services/worker/vendors/trec-car-tools/README.mkd
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/README.mkd
rename to services/worker/vendors/trec-car-tools/README.mkd
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/Makefile b/services/worker/vendors/trec-car-tools/python3/Makefile
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/Makefile
rename to services/worker/vendors/trec-car-tools/python3/Makefile
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/README.mkd b/services/worker/vendors/trec-car-tools/python3/README.mkd
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/README.mkd
rename to services/worker/vendors/trec-car-tools/python3/README.mkd
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/annotated_content.py b/services/worker/vendors/trec-car-tools/python3/annotated_content.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/annotated_content.py
rename to services/worker/vendors/trec-car-tools/python3/annotated_content.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/trec_car/__init__.py b/services/worker/vendors/trec-car-tools/python3/build/lib/trec_car/__init__.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/trec_car/__init__.py
rename to services/worker/vendors/trec-car-tools/python3/build/lib/trec_car/__init__.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/trec_car/format_runs.py b/services/worker/vendors/trec-car-tools/python3/build/lib/trec_car/format_runs.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/trec_car/format_runs.py
rename to services/worker/vendors/trec-car-tools/python3/build/lib/trec_car/format_runs.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/trec_car/read_data.py b/services/worker/vendors/trec-car-tools/python3/build/lib/trec_car/read_data.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/trec_car/read_data.py
rename to services/worker/vendors/trec-car-tools/python3/build/lib/trec_car/read_data.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/conf.py b/services/worker/vendors/trec-car-tools/python3/conf.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/conf.py
rename to services/worker/vendors/trec-car-tools/python3/conf.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/format_runs_test.py b/services/worker/vendors/trec-car-tools/python3/format_runs_test.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/format_runs_test.py
rename to services/worker/vendors/trec-car-tools/python3/format_runs_test.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/index.rst b/services/worker/vendors/trec-car-tools/python3/index.rst
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/index.rst
rename to services/worker/vendors/trec-car-tools/python3/index.rst
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/read_data_test.py b/services/worker/vendors/trec-car-tools/python3/read_data_test.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/read_data_test.py
rename to services/worker/vendors/trec-car-tools/python3/read_data_test.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/requirements.txt b/services/worker/vendors/trec-car-tools/python3/requirements.txt
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/requirements.txt
rename to services/worker/vendors/trec-car-tools/python3/requirements.txt
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/setup.py b/services/worker/vendors/trec-car-tools/python3/setup.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/setup.py
rename to services/worker/vendors/trec-car-tools/python3/setup.py
diff --git a/workers/datasets_based/vendors/trec-car-tools/python3/test.py b/services/worker/vendors/trec-car-tools/python3/test.py
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/python3/test.py
rename to services/worker/vendors/trec-car-tools/python3/test.py
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car/__init__.py b/services/worker/vendors/trec-car-tools/python3/trec_car/__init__.py
new file mode 100644
index 00000000..ccb9b451
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car/__init__.py
@@ -0,0 +1,5 @@
+"""__init__ module for trec-car-tools, imports all necessary functions for reading cbor data provided in the TREC CAR"""
+
+__version__ = 1.0
+
+__all__ = ['read_data', 'format_runs']
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car/format_runs.py b/services/worker/vendors/trec-car-tools/python3/trec_car/format_runs.py
new file mode 100644
index 00000000..cc859620
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car/format_runs.py
@@ -0,0 +1,59 @@
+import csv
+import urllib.parse
+from typing import *
+
+
+
+def encode_section_path(page_id, section_path):
+ elements = [page_id] + section_path
+
+ return '/'.join([urllib.parse.quote(elem) for elem in elements])
+ # return urllib.parse.urlencode({'page':page_id, 'sectionpath':section_path})
+
+def encode_page_only(page_id):
+ return urllib.parse.quote(page_id)
+
+
+class RankingEntry(object):
+ """
+ A paragraph within a Wikipedia page.
+
+ Attributes:
+ paragraph The content of the Paragraph (which in turn contain a list of ParaBodys)
+ """
+ def __init__(self, query_id:str, paragraph_id:str, rank:int, score:float, exp_name:str=None, paragraph_content:str=None):
+ assert(rank > 0)
+ self.query_id = query_id
+ self.paragraph_id = paragraph_id
+ self.rank = rank
+ self.score = score
+ self.exp_name = exp_name
+ self.paragraph_content = paragraph_content
+
+ def to_trec_eval_row(self, alternative_exp_name=None, page_only=False):
+ exp_name_ = alternative_exp_name if alternative_exp_name is not None \
+ else self.exp_name
+ return [self.query_id, 'Q0', self.paragraph_id, self.rank, self.score, exp_name_]
+
+#
+# csv.register_dialect(
+# 'trec_eval',
+# delimiter = ' ',
+# quotechar = '"',
+# doublequote = False,
+# skipinitialspace = False,
+# lineterminator = '\n',
+# quoting = csv.QUOTE_NONE)
+#
+#
+# def configure_csv_writer(fileobj):
+# 'Convenience method to create a csv writer with the trec_eval_dialect'
+# return csv.writer(fileobj, dialect='trec_eval')
+#
+
+def format_run(writer, ranking_of_paragraphs, exp_name=None):
+ 'write one ranking to the csv writer'
+ for elem in ranking_of_paragraphs:
+ # query-number Q0 document-id rank score Exp
+ writer.write(" ".join([str(x) for x in elem.to_trec_eval_row(exp_name)]))
+ writer.write("\n")
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car/read_data.py b/services/worker/vendors/trec-car-tools/python3/trec_car/read_data.py
new file mode 100644
index 00000000..6ba92e94
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car/read_data.py
@@ -0,0 +1,778 @@
+# Use python 3.6 or higher
+# obsolete: conda install -c auto cbor=0.1.4
+
+from __future__ import print_function
+
+from abc import abstractmethod
+
+import cbor
+import itertools
+import typing
+
+PageId = str
+PageName = str
+
+class CborElementNotDefinedException(Exception):
+ def __init__(self, cbor):
+ self.cbor = cbor
+ Exception.__init__(self, 'unknown Cbor element encountrered: %s' % str(cbor))
+
+class WrongCarFileException(Exception):
+ def __init__(self, file_type, expected_file_types):
+ self.file_type = file_type
+ self.expected_file_types = expected_file_types
+ Exception.__init__(self, 'Open method does not support CAR file type: %s. Instead expect following CAR file types: %s' % (str(file_type), str(expected_file_types)))
+
+class BrokenCborFileException(Exception):
+ def __init__(self):
+ Exception.__init__(self, 'Corrupt, incomplete, or otherwise broken CBOR file. Please re-download or contact the organizers or use appropriate reader to open this file.')
+
+
+class Page(object):
+ """
+ The name and skeleton of a Wikipedia page.
+
+ .. attribute:: page_name
+
+ :rtype: PageName
+
+ The name of the page.
+
+ .. attribute:: skeleton
+
+ :rtype: typing.List[PageSkeleton]
+
+ The contents of the page
+
+ .. attribute:: page_type
+
+ :rtype: PageType
+
+ Type about the page
+
+ .. attribute:: page_meta
+
+ :rtype: PageMetadata
+
+ Metadata about the page
+ """
+ def __init__(self, page_name, page_id, skeleton, page_type, page_meta):
+ self.page_name = page_name
+ self.page_id = page_id
+ self.skeleton = list(skeleton)
+ self.child_sections = [child for child in self.skeleton if isinstance(child, Section)]
+ self.page_type = page_type
+ self.page_meta = page_meta
+
+ def deep_headings_list(self):
+ return [child.nested_headings() for child in self.child_sections]
+
+ def flat_headings_list(self):
+ """ return
+ Returns a flat list of headings contained by the :class:`Page`.
+
+ :rtype: typing.List[Section]
+ """
+ def flatten(prefix, headings):
+ for section, children in headings:
+ new_prefix = prefix + [section]
+ if len(children)>0 :
+ yield new_prefix
+ yield from flatten(new_prefix, children)
+ else:
+ yield new_prefix
+
+ deep_headings = self.deep_headings_list()
+ return list(flatten([], deep_headings))
+
+ def get_infoboxes(self):
+ toplevel_infoboxes = [child for child in self.skeleton if isinstance(child, InfoBox)]
+ section_infoboxes = [section.get_infoboxes()
+ for sections
+ in self.flat_headings_list()
+ for section in sections]
+ return toplevel_infoboxes + list(itertools.chain.from_iterable(section_infoboxes))
+
+
+ @staticmethod
+ def from_cbor(cbor):
+
+ if not (cbor[0] == 0 or cbor[0] == 1): # tag
+ raise CborElementNotDefinedException(cbor)
+ pagename = cbor[1]
+ pageId = cbor[2].decode('ascii')
+
+ if len(cbor)==4:
+ return Page(pagename, pageId, map(PageSkeleton.from_cbor, cbor[3]), ArticlePage, PageMetadata.default())
+ else:
+ page_type = PageType.from_cbor(cbor[4])
+ return Page(pagename, pageId, map(PageSkeleton.from_cbor, cbor[3]), page_type, PageMetadata.from_cbor(cbor[5]))
+
+ def __str__(self):
+ return "Page(%s)" % self.page_name
+
+ def to_string(self):
+ """
+ Render a string representation of the page.
+
+ :rtype: str
+ """
+ return self.page_name + self.page_meta +\
+ '\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~' + '\n'.join(str(s) for s in self.skeleton)
+
+ def nested_headings(self):
+ """
+ Each heading recursively represented by a pair of ``(heading,
+ list_of_child_sections)``.
+
+ :rtype: typing.List[typing.Tuple[Section, typing.List[Section]]]
+ """
+ result = [child.nested_headings() for child in self.child_sections]
+ return result
+
+ def outline(self):
+ return self.child_sections
+
+ def get_text(self):
+ """Include all visible text below this elements. Includes Captions of images, but no headings and no infoboxes. See `get_text_with_headings` for a version that includes headings."""
+ return '\n'.join(skel.get_text() for skel in self.skeleton)
+
+
+ def get_text_with_headings(self, include_heading = False):
+ """Include all visible text below this elements. While the heading of this element is excluded, headings of subsections will be included. Captions of images are excluded."""
+ return '\n'.join(skel.get_text_with_headings(include_heading = True) for skel in self.skeleton)
+
+class PageType(object):
+ """
+ An abstract base class representing the various types of pages.
+
+ Subclasses include
+
+ * :class:`ArticlePage`
+ * :class:`CategoryPage`
+ * :class:`DisambiguationPage`
+ * :class:`RedirectPage`
+ """
+ @staticmethod
+ def from_cbor(cbor):
+ typetag = cbor[0]
+ if typetag == 0: return ArticlePage()
+ elif typetag == 1: return CategoryPage()
+ elif typetag == 2: return DisambiguationPage()
+ elif typetag == 3:
+ target = cbor[1]
+ if type(target) == list: # TODO this is almost certainly wrong
+ targetPage = target[1]
+ else:
+ targetPage = target.decode('ascii')
+ return RedirectPage(targetPage)
+ else:
+ raise CborElementNotDefinedException(cbor)
+
+class ArticlePage(PageType):
+ ''
+ def __init__(self):
+ pass
+ def __str__(self): return "ArticlePage"
+
+class CategoryPage(PageType):
+ def __init__(self):
+ pass
+ def __str__(self): return "CategoryPage"
+
+class DisambiguationPage(PageType):
+ def __init__(self):
+ pass
+ def __str__(self): return "Disambiguation Page"
+
+class RedirectPage(PageType):
+ """
+ .. attribute:: targetPage
+
+ :rtype: PageId
+
+ The target of the redirect.
+ """
+ def __init__(self, targetPage):
+ self.targetPage = targetPage
+ def __str__(self):
+ return "RedirectPage " + self.targetPage
+
+class PageMetadata(object):
+ """
+ Meta data for a page
+
+ .. attribute:: redirectNames
+
+ :rtype: PageName
+
+ Names of pages which redirect to this page
+
+ .. attribute:: disambiguationNames
+
+ :rtype: PageName
+
+ Names of disambiguation pages which link to this page
+
+ .. attribute:: disambiguationId
+
+ :rtype: PageId
+
+ Page IDs of disambiguation pages which link to this page
+
+ .. attribute:: categoryNames
+
+ :rtype: str
+
+ Page names of categories to which this page belongs
+
+ .. attribute:: categoryIds
+
+ :rtype: str
+
+ Page IDs of categories to which this page belongs
+
+ .. attribute:: inlinkIds
+
+ :rtype: str
+
+ Page IDs of pages containing inlinks
+
+ .. attribute:: inlinkAnchors
+ inlinkAnchor frequencies
+
+ :rtype: str
+
+ (Anchor text, frequency) of pages containing inlinks
+ """
+ def __init__(self, redirectNames, disambiguationNames, disambiguationIds, categoryNames, categoryIds, inlinkIds,
+ inlinkAnchors):
+ self.inlinkAnchors = inlinkAnchors
+ self.inlinkIds = inlinkIds
+ self.categoryIds = categoryIds
+ self.categoryNames = categoryNames
+ self.disambiguationIds = disambiguationIds
+ self.disambiguationNames = disambiguationNames
+ self.redirectNames = redirectNames
+
+ @staticmethod
+ def default():
+ return PageMetadata(None, None, None, None, None, None, None)
+
+ def __str__(self):
+ redirStr = ("" if self.redirectNames is None else (" redirected = "+", ".join([name for name in self.redirectNames])))
+ disamStr = ("" if self.disambiguationNames is None else (" disambiguated = "+", ".join([name for name in self.disambiguationNames])))
+ catStr = ("" if self.redirectNames is None else (" categories = "+", ".join([name for name in (self.categoryNames or [])])))
+ inlinkStr = ("" if self.inlinkIds is None else (" inlinks = "+", ".join([name for name in self.inlinkIds])))
+ # inlinkAnchorStr = str (self.inlinkAnchors)
+ inlinkAnchorStr = ("" if self.inlinkAnchors is None else
+ (" inlinkAnchors = "+", ".join(
+ [ ("%s: %d" % (name, freq)) for (name, freq) in self.inlinkAnchors]
+ # [ ("%s: " % (name)) for (name, freq) in self.inlinkAnchors] \
+ )))
+ return "%s \n%s \n%s \n%s \n%s\n" % (redirStr, disamStr, catStr, inlinkStr, inlinkAnchorStr)
+
+ @staticmethod
+ def from_cbor(cbor):
+ redirectNames=None
+ disambiguationNames=None
+ disambiguationIds=None
+ categoryNames=None
+ categoryIds=None
+ inlinkIds=None
+ inlinkAnchors=None
+
+ def decodeListOfIdList(cbor):
+ if len(cbor)==0: return None
+ else:
+ return [elem.decode('ascii') for elem in cbor]
+
+ def decodeListOfNameList(cbor):
+ if len(cbor)==0: return None
+ else:
+ return cbor
+
+ def decodeListOfNameIntList(cbor):
+ if len(cbor)==0: return None
+ else:
+ # need to convert list of pair-lists to lists of pair-tuples
+ return [(elem[0], elem[1]) for elem in cbor]
+
+ for i in range(0, len(cbor), 2):
+ tag = cbor[i][0]
+ cbor_data = cbor[i+1]
+
+ if tag == 0:
+ redirectNames = decodeListOfNameList(cbor_data)
+ elif tag == 1:
+ disambiguationNames=decodeListOfNameList(cbor_data)
+ elif tag == 2:
+ disambiguationIds=decodeListOfIdList(cbor_data)
+ elif tag == 3:
+ categoryNames=decodeListOfNameList(cbor_data)
+ elif tag == 4:
+ categoryIds=decodeListOfIdList(cbor_data)
+ elif tag == 5:
+ inlinkIds=decodeListOfIdList(cbor_data)
+
+ elif tag == 6:
+ # compatability with v1.6
+ inlinkAnchors = [(anchor, 1) for anchor in decodeListOfNameList(cbor_data)]
+ elif tag == 7:
+ # compatability with v2.0
+ inlinkAnchors = decodeListOfNameIntList(cbor_data)
+ i+=2
+
+ return PageMetadata(redirectNames, disambiguationNames, disambiguationIds, categoryNames, categoryIds, inlinkIds, inlinkAnchors)
+
+class PageSkeleton(object):
+ """
+ An abstract superclass for the various types of page elements. Subclasses include:
+
+ * :class:`Section`
+ * :class:`Para`
+ * :class:`Image`
+
+ """
+ @staticmethod
+ def from_cbor(cbor):
+ tag = cbor[0]
+ if tag == 0: # section
+ heading = cbor[1]
+ headingId = cbor[2].decode('ascii')
+ return Section(heading, headingId, map(PageSkeleton.from_cbor, cbor[3]))
+ elif tag == 1: # para-wrapper
+ return Para(Paragraph.from_cbor(cbor[1]))
+ elif tag == 2: #images
+ imageUrl = cbor[1]
+ caption = [PageSkeleton.from_cbor(elem) for elem in cbor[2]]
+ return Image(imageUrl, caption=caption)
+ elif tag == 3: # paragraph
+ level = cbor[1]
+ body = Paragraph.from_cbor(cbor[2])
+ return List(level, body)
+ elif tag == 4: # infobox
+ infobox_title = cbor[1]
+ cbor_entries = cbor[2]
+ entries = [ (kv[0], PageSkeleton.from_cbor(kv[1][0])) for kv in cbor_entries if kv[1] and kv[1][0]] # if no value is defined kv[1] will be null.
+ return InfoBox(infobox_title, entries)
+ else:
+ raise CborElementNotDefinedException(cbor)
+
+ def get_text(self):
+ """Includes visible text of this element and below. Headings are excluded. Image Captions are included. Infoboxes are ignored. (For a version with headers and no captions see `get_text_with_headings` """
+ raise NotImplementedError
+
+ def get_text_with_headings(self, include_heading = False):
+ """Include all visible text below this elements. While the heading of this element is excluded, headings of subsections will be included. Captions of images are excluded."""
+ raise NotImplementedError
+
+
+class Section(PageSkeleton):
+ """
+ A section of a Wikipedia page.
+
+ .. attribute:: heading
+
+ :rtype: str
+
+ The section heading.
+
+ .. attribute:: headingId
+
+ :rtype: str
+
+ The unique identifier of a section heading.
+
+ .. attribute:: children
+
+ :rtype: typing.List[PageSkeleton]
+
+ The :class:`PageSkeleton` elements contained by the section.
+ """
+ def __init__(self, heading, headingId, children):
+ self.heading = heading
+ self.headingId = headingId
+ self.children = list(children)
+ self.child_sections = [child for child in self.children if isinstance(child, Section)]
+
+ def str_(self, level):
+ bar = "".join("="*level)
+ children = "".join(c.str_(level=level+1) for c in self.children)
+ return "\n%s %s %s\n\n%s" % (bar, self.heading, bar, children)
+
+ def __str__(self):
+ return self.str_(level=1)
+
+ def __getitem__(self, idx):
+ return self.children[idx]
+
+ def nested_headings(self):
+ return (self, [child.nested_headings() for child in self.child_sections])
+
+ def get_text(self):
+ return '\n'.join(child.get_text() for child in self.children)
+
+ def get_text_with_headings(self, include_heading = False):
+ opt_heading = self.heading + "\n" if include_heading else ""
+ return opt_heading + '\n'.join(child.get_text_with_headings(include_heading = True) for child in self.children)
+
+
+ def get_infoboxes(self):
+ return [child for child in self.children if isinstance(child, InfoBox)]
+
+class Para(PageSkeleton):
+ """
+ A paragraph within a Wikipedia page.
+
+ .. attribute:: paragraph
+
+ :rtype: Paragraph
+
+ The content of the Paragraph (which in turn contain a list of :class:`ParaBody`\ s)
+ """
+ def __init__(self, paragraph):
+ self.paragraph = paragraph
+
+ def str_(self, level=None):
+ return str(self.paragraph)
+
+ def __str__(self):
+ return self.str_()
+
+ def get_text(self):
+ return self.paragraph.get_text()
+
+ def get_text_with_headings(self, include_heading = False):
+ return self.get_text()
+
+class Image(PageSkeleton):
+ """
+ An image within a Wikipedia page.
+
+ .. attribute:: caption
+
+ :rtype: str
+
+ PageSkeleton representing the caption of the image
+
+ .. attribute:: imageurl
+
+ :rtype: str
+
+ URL to the image; spaces need to be replaced with underscores, Wikimedia
+ Commons namespace needs to be prefixed
+ """
+ def __init__(self, imageurl, caption):
+ self.caption = caption
+ self.imageurl = imageurl
+
+ def str_(self, level=None):
+ return str("!["+self.imageurl+"]. Caption: "+(''.join([str(skel) for skel in self.caption])))
+
+ def __str__(self):
+ return self.str_()
+
+ def get_text(self):
+ return '\n'.join(skel.get_text() for skel in self.caption)
+
+ def get_text_with_headings(self, include_heading = False):
+ return ''
+
+class List(PageSkeleton):
+ """
+ An list element within a Wikipedia page.
+
+ .. attribute:: level
+
+ :rtype: int
+
+ The list nesting level
+
+ .. attribute:: body
+
+ A :class:`Paragraph` containing the list element contents.
+ """
+ def __init__(self, level, body):
+ self.level = level
+ self.body = body
+
+ def str_(self, level=None):
+ return str("*" * self.level + " " + str(self.body) + '\n')
+
+ def __str__(self):
+ return self.str_()
+
+
+ def get_text(self):
+ return self.body.get_text()
+
+ def get_text_with_headings(self, include_heading = False):
+ return self.get_text()
+
+
+class InfoBox(PageSkeleton):
+ def __init__(self, infobox_type, entries):
+ """
+ An list element within a Wikipedia page.
+
+ .. attribute:: infobox_type
+
+ :rtype: str
+
+ The title/type of the infobox
+
+ .. attribute:: entries
+
+ Key-value pair, where key is a string, and value is a :class:`PageSkeleton` containing the value. Values are often paragraphs or images, but they can also be lists.
+ """
+ self.title = infobox_type
+ self.entries = entries
+
+ def str_(self, level=None):
+ return self.title+ "\n"+ ("\n".join([key+": "+str(values) for (key,values) in self.entries]))
+
+ def __str__(self):
+ return self.str_()
+
+
+
+ def get_text(self):
+ return ''
+
+ def get_text_with_headings(self, include_heading = False):
+ return ''
+
+
+class Paragraph(object):
+ """
+ A paragraph.
+ """
+ def __init__(self, para_id, bodies):
+ self.para_id = para_id
+ self.bodies = list(bodies)
+
+ @staticmethod
+ def from_cbor(cbor):
+ if (not cbor[0] == 0):
+ raise CborElementNotDefinedException(cbor)
+
+ paragraphId = cbor[1].decode('ascii')
+ return Paragraph(paragraphId, map(ParaBody.from_cbor, cbor[2]))
+
+ def get_text(self):
+ """
+ Get all of the contained text.
+
+ :rtype: str
+ """
+ return ''.join([body.get_text() for body in self.bodies])
+
+ def str_(self, level=None):
+ return ' '.join(str(body) for body in self.bodies)
+
+ def __str__(self):
+ return self.str_()
+
+
+class ParaBody(object):
+ """
+ An abstract superclass representing a bit of :class:`Paragraph` content.
+ """
+ @staticmethod
+ def from_cbor(cbor):
+ tag = cbor[0]
+ if tag == 0:
+ return ParaText(cbor[1])
+ elif tag == 1:
+ cbor_ = cbor[1]
+ linkSection = None
+ if len(cbor_[2]) == 1:
+ linkSection = cbor_[2][0]
+ linkTargetId = cbor_[3].decode('ascii')
+ return ParaLink(cbor_[1], linkSection, linkTargetId, cbor_[4])
+ else:
+ raise CborElementNotDefinedException(cbor)
+
+ @abstractmethod
+ def get_text(self):
+ """
+ Get all of the text within a :class:`ParaBody`.
+
+ :rtype: str
+ """
+ raise NotImplementedError
+
+class ParaText(ParaBody):
+ """
+ A bit of plain text from a paragraph.
+
+ .. attribute:: text
+
+ :rtype: str
+
+ The text
+ """
+ def __init__(self, text):
+ self.text = text
+
+ def get_text(self):
+ return self.text
+
+ def str_(self, level=None):
+ return self.text
+
+ def __str__(self):
+ return self.str_()
+
+
+class ParaLink(ParaBody):
+ """
+ A link within a paragraph.
+
+ .. attribute:: page
+
+ :rtype: PageName
+
+ The page name of the link target
+
+ .. attribute:: pageid
+
+ :rtype: PageId
+
+ The link target as trec-car identifer
+
+ .. attribute:: link_section
+
+ :rtype: str
+
+ Section anchor of link target (i.e. the part after the ``#`` in the
+ URL), or ``None``.
+
+ .. attribute:: anchor_text
+
+ :rtype: str
+
+ The anchor text of the link
+ """
+ def __init__(self, page, link_section, pageid, anchor_text):
+ self.page = page
+ self.pageid = pageid
+ self.link_section = link_section
+ self.anchor_text = anchor_text
+
+ def get_text(self):
+ return self.anchor_text
+
+ def str_(self, level=None):
+ return "[%s](%s)" % (self.anchor_text, self.page)
+
+ def __str__(self):
+ return self.str_()
+
+
+def _iter_with_header(file, parse, expected_file_types):
+ maybe_hdr = cbor.load(file)
+ if isinstance(maybe_hdr, list) and maybe_hdr[0] == 'CAR':
+ # we have a header
+ file_type = maybe_hdr[1][0]
+ if not file_type in expected_file_types:
+ # print( 'File type tag is expected to be ', (" ".join(expected_file_types)), 'but given file is of type ', file_type)
+ # print('Did not expect file of type', file_type)
+ raise WrongCarFileException(file_type, expected_file_types)
+
+ # read beginning of variable-length list
+ if (not file.read(1) == b'\x9f'):
+ raise BrokenCborFileException()
+ else:
+ yield parse(maybe_hdr)
+
+ while True:
+ try:
+ # Check for break symbol
+ if (peek_for_break(file)):
+ break
+
+ yield parse(cbor.load(file))
+ except EOFError:
+ break
+
+def peek_for_break(cbor):
+ b = cbor.peek(1)
+ return b[0:1] == b'\xff'
+
+
+def iter_annotations(file):
+ """
+ Iterate over the :class:`Page`\ s of an annotations file.
+
+ :type file: typing.BinaryIO
+ :rtype: typing.Iterator[Page]
+ """
+ return _iter_with_header(file, Page.from_cbor, [0,1])
+ # return TrecCarHeader.from_cbor(file)
+
+
+
+def iter_pages(file):
+ """
+ Iterate over the :class:`Page`\ s of an annotations file.
+
+ :type file: typing.BinaryIO
+ :rtype: typing.Iterator[Page]
+ """
+ return _iter_with_header(file, Page.from_cbor, [0])
+
+
+
+def iter_outlines(file):
+ """
+ Iterate over the :class:`Page`\ s of an annotations file.
+
+ :type file: typing.BinaryIO
+ :rtype: typing.Iterator[Page]
+ """
+ return _iter_with_header(file, Page.from_cbor, [1])
+
+
+def iter_paragraphs(file):
+ """
+ Iterate over the :class:`Paragraph`\ s of an paragraphs file.
+
+ :type file: typing.BinaryIO
+ :rtype: typing.Iterator[Paragraph]
+ """
+ return _iter_with_header(file, Paragraph.from_cbor, [2])
+
+def dump_annotations(file):
+ for page in iter_annotations(file):
+ print(page.to_string())
+
+def with_toc(read_val):
+ class AnnotationsFile(object):
+ def __init__(self, fname):
+ """
+ Read annotations from a file.
+
+ Arguments:
+ fname The name of the CBOR file. A table-of-contents file is
+ also expected to be present.
+ """
+ self.cbor = open(fname, 'rb')
+ self.toc = cbor.load(open(fname+'.toc', 'rb'))
+
+ def keys(self):
+ """ The page names contained in an annotations file. """
+ return self.toc.keys()
+
+ def get(self, page):
+ """ Lookup a page by name. Returns a Page or None """
+ offset = self.toc.get(page)
+ if offset is not None:
+ self.cbor.seek(offset)
+ return read_val(cbor.load(self.cbor))
+ return None
+ return AnnotationsFile
+
+AnnotationsFile = with_toc(Page.from_cbor)
+ParagraphsFile = with_toc(Paragraph.from_cbor)
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/PKG-INFO b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/PKG-INFO
new file mode 100644
index 00000000..b021b26c
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/PKG-INFO
@@ -0,0 +1,14 @@
+Metadata-Version: 2.1
+Name: trec-car-tools
+Version: 2.5.4
+Summary: Support tools for TREC CAR participants. Also see trec-car.cs.unh.edu
+Home-page: https://github.com/TREMA-UNH/trec-car-tools/python3
+Author: laura-dietz
+Author-email: [email protected]
+License: BSD 3-Clause
+Keywords: wikipedia,complex answer retrieval,trec car
+Classifier: Programming Language :: Python :: 3
+Classifier: Programming Language :: Python :: 3.6
+Classifier: Programming Language :: Python :: 3.7
+Classifier: Programming Language :: Python :: 3.8
+Requires-Python: >=3.6
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/SOURCES.txt b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/SOURCES.txt
new file mode 100644
index 00000000..3f84c7a8
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/SOURCES.txt
@@ -0,0 +1,9 @@
+setup.py
+trec_car/__init__.py
+trec_car/format_runs.py
+trec_car/read_data.py
+trec_car_tools.egg-info/PKG-INFO
+trec_car_tools.egg-info/SOURCES.txt
+trec_car_tools.egg-info/dependency_links.txt
+trec_car_tools.egg-info/requires.txt
+trec_car_tools.egg-info/top_level.txt
\ No newline at end of file
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/dependency_links.txt b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/dependency_links.txt
new file mode 100644
index 00000000..8b137891
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/dependency_links.txt
@@ -0,0 +1 @@
+
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/requires.txt b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/requires.txt
new file mode 100644
index 00000000..0b814045
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/requires.txt
@@ -0,0 +1,2 @@
+cbor>=1.0.0
+numpy>=1.11.2
diff --git a/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/top_level.txt b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/top_level.txt
new file mode 100644
index 00000000..f0eefdfa
--- /dev/null
+++ b/services/worker/vendors/trec-car-tools/python3/trec_car_tools.egg-info/top_level.txt
@@ -0,0 +1 @@
+trec_car
diff --git a/workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/pom.xml b/services/worker/vendors/trec-car-tools/trec-car-tools-example/pom.xml
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/pom.xml
rename to services/worker/vendors/trec-car-tools/trec-car-tools-example/pom.xml
diff --git a/workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarBuildLuceneIndex.java b/services/worker/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarBuildLuceneIndex.java
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarBuildLuceneIndex.java
rename to services/worker/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarBuildLuceneIndex.java
diff --git a/workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarQueryLuceneIndex.java b/services/worker/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarQueryLuceneIndex.java
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarQueryLuceneIndex.java
rename to services/worker/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarQueryLuceneIndex.java
diff --git a/workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarToolsExample.java b/services/worker/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarToolsExample.java
similarity index 100%
rename from workers/datasets_based/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarToolsExample.java
rename to services/worker/vendors/trec-car-tools/trec-car-tools-example/src/main/java/edu/unh/cs/TrecCarToolsExample.java
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index 3bbca9a0..676cc299 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -76 +76 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
@@ -94 +94 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
@@ -112 +112 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
@@ -130 +130 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
@@ -157 +157 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
@@ -183 +183 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
@@ -198 +198 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
@@ -213 +213 @@ services:
- dockerfile: workers/datasets_based/Dockerfile
+ dockerfile: services/worker/Dockerfile
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
index 33a4e81a..4b488ce9 100644
--- a/tools/docker-compose-dev-base.yml
+++ b/tools/docker-compose-dev-base.yml
@@ -34 +34 @@ services:
- - ../workers/datasets_based/src:/src/workers/datasets_based/src
+ - ../services/worker/src:/src/services/worker/src
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 33384dab..208b0a4d 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -76 +76 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
@@ -94 +94 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
@@ -112 +112 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
@@ -130 +130 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
@@ -157 +157 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
@@ -183 +183 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
@@ -198 +198 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
@@ -213 +213 @@ services:
- dockerfile: workers/datasets_based/dev.Dockerfile
+ dockerfile: services/worker/dev.Dockerfile
|
|
67c2eee52b595f51005e6448722b0ac586cd64b5
|
Quentin Lhoest
| 2023-02-10T22:12:27 |
add admin ui url (#801)
|
diff --git a/front/admin_ui/README.md b/front/admin_ui/README.md
index 2d88a546..e4eddbb6 100644
--- a/front/admin_ui/README.md
+++ b/front/admin_ui/README.md
@@ -2,0 +3,2 @@
+Deployed at (internal) https://huggingface.co/spaces/datasets-maintainers/datasets-server-admin-ui
+
|
|
6666942a0d1194d6d48fdca42377d09359bd0a21
|
Sylvain Lesage
| 2023-02-10T17:20:41 |
Check dataset connection before migration job (and other apps) (#792)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/main.py b/jobs/mongodb_migration/src/mongodb_migration/main.py
index 82b362e6..0f15ca0c 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/main.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/main.py
@@ -3,0 +4 @@
+import logging
@@ -14 +15,2 @@ from mongodb_migration.resources import MigrationsMongoResource
-if __name__ == "__main__":
+
+def run_job() -> None:
@@ -21,2 +23,6 @@ if __name__ == "__main__":
- CacheMongoResource(database=job_config.cache.mongo_database, host=job_config.cache.mongo_url),
- QueueMongoResource(database=job_config.queue.mongo_database, host=job_config.queue.mongo_url),
+ CacheMongoResource(
+ database=job_config.cache.mongo_database, host=job_config.cache.mongo_url
+ ) as cache_resource,
+ QueueMongoResource(
+ database=job_config.queue.mongo_database, host=job_config.queue.mongo_url
+ ) as queue_resource,
@@ -25 +31 @@ if __name__ == "__main__":
- ),
+ ) as migrations_database_resource,
@@ -26,0 +33,15 @@ if __name__ == "__main__":
+ if not cache_resource.is_available():
+ logging.warning(
+ "The connection to the cache database could not be established. The migration job is skipped."
+ )
+ return
+ if queue_resource.is_available() is False:
+ logging.warning(
+ "The connection to the queue database could not be established. The migration job is skipped."
+ )
+ return
+ if migrations_database_resource.is_available() is False:
+ logging.warning(
+ "The connection to the migrations database could not be established. The migration job is skipped."
+ )
+ return
@@ -28,5 +49,9 @@ if __name__ == "__main__":
- try:
- Plan(collected_migrations=collected_migrations).execute()
- sys.exit(0)
- except Exception:
- sys.exit(1)
+ Plan(collected_migrations=collected_migrations).execute()
+
+
+if __name__ == "__main__":
+ try:
+ run_job()
+ sys.exit(0)
+ except Exception:
+ sys.exit(1)
diff --git a/jobs/mongodb_migration/tests/test_resources.py b/jobs/mongodb_migration/tests/test_resources.py
index e8ff0523..e346a9af 100644
--- a/jobs/mongodb_migration/tests/test_resources.py
+++ b/jobs/mongodb_migration/tests/test_resources.py
@@ -19,0 +20 @@ def test_cache_database(mongo_host: str) -> None:
+ assert resource.is_available()
diff --git a/libs/libcommon/src/libcommon/resources.py b/libs/libcommon/src/libcommon/resources.py
index 2eeeece4..375215b7 100644
--- a/libs/libcommon/src/libcommon/resources.py
+++ b/libs/libcommon/src/libcommon/resources.py
@@ -6,0 +7,2 @@ from mongoengine.connection import ConnectionFailure, connect, disconnect
+from pymongo import MongoClient # type: ignore
+from pymongo.errors import ServerSelectionTimeoutError
@@ -16 +18,2 @@ class Resource:
- The method allocate() is called when the resource is created. The method release() allows to free the resource.
+ The method allocate() is called when the resource is created.
+ The method release() allows to free the resource.
@@ -24 +27 @@ class Resource:
- Resources should be inherited from this class and implement the allocate() and release() methods.
+ Resources should be inherited from this class and must implement the allocate(), check() and release() methods.
@@ -51,0 +55,2 @@ class MongoResource(Resource):
+ The method is_available() allows to check if the resource is available. It's not called automatically.
+
@@ -64,0 +70,2 @@ class MongoResource(Resource):
+ _client: MongoClient = field(init=False)
+
@@ -67 +74 @@ class MongoResource(Resource):
- connect(
+ self._client = connect(
@@ -75,0 +83,8 @@ class MongoResource(Resource):
+ def is_available(self) -> bool:
+ """Check if the connection is available."""
+ try:
+ self._client.is_mongos
+ return True
+ except ServerSelectionTimeoutError:
+ return False
+
diff --git a/libs/libcommon/src/libcommon/storage.py b/libs/libcommon/src/libcommon/storage.py
index a6dba310..a4197e8b 100644
--- a/libs/libcommon/src/libcommon/storage.py
+++ b/libs/libcommon/src/libcommon/storage.py
@@ -6,0 +7 @@ from os import PathLike, makedirs
+from pathlib import Path
@@ -50,0 +52,12 @@ def init_assets_dir(directory: Optional[StrPath] = None) -> StrPath:
+def exists(path: StrPath) -> bool:
+ """Check if a path exists.
+
+ Args:
+ path (Union[str, PathLike[str]]): The path to check.
+
+ Returns:
+ bool: True if the path exists, False otherwise.
+ """
+ return Path(path).exists()
+
+
diff --git a/libs/libcommon/tests/test_resources.py b/libs/libcommon/tests/test_resources.py
index 98679570..e93cf39b 100644
--- a/libs/libcommon/tests/test_resources.py
+++ b/libs/libcommon/tests/test_resources.py
@@ -30,0 +31 @@ def test_database_resource(queue_mongo_host: str) -> None:
+ assert resource_1.is_available()
@@ -44,0 +46 @@ def test_database_resource(queue_mongo_host: str) -> None:
+ assert resource_2.is_available()
@@ -77,0 +80 @@ def test_database_resource_errors(
+ assert not resource.is_available()
@@ -80,0 +84 @@ def test_database_resource_errors(
+ assert resource.is_available()
@@ -97,0 +102 @@ def test_cache_database(cache_mongo_host: str) -> None:
+ assert resource.is_available()
@@ -110,0 +116 @@ def test_queue_database(queue_mongo_host: str) -> None:
+ assert resource.is_available()
diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py
index 6c7cd4d5..513d2445 100644
--- a/services/admin/src/admin/app.py
+++ b/services/admin/src/admin/app.py
@@ -8 +8 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource
-from libcommon.storage import init_assets_dir
+from libcommon.storage import exists, init_assets_dir
@@ -35,0 +36,2 @@ def create_app() -> Starlette:
+ if not exists(assets_directory):
+ raise RuntimeError("The assets storage directory could not be accessed. Exiting.")
@@ -41,4 +43,7 @@ def create_app() -> Starlette:
- resources: list[Resource] = [
- CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url),
- QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url),
- ]
+ cache_resource = CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url)
+ queue_resource = QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url)
+ resources: list[Resource] = [cache_resource, queue_resource]
+ if not cache_resource.is_available():
+ raise RuntimeError("The connection to the cache database could not be established. Exiting.")
+ if not queue_resource.is_available():
+ raise RuntimeError("The connection to the queue database could not be established. Exiting.")
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 26e66a2b..107c6ab9 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -46,4 +46,7 @@ def create_app() -> Starlette:
- resources: list[Resource] = [
- CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url),
- QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url),
- ]
+ cache_resource = CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url)
+ queue_resource = QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url)
+ resources: list[Resource] = [cache_resource, queue_resource]
+ if not cache_resource.is_available():
+ raise RuntimeError("The connection to the cache database could not be established. Exiting.")
+ if not queue_resource.is_available():
+ raise RuntimeError("The connection to the queue database could not be established. Exiting.")
diff --git a/workers/datasets_based/src/datasets_based/main.py b/workers/datasets_based/src/datasets_based/main.py
index 0c5bb1ff..08f64331 100644
--- a/workers/datasets_based/src/datasets_based/main.py
+++ b/workers/datasets_based/src/datasets_based/main.py
@@ -31,2 +31,6 @@ if __name__ == "__main__":
- CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url),
- QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url),
+ CacheMongoResource(
+ database=app_config.cache.mongo_database, host=app_config.cache.mongo_url
+ ) as cache_resource,
+ QueueMongoResource(
+ database=app_config.queue.mongo_database, host=app_config.queue.mongo_url
+ ) as queue_resource,
@@ -33,0 +38,5 @@ if __name__ == "__main__":
+ if not cache_resource.is_available():
+ raise RuntimeError("The connection to the cache database could not be established. Exiting.")
+ if not queue_resource.is_available():
+ raise RuntimeError("The connection to the queue database could not be established. Exiting.")
+
|
|
f0d048e67f94fdfd422f5c31dbde2fe591ef68a8
|
Sylvain Lesage
| 2023-02-10T16:31:03 |
Upgrade dependencies, fix kenlm (#803)
|
diff --git a/front/admin_ui/.python-version b/front/admin_ui/.python-version
new file mode 100644
index 00000000..b326afbc
--- /dev/null
+++ b/front/admin_ui/.python-version
@@ -0,0 +1 @@
+3.9.15
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index 49b44cd7..f090ec88 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -5 +5 @@ name = "aiofiles"
-version = "22.1.0"
+version = "23.1.0"
@@ -11,2 +11,2 @@ files = [
- {file = "aiofiles-22.1.0-py3-none-any.whl", hash = "sha256:1142fa8e80dbae46bb6339573ad4c8c0841358f79c6eb50a493dceca14621bad"},
- {file = "aiofiles-22.1.0.tar.gz", hash = "sha256:9107f1ca0b2a5553987a94a3c9959fe5b491fdf731389aa5b7b1bd0733e32de6"},
+ {file = "aiofiles-23.1.0-py3-none-any.whl", hash = "sha256:9312414ae06472eb6f1d163f555e466a23aed1c8f60c30cccf7121dba2e53eb2"},
+ {file = "aiofiles-23.1.0.tar.gz", hash = "sha256:edd247df9a19e0db16534d4baaf536d6609a43e1de5401d7a4c1c148753a1635"},
@@ -458 +458 @@ name = "fastapi"
-version = "0.89.1"
+version = "0.90.1"
@@ -464,2 +464,2 @@ files = [
- {file = "fastapi-0.89.1-py3-none-any.whl", hash = "sha256:f9773ea22290635b2f48b4275b2bf69a8fa721fda2e38228bed47139839dc877"},
- {file = "fastapi-0.89.1.tar.gz", hash = "sha256:15d9271ee52b572a015ca2ae5c72e1ce4241dd8532a534ad4f7ec70c376a580f"},
+ {file = "fastapi-0.90.1-py3-none-any.whl", hash = "sha256:d4e4bd820204eeaa19879bf129bbce73db09bdc209d5d83a064b40b2b00557b0"},
+ {file = "fastapi-0.90.1.tar.gz", hash = "sha256:d17e85deb3a350b731467e7bf035e158faffa381310a1b7356421e916fcc64f2"},
@@ -470 +470 @@ pydantic = ">=1.6.2,<1.7 || >1.7,<1.7.1 || >1.7.1,<1.7.2 || >1.7.2,<1.7.3 || >1.
-starlette = "0.22.0"
+starlette = ">=0.22.0,<0.24.0"
@@ -1109 +1109 @@ name = "numpy"
-version = "1.24.1"
+version = "1.24.2"
@@ -1115,28 +1115,28 @@ files = [
- {file = "numpy-1.24.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:179a7ef0889ab769cc03573b6217f54c8bd8e16cef80aad369e1e8185f994cd7"},
- {file = "numpy-1.24.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b09804ff570b907da323b3d762e74432fb07955701b17b08ff1b5ebaa8cfe6a9"},
- {file = "numpy-1.24.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f1b739841821968798947d3afcefd386fa56da0caf97722a5de53e07c4ccedc7"},
- {file = "numpy-1.24.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e3463e6ac25313462e04aea3fb8a0a30fb906d5d300f58b3bc2c23da6a15398"},
- {file = "numpy-1.24.1-cp310-cp310-win32.whl", hash = "sha256:b31da69ed0c18be8b77bfce48d234e55d040793cebb25398e2a7d84199fbc7e2"},
- {file = "numpy-1.24.1-cp310-cp310-win_amd64.whl", hash = "sha256:b07b40f5fb4fa034120a5796288f24c1fe0e0580bbfff99897ba6267af42def2"},
- {file = "numpy-1.24.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7094891dcf79ccc6bc2a1f30428fa5edb1e6fb955411ffff3401fb4ea93780a8"},
- {file = "numpy-1.24.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:28e418681372520c992805bb723e29d69d6b7aa411065f48216d8329d02ba032"},
- {file = "numpy-1.24.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e274f0f6c7efd0d577744f52032fdd24344f11c5ae668fe8d01aac0422611df1"},
- {file = "numpy-1.24.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0044f7d944ee882400890f9ae955220d29b33d809a038923d88e4e01d652acd9"},
- {file = "numpy-1.24.1-cp311-cp311-win32.whl", hash = "sha256:442feb5e5bada8408e8fcd43f3360b78683ff12a4444670a7d9e9824c1817d36"},
- {file = "numpy-1.24.1-cp311-cp311-win_amd64.whl", hash = "sha256:de92efa737875329b052982e37bd4371d52cabf469f83e7b8be9bb7752d67e51"},
- {file = "numpy-1.24.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b162ac10ca38850510caf8ea33f89edcb7b0bb0dfa5592d59909419986b72407"},
- {file = "numpy-1.24.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:26089487086f2648944f17adaa1a97ca6aee57f513ba5f1c0b7ebdabbe2b9954"},
- {file = "numpy-1.24.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:caf65a396c0d1f9809596be2e444e3bd4190d86d5c1ce21f5fc4be60a3bc5b36"},
- {file = "numpy-1.24.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b0677a52f5d896e84414761531947c7a330d1adc07c3a4372262f25d84af7bf7"},
- {file = "numpy-1.24.1-cp38-cp38-win32.whl", hash = "sha256:dae46bed2cb79a58d6496ff6d8da1e3b95ba09afeca2e277628171ca99b99db1"},
- {file = "numpy-1.24.1-cp38-cp38-win_amd64.whl", hash = "sha256:6ec0c021cd9fe732e5bab6401adea5a409214ca5592cd92a114f7067febcba0c"},
- {file = "numpy-1.24.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:28bc9750ae1f75264ee0f10561709b1462d450a4808cd97c013046073ae64ab6"},
- {file = "numpy-1.24.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:84e789a085aabef2f36c0515f45e459f02f570c4b4c4c108ac1179c34d475ed7"},
- {file = "numpy-1.24.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8e669fbdcdd1e945691079c2cae335f3e3a56554e06bbd45d7609a6cf568c700"},
- {file = "numpy-1.24.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ef85cf1f693c88c1fd229ccd1055570cb41cdf4875873b7728b6301f12cd05bf"},
- {file = "numpy-1.24.1-cp39-cp39-win32.whl", hash = "sha256:87a118968fba001b248aac90e502c0b13606721b1343cdaddbc6e552e8dfb56f"},
- {file = "numpy-1.24.1-cp39-cp39-win_amd64.whl", hash = "sha256:ddc7ab52b322eb1e40521eb422c4e0a20716c271a306860979d450decbb51b8e"},
- {file = "numpy-1.24.1-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:ed5fb71d79e771ec930566fae9c02626b939e37271ec285e9efaf1b5d4370e7d"},
- {file = "numpy-1.24.1-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad2925567f43643f51255220424c23d204024ed428afc5aad0f86f3ffc080086"},
- {file = "numpy-1.24.1-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cfa1161c6ac8f92dea03d625c2d0c05e084668f4a06568b77a25a89111621566"},
- {file = "numpy-1.24.1.tar.gz", hash = "sha256:2386da9a471cc00a1f47845e27d916d5ec5346ae9696e01a8a34760858fe9dd2"},
+ {file = "numpy-1.24.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eef70b4fc1e872ebddc38cddacc87c19a3709c0e3e5d20bf3954c147b1dd941d"},
+ {file = "numpy-1.24.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e8d2859428712785e8a8b7d2b3ef0a1d1565892367b32f915c4a4df44d0e64f5"},
+ {file = "numpy-1.24.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6524630f71631be2dabe0c541e7675db82651eb998496bbe16bc4f77f0772253"},
+ {file = "numpy-1.24.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a51725a815a6188c662fb66fb32077709a9ca38053f0274640293a14fdd22978"},
+ {file = "numpy-1.24.2-cp310-cp310-win32.whl", hash = "sha256:2620e8592136e073bd12ee4536149380695fbe9ebeae845b81237f986479ffc9"},
+ {file = "numpy-1.24.2-cp310-cp310-win_amd64.whl", hash = "sha256:97cf27e51fa078078c649a51d7ade3c92d9e709ba2bfb97493007103c741f1d0"},
+ {file = "numpy-1.24.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7de8fdde0003f4294655aa5d5f0a89c26b9f22c0a58790c38fae1ed392d44a5a"},
+ {file = "numpy-1.24.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4173bde9fa2a005c2c6e2ea8ac1618e2ed2c1c6ec8a7657237854d42094123a0"},
+ {file = "numpy-1.24.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4cecaed30dc14123020f77b03601559fff3e6cd0c048f8b5289f4eeabb0eb281"},
+ {file = "numpy-1.24.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9a23f8440561a633204a67fb44617ce2a299beecf3295f0d13c495518908e910"},
+ {file = "numpy-1.24.2-cp311-cp311-win32.whl", hash = "sha256:e428c4fbfa085f947b536706a2fc349245d7baa8334f0c5723c56a10595f9b95"},
+ {file = "numpy-1.24.2-cp311-cp311-win_amd64.whl", hash = "sha256:557d42778a6869c2162deb40ad82612645e21d79e11c1dc62c6e82a2220ffb04"},
+ {file = "numpy-1.24.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d0a2db9d20117bf523dde15858398e7c0858aadca7c0f088ac0d6edd360e9ad2"},
+ {file = "numpy-1.24.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:c72a6b2f4af1adfe193f7beb91ddf708ff867a3f977ef2ec53c0ffb8283ab9f5"},
+ {file = "numpy-1.24.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c29e6bd0ec49a44d7690ecb623a8eac5ab8a923bce0bea6293953992edf3a76a"},
+ {file = "numpy-1.24.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2eabd64ddb96a1239791da78fa5f4e1693ae2dadc82a76bc76a14cbb2b966e96"},
+ {file = "numpy-1.24.2-cp38-cp38-win32.whl", hash = "sha256:e3ab5d32784e843fc0dd3ab6dcafc67ef806e6b6828dc6af2f689be0eb4d781d"},
+ {file = "numpy-1.24.2-cp38-cp38-win_amd64.whl", hash = "sha256:76807b4063f0002c8532cfeac47a3068a69561e9c8715efdad3c642eb27c0756"},
+ {file = "numpy-1.24.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4199e7cfc307a778f72d293372736223e39ec9ac096ff0a2e64853b866a8e18a"},
+ {file = "numpy-1.24.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:adbdce121896fd3a17a77ab0b0b5eedf05a9834a18699db6829a64e1dfccca7f"},
+ {file = "numpy-1.24.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:889b2cc88b837d86eda1b17008ebeb679d82875022200c6e8e4ce6cf549b7acb"},
+ {file = "numpy-1.24.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f64bb98ac59b3ea3bf74b02f13836eb2e24e48e0ab0145bbda646295769bd780"},
+ {file = "numpy-1.24.2-cp39-cp39-win32.whl", hash = "sha256:63e45511ee4d9d976637d11e6c9864eae50e12dc9598f531c035265991910468"},
+ {file = "numpy-1.24.2-cp39-cp39-win_amd64.whl", hash = "sha256:a77d3e1163a7770164404607b7ba3967fb49b24782a6ef85d9b5f54126cc39e5"},
+ {file = "numpy-1.24.2-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92011118955724465fb6853def593cf397b4a1367495e0b59a7e69d40c4eb71d"},
+ {file = "numpy-1.24.2-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9006288bcf4895917d02583cf3411f98631275bc67cce355a7f39f8c14338fa"},
+ {file = "numpy-1.24.2-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:150947adbdfeceec4e5926d956a06865c1c690f2fd902efede4ca6fe2e657c3f"},
+ {file = "numpy-1.24.2.tar.gz", hash = "sha256:003a9f530e880cb2cd177cba1af7220b9aa42def9c4afc2a2fc3ee6be7eb2b22"},
@@ -1147 +1147 @@ name = "orjson"
-version = "3.8.5"
+version = "3.8.6"
@@ -1153,44 +1153,44 @@ files = [
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:143639b9898b094883481fac37733231da1c2ae3aec78a1dd8d3b58c9c9fceef"},
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:31f43e63e0d94784c55e86bd376df3f80b574bea8c0bc5ecd8041009fa8ec78a"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c802ea6d4a0d40f096aceb5e7ef0a26c23d276cb9334e1cadcf256bb090b6426"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bf298b55b371c2772420c5ace4d47b0a3ea1253667e20ded3c363160fd0575f6"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68cb4a8501a463771d55bb22fc72795ec7e21d71ab083e000a2c3b651b6fb2af"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:4f1427952b3bd92bfb63a61b7ffc33a9f54ec6de296fa8d924cbeba089866acb"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c0a9f329468c8eb000742455b83546849bcd69495d6baa6e171c7ee8600a47bd"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6535d527aa1e4a757a6ce9b61f3dd74edc762e7d2c6991643aae7c560c8440bd"},
- {file = "orjson-3.8.5-cp310-none-win_amd64.whl", hash = "sha256:2eee64c028adf6378dd714c8debc96d5b92b6bb4862debb65ca868e59bac6c63"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:f5745ff473dd5c6718bf8c8d5bc183f638b4f3e03c7163ffcda4d4ef453f42ff"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:544f1240b295083697027a5093ec66763218ff16f03521d5020e7a436d2e417b"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c85c9c6bab97a831e7741089057347d99901b4db2451a076ca8adedc7d96297f"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9bae7347764e7be6dada980fd071e865544c98317ab61af575c9cc5e1dc7e3fe"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c67f6f6e9d26a06b63126112a7bc8d8529df048d31df2a257a8484b76adf3e5d"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:758238364142fcbeca34c968beefc0875ffa10aa2f797c82f51cfb1d22d0934e"},
- {file = "orjson-3.8.5-cp311-none-win_amd64.whl", hash = "sha256:cc7579240fb88a626956a6cb4a181a11b62afbc409ce239a7b866568a2412fa2"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:79aa3e47cbbd4eedbbde4f988f766d6cf38ccb51d52cfabfeb6b8d1b58654d25"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:2544cd0d089faa862f5a39f508ee667419e3f9e11f119a6b1505cfce0eb26601"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f2be0025ca7e460bcacb250aba8ce0239be62957d58cf34045834cc9302611d3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b57bf72902d818506906e49c677a791f90dbd7f0997d60b14bc6c1ce4ce4cf9"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:93ae9832a11c6a9efa8c14224e5caf6e35046efd781de14e59eb69ab4e561cf3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:0e28330cc6d51741cad0edd1b57caf6c5531aff30afe41402acde0a03246b8ed"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:155954d725627b5480e6cc1ca488afb4fa685099a4ace5f5bf21a182fabf6706"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ece1b6ef9312df5d5274ca6786e613b7da7de816356e36bcad9ea8a73d15ab71"},
- {file = "orjson-3.8.5-cp37-none-win_amd64.whl", hash = "sha256:6f58d1f0702332496bc1e2d267c7326c851991b62cf6395370d59c47f9890007"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:933f4ab98362f46a59a6d0535986e1f0cae2f6b42435e24a55922b4bc872af0c"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:47a7ca236b25a138a74b2cb5169adcdc5b2b8abdf661de438ba65967a2cde9dc"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b573ca942c626fcf8a86be4f180b86b2498b18ae180f37b4180c2aced5808710"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a9bab11611d5452efe4ae5315f5eb806f66104c08a089fb84c648d2e8e00f106"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eee2f5f6476617d01ca166266d70fd5605d3397a41f067022ce04a2e1ced4c8d"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:ec0b0b6cd0b84f03537f22b719aca705b876c54ab5cf3471d551c9644127284f"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:df3287dc304c8c4556dc85c4ab89eb333307759c1863f95e72e555c0cfce3e01"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:09f40add3c2d208e20f8bf185df38f992bf5092202d2d30eced8f6959963f1d5"},
- {file = "orjson-3.8.5-cp38-none-win_amd64.whl", hash = "sha256:232ec1df0d708f74e0dd1fccac1e9a7008cd120d48fe695e8f0c9d80771da430"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:8fba3e7aede3e88a01e94e6fe63d4580162b212e6da27ae85af50a1787e41416"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:85e22c358cab170c8604e9edfffcc45dd7b0027ce57ed6bcacb556e8bfbbb704"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:eeab1d8247507a75926adf3ca995c74e91f5db1f168815bf3e774f992ba52b50"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:daaaef15a41e9e8cadc7677cefe00065ae10bce914eefe8da1cd26b3d063970b"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ccc9f52cf46bd353c6ae1153eaf9d18257ddc110d135198b0cd8718474685ce"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:d48c182c7ff4ea0787806de8a2f9298ca44fd0068ecd5f23a4b2d8e03c745cb6"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:1848e3b4cc09cc82a67262ae56e2a772b0548bb5a6f9dcaee10dcaaf0a5177b7"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:38480031bc8add58effe802291e4abf7042ef72ae1a4302efe9a36c8f8bfbfcc"},
- {file = "orjson-3.8.5-cp39-none-win_amd64.whl", hash = "sha256:0e9a1c2e649cbaed410c882cedc8f3b993d8f1426d9327f31762d3f46fe7cc88"},
- {file = "orjson-3.8.5.tar.gz", hash = "sha256:77a3b2bd0c4ef7723ea09081e3329dac568a62463aed127c1501441b07ffc64b"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"},
+ {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"},
+ {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"},
+ {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"},
+ {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"},
+ {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"},
+ {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"},
@@ -1271,7 +1270,0 @@ files = [
- {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
- {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
- {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
- {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
- {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
- {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
- {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -1664 +1657 @@ name = "starlette"
-version = "0.22.0"
+version = "0.23.1"
@@ -1670,2 +1663,2 @@ files = [
- {file = "starlette-0.22.0-py3-none-any.whl", hash = "sha256:b5eda991ad5f0ee5d8ce4c4540202a573bb6691ecd0c712262d0bc85cf8f2c50"},
- {file = "starlette-0.22.0.tar.gz", hash = "sha256:b092cbc365bea34dd6840b42861bdabb2f507f8671e642e8272d2442e08ea4ff"},
+ {file = "starlette-0.23.1-py3-none-any.whl", hash = "sha256:ec69736c90be8dbfc6ec6800ba6feb79c8c44f9b1706c0b2bb27f936bcf362cc"},
+ {file = "starlette-0.23.1.tar.gz", hash = "sha256:8510e5b3d670326326c5c1d4cb657cc66832193fe5d5b7015a51c7b1e1b1bf42"},
diff --git a/front/admin_ui/poetry.toml b/front/admin_ui/poetry.toml
new file mode 100644
index 00000000..5fcef8cd
--- /dev/null
+++ b/front/admin_ui/poetry.toml
@@ -0,0 +1,3 @@
+[virtualenvs]
+in-project = true
+prefer-active-python = true
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index ca505887..12a65d67 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -528 +528 @@ name = "isort"
-version = "5.11.4"
+version = "5.12.0"
@@ -532 +532 @@ optional = false
-python-versions = ">=3.7.0"
+python-versions = ">=3.8.0"
@@ -534,2 +534,2 @@ files = [
- {file = "isort-5.11.4-py3-none-any.whl", hash = "sha256:c033fd0edb91000a7f09527fe5c75321878f98322a77ddcc81adbd83724afb7b"},
- {file = "isort-5.11.4.tar.gz", hash = "sha256:6db30c5ded9815d813932c04c2f85a360bcdd35fed496f4d8f35495ef0a261b6"},
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
@@ -539,2 +539,2 @@ files = [
-colors = ["colorama (>=0.4.3,<0.5.0)"]
-pipfile-deprecated-finder = ["pipreqs", "requirementslib"]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
@@ -781 +781 @@ name = "mypy-extensions"
-version = "0.4.3"
+version = "0.4.4"
@@ -785 +785 @@ optional = false
-python-versions = "*"
+python-versions = ">=2.7"
@@ -787,2 +787 @@ files = [
- {file = "mypy_extensions-0.4.3-py2.py3-none-any.whl", hash = "sha256:090fedd75945a69ae91ce1303b5824f428daf5a028d2f6ab8a299250a846f15d"},
- {file = "mypy_extensions-0.4.3.tar.gz", hash = "sha256:2d82818f5bb3e369420cb3c4060a7970edba416647068eb4c5343488a6c604a8"},
+ {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
@@ -793 +792 @@ name = "orjson"
-version = "3.8.5"
+version = "3.8.6"
@@ -799,44 +798,44 @@ files = [
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:143639b9898b094883481fac37733231da1c2ae3aec78a1dd8d3b58c9c9fceef"},
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:31f43e63e0d94784c55e86bd376df3f80b574bea8c0bc5ecd8041009fa8ec78a"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c802ea6d4a0d40f096aceb5e7ef0a26c23d276cb9334e1cadcf256bb090b6426"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bf298b55b371c2772420c5ace4d47b0a3ea1253667e20ded3c363160fd0575f6"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68cb4a8501a463771d55bb22fc72795ec7e21d71ab083e000a2c3b651b6fb2af"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:4f1427952b3bd92bfb63a61b7ffc33a9f54ec6de296fa8d924cbeba089866acb"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c0a9f329468c8eb000742455b83546849bcd69495d6baa6e171c7ee8600a47bd"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6535d527aa1e4a757a6ce9b61f3dd74edc762e7d2c6991643aae7c560c8440bd"},
- {file = "orjson-3.8.5-cp310-none-win_amd64.whl", hash = "sha256:2eee64c028adf6378dd714c8debc96d5b92b6bb4862debb65ca868e59bac6c63"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:f5745ff473dd5c6718bf8c8d5bc183f638b4f3e03c7163ffcda4d4ef453f42ff"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:544f1240b295083697027a5093ec66763218ff16f03521d5020e7a436d2e417b"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c85c9c6bab97a831e7741089057347d99901b4db2451a076ca8adedc7d96297f"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9bae7347764e7be6dada980fd071e865544c98317ab61af575c9cc5e1dc7e3fe"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c67f6f6e9d26a06b63126112a7bc8d8529df048d31df2a257a8484b76adf3e5d"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:758238364142fcbeca34c968beefc0875ffa10aa2f797c82f51cfb1d22d0934e"},
- {file = "orjson-3.8.5-cp311-none-win_amd64.whl", hash = "sha256:cc7579240fb88a626956a6cb4a181a11b62afbc409ce239a7b866568a2412fa2"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:79aa3e47cbbd4eedbbde4f988f766d6cf38ccb51d52cfabfeb6b8d1b58654d25"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:2544cd0d089faa862f5a39f508ee667419e3f9e11f119a6b1505cfce0eb26601"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f2be0025ca7e460bcacb250aba8ce0239be62957d58cf34045834cc9302611d3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b57bf72902d818506906e49c677a791f90dbd7f0997d60b14bc6c1ce4ce4cf9"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:93ae9832a11c6a9efa8c14224e5caf6e35046efd781de14e59eb69ab4e561cf3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:0e28330cc6d51741cad0edd1b57caf6c5531aff30afe41402acde0a03246b8ed"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:155954d725627b5480e6cc1ca488afb4fa685099a4ace5f5bf21a182fabf6706"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ece1b6ef9312df5d5274ca6786e613b7da7de816356e36bcad9ea8a73d15ab71"},
- {file = "orjson-3.8.5-cp37-none-win_amd64.whl", hash = "sha256:6f58d1f0702332496bc1e2d267c7326c851991b62cf6395370d59c47f9890007"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:933f4ab98362f46a59a6d0535986e1f0cae2f6b42435e24a55922b4bc872af0c"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:47a7ca236b25a138a74b2cb5169adcdc5b2b8abdf661de438ba65967a2cde9dc"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b573ca942c626fcf8a86be4f180b86b2498b18ae180f37b4180c2aced5808710"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a9bab11611d5452efe4ae5315f5eb806f66104c08a089fb84c648d2e8e00f106"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eee2f5f6476617d01ca166266d70fd5605d3397a41f067022ce04a2e1ced4c8d"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:ec0b0b6cd0b84f03537f22b719aca705b876c54ab5cf3471d551c9644127284f"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:df3287dc304c8c4556dc85c4ab89eb333307759c1863f95e72e555c0cfce3e01"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:09f40add3c2d208e20f8bf185df38f992bf5092202d2d30eced8f6959963f1d5"},
- {file = "orjson-3.8.5-cp38-none-win_amd64.whl", hash = "sha256:232ec1df0d708f74e0dd1fccac1e9a7008cd120d48fe695e8f0c9d80771da430"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:8fba3e7aede3e88a01e94e6fe63d4580162b212e6da27ae85af50a1787e41416"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:85e22c358cab170c8604e9edfffcc45dd7b0027ce57ed6bcacb556e8bfbbb704"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:eeab1d8247507a75926adf3ca995c74e91f5db1f168815bf3e774f992ba52b50"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:daaaef15a41e9e8cadc7677cefe00065ae10bce914eefe8da1cd26b3d063970b"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ccc9f52cf46bd353c6ae1153eaf9d18257ddc110d135198b0cd8718474685ce"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:d48c182c7ff4ea0787806de8a2f9298ca44fd0068ecd5f23a4b2d8e03c745cb6"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:1848e3b4cc09cc82a67262ae56e2a772b0548bb5a6f9dcaee10dcaaf0a5177b7"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:38480031bc8add58effe802291e4abf7042ef72ae1a4302efe9a36c8f8bfbfcc"},
- {file = "orjson-3.8.5-cp39-none-win_amd64.whl", hash = "sha256:0e9a1c2e649cbaed410c882cedc8f3b993d8f1426d9327f31762d3f46fe7cc88"},
- {file = "orjson-3.8.5.tar.gz", hash = "sha256:77a3b2bd0c4ef7723ea09081e3329dac568a62463aed127c1501441b07ffc64b"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"},
+ {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"},
+ {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"},
+ {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"},
+ {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"},
+ {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"},
+ {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"},
@@ -899 +898 @@ name = "pip"
-version = "22.3.1"
+version = "23.0"
@@ -905,2 +904,2 @@ files = [
- {file = "pip-22.3.1-py3-none-any.whl", hash = "sha256:908c78e6bc29b676ede1c4d57981d490cb892eb45cd8c214ab6298125119e077"},
- {file = "pip-22.3.1.tar.gz", hash = "sha256:65fd48317359f3af8e593943e6ae1506b66325085ea64b706a998c6e83eeaf38"},
+ {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
+ {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
@@ -975 +974 @@ name = "platformdirs"
-version = "2.6.2"
+version = "3.0.0"
@@ -981,2 +980,2 @@ files = [
- {file = "platformdirs-2.6.2-py3-none-any.whl", hash = "sha256:83c8f6d04389165de7c9b6f0c682439697887bca0aa2f1c87ef1826be3584490"},
- {file = "platformdirs-2.6.2.tar.gz", hash = "sha256:e1fea1fe471b9ff8332e229df3cb7de4f53eeea4998d3b6bfff542115e998bd2"},
+ {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"},
+ {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"},
@@ -986,2 +985,2 @@ files = [
-docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"]
-test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
+docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
@@ -1384 +1383 @@ name = "rich"
-version = "13.2.0"
+version = "13.3.1"
@@ -1390,2 +1389,2 @@ files = [
- {file = "rich-13.2.0-py3-none-any.whl", hash = "sha256:7c963f0d03819221e9ac561e1bc866e3f95a02248c1234daa48954e6d381c003"},
- {file = "rich-13.2.0.tar.gz", hash = "sha256:f1a00cdd3eebf999a15d85ec498bfe0b1a77efe9b34f645768a54132ef444ac5"},
+ {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"},
+ {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"},
@@ -1396 +1395 @@ markdown-it-py = ">=2.1.0,<3.0.0"
-pygments = ">=2.6.0,<3.0.0"
+pygments = ">=2.14.0,<3.0.0"
@@ -1399 +1398 @@ pygments = ">=2.6.0,<3.0.0"
-jupyter = ["ipywidgets (>=7.5.1,<8.0.0)"]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
@@ -1403 +1402 @@ name = "setuptools"
-version = "66.1.1"
+version = "67.2.0"
@@ -1409,2 +1408,2 @@ files = [
- {file = "setuptools-66.1.1-py3-none-any.whl", hash = "sha256:6f590d76b713d5de4e49fe4fbca24474469f53c83632d5d0fd056f7ff7e8112b"},
- {file = "setuptools-66.1.1.tar.gz", hash = "sha256:ac4008d396bc9cd983ea483cb7139c0240a07bbc74ffb6232fceffedc6cf03a8"},
+ {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
+ {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index d2090897..29e1eb92 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -528 +528 @@ name = "isort"
-version = "5.11.4"
+version = "5.12.0"
@@ -532 +532 @@ optional = false
-python-versions = ">=3.7.0"
+python-versions = ">=3.8.0"
@@ -534,2 +534,2 @@ files = [
- {file = "isort-5.11.4-py3-none-any.whl", hash = "sha256:c033fd0edb91000a7f09527fe5c75321878f98322a77ddcc81adbd83724afb7b"},
- {file = "isort-5.11.4.tar.gz", hash = "sha256:6db30c5ded9815d813932c04c2f85a360bcdd35fed496f4d8f35495ef0a261b6"},
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
@@ -539,2 +539,2 @@ files = [
-colors = ["colorama (>=0.4.3,<0.5.0)"]
-pipfile-deprecated-finder = ["pipreqs", "requirementslib"]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
@@ -757 +757 @@ name = "mypy-extensions"
-version = "0.4.3"
+version = "0.4.4"
@@ -761 +761 @@ optional = false
-python-versions = "*"
+python-versions = ">=2.7"
@@ -763,2 +763 @@ files = [
- {file = "mypy_extensions-0.4.3-py2.py3-none-any.whl", hash = "sha256:090fedd75945a69ae91ce1303b5824f428daf5a028d2f6ab8a299250a846f15d"},
- {file = "mypy_extensions-0.4.3.tar.gz", hash = "sha256:2d82818f5bb3e369420cb3c4060a7970edba416647068eb4c5343488a6c604a8"},
+ {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
@@ -769 +768 @@ name = "orjson"
-version = "3.8.5"
+version = "3.8.6"
@@ -775,44 +774,44 @@ files = [
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:143639b9898b094883481fac37733231da1c2ae3aec78a1dd8d3b58c9c9fceef"},
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:31f43e63e0d94784c55e86bd376df3f80b574bea8c0bc5ecd8041009fa8ec78a"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c802ea6d4a0d40f096aceb5e7ef0a26c23d276cb9334e1cadcf256bb090b6426"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bf298b55b371c2772420c5ace4d47b0a3ea1253667e20ded3c363160fd0575f6"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68cb4a8501a463771d55bb22fc72795ec7e21d71ab083e000a2c3b651b6fb2af"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:4f1427952b3bd92bfb63a61b7ffc33a9f54ec6de296fa8d924cbeba089866acb"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c0a9f329468c8eb000742455b83546849bcd69495d6baa6e171c7ee8600a47bd"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6535d527aa1e4a757a6ce9b61f3dd74edc762e7d2c6991643aae7c560c8440bd"},
- {file = "orjson-3.8.5-cp310-none-win_amd64.whl", hash = "sha256:2eee64c028adf6378dd714c8debc96d5b92b6bb4862debb65ca868e59bac6c63"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:f5745ff473dd5c6718bf8c8d5bc183f638b4f3e03c7163ffcda4d4ef453f42ff"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:544f1240b295083697027a5093ec66763218ff16f03521d5020e7a436d2e417b"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c85c9c6bab97a831e7741089057347d99901b4db2451a076ca8adedc7d96297f"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9bae7347764e7be6dada980fd071e865544c98317ab61af575c9cc5e1dc7e3fe"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c67f6f6e9d26a06b63126112a7bc8d8529df048d31df2a257a8484b76adf3e5d"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:758238364142fcbeca34c968beefc0875ffa10aa2f797c82f51cfb1d22d0934e"},
- {file = "orjson-3.8.5-cp311-none-win_amd64.whl", hash = "sha256:cc7579240fb88a626956a6cb4a181a11b62afbc409ce239a7b866568a2412fa2"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:79aa3e47cbbd4eedbbde4f988f766d6cf38ccb51d52cfabfeb6b8d1b58654d25"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:2544cd0d089faa862f5a39f508ee667419e3f9e11f119a6b1505cfce0eb26601"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f2be0025ca7e460bcacb250aba8ce0239be62957d58cf34045834cc9302611d3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b57bf72902d818506906e49c677a791f90dbd7f0997d60b14bc6c1ce4ce4cf9"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:93ae9832a11c6a9efa8c14224e5caf6e35046efd781de14e59eb69ab4e561cf3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:0e28330cc6d51741cad0edd1b57caf6c5531aff30afe41402acde0a03246b8ed"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:155954d725627b5480e6cc1ca488afb4fa685099a4ace5f5bf21a182fabf6706"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ece1b6ef9312df5d5274ca6786e613b7da7de816356e36bcad9ea8a73d15ab71"},
- {file = "orjson-3.8.5-cp37-none-win_amd64.whl", hash = "sha256:6f58d1f0702332496bc1e2d267c7326c851991b62cf6395370d59c47f9890007"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:933f4ab98362f46a59a6d0535986e1f0cae2f6b42435e24a55922b4bc872af0c"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:47a7ca236b25a138a74b2cb5169adcdc5b2b8abdf661de438ba65967a2cde9dc"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b573ca942c626fcf8a86be4f180b86b2498b18ae180f37b4180c2aced5808710"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a9bab11611d5452efe4ae5315f5eb806f66104c08a089fb84c648d2e8e00f106"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eee2f5f6476617d01ca166266d70fd5605d3397a41f067022ce04a2e1ced4c8d"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:ec0b0b6cd0b84f03537f22b719aca705b876c54ab5cf3471d551c9644127284f"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:df3287dc304c8c4556dc85c4ab89eb333307759c1863f95e72e555c0cfce3e01"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:09f40add3c2d208e20f8bf185df38f992bf5092202d2d30eced8f6959963f1d5"},
- {file = "orjson-3.8.5-cp38-none-win_amd64.whl", hash = "sha256:232ec1df0d708f74e0dd1fccac1e9a7008cd120d48fe695e8f0c9d80771da430"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:8fba3e7aede3e88a01e94e6fe63d4580162b212e6da27ae85af50a1787e41416"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:85e22c358cab170c8604e9edfffcc45dd7b0027ce57ed6bcacb556e8bfbbb704"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:eeab1d8247507a75926adf3ca995c74e91f5db1f168815bf3e774f992ba52b50"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:daaaef15a41e9e8cadc7677cefe00065ae10bce914eefe8da1cd26b3d063970b"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ccc9f52cf46bd353c6ae1153eaf9d18257ddc110d135198b0cd8718474685ce"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:d48c182c7ff4ea0787806de8a2f9298ca44fd0068ecd5f23a4b2d8e03c745cb6"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:1848e3b4cc09cc82a67262ae56e2a772b0548bb5a6f9dcaee10dcaaf0a5177b7"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:38480031bc8add58effe802291e4abf7042ef72ae1a4302efe9a36c8f8bfbfcc"},
- {file = "orjson-3.8.5-cp39-none-win_amd64.whl", hash = "sha256:0e9a1c2e649cbaed410c882cedc8f3b993d8f1426d9327f31762d3f46fe7cc88"},
- {file = "orjson-3.8.5.tar.gz", hash = "sha256:77a3b2bd0c4ef7723ea09081e3329dac568a62463aed127c1501441b07ffc64b"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"},
+ {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"},
+ {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"},
+ {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"},
+ {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"},
+ {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"},
+ {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"},
@@ -875 +874 @@ name = "pip"
-version = "22.3.1"
+version = "23.0"
@@ -881,2 +880,2 @@ files = [
- {file = "pip-22.3.1-py3-none-any.whl", hash = "sha256:908c78e6bc29b676ede1c4d57981d490cb892eb45cd8c214ab6298125119e077"},
- {file = "pip-22.3.1.tar.gz", hash = "sha256:65fd48317359f3af8e593943e6ae1506b66325085ea64b706a998c6e83eeaf38"},
+ {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
+ {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
@@ -951 +950 @@ name = "platformdirs"
-version = "2.6.2"
+version = "3.0.0"
@@ -957,2 +956,2 @@ files = [
- {file = "platformdirs-2.6.2-py3-none-any.whl", hash = "sha256:83c8f6d04389165de7c9b6f0c682439697887bca0aa2f1c87ef1826be3584490"},
- {file = "platformdirs-2.6.2.tar.gz", hash = "sha256:e1fea1fe471b9ff8332e229df3cb7de4f53eeea4998d3b6bfff542115e998bd2"},
+ {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"},
+ {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"},
@@ -962,2 +961,2 @@ files = [
-docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"]
-test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
+docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
@@ -1360 +1359 @@ name = "rich"
-version = "13.2.0"
+version = "13.3.1"
@@ -1366,2 +1365,2 @@ files = [
- {file = "rich-13.2.0-py3-none-any.whl", hash = "sha256:7c963f0d03819221e9ac561e1bc866e3f95a02248c1234daa48954e6d381c003"},
- {file = "rich-13.2.0.tar.gz", hash = "sha256:f1a00cdd3eebf999a15d85ec498bfe0b1a77efe9b34f645768a54132ef444ac5"},
+ {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"},
+ {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"},
@@ -1372 +1371 @@ markdown-it-py = ">=2.1.0,<3.0.0"
-pygments = ">=2.6.0,<3.0.0"
+pygments = ">=2.14.0,<3.0.0"
@@ -1375 +1374 @@ pygments = ">=2.6.0,<3.0.0"
-jupyter = ["ipywidgets (>=7.5.1,<8.0.0)"]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
@@ -1379 +1378 @@ name = "setuptools"
-version = "66.1.1"
+version = "67.2.0"
@@ -1385,2 +1384,2 @@ files = [
- {file = "setuptools-66.1.1-py3-none-any.whl", hash = "sha256:6f590d76b713d5de4e49fe4fbca24474469f53c83632d5d0fd056f7ff7e8112b"},
- {file = "setuptools-66.1.1.tar.gz", hash = "sha256:ac4008d396bc9cd983ea483cb7139c0240a07bbc74ffb6232fceffedc6cf03a8"},
+ {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
+ {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index d598f491..47bc9eb4 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -576 +576 @@ name = "isort"
-version = "5.11.4"
+version = "5.12.0"
@@ -580 +580 @@ optional = false
-python-versions = ">=3.7.0"
+python-versions = ">=3.8.0"
@@ -582,2 +582,2 @@ files = [
- {file = "isort-5.11.4-py3-none-any.whl", hash = "sha256:c033fd0edb91000a7f09527fe5c75321878f98322a77ddcc81adbd83724afb7b"},
- {file = "isort-5.11.4.tar.gz", hash = "sha256:6db30c5ded9815d813932c04c2f85a360bcdd35fed496f4d8f35495ef0a261b6"},
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
@@ -587,2 +587,2 @@ files = [
-colors = ["colorama (>=0.4.3,<0.5.0)"]
-pipfile-deprecated-finder = ["pipreqs", "requirementslib"]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
@@ -829 +829 @@ name = "mypy-extensions"
-version = "0.4.3"
+version = "0.4.4"
@@ -833 +833 @@ optional = false
-python-versions = "*"
+python-versions = ">=2.7"
@@ -835,2 +835 @@ files = [
- {file = "mypy_extensions-0.4.3-py2.py3-none-any.whl", hash = "sha256:090fedd75945a69ae91ce1303b5824f428daf5a028d2f6ab8a299250a846f15d"},
- {file = "mypy_extensions-0.4.3.tar.gz", hash = "sha256:2d82818f5bb3e369420cb3c4060a7970edba416647068eb4c5343488a6c604a8"},
+ {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
@@ -841 +840 @@ name = "orjson"
-version = "3.8.5"
+version = "3.8.6"
@@ -847,44 +846,44 @@ files = [
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:143639b9898b094883481fac37733231da1c2ae3aec78a1dd8d3b58c9c9fceef"},
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:31f43e63e0d94784c55e86bd376df3f80b574bea8c0bc5ecd8041009fa8ec78a"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c802ea6d4a0d40f096aceb5e7ef0a26c23d276cb9334e1cadcf256bb090b6426"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bf298b55b371c2772420c5ace4d47b0a3ea1253667e20ded3c363160fd0575f6"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68cb4a8501a463771d55bb22fc72795ec7e21d71ab083e000a2c3b651b6fb2af"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:4f1427952b3bd92bfb63a61b7ffc33a9f54ec6de296fa8d924cbeba089866acb"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c0a9f329468c8eb000742455b83546849bcd69495d6baa6e171c7ee8600a47bd"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6535d527aa1e4a757a6ce9b61f3dd74edc762e7d2c6991643aae7c560c8440bd"},
- {file = "orjson-3.8.5-cp310-none-win_amd64.whl", hash = "sha256:2eee64c028adf6378dd714c8debc96d5b92b6bb4862debb65ca868e59bac6c63"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:f5745ff473dd5c6718bf8c8d5bc183f638b4f3e03c7163ffcda4d4ef453f42ff"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:544f1240b295083697027a5093ec66763218ff16f03521d5020e7a436d2e417b"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c85c9c6bab97a831e7741089057347d99901b4db2451a076ca8adedc7d96297f"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9bae7347764e7be6dada980fd071e865544c98317ab61af575c9cc5e1dc7e3fe"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c67f6f6e9d26a06b63126112a7bc8d8529df048d31df2a257a8484b76adf3e5d"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:758238364142fcbeca34c968beefc0875ffa10aa2f797c82f51cfb1d22d0934e"},
- {file = "orjson-3.8.5-cp311-none-win_amd64.whl", hash = "sha256:cc7579240fb88a626956a6cb4a181a11b62afbc409ce239a7b866568a2412fa2"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:79aa3e47cbbd4eedbbde4f988f766d6cf38ccb51d52cfabfeb6b8d1b58654d25"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:2544cd0d089faa862f5a39f508ee667419e3f9e11f119a6b1505cfce0eb26601"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f2be0025ca7e460bcacb250aba8ce0239be62957d58cf34045834cc9302611d3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b57bf72902d818506906e49c677a791f90dbd7f0997d60b14bc6c1ce4ce4cf9"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:93ae9832a11c6a9efa8c14224e5caf6e35046efd781de14e59eb69ab4e561cf3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:0e28330cc6d51741cad0edd1b57caf6c5531aff30afe41402acde0a03246b8ed"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:155954d725627b5480e6cc1ca488afb4fa685099a4ace5f5bf21a182fabf6706"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ece1b6ef9312df5d5274ca6786e613b7da7de816356e36bcad9ea8a73d15ab71"},
- {file = "orjson-3.8.5-cp37-none-win_amd64.whl", hash = "sha256:6f58d1f0702332496bc1e2d267c7326c851991b62cf6395370d59c47f9890007"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:933f4ab98362f46a59a6d0535986e1f0cae2f6b42435e24a55922b4bc872af0c"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:47a7ca236b25a138a74b2cb5169adcdc5b2b8abdf661de438ba65967a2cde9dc"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b573ca942c626fcf8a86be4f180b86b2498b18ae180f37b4180c2aced5808710"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a9bab11611d5452efe4ae5315f5eb806f66104c08a089fb84c648d2e8e00f106"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eee2f5f6476617d01ca166266d70fd5605d3397a41f067022ce04a2e1ced4c8d"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:ec0b0b6cd0b84f03537f22b719aca705b876c54ab5cf3471d551c9644127284f"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:df3287dc304c8c4556dc85c4ab89eb333307759c1863f95e72e555c0cfce3e01"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:09f40add3c2d208e20f8bf185df38f992bf5092202d2d30eced8f6959963f1d5"},
- {file = "orjson-3.8.5-cp38-none-win_amd64.whl", hash = "sha256:232ec1df0d708f74e0dd1fccac1e9a7008cd120d48fe695e8f0c9d80771da430"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:8fba3e7aede3e88a01e94e6fe63d4580162b212e6da27ae85af50a1787e41416"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:85e22c358cab170c8604e9edfffcc45dd7b0027ce57ed6bcacb556e8bfbbb704"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:eeab1d8247507a75926adf3ca995c74e91f5db1f168815bf3e774f992ba52b50"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:daaaef15a41e9e8cadc7677cefe00065ae10bce914eefe8da1cd26b3d063970b"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ccc9f52cf46bd353c6ae1153eaf9d18257ddc110d135198b0cd8718474685ce"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:d48c182c7ff4ea0787806de8a2f9298ca44fd0068ecd5f23a4b2d8e03c745cb6"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:1848e3b4cc09cc82a67262ae56e2a772b0548bb5a6f9dcaee10dcaaf0a5177b7"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:38480031bc8add58effe802291e4abf7042ef72ae1a4302efe9a36c8f8bfbfcc"},
- {file = "orjson-3.8.5-cp39-none-win_amd64.whl", hash = "sha256:0e9a1c2e649cbaed410c882cedc8f3b993d8f1426d9327f31762d3f46fe7cc88"},
- {file = "orjson-3.8.5.tar.gz", hash = "sha256:77a3b2bd0c4ef7723ea09081e3329dac568a62463aed127c1501441b07ffc64b"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"},
+ {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"},
+ {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"},
+ {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"},
+ {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"},
+ {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"},
+ {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"},
@@ -947 +946 @@ name = "pip"
-version = "22.3.1"
+version = "23.0"
@@ -953,2 +952,2 @@ files = [
- {file = "pip-22.3.1-py3-none-any.whl", hash = "sha256:908c78e6bc29b676ede1c4d57981d490cb892eb45cd8c214ab6298125119e077"},
- {file = "pip-22.3.1.tar.gz", hash = "sha256:65fd48317359f3af8e593943e6ae1506b66325085ea64b706a998c6e83eeaf38"},
+ {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
+ {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
@@ -1023 +1022 @@ name = "platformdirs"
-version = "2.6.2"
+version = "3.0.0"
@@ -1029,2 +1028,2 @@ files = [
- {file = "platformdirs-2.6.2-py3-none-any.whl", hash = "sha256:83c8f6d04389165de7c9b6f0c682439697887bca0aa2f1c87ef1826be3584490"},
- {file = "platformdirs-2.6.2.tar.gz", hash = "sha256:e1fea1fe471b9ff8332e229df3cb7de4f53eeea4998d3b6bfff542115e998bd2"},
+ {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"},
+ {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"},
@@ -1034,2 +1033,2 @@ files = [
-docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"]
-test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
+docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
@@ -1466 +1465 @@ name = "rich"
-version = "13.2.0"
+version = "13.3.1"
@@ -1472,2 +1471,2 @@ files = [
- {file = "rich-13.2.0-py3-none-any.whl", hash = "sha256:7c963f0d03819221e9ac561e1bc866e3f95a02248c1234daa48954e6d381c003"},
- {file = "rich-13.2.0.tar.gz", hash = "sha256:f1a00cdd3eebf999a15d85ec498bfe0b1a77efe9b34f645768a54132ef444ac5"},
+ {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"},
+ {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"},
@@ -1478 +1477 @@ markdown-it-py = ">=2.1.0,<3.0.0"
-pygments = ">=2.6.0,<3.0.0"
+pygments = ">=2.14.0,<3.0.0"
@@ -1481 +1480 @@ pygments = ">=2.6.0,<3.0.0"
-jupyter = ["ipywidgets (>=7.5.1,<8.0.0)"]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
@@ -1485 +1484 @@ name = "setuptools"
-version = "66.1.1"
+version = "67.2.0"
@@ -1491,2 +1490,2 @@ files = [
- {file = "setuptools-66.1.1-py3-none-any.whl", hash = "sha256:6f590d76b713d5de4e49fe4fbca24474469f53c83632d5d0fd056f7ff7e8112b"},
- {file = "setuptools-66.1.1.tar.gz", hash = "sha256:ac4008d396bc9cd983ea483cb7139c0240a07bbc74ffb6232fceffedc6cf03a8"},
+ {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
+ {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index 7beb4d4b..dd6111df 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -576 +576 @@ name = "isort"
-version = "5.11.4"
+version = "5.12.0"
@@ -580 +580 @@ optional = false
-python-versions = ">=3.7.0"
+python-versions = ">=3.8.0"
@@ -582,2 +582,2 @@ files = [
- {file = "isort-5.11.4-py3-none-any.whl", hash = "sha256:c033fd0edb91000a7f09527fe5c75321878f98322a77ddcc81adbd83724afb7b"},
- {file = "isort-5.11.4.tar.gz", hash = "sha256:6db30c5ded9815d813932c04c2f85a360bcdd35fed496f4d8f35495ef0a261b6"},
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
@@ -587,2 +587,2 @@ files = [
-colors = ["colorama (>=0.4.3,<0.5.0)"]
-pipfile-deprecated-finder = ["pipreqs", "requirementslib"]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
@@ -909 +909 @@ name = "mypy-extensions"
-version = "0.4.3"
+version = "0.4.4"
@@ -913 +913 @@ optional = false
-python-versions = "*"
+python-versions = ">=2.7"
@@ -915,2 +915 @@ files = [
- {file = "mypy_extensions-0.4.3-py2.py3-none-any.whl", hash = "sha256:090fedd75945a69ae91ce1303b5824f428daf5a028d2f6ab8a299250a846f15d"},
- {file = "mypy_extensions-0.4.3.tar.gz", hash = "sha256:2d82818f5bb3e369420cb3c4060a7970edba416647068eb4c5343488a6c604a8"},
+ {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
@@ -921 +920 @@ name = "orjson"
-version = "3.8.5"
+version = "3.8.6"
@@ -927,44 +926,44 @@ files = [
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:143639b9898b094883481fac37733231da1c2ae3aec78a1dd8d3b58c9c9fceef"},
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:31f43e63e0d94784c55e86bd376df3f80b574bea8c0bc5ecd8041009fa8ec78a"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c802ea6d4a0d40f096aceb5e7ef0a26c23d276cb9334e1cadcf256bb090b6426"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bf298b55b371c2772420c5ace4d47b0a3ea1253667e20ded3c363160fd0575f6"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68cb4a8501a463771d55bb22fc72795ec7e21d71ab083e000a2c3b651b6fb2af"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:4f1427952b3bd92bfb63a61b7ffc33a9f54ec6de296fa8d924cbeba089866acb"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c0a9f329468c8eb000742455b83546849bcd69495d6baa6e171c7ee8600a47bd"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6535d527aa1e4a757a6ce9b61f3dd74edc762e7d2c6991643aae7c560c8440bd"},
- {file = "orjson-3.8.5-cp310-none-win_amd64.whl", hash = "sha256:2eee64c028adf6378dd714c8debc96d5b92b6bb4862debb65ca868e59bac6c63"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:f5745ff473dd5c6718bf8c8d5bc183f638b4f3e03c7163ffcda4d4ef453f42ff"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:544f1240b295083697027a5093ec66763218ff16f03521d5020e7a436d2e417b"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c85c9c6bab97a831e7741089057347d99901b4db2451a076ca8adedc7d96297f"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9bae7347764e7be6dada980fd071e865544c98317ab61af575c9cc5e1dc7e3fe"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c67f6f6e9d26a06b63126112a7bc8d8529df048d31df2a257a8484b76adf3e5d"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:758238364142fcbeca34c968beefc0875ffa10aa2f797c82f51cfb1d22d0934e"},
- {file = "orjson-3.8.5-cp311-none-win_amd64.whl", hash = "sha256:cc7579240fb88a626956a6cb4a181a11b62afbc409ce239a7b866568a2412fa2"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:79aa3e47cbbd4eedbbde4f988f766d6cf38ccb51d52cfabfeb6b8d1b58654d25"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:2544cd0d089faa862f5a39f508ee667419e3f9e11f119a6b1505cfce0eb26601"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f2be0025ca7e460bcacb250aba8ce0239be62957d58cf34045834cc9302611d3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b57bf72902d818506906e49c677a791f90dbd7f0997d60b14bc6c1ce4ce4cf9"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:93ae9832a11c6a9efa8c14224e5caf6e35046efd781de14e59eb69ab4e561cf3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:0e28330cc6d51741cad0edd1b57caf6c5531aff30afe41402acde0a03246b8ed"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:155954d725627b5480e6cc1ca488afb4fa685099a4ace5f5bf21a182fabf6706"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ece1b6ef9312df5d5274ca6786e613b7da7de816356e36bcad9ea8a73d15ab71"},
- {file = "orjson-3.8.5-cp37-none-win_amd64.whl", hash = "sha256:6f58d1f0702332496bc1e2d267c7326c851991b62cf6395370d59c47f9890007"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:933f4ab98362f46a59a6d0535986e1f0cae2f6b42435e24a55922b4bc872af0c"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:47a7ca236b25a138a74b2cb5169adcdc5b2b8abdf661de438ba65967a2cde9dc"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b573ca942c626fcf8a86be4f180b86b2498b18ae180f37b4180c2aced5808710"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a9bab11611d5452efe4ae5315f5eb806f66104c08a089fb84c648d2e8e00f106"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eee2f5f6476617d01ca166266d70fd5605d3397a41f067022ce04a2e1ced4c8d"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:ec0b0b6cd0b84f03537f22b719aca705b876c54ab5cf3471d551c9644127284f"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:df3287dc304c8c4556dc85c4ab89eb333307759c1863f95e72e555c0cfce3e01"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:09f40add3c2d208e20f8bf185df38f992bf5092202d2d30eced8f6959963f1d5"},
- {file = "orjson-3.8.5-cp38-none-win_amd64.whl", hash = "sha256:232ec1df0d708f74e0dd1fccac1e9a7008cd120d48fe695e8f0c9d80771da430"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:8fba3e7aede3e88a01e94e6fe63d4580162b212e6da27ae85af50a1787e41416"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:85e22c358cab170c8604e9edfffcc45dd7b0027ce57ed6bcacb556e8bfbbb704"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:eeab1d8247507a75926adf3ca995c74e91f5db1f168815bf3e774f992ba52b50"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:daaaef15a41e9e8cadc7677cefe00065ae10bce914eefe8da1cd26b3d063970b"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ccc9f52cf46bd353c6ae1153eaf9d18257ddc110d135198b0cd8718474685ce"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:d48c182c7ff4ea0787806de8a2f9298ca44fd0068ecd5f23a4b2d8e03c745cb6"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:1848e3b4cc09cc82a67262ae56e2a772b0548bb5a6f9dcaee10dcaaf0a5177b7"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:38480031bc8add58effe802291e4abf7042ef72ae1a4302efe9a36c8f8bfbfcc"},
- {file = "orjson-3.8.5-cp39-none-win_amd64.whl", hash = "sha256:0e9a1c2e649cbaed410c882cedc8f3b993d8f1426d9327f31762d3f46fe7cc88"},
- {file = "orjson-3.8.5.tar.gz", hash = "sha256:77a3b2bd0c4ef7723ea09081e3329dac568a62463aed127c1501441b07ffc64b"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"},
+ {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"},
+ {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"},
+ {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"},
+ {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"},
+ {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"},
+ {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"},
@@ -1027 +1026 @@ name = "pip"
-version = "22.3.1"
+version = "23.0"
@@ -1033,2 +1032,2 @@ files = [
- {file = "pip-22.3.1-py3-none-any.whl", hash = "sha256:908c78e6bc29b676ede1c4d57981d490cb892eb45cd8c214ab6298125119e077"},
- {file = "pip-22.3.1.tar.gz", hash = "sha256:65fd48317359f3af8e593943e6ae1506b66325085ea64b706a998c6e83eeaf38"},
+ {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
+ {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
@@ -1103 +1102 @@ name = "platformdirs"
-version = "2.6.2"
+version = "3.0.0"
@@ -1109,2 +1108,2 @@ files = [
- {file = "platformdirs-2.6.2-py3-none-any.whl", hash = "sha256:83c8f6d04389165de7c9b6f0c682439697887bca0aa2f1c87ef1826be3584490"},
- {file = "platformdirs-2.6.2.tar.gz", hash = "sha256:e1fea1fe471b9ff8332e229df3cb7de4f53eeea4998d3b6bfff542115e998bd2"},
+ {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"},
+ {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"},
@@ -1114,2 +1113,2 @@ files = [
-docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"]
-test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
+docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
@@ -1579 +1578 @@ name = "rich"
-version = "13.2.0"
+version = "13.3.1"
@@ -1585,2 +1584,2 @@ files = [
- {file = "rich-13.2.0-py3-none-any.whl", hash = "sha256:7c963f0d03819221e9ac561e1bc866e3f95a02248c1234daa48954e6d381c003"},
- {file = "rich-13.2.0.tar.gz", hash = "sha256:f1a00cdd3eebf999a15d85ec498bfe0b1a77efe9b34f645768a54132ef444ac5"},
+ {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"},
+ {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"},
@@ -1591 +1590 @@ markdown-it-py = ">=2.1.0,<3.0.0"
-pygments = ">=2.6.0,<3.0.0"
+pygments = ">=2.14.0,<3.0.0"
@@ -1594 +1593 @@ pygments = ">=2.6.0,<3.0.0"
-jupyter = ["ipywidgets (>=7.5.1,<8.0.0)"]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
@@ -1598 +1597 @@ name = "setuptools"
-version = "66.1.1"
+version = "67.2.0"
@@ -1604,2 +1603,2 @@ files = [
- {file = "setuptools-66.1.1-py3-none-any.whl", hash = "sha256:6f590d76b713d5de4e49fe4fbca24474469f53c83632d5d0fd056f7ff7e8112b"},
- {file = "setuptools-66.1.1.tar.gz", hash = "sha256:ac4008d396bc9cd983ea483cb7139c0240a07bbc74ffb6232fceffedc6cf03a8"},
+ {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
+ {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
diff --git a/workers/datasets_based/poetry.lock b/workers/datasets_based/poetry.lock
index 80a236e5..2f85f69a 100644
--- a/workers/datasets_based/poetry.lock
+++ b/workers/datasets_based/poetry.lock
@@ -299 +299 @@ name = "beautifulsoup4"
-version = "4.11.1"
+version = "4.11.2"
@@ -305,2 +305,2 @@ files = [
- {file = "beautifulsoup4-4.11.1-py3-none-any.whl", hash = "sha256:58d5c3d29f5a36ffeb94f02f0d786cd53014cf9b3b3951d42e0080d8a9498d30"},
- {file = "beautifulsoup4-4.11.1.tar.gz", hash = "sha256:ad9aa55b65ef2808eb405f46cf74df7fcb7044d5cbc26487f96eb2ef2e436693"},
+ {file = "beautifulsoup4-4.11.2-py3-none-any.whl", hash = "sha256:0e79446b10b3ecb499c1556f7e228a53e64a2bfcebd455f370d8927cb5b59e39"},
+ {file = "beautifulsoup4-4.11.2.tar.gz", hash = "sha256:bc4bdda6717de5a2987436fb8d72f45dc90dd856bdfd512a1314ce90349a0106"},
@@ -1030 +1030 @@ name = "fastavro"
-version = "1.7.0"
+version = "1.7.1"
@@ -1036,21 +1036,21 @@ files = [
- {file = "fastavro-1.7.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:ab3387a06e272980fa034f5c62f7063977b77df6416d3d30a4d3b49cc8827566"},
- {file = "fastavro-1.7.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:216132bc54da19e97e1531dd69c86282408d4c797749d83b01b3a00862a180de"},
- {file = "fastavro-1.7.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c20cf6cd8098bb93c2cffd53d03ccea1dcf9ec594a5c83963acf29a2882f8693"},
- {file = "fastavro-1.7.0-cp310-cp310-win_amd64.whl", hash = "sha256:54b60e79506a456bcfc940560fa2c73a7a8e3ddc58a1ae4d94fdd99f6b02aef0"},
- {file = "fastavro-1.7.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:4da6d83abd04a804310667f8d1fc23d44e9ed91ed9a9bc9c1fcd906e0c403b12"},
- {file = "fastavro-1.7.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7c664302f94064adeb93403c61c74e07b9710526403eba3b59169f99bb99c55c"},
- {file = "fastavro-1.7.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7aab162f02e64bf82d0282430a3c6ec7a36982b1c5d479e7dcc278e6d62a84b8"},
- {file = "fastavro-1.7.0-cp311-cp311-win_amd64.whl", hash = "sha256:63d0d2a2bb3e85d006c834633be51b105a50b0dc7cc8423b06f30601b532adf4"},
- {file = "fastavro-1.7.0-cp37-cp37m-macosx_10_15_x86_64.whl", hash = "sha256:a07a3245049529d6d9028d664361cc4990e74d035d2303875295e2f7b97eba09"},
- {file = "fastavro-1.7.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7df57d31cb58770a9066790250e9f4ec91242c69e1dc62ea732a6fb2406a8f96"},
- {file = "fastavro-1.7.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b905634b5763ca08c9f7b51486e2c3ae7907f5d9bc48208c69b16ccbe8455e90"},
- {file = "fastavro-1.7.0-cp37-cp37m-win_amd64.whl", hash = "sha256:749206f1cec3d7429546e49db5708f4571497d35181b6b334c4844133f230515"},
- {file = "fastavro-1.7.0-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:a0ebd5c1269085179de4b3f072de274fb66a471ecbc5245bd8684e6f94943c2f"},
- {file = "fastavro-1.7.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a7e35b94b692f8cca0096c89abf1937efed66252dea0b3b3165babfb3c289fb7"},
- {file = "fastavro-1.7.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:202424a5a7831b773f0f2cc2f82e89ed1726051fd5994f13dc678514144e10d4"},
- {file = "fastavro-1.7.0-cp38-cp38-win_amd64.whl", hash = "sha256:601f8c2ec166bd721f4b12eafe195dd1373d3f8dce4fe2425abd2df3e3968ac7"},
- {file = "fastavro-1.7.0-cp39-cp39-macosx_10_15_x86_64.whl", hash = "sha256:91cc9be740abca894082af4fe3ab9db057d4e5fa783cfa9a94c02ec041bf4346"},
- {file = "fastavro-1.7.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9e4463b00242e4baf52d499aeefab46a26d9dd18d808d4219cd4d21089da540e"},
- {file = "fastavro-1.7.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f7348d318858203bd108e6bcde177d8a6f0590b52bc624d815f26fb6c37029bb"},
- {file = "fastavro-1.7.0-cp39-cp39-win_amd64.whl", hash = "sha256:492e8902063aa1c73170e11c18495fcaa86b71eae3513ef83ba438ca02b16b34"},
- {file = "fastavro-1.7.0.tar.gz", hash = "sha256:4b1205f46489b4032d3155c1ab44d9824be0c7454df98d3a5bd22b78b98f23c8"},
+ {file = "fastavro-1.7.1-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:f8f7f11af8c2c074341217d6247b7ba09cadcd55f899e046c14e3a44afa5fc95"},
+ {file = "fastavro-1.7.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:11159cf23f5ff4b752b028a77bd2a7941599932527e8a6512779e25b0503f037"},
+ {file = "fastavro-1.7.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:86ce89743711355eee7d3fec34847de1ab574f4567aa4a667e966711bb2e0cd9"},
+ {file = "fastavro-1.7.1-cp310-cp310-win_amd64.whl", hash = "sha256:2e3f5ae42e033dbb6e0efa788c4b8a4e5a59bc2b9cb83f717b6d85176727faed"},
+ {file = "fastavro-1.7.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:8b671a97e864e4c024061c0d6f93f2768ba0595f917317ca4fe3e99dca6fcf3d"},
+ {file = "fastavro-1.7.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e9be910ca1645e0e5c4192d667cfdfa58dff4f8690db5af1ae33602643d41a78"},
+ {file = "fastavro-1.7.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6f233a1d0d95265f81a25c274f55017bd39d9b8f7f1387a4235bf8e01841a9ff"},
+ {file = "fastavro-1.7.1-cp311-cp311-win_amd64.whl", hash = "sha256:31c924ed2fdad56182b543941cdec9cc384443cc3ca9462e0580afeb4dc64f4b"},
+ {file = "fastavro-1.7.1-cp37-cp37m-macosx_10_15_x86_64.whl", hash = "sha256:8ad522c7f2c7791cfe54a7e87af8ac09634872f4fdeef28deec97fab8de38b24"},
+ {file = "fastavro-1.7.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:74fe0fb24a489e2fd567775935eb55a461fc6c4da8b5e3467245752ac2098284"},
+ {file = "fastavro-1.7.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f7a744569b4b44ea1f4f87e7ea8298e1e2bf23779aa6ef255b95f9f38faad48"},
+ {file = "fastavro-1.7.1-cp37-cp37m-win_amd64.whl", hash = "sha256:b793252a6325341890bbbbdca2804d4db3d5d29ff7f15a58fcf84dda440808fa"},
+ {file = "fastavro-1.7.1-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:a107a8b333628c0f8a2ece5d5a57c69923719a139b106ace4050206250df4b13"},
+ {file = "fastavro-1.7.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:182671595683054ec234beec03f485b5c889c21c08e429577ae7929480703409"},
+ {file = "fastavro-1.7.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:980ce7fdccf328d287e8b789f4e3b422f64c84ed9cd81c05dd7c560c3d8076b1"},
+ {file = "fastavro-1.7.1-cp38-cp38-win_amd64.whl", hash = "sha256:a5f104bc5b4986bbbcab170918c4d8ffa4f8efa3ebe8ec190954178630074d5a"},
+ {file = "fastavro-1.7.1-cp39-cp39-macosx_11_0_x86_64.whl", hash = "sha256:5ff701d6b228218a3d9c09b392205dd0899afa501a4f14724bef0ce13a719700"},
+ {file = "fastavro-1.7.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb2d8ea7b21493bb18ff7c68401edbc663bbd8a57016d6cf2c4b0a2dc4464e7"},
+ {file = "fastavro-1.7.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e15e8163581983f07a3156902761cf746bbe1e646abd9553cc9a1cede6e23ae9"},
+ {file = "fastavro-1.7.1-cp39-cp39-win_amd64.whl", hash = "sha256:4d56cc05ceb1c9f4de19c1f330a4f140af3b944834d63cd0e11517deaea21ee1"},
+ {file = "fastavro-1.7.1.tar.gz", hash = "sha256:4b8bcae4ed6343af186e638061cdfbc5331cdb5e026d055099c91d4f07be838c"},
@@ -1670 +1670 @@ name = "isort"
-version = "5.11.4"
+version = "5.12.0"
@@ -1674 +1674 @@ optional = false
-python-versions = ">=3.7.0"
+python-versions = ">=3.8.0"
@@ -1676,2 +1676,2 @@ files = [
- {file = "isort-5.11.4-py3-none-any.whl", hash = "sha256:c033fd0edb91000a7f09527fe5c75321878f98322a77ddcc81adbd83724afb7b"},
- {file = "isort-5.11.4.tar.gz", hash = "sha256:6db30c5ded9815d813932c04c2f85a360bcdd35fed496f4d8f35495ef0a261b6"},
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
@@ -1681,2 +1681,2 @@ files = [
-colors = ["colorama (>=0.4.3,<0.5.0)"]
-pipfile-deprecated-finder = ["pipreqs", "requirementslib"]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
@@ -1720,3 +1720,2 @@ python-versions = "*"
-files = [
- {file = "master.zip", hash = "sha256:3761fa0854068bee1fef62deb3680365ea47fb7c502de5a41c401fb8d34f8dfd"},
-]
+files = []
+develop = false
@@ -1725,2 +1724,5 @@ files = [
-type = "url"
-url = "https://github.com/kpu/kenlm/archive/master.zip"
+type = "git"
+url = "https://github.com/kpu/kenlm"
+reference = "master"
+resolved_reference = "bcd4af619a2fa45f5876d8855f7876cc09f663af"
+
@@ -2377 +2379 @@ name = "mypy-extensions"
-version = "0.4.3"
+version = "0.4.4"
@@ -2381 +2383 @@ optional = false
-python-versions = "*"
+python-versions = ">=2.7"
@@ -2383,2 +2385 @@ files = [
- {file = "mypy_extensions-0.4.3-py2.py3-none-any.whl", hash = "sha256:090fedd75945a69ae91ce1303b5824f428daf5a028d2f6ab8a299250a846f15d"},
- {file = "mypy_extensions-0.4.3.tar.gz", hash = "sha256:2d82818f5bb3e369420cb3c4060a7970edba416647068eb4c5343488a6c604a8"},
+ {file = "mypy_extensions-0.4.4.tar.gz", hash = "sha256:c8b707883a96efe9b4bb3aaf0dcc07e7e217d7d8368eec4db4049ee9e142f4fd"},
@@ -2539 +2540 @@ name = "openpyxl"
-version = "3.0.10"
+version = "3.1.0"
@@ -2545,2 +2546,2 @@ files = [
- {file = "openpyxl-3.0.10-py2.py3-none-any.whl", hash = "sha256:0ab6d25d01799f97a9464630abacbb34aafecdcaa0ef3cba6d6b3499867d0355"},
- {file = "openpyxl-3.0.10.tar.gz", hash = "sha256:e47805627aebcf860edb4edf7987b1309c1b3632f3750538ed962bbcc3bd7449"},
+ {file = "openpyxl-3.1.0-py2.py3-none-any.whl", hash = "sha256:24d7d361025d186ba91eff58135d50855cf035a84371b891e58fb6eb5125660f"},
+ {file = "openpyxl-3.1.0.tar.gz", hash = "sha256:eccedbe1cdd8b2494057e73959b496821141038dbb7eb9266ea59e3f34208231"},
@@ -2573 +2574 @@ name = "orjson"
-version = "3.8.5"
+version = "3.8.6"
@@ -2579,44 +2580,44 @@ files = [
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:143639b9898b094883481fac37733231da1c2ae3aec78a1dd8d3b58c9c9fceef"},
- {file = "orjson-3.8.5-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:31f43e63e0d94784c55e86bd376df3f80b574bea8c0bc5ecd8041009fa8ec78a"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c802ea6d4a0d40f096aceb5e7ef0a26c23d276cb9334e1cadcf256bb090b6426"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bf298b55b371c2772420c5ace4d47b0a3ea1253667e20ded3c363160fd0575f6"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68cb4a8501a463771d55bb22fc72795ec7e21d71ab083e000a2c3b651b6fb2af"},
- {file = "orjson-3.8.5-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:4f1427952b3bd92bfb63a61b7ffc33a9f54ec6de296fa8d924cbeba089866acb"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c0a9f329468c8eb000742455b83546849bcd69495d6baa6e171c7ee8600a47bd"},
- {file = "orjson-3.8.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6535d527aa1e4a757a6ce9b61f3dd74edc762e7d2c6991643aae7c560c8440bd"},
- {file = "orjson-3.8.5-cp310-none-win_amd64.whl", hash = "sha256:2eee64c028adf6378dd714c8debc96d5b92b6bb4862debb65ca868e59bac6c63"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:f5745ff473dd5c6718bf8c8d5bc183f638b4f3e03c7163ffcda4d4ef453f42ff"},
- {file = "orjson-3.8.5-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:544f1240b295083697027a5093ec66763218ff16f03521d5020e7a436d2e417b"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c85c9c6bab97a831e7741089057347d99901b4db2451a076ca8adedc7d96297f"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9bae7347764e7be6dada980fd071e865544c98317ab61af575c9cc5e1dc7e3fe"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c67f6f6e9d26a06b63126112a7bc8d8529df048d31df2a257a8484b76adf3e5d"},
- {file = "orjson-3.8.5-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:758238364142fcbeca34c968beefc0875ffa10aa2f797c82f51cfb1d22d0934e"},
- {file = "orjson-3.8.5-cp311-none-win_amd64.whl", hash = "sha256:cc7579240fb88a626956a6cb4a181a11b62afbc409ce239a7b866568a2412fa2"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:79aa3e47cbbd4eedbbde4f988f766d6cf38ccb51d52cfabfeb6b8d1b58654d25"},
- {file = "orjson-3.8.5-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:2544cd0d089faa862f5a39f508ee667419e3f9e11f119a6b1505cfce0eb26601"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f2be0025ca7e460bcacb250aba8ce0239be62957d58cf34045834cc9302611d3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b57bf72902d818506906e49c677a791f90dbd7f0997d60b14bc6c1ce4ce4cf9"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:93ae9832a11c6a9efa8c14224e5caf6e35046efd781de14e59eb69ab4e561cf3"},
- {file = "orjson-3.8.5-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:0e28330cc6d51741cad0edd1b57caf6c5531aff30afe41402acde0a03246b8ed"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:155954d725627b5480e6cc1ca488afb4fa685099a4ace5f5bf21a182fabf6706"},
- {file = "orjson-3.8.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ece1b6ef9312df5d5274ca6786e613b7da7de816356e36bcad9ea8a73d15ab71"},
- {file = "orjson-3.8.5-cp37-none-win_amd64.whl", hash = "sha256:6f58d1f0702332496bc1e2d267c7326c851991b62cf6395370d59c47f9890007"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:933f4ab98362f46a59a6d0535986e1f0cae2f6b42435e24a55922b4bc872af0c"},
- {file = "orjson-3.8.5-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:47a7ca236b25a138a74b2cb5169adcdc5b2b8abdf661de438ba65967a2cde9dc"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b573ca942c626fcf8a86be4f180b86b2498b18ae180f37b4180c2aced5808710"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a9bab11611d5452efe4ae5315f5eb806f66104c08a089fb84c648d2e8e00f106"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eee2f5f6476617d01ca166266d70fd5605d3397a41f067022ce04a2e1ced4c8d"},
- {file = "orjson-3.8.5-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:ec0b0b6cd0b84f03537f22b719aca705b876c54ab5cf3471d551c9644127284f"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:df3287dc304c8c4556dc85c4ab89eb333307759c1863f95e72e555c0cfce3e01"},
- {file = "orjson-3.8.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:09f40add3c2d208e20f8bf185df38f992bf5092202d2d30eced8f6959963f1d5"},
- {file = "orjson-3.8.5-cp38-none-win_amd64.whl", hash = "sha256:232ec1df0d708f74e0dd1fccac1e9a7008cd120d48fe695e8f0c9d80771da430"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:8fba3e7aede3e88a01e94e6fe63d4580162b212e6da27ae85af50a1787e41416"},
- {file = "orjson-3.8.5-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:85e22c358cab170c8604e9edfffcc45dd7b0027ce57ed6bcacb556e8bfbbb704"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:eeab1d8247507a75926adf3ca995c74e91f5db1f168815bf3e774f992ba52b50"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:daaaef15a41e9e8cadc7677cefe00065ae10bce914eefe8da1cd26b3d063970b"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ccc9f52cf46bd353c6ae1153eaf9d18257ddc110d135198b0cd8718474685ce"},
- {file = "orjson-3.8.5-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:d48c182c7ff4ea0787806de8a2f9298ca44fd0068ecd5f23a4b2d8e03c745cb6"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:1848e3b4cc09cc82a67262ae56e2a772b0548bb5a6f9dcaee10dcaaf0a5177b7"},
- {file = "orjson-3.8.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:38480031bc8add58effe802291e4abf7042ef72ae1a4302efe9a36c8f8bfbfcc"},
- {file = "orjson-3.8.5-cp39-none-win_amd64.whl", hash = "sha256:0e9a1c2e649cbaed410c882cedc8f3b993d8f1426d9327f31762d3f46fe7cc88"},
- {file = "orjson-3.8.5.tar.gz", hash = "sha256:77a3b2bd0c4ef7723ea09081e3329dac568a62463aed127c1501441b07ffc64b"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"},
+ {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"},
+ {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"},
+ {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"},
+ {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"},
+ {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"},
+ {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"},
+ {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"},
+ {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"},
+ {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"},
+ {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"},
+ {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"},
+ {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"},
+ {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"},
+ {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"},
+ {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"},
+ {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"},
+ {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"},
+ {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"},
+ {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"},
+ {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"},
@@ -2700 +2701 @@ name = "pandas-stubs"
-version = "1.5.2.230105"
+version = "1.5.3.230203"
@@ -2706,2 +2707,2 @@ files = [
- {file = "pandas_stubs-1.5.2.230105-py3-none-any.whl", hash = "sha256:b2874d26eabcaecb83aa8435d47e41f1b5f90ad04f3e1e70c3c8604027490225"},
- {file = "pandas_stubs-1.5.2.230105.tar.gz", hash = "sha256:c78e433aca82577ce4a9fefc0d78cd26828dd95f8b1612416babff1b6957a22a"},
+ {file = "pandas_stubs-1.5.3.230203-py3-none-any.whl", hash = "sha256:04da026f9933e1534ba480a2b06139a2dbd624c57ce0697e5873460d7815c67d"},
+ {file = "pandas_stubs-1.5.3.230203.tar.gz", hash = "sha256:7a4eedb210c77e0911fe32a4178673ba68ef569c034117ea19ada79451db4af9"},
@@ -2767,7 +2767,0 @@ files = [
- {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"},
- {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"},
- {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"},
- {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"},
- {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"},
- {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"},
- {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"},
@@ -2845 +2839 @@ name = "pip"
-version = "22.3.1"
+version = "23.0"
@@ -2851,2 +2845,2 @@ files = [
- {file = "pip-22.3.1-py3-none-any.whl", hash = "sha256:908c78e6bc29b676ede1c4d57981d490cb892eb45cd8c214ab6298125119e077"},
- {file = "pip-22.3.1.tar.gz", hash = "sha256:65fd48317359f3af8e593943e6ae1506b66325085ea64b706a998c6e83eeaf38"},
+ {file = "pip-23.0-py3-none-any.whl", hash = "sha256:b5f88adff801f5ef052bcdef3daa31b55eb67b0fccd6d0106c206fa248e0463c"},
+ {file = "pip-23.0.tar.gz", hash = "sha256:aee438284e82c8def684b0bcc50b1f6ed5e941af97fa940e83e2e8ef1a59da9b"},
@@ -2921 +2915 @@ name = "platformdirs"
-version = "2.6.2"
+version = "3.0.0"
@@ -2927,2 +2921,2 @@ files = [
- {file = "platformdirs-2.6.2-py3-none-any.whl", hash = "sha256:83c8f6d04389165de7c9b6f0c682439697887bca0aa2f1c87ef1826be3584490"},
- {file = "platformdirs-2.6.2.tar.gz", hash = "sha256:e1fea1fe471b9ff8332e229df3cb7de4f53eeea4998d3b6bfff542115e998bd2"},
+ {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"},
+ {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"},
@@ -2932,2 +2926,2 @@ files = [
-docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"]
-test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
+docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"]
@@ -3287 +3281 @@ name = "pycryptodomex"
-version = "3.16.0"
+version = "3.17"
@@ -3293,26 +3287,33 @@ files = [
- {file = "pycryptodomex-3.16.0-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:b3d04c00d777c36972b539fb79958790126847d84ec0129fce1efef250bfe3ce"},
- {file = "pycryptodomex-3.16.0-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:e5a670919076b71522c7d567a9043f66f14b202414a63c3a078b5831ae342c03"},
- {file = "pycryptodomex-3.16.0-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:ce338a9703f54b2305a408fc9890eb966b727ce72b69f225898bb4e9d9ed3f1f"},
- {file = "pycryptodomex-3.16.0-cp27-cp27m-manylinux2014_aarch64.whl", hash = "sha256:a1c0ae7123448ecb034c75c713189cb00ebe2d415b11682865b6c54d200d9c93"},
- {file = "pycryptodomex-3.16.0-cp27-cp27m-win32.whl", hash = "sha256:8851585ff19871e5d69e1790f4ca5f6fd1699d6b8b14413b472a4c0dbc7ea780"},
- {file = "pycryptodomex-3.16.0-cp27-cp27m-win_amd64.whl", hash = "sha256:8dd2d9e3c617d0712ed781a77efd84ea579e76c5f9b2a4bc0b684ebeddf868b2"},
- {file = "pycryptodomex-3.16.0-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:2ad9bb86b355b6104796567dd44c215b3dc953ef2fae5e0bdfb8516731df92cf"},
- {file = "pycryptodomex-3.16.0-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:e25a2f5667d91795f9417cb856f6df724ccdb0cdd5cbadb212ee9bf43946e9f8"},
- {file = "pycryptodomex-3.16.0-cp27-cp27mu-manylinux2014_aarch64.whl", hash = "sha256:b0789a8490114a2936ed77c87792cfe77582c829cb43a6d86ede0f9624ba8aa3"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-macosx_10_9_x86_64.whl", hash = "sha256:0da835af786fdd1c9930994c78b23e88d816dc3f99aa977284a21bbc26d19735"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-manylinux2014_aarch64.whl", hash = "sha256:22aed0868622d95179217c298e37ed7410025c7b29dac236d3230617d1e4ed56"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:1619087fb5b31510b0b0b058a54f001a5ffd91e6ffee220d9913064519c6a69d"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:70288d9bfe16b2fd0d20b6c365db614428f1bcde7b20d56e74cf88ade905d9eb"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:7993d26dae4d83b8f4ce605bb0aecb8bee330bb3c95475ef06f3694403621e71"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-musllinux_1_1_i686.whl", hash = "sha256:1cda60207be8c1cf0b84b9138f9e3ca29335013d2b690774a5e94678ff29659a"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:04610536921c1ec7adba158ef570348550c9f3a40bc24be9f8da2ef7ab387981"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-win32.whl", hash = "sha256:daa67f5ebb6fbf1ee9c90decaa06ca7fc88a548864e5e484d52b0920a57fe8a5"},
- {file = "pycryptodomex-3.16.0-cp35-abi3-win_amd64.whl", hash = "sha256:231dc8008cbdd1ae0e34645d4523da2dbc7a88c325f0d4a59635a86ee25b41dd"},
- {file = "pycryptodomex-3.16.0-pp27-pypy_73-macosx_10_9_x86_64.whl", hash = "sha256:4dbbe18cc232b5980c7633972ae5417d0df76fe89e7db246eefd17ef4d8e6d7a"},
- {file = "pycryptodomex-3.16.0-pp27-pypy_73-manylinux2010_x86_64.whl", hash = "sha256:893f8a97d533c66cc3a56e60dd3ed40a3494ddb4aafa7e026429a08772f8a849"},
- {file = "pycryptodomex-3.16.0-pp27-pypy_73-win32.whl", hash = "sha256:6a465e4f856d2a4f2a311807030c89166529ccf7ccc65bef398de045d49144b6"},
- {file = "pycryptodomex-3.16.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:ba57ac7861fd2c837cdb33daf822f2a052ff57dd769a2107807f52a36d0e8d38"},
- {file = "pycryptodomex-3.16.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:f2b971a7b877348a27dcfd0e772a0343fb818df00b74078e91c008632284137d"},
- {file = "pycryptodomex-3.16.0-pp37-pypy37_pp73-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:e2453162f473c1eae4826eb10cd7bce19b5facac86d17fb5f29a570fde145abd"},
- {file = "pycryptodomex-3.16.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:0ba28aa97cdd3ff5ed1a4f2b7f5cd04e721166bd75bd2b929e2734433882b583"},
- {file = "pycryptodomex-3.16.0.tar.gz", hash = "sha256:e9ba9d8ed638733c9e95664470b71d624a6def149e2db6cc52c1aca5a6a2df1d"},
+ {file = "pycryptodomex-3.17-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:12056c38e49d972f9c553a3d598425f8a1c1d35b2e4330f89d5ff1ffb70de041"},
+ {file = "pycryptodomex-3.17-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:ab33c2d9f275e05e235dbca1063753b5346af4a5cac34a51fa0da0d4edfb21d7"},
+ {file = "pycryptodomex-3.17-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:caa937ff29d07a665dfcfd7a84f0d4207b2ebf483362fa9054041d67fdfacc20"},
+ {file = "pycryptodomex-3.17-cp27-cp27m-manylinux2014_aarch64.whl", hash = "sha256:db23d7341e21b273d2440ec6faf6c8b1ca95c8894da612e165be0b89a8688340"},
+ {file = "pycryptodomex-3.17-cp27-cp27m-musllinux_1_1_aarch64.whl", hash = "sha256:f854c8476512cebe6a8681cc4789e4fcff6019c17baa0fd72b459155dc605ab4"},
+ {file = "pycryptodomex-3.17-cp27-cp27m-win32.whl", hash = "sha256:a57e3257bacd719769110f1f70dd901c5b6955e9596ad403af11a3e6e7e3311c"},
+ {file = "pycryptodomex-3.17-cp27-cp27m-win_amd64.whl", hash = "sha256:d38ab9e53b1c09608ba2d9b8b888f1e75d6f66e2787e437adb1fecbffec6b112"},
+ {file = "pycryptodomex-3.17-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:3c2516b42437ae6c7a29ef3ddc73c8d4714e7b6df995b76be4695bbe4b3b5cd2"},
+ {file = "pycryptodomex-3.17-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:5c23482860302d0d9883404eaaa54b0615eefa5274f70529703e2c43cc571827"},
+ {file = "pycryptodomex-3.17-cp27-cp27mu-manylinux2014_aarch64.whl", hash = "sha256:7a8dc3ee7a99aae202a4db52de5a08aa4d01831eb403c4d21da04ec2f79810db"},
+ {file = "pycryptodomex-3.17-cp27-cp27mu-musllinux_1_1_aarch64.whl", hash = "sha256:7cc28dd33f1f3662d6da28ead4f9891035f63f49d30267d3b41194c8778997c8"},
+ {file = "pycryptodomex-3.17-cp35-abi3-macosx_10_9_universal2.whl", hash = "sha256:2d4d395f109faba34067a08de36304e846c791808524614c731431ee048fe70a"},
+ {file = "pycryptodomex-3.17-cp35-abi3-macosx_10_9_x86_64.whl", hash = "sha256:55eed98b4150a744920597c81b3965b632038781bab8a08a12ea1d004213c600"},
+ {file = "pycryptodomex-3.17-cp35-abi3-manylinux2014_aarch64.whl", hash = "sha256:7fa0b52df90343fafe319257b31d909be1d2e8852277fb0376ba89d26d2921db"},
+ {file = "pycryptodomex-3.17-cp35-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:78f0ddd4adc64baa39b416f3637aaf99f45acb0bcdc16706f0cc7ebfc6f10109"},
+ {file = "pycryptodomex-3.17-cp35-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a4fa037078e92c7cc49f6789a8bac3de06856740bb2038d05f2d9a2e4b165d59"},
+ {file = "pycryptodomex-3.17-cp35-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:88b0d5bb87eaf2a31e8a759302b89cf30c97f2f8ca7d83b8c9208abe8acb447a"},
+ {file = "pycryptodomex-3.17-cp35-abi3-musllinux_1_1_i686.whl", hash = "sha256:6feedf4b0e36b395329b4186a805f60f900129cdf0170e120ecabbfcb763995d"},
+ {file = "pycryptodomex-3.17-cp35-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:7a6651a07f67c28b6e978d63aa3a3fccea0feefed9a8453af3f7421a758461b7"},
+ {file = "pycryptodomex-3.17-cp35-abi3-win32.whl", hash = "sha256:32e764322e902bbfac49ca1446604d2839381bbbdd5a57920c9daaf2e0b778df"},
+ {file = "pycryptodomex-3.17-cp35-abi3-win_amd64.whl", hash = "sha256:4b51e826f0a04d832eda0790bbd0665d9bfe73e5a4d8ea93b6a9b38beeebe935"},
+ {file = "pycryptodomex-3.17-pp27-pypy_73-macosx_10_9_x86_64.whl", hash = "sha256:d4cf0128da167562c49b0e034f09e9cedd733997354f2314837c2fa461c87bb1"},
+ {file = "pycryptodomex-3.17-pp27-pypy_73-manylinux2010_x86_64.whl", hash = "sha256:c92537b596bd5bffb82f8964cabb9fef1bca8a28a9e0a69ffd3ec92a4a7ad41b"},
+ {file = "pycryptodomex-3.17-pp27-pypy_73-win32.whl", hash = "sha256:599bb4ae4bbd614ca05f49bd4e672b7a250b80b13ae1238f05fd0f09d87ed80a"},
+ {file = "pycryptodomex-3.17-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4c4674f4b040321055c596aac926d12f7f6859dfe98cd12f4d9453b43ab6adc8"},
+ {file = "pycryptodomex-3.17-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:67a3648025e4ddb72d43addab764336ba2e670c8377dba5dd752e42285440d31"},
+ {file = "pycryptodomex-3.17-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:40e8a11f578bd0851b02719c862d55d3ee18d906c8b68a9c09f8c564d6bb5b92"},
+ {file = "pycryptodomex-3.17-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:23d83b610bd97704f0cd3acc48d99b76a15c8c1540d8665c94d514a49905bad7"},
+ {file = "pycryptodomex-3.17-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:fd29d35ac80755e5c0a99d96b44fb9abbd7e871849581ea6a4cb826d24267537"},
+ {file = "pycryptodomex-3.17-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:64b876d57cb894b31056ad8dd6a6ae1099b117ae07a3d39707221133490e5715"},
+ {file = "pycryptodomex-3.17-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ee8bf4fdcad7d66beb744957db8717afc12d176e3fd9c5d106835133881a049b"},
+ {file = "pycryptodomex-3.17-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:c84689c73358dfc23f9fdcff2cb9e7856e65e2ce3b5ed8ff630d4c9bdeb1867b"},
+ {file = "pycryptodomex-3.17.tar.gz", hash = "sha256:0af93aad8d62e810247beedef0261c148790c52f3cd33643791cc6396dd217c1"},
@@ -4125 +4126 @@ name = "rich"
-version = "13.2.0"
+version = "13.3.1"
@@ -4131,2 +4132,2 @@ files = [
- {file = "rich-13.2.0-py3-none-any.whl", hash = "sha256:7c963f0d03819221e9ac561e1bc866e3f95a02248c1234daa48954e6d381c003"},
- {file = "rich-13.2.0.tar.gz", hash = "sha256:f1a00cdd3eebf999a15d85ec498bfe0b1a77efe9b34f645768a54132ef444ac5"},
+ {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"},
+ {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"},
@@ -4137 +4138 @@ markdown-it-py = ">=2.1.0,<3.0.0"
-pygments = ">=2.6.0,<3.0.0"
+pygments = ">=2.14.0,<3.0.0"
@@ -4140 +4141 @@ pygments = ">=2.6.0,<3.0.0"
-jupyter = ["ipywidgets (>=7.5.1,<8.0.0)"]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
@@ -4241 +4242 @@ name = "setuptools"
-version = "66.1.1"
+version = "67.2.0"
@@ -4247,2 +4248,2 @@ files = [
- {file = "setuptools-66.1.1-py3-none-any.whl", hash = "sha256:6f590d76b713d5de4e49fe4fbca24474469f53c83632d5d0fd056f7ff7e8112b"},
- {file = "setuptools-66.1.1.tar.gz", hash = "sha256:ac4008d396bc9cd983ea483cb7139c0240a07bbc74ffb6232fceffedc6cf03a8"},
+ {file = "setuptools-67.2.0-py3-none-any.whl", hash = "sha256:16ccf598aab3b506593c17378473978908a2734d7336755a8769b480906bec1c"},
+ {file = "setuptools-67.2.0.tar.gz", hash = "sha256:b440ee5f7e607bb8c9de15259dba2583dd41a38879a7abc1d43a71c59524da48"},
@@ -4593,2 +4593,0 @@ files = [
- {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"},
- {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"},
@@ -4600 +4598,0 @@ files = [
- {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"},
@@ -4609 +4606,0 @@ files = [
- {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"},
@@ -4728 +4725 @@ name = "transformers"
-version = "4.26.0"
+version = "4.26.1"
@@ -4734,2 +4731,2 @@ files = [
- {file = "transformers-4.26.0-py3-none-any.whl", hash = "sha256:6a902eee6098d9a737faadf185b8df5a169acc695ebbde5a81b90528f43e665f"},
- {file = "transformers-4.26.0.tar.gz", hash = "sha256:d7859bd83829a3682ca632197ee5c72556e1063d199ab84eec35c4f23b3d73a3"},
+ {file = "transformers-4.26.1-py3-none-any.whl", hash = "sha256:dae2fa15290c1f526e1b629b0e235eea5e4c04078fcaf1f197a70d51b4f65df2"},
+ {file = "transformers-4.26.1.tar.gz", hash = "sha256:32dc474157367f8e551f470af0136a1ddafc9e18476400c3869f1ef4f0c12042"},
@@ -4898 +4895 @@ name = "types-requests"
-version = "2.28.11.8"
+version = "2.28.11.12"
@@ -4904,2 +4901,2 @@ files = [
- {file = "types-requests-2.28.11.8.tar.gz", hash = "sha256:e67424525f84adfbeab7268a159d3c633862dafae15c5b19547ce1b55954f0a3"},
- {file = "types_requests-2.28.11.8-py3-none-any.whl", hash = "sha256:61960554baca0008ae7e2db2bd3b322ca9a144d3e80ce270f5fb640817e40994"},
+ {file = "types-requests-2.28.11.12.tar.gz", hash = "sha256:fd530aab3fc4f05ee36406af168f0836e6f00f1ee51a0b96b7311f82cb675230"},
+ {file = "types_requests-2.28.11.12-py3-none-any.whl", hash = "sha256:dbc2933635860e553ffc59f5e264264981358baffe6342b925e3eb8261f866ee"},
@@ -4913 +4910 @@ name = "types-urllib3"
-version = "1.26.25.4"
+version = "1.26.25.5"
@@ -4919,2 +4916,2 @@ files = [
- {file = "types-urllib3-1.26.25.4.tar.gz", hash = "sha256:eec5556428eec862b1ac578fb69aab3877995a99ffec9e5a12cf7fbd0cc9daee"},
- {file = "types_urllib3-1.26.25.4-py3-none-any.whl", hash = "sha256:ed6b9e8a8be488796f72306889a06a3fc3cb1aa99af02ab8afb50144d7317e49"},
+ {file = "types-urllib3-1.26.25.5.tar.gz", hash = "sha256:5630e578246d170d91ebe3901788cd28d53c4e044dc2e2488e3b0d55fb6895d8"},
+ {file = "types_urllib3-1.26.25.5-py3-none-any.whl", hash = "sha256:e8f25c8bb85cde658c72ee931e56e7abd28803c26032441eea9ff4a4df2b0c31"},
@@ -5355 +5352 @@ name = "zipp"
-version = "3.11.0"
+version = "3.13.0"
@@ -5361,2 +5358,2 @@ files = [
- {file = "zipp-3.11.0-py3-none-any.whl", hash = "sha256:83a28fcb75844b5c0cdaf5aa4003c2d728c77e05f5aeabe8e95e56727005fbaa"},
- {file = "zipp-3.11.0.tar.gz", hash = "sha256:a7a22e05929290a67401440b39690ae6563279bced5f314609d9d03798f56766"},
+ {file = "zipp-3.13.0-py3-none-any.whl", hash = "sha256:e8b2a36ea17df80ffe9e2c4fda3f693c3dad6df1697d3cd3af232db680950b0b"},
+ {file = "zipp-3.13.0.tar.gz", hash = "sha256:23f70e964bc11a34cef175bc90ba2914e1e4545ea1e3e2f67c079671883f9cb6"},
@@ -5366 +5363 @@ files = [
-docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (>=3.5)"]
+docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-lint"]
@@ -5439 +5436 @@ python-versions = "3.9.15"
-content-hash = "1d75ae381d53e8abcc1896f6f33c71afadc51f9ea0d809e9effadacf716fa628"
+content-hash = "84b6ae8841c72d6f4955e9255622d2b826171732e65fb8104d8cb1ecd8556413"
diff --git a/workers/datasets_based/pyproject.toml b/workers/datasets_based/pyproject.toml
index 76e98168..946ae1e0 100644
--- a/workers/datasets_based/pyproject.toml
+++ b/workers/datasets_based/pyproject.toml
@@ -18 +18 @@ huggingface-hub = "^0.12.0"
-kenlm = { url = "https://github.com/kpu/kenlm/archive/master.zip" }
+kenlm = { git = "https://github.com/kpu/kenlm", branch = "master" }
|
|
6023542baef3b66e06a93c3c99b3183280cf8aec
|
Sylvain Lesage
| 2023-02-10T16:29:42 |
use classmethod for factories instead of staticmethod (#791)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/config.py b/jobs/mongodb_migration/src/mongodb_migration/config.py
index 8047c2a4..dcecb7a6 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/config.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/config.py
@@ -18,2 +18,2 @@ class DatabaseMigrationsConfig:
- @staticmethod
- def from_env() -> "DatabaseMigrationsConfig":
+ @classmethod
+ def from_env(cls) -> "DatabaseMigrationsConfig":
@@ -22 +22 @@ class DatabaseMigrationsConfig:
- return DatabaseMigrationsConfig(
+ return cls(
@@ -35,3 +35,3 @@ class JobConfig:
- @staticmethod
- def from_env() -> "JobConfig":
- return JobConfig(
+ @classmethod
+ def from_env(cls) -> "JobConfig":
+ return cls(
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index b41dc37d..d13466d7 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -22,2 +22,2 @@ class AssetsConfig:
- @staticmethod
- def from_env() -> "AssetsConfig":
+ @classmethod
+ def from_env(cls) -> "AssetsConfig":
@@ -26 +26 @@ class AssetsConfig:
- return AssetsConfig(
+ return cls(
@@ -43,2 +43,2 @@ class CommonConfig:
- @staticmethod
- def from_env() -> "CommonConfig":
+ @classmethod
+ def from_env(cls) -> "CommonConfig":
@@ -47 +47 @@ class CommonConfig:
- return CommonConfig(
+ return cls(
@@ -63,2 +63,2 @@ class CacheConfig:
- @staticmethod
- def from_env() -> "CacheConfig":
+ @classmethod
+ def from_env(cls) -> "CacheConfig":
@@ -67 +67 @@ class CacheConfig:
- return CacheConfig(
+ return cls(
@@ -84,2 +84,2 @@ class QueueConfig:
- @staticmethod
- def from_env() -> "QueueConfig":
+ @classmethod
+ def from_env(cls) -> "QueueConfig":
@@ -88 +88 @@ class QueueConfig:
- return QueueConfig(
+ return cls(
@@ -110,2 +110,2 @@ class ProcessingGraphConfig:
- @staticmethod
- def from_env() -> "ProcessingGraphConfig":
+ @classmethod
+ def from_env(cls) -> "ProcessingGraphConfig":
@@ -113 +113 @@ class ProcessingGraphConfig:
- return ProcessingGraphConfig()
+ return cls()
diff --git a/services/admin/src/admin/config.py b/services/admin/src/admin/config.py
index 34233935..64313c20 100644
--- a/services/admin/src/admin/config.py
+++ b/services/admin/src/admin/config.py
@@ -27,2 +27,2 @@ class UvicornConfig:
- @staticmethod
- def from_env() -> "UvicornConfig":
+ @classmethod
+ def from_env(cls) -> "UvicornConfig":
@@ -31 +31 @@ class UvicornConfig:
- return UvicornConfig(
+ return cls(
@@ -55,2 +55,2 @@ class AdminConfig:
- @staticmethod
- def from_env(common_config: CommonConfig) -> "AdminConfig":
+ @classmethod
+ def from_env(cls, common_config: CommonConfig) -> "AdminConfig":
@@ -61 +61 @@ class AdminConfig:
- return AdminConfig(
+ return cls(
@@ -84,2 +84,2 @@ class AppConfig:
- @staticmethod
- def from_env() -> "AppConfig":
+ @classmethod
+ def from_env(cls) -> "AppConfig":
@@ -87 +87 @@ class AppConfig:
- return AppConfig(
+ return cls(
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index f5b58c8e..2cf5637c 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -26,2 +26,2 @@ class UvicornConfig:
- @staticmethod
- def from_env() -> "UvicornConfig":
+ @classmethod
+ def from_env(cls) -> "UvicornConfig":
@@ -30 +30 @@ class UvicornConfig:
- return UvicornConfig(
+ return cls(
@@ -50,2 +50,2 @@ class ApiConfig:
- @staticmethod
- def from_env(common_config: CommonConfig) -> "ApiConfig":
+ @classmethod
+ def from_env(cls, common_config: CommonConfig) -> "ApiConfig":
@@ -56 +56 @@ class ApiConfig:
- return ApiConfig(
+ return cls(
@@ -72,2 +72,2 @@ class AppConfig:
- @staticmethod
- def from_env() -> "AppConfig":
+ @classmethod
+ def from_env(cls) -> "AppConfig":
@@ -75 +75 @@ class AppConfig:
- return AppConfig(
+ return cls(
diff --git a/workers/datasets_based/src/datasets_based/config.py b/workers/datasets_based/src/datasets_based/config.py
index 07dc08fe..fc3dd81d 100644
--- a/workers/datasets_based/src/datasets_based/config.py
+++ b/workers/datasets_based/src/datasets_based/config.py
@@ -34,2 +34,2 @@ class WorkerLoopConfig:
- @staticmethod
- def from_env() -> "WorkerLoopConfig":
+ @classmethod
+ def from_env(cls) -> "WorkerLoopConfig":
@@ -38 +38 @@ class WorkerLoopConfig:
- return WorkerLoopConfig(
+ return cls(
@@ -58,2 +58,2 @@ class DatasetsBasedConfig:
- @staticmethod
- def from_env() -> "DatasetsBasedConfig":
+ @classmethod
+ def from_env(cls) -> "DatasetsBasedConfig":
@@ -62 +62 @@ class DatasetsBasedConfig:
- return DatasetsBasedConfig(
+ return cls(
@@ -84,2 +84,2 @@ class FirstRowsConfig:
- @staticmethod
- def from_env() -> "FirstRowsConfig":
+ @classmethod
+ def from_env(cls) -> "FirstRowsConfig":
@@ -88 +88 @@ class FirstRowsConfig:
- return FirstRowsConfig(
+ return cls(
@@ -116,2 +116,2 @@ class ParquetAndDatasetInfoConfig:
- @staticmethod
- def from_env() -> "ParquetAndDatasetInfoConfig":
+ @classmethod
+ def from_env(cls) -> "ParquetAndDatasetInfoConfig":
@@ -120 +120 @@ class ParquetAndDatasetInfoConfig:
- return ParquetAndDatasetInfoConfig(
+ return cls(
@@ -141,2 +141,2 @@ class NumbaConfig:
- @staticmethod
- def from_env() -> "NumbaConfig":
+ @classmethod
+ def from_env(cls) -> "NumbaConfig":
@@ -145 +145 @@ class NumbaConfig:
- return NumbaConfig(path=env.str(name="NUMBA_CACHE_DIR", default=NUMBA_CACHE_DIR))
+ return cls(path=env.str(name="NUMBA_CACHE_DIR", default=NUMBA_CACHE_DIR))
@@ -159,3 +159,3 @@ class AppConfig:
- @staticmethod
- def from_env() -> "AppConfig":
- return AppConfig(
+ @classmethod
+ def from_env(cls) -> "AppConfig":
+ return cls(
|
|
a7f3ecd180bc556a8389511c936a96817bbdd82d
|
Sylvain Lesage
| 2023-02-10T16:07:28 |
feat: 🎸 ensure immutability of the configs (#790)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/config.py b/jobs/mongodb_migration/src/mongodb_migration/config.py
index 98cfea0a..8047c2a4 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/config.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/config.py
@@ -13 +13 @@ DATABASE_MIGRATIONS_MONGO_URL = "mongodb://localhost:27017"
-@dataclass
+@dataclass(frozen=True)
@@ -28 +28 @@ class DatabaseMigrationsConfig:
-@dataclass
+@dataclass(frozen=True)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 0a75c74d..b41dc37d 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -17 +17 @@ ASSETS_STORE_DIRECTORY = None
-@dataclass
+@dataclass(frozen=True)
@@ -37 +37 @@ COMMON_LOG_LEVEL = logging.INFO
-@dataclass
+@dataclass(frozen=True)
@@ -58 +58 @@ CACHE_MONGO_URL = "mongodb://localhost:27017"
-@dataclass
+@dataclass(frozen=True)
@@ -78 +78 @@ QUEUE_MONGO_URL = "mongodb://localhost:27017"
-@dataclass
+@dataclass(frozen=True)
@@ -95 +95 @@ class QueueConfig:
-@dataclass
+@dataclass(frozen=True)
diff --git a/services/admin/src/admin/config.py b/services/admin/src/admin/config.py
index d1a226e1..34233935 100644
--- a/services/admin/src/admin/config.py
+++ b/services/admin/src/admin/config.py
@@ -21 +21 @@ ADMIN_UVICORN_PORT = 8000
-@dataclass
+@dataclass(frozen=True)
@@ -46 +46 @@ ADMIN_MAX_AGE = 10
-@dataclass
+@dataclass(frozen=True)
@@ -75 +75 @@ class AdminConfig:
-@dataclass
+@dataclass(frozen=True)
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index f1381415..f5b58c8e 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -20 +20 @@ API_UVICORN_PORT = 8000
-@dataclass
+@dataclass(frozen=True)
@@ -43 +43 @@ API_MAX_AGE_SHORT = 10 # 10 seconds
-@dataclass
+@dataclass(frozen=True)
@@ -64 +64 @@ class ApiConfig:
-@dataclass
+@dataclass(frozen=True)
diff --git a/workers/datasets_based/src/datasets_based/config.py b/workers/datasets_based/src/datasets_based/config.py
index 3bbf8e05..07dc08fe 100644
--- a/workers/datasets_based/src/datasets_based/config.py
+++ b/workers/datasets_based/src/datasets_based/config.py
@@ -26 +26 @@ def get_empty_str_list() -> List[str]:
-@dataclass
+@dataclass(frozen=True)
@@ -52 +52 @@ DATASETS_BASED_CONTENT_MAX_BYTES = 10_000_000
-@dataclass
+@dataclass(frozen=True)
@@ -76 +76 @@ FIRST_ROWS_COLUMNS_MAX_NUMBER = 1_000
-@dataclass
+@dataclass(frozen=True)
@@ -105 +105 @@ PARQUET_AND_DATASET_INFO_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s"
-@dataclass
+@dataclass(frozen=True)
@@ -137 +137 @@ NUMBA_CACHE_DIR: Optional[str] = None
-@dataclass
+@dataclass(frozen=True)
@@ -148 +148 @@ class NumbaConfig:
-@dataclass
+@dataclass(frozen=True)
|
|
761d16a6a0a87517f8a754f37d965f900f3323d7
|
Sylvain Lesage
| 2023-02-10T15:08:39 |
feat: 🎸 add concept of Resource (#784)
|
diff --git a/.github/workflows/_unit-tests-python.yml b/.github/workflows/_unit-tests-python.yml
index 6b29be44..446d3046 100644
--- a/.github/workflows/_unit-tests-python.yml
+++ b/.github/workflows/_unit-tests-python.yml
@@ -51,0 +52 @@ jobs:
+ DATABASE_MIGRATIONS_MONGO_URL: mongodb://localhost:${{ env.mongo-port }}
diff --git a/chart/templates/jobs/mongodb-migration/_container.tpl b/chart/templates/jobs/mongodb-migration/_container.tpl
index a1c90051..0ce5725d 100644
--- a/chart/templates/jobs/mongodb-migration/_container.tpl
+++ b/chart/templates/jobs/mongodb-migration/_container.tpl
@@ -12 +12 @@
- - name: MONGODB_MIGRATION_MONGO_DATABASE
+ - name: DATABASE_MIGRATIONS_MONGO_DATABASE
@@ -14 +14 @@
- - name: MONGODB_MIGRATION_MONGO_URL
+ - name: DATABASE_MIGRATIONS_MONGO_URL
diff --git a/chart/templates/worker/config-names/_container.tpl b/chart/templates/worker/config-names/_container.tpl
index bc2f8550..3a87636b 100644
--- a/chart/templates/worker/config-names/_container.tpl
+++ b/chart/templates/worker/config-names/_container.tpl
@@ -25,0 +26 @@
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
diff --git a/chart/templates/worker/config-names/deployment.yaml b/chart/templates/worker/config-names/deployment.yaml
index a58f89ae..72c63c12 100644
--- a/chart/templates/worker/config-names/deployment.yaml
+++ b/chart/templates/worker/config-names/deployment.yaml
@@ -23,0 +24 @@ spec:
+ {{ include "initContainerAssets" . | nindent 8 }}
diff --git a/chart/templates/worker/dataset-info/_container.tpl b/chart/templates/worker/dataset-info/_container.tpl
index 6091e71a..03bb31c0 100644
--- a/chart/templates/worker/dataset-info/_container.tpl
+++ b/chart/templates/worker/dataset-info/_container.tpl
@@ -21,0 +22,2 @@
+ volumeMounts:
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
diff --git a/chart/templates/worker/dataset-info/deployment.yaml b/chart/templates/worker/dataset-info/deployment.yaml
index cb750c10..f17170e1 100644
--- a/chart/templates/worker/dataset-info/deployment.yaml
+++ b/chart/templates/worker/dataset-info/deployment.yaml
@@ -22,0 +23,2 @@ spec:
+ initContainers:
+ {{ include "initContainerAssets" . | nindent 8 }}
diff --git a/chart/templates/worker/parquet-and-dataset-info/_container.tpl b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
index b926df09..2a0f4b85 100644
--- a/chart/templates/worker/parquet-and-dataset-info/_container.tpl
+++ b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
@@ -49,0 +50 @@
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
diff --git a/chart/templates/worker/parquet-and-dataset-info/deployment.yaml b/chart/templates/worker/parquet-and-dataset-info/deployment.yaml
index 8faa05e3..671a21d5 100644
--- a/chart/templates/worker/parquet-and-dataset-info/deployment.yaml
+++ b/chart/templates/worker/parquet-and-dataset-info/deployment.yaml
@@ -23 +23,3 @@ spec:
- initContainers: {{ include "initContainerCache" . | nindent 8 }}
+ initContainers:
+ {{ include "initContainerAssets" . | nindent 8 }}
+ {{ include "initContainerCache" . | nindent 8 }}
diff --git a/chart/templates/worker/parquet/_container.tpl b/chart/templates/worker/parquet/_container.tpl
index ace15115..ec04cbea 100644
--- a/chart/templates/worker/parquet/_container.tpl
+++ b/chart/templates/worker/parquet/_container.tpl
@@ -21,0 +22,2 @@
+ volumeMounts:
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
diff --git a/chart/templates/worker/parquet/deployment.yaml b/chart/templates/worker/parquet/deployment.yaml
index 980b3459..126e8fc6 100644
--- a/chart/templates/worker/parquet/deployment.yaml
+++ b/chart/templates/worker/parquet/deployment.yaml
@@ -22,0 +23,2 @@ spec:
+ initContainers:
+ {{ include "initContainerAssets" . | nindent 8 }}
diff --git a/chart/templates/worker/sizes/_container.tpl b/chart/templates/worker/sizes/_container.tpl
index b60db74f..a575818d 100644
--- a/chart/templates/worker/sizes/_container.tpl
+++ b/chart/templates/worker/sizes/_container.tpl
@@ -21,0 +22,2 @@
+ volumeMounts:
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
diff --git a/chart/templates/worker/sizes/deployment.yaml b/chart/templates/worker/sizes/deployment.yaml
index 5e0a1494..6dbd20b7 100644
--- a/chart/templates/worker/sizes/deployment.yaml
+++ b/chart/templates/worker/sizes/deployment.yaml
@@ -22,0 +23,2 @@ spec:
+ initContainers:
+ {{ include "initContainerAssets" . | nindent 8 }}
diff --git a/chart/templates/worker/split-names/_container.tpl b/chart/templates/worker/split-names/_container.tpl
index 67879b34..53f0d88c 100644
--- a/chart/templates/worker/split-names/_container.tpl
+++ b/chart/templates/worker/split-names/_container.tpl
@@ -25,0 +26 @@
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
diff --git a/chart/templates/worker/split-names/deployment.yaml b/chart/templates/worker/split-names/deployment.yaml
index 06ab6d03..df03783b 100644
--- a/chart/templates/worker/split-names/deployment.yaml
+++ b/chart/templates/worker/split-names/deployment.yaml
@@ -23,0 +24 @@ spec:
+ {{ include "initContainerAssets" . | nindent 8 }}
diff --git a/chart/templates/worker/splits/_container.tpl b/chart/templates/worker/splits/_container.tpl
index a3715380..d862a20a 100644
--- a/chart/templates/worker/splits/_container.tpl
+++ b/chart/templates/worker/splits/_container.tpl
@@ -25,0 +26 @@
+ {{ include "volumeMountAssetsRW" . | nindent 2 }}
diff --git a/chart/templates/worker/splits/deployment.yaml b/chart/templates/worker/splits/deployment.yaml
index 0a264547..cd1872a9 100644
--- a/chart/templates/worker/splits/deployment.yaml
+++ b/chart/templates/worker/splits/deployment.yaml
@@ -23 +23,3 @@ spec:
- initContainers: {{ include "initContainerCache" . | nindent 8 }}
+ initContainers:
+ {{ include "initContainerAssets" . | nindent 8 }}
+ {{ include "initContainerCache" . | nindent 8 }}
diff --git a/jobs/mongodb_migration/Makefile b/jobs/mongodb_migration/Makefile
index 71d907fd..d361e8f2 100644
--- a/jobs/mongodb_migration/Makefile
+++ b/jobs/mongodb_migration/Makefile
@@ -3 +3 @@ export COMPOSE_PROJECT_NAME := mongodb_migration
-export MONGO_PORT := 27060
+export MONGO_PORT := 27070
@@ -6 +6 @@ export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT}
-export MONGODB_MIGRATION_MONGO_URL := mongodb://localhost:${MONGO_PORT}
+export DATABASE_MIGRATIONS_MONGO_URL := mongodb://localhost:${MONGO_PORT}
diff --git a/jobs/mongodb_migration/README.md b/jobs/mongodb_migration/README.md
index ecac3b78..2e0539fe 100644
--- a/jobs/mongodb_migration/README.md
+++ b/jobs/mongodb_migration/README.md
@@ -11 +11 @@ The script can be configured using environment variables. They are grouped by sc
-Set environment variables to configure the job (`MONGODB_MIGRATION_` prefix):
+Set environment variables to configure the job (`DATABASE_MIGRATIONS_` prefix):
@@ -13,2 +13,2 @@ Set environment variables to configure the job (`MONGODB_MIGRATION_` prefix):
-- `MONGODB_MIGRATION_MONGO_DATABASE`: the name of the database used for storing the migrations history. Defaults to `"datasets_server_maintenance"`.
-- `MONGODB_MIGRATION_MONGO_URL`: the URL used to connect to the mongo db server. Defaults to `"mongodb://localhost:27017"`.
+- `DATABASE_MIGRATIONS_MONGO_DATABASE`: the name of the database used for storing the migrations history. Defaults to `"datasets_server_maintenance"`.
+- `DATABASE_MIGRATIONS_MONGO_URL`: the URL used to connect to the mongo db server. Defaults to `"mongodb://localhost:27017"`.
diff --git a/jobs/mongodb_migration/src/mongodb_migration/config.py b/jobs/mongodb_migration/src/mongodb_migration/config.py
index cbf68dca..98cfea0a 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/config.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/config.py
@@ -9,4 +9,2 @@ from libcommon.config import CacheConfig, CommonConfig, QueueConfig
-from mongodb_migration.database_migrations import connect_to_database
-
-MONGODB_MIGRATION_MONGO_DATABASE = "datasets_server_maintenance"
-MONGO_DATABASE_MONGO_URL = "mongodb://localhost:27017"
+DATABASE_MIGRATIONS_MONGO_DATABASE = "datasets_server_maintenance"
+DATABASE_MIGRATIONS_MONGO_URL = "mongodb://localhost:27017"
@@ -16,6 +14,3 @@ MONGO_DATABASE_MONGO_URL = "mongodb://localhost:27017"
-class MongodbMigrationConfig:
- mongo_database: str = MONGODB_MIGRATION_MONGO_DATABASE
- mongo_url: str = MONGO_DATABASE_MONGO_URL
-
- def __post_init__(self):
- connect_to_database(database=self.mongo_database, host=self.mongo_url)
+class DatabaseMigrationsConfig:
+ mongo_database: str = DATABASE_MIGRATIONS_MONGO_DATABASE
+ mongo_url: str = DATABASE_MIGRATIONS_MONGO_URL
@@ -24 +19 @@ class MongodbMigrationConfig:
- def from_env() -> "MongodbMigrationConfig":
+ def from_env() -> "DatabaseMigrationsConfig":
@@ -26,4 +21,4 @@ class MongodbMigrationConfig:
- with env.prefixed("MONGODB_MIGRATION_"):
- return MongodbMigrationConfig(
- mongo_database=env.str(name="MONGO_DATABASE", default=MONGODB_MIGRATION_MONGO_DATABASE),
- mongo_url=env.str(name="MONGO_URL", default=MONGO_DATABASE_MONGO_URL),
+ with env.prefixed("DATABASE_MIGRATIONS_"):
+ return DatabaseMigrationsConfig(
+ mongo_database=env.str(name="MONGO_DATABASE", default=DATABASE_MIGRATIONS_MONGO_DATABASE),
+ mongo_url=env.str(name="MONGO_URL", default=DATABASE_MIGRATIONS_MONGO_URL),
@@ -37 +32 @@ class JobConfig:
- mongodb_migration: MongodbMigrationConfig = field(default_factory=MongodbMigrationConfig)
+ database_migrations: DatabaseMigrationsConfig = field(default_factory=DatabaseMigrationsConfig)
@@ -45 +40 @@ class JobConfig:
- mongodb_migration=MongodbMigrationConfig.from_env(),
+ database_migrations=DatabaseMigrationsConfig.from_env(),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/constants.py b/jobs/mongodb_migration/src/mongodb_migration/constants.py
new file mode 100644
index 00000000..bf3b9f23
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/constants.py
@@ -0,0 +1,4 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+DATABASE_MIGRATIONS_MONGOENGINE_ALIAS = "maintenance"
diff --git a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py
index c6590d7d..3ea56734 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py
@@ -7 +7 @@ from typing import Generic, Type, TypeVar
-from mongoengine import Document, DoesNotExist, connect
+from mongoengine import Document, DoesNotExist
@@ -10,0 +11,2 @@ from mongoengine.queryset.queryset import QuerySet
+from mongodb_migration.constants import DATABASE_MIGRATIONS_MONGOENGINE_ALIAS
+
@@ -30,6 +31,0 @@ class QuerySetManager(Generic[U]):
-DATABASE_ALIAS = "maintenance"
-
-
-def connect_to_database(database: str, host: str) -> None:
- connect(db=database, alias=DATABASE_ALIAS, host=host)
-
@@ -47 +43 @@ class DatabaseMigration(Document):
- "db_alias": DATABASE_ALIAS,
+ "db_alias": DATABASE_MIGRATIONS_MONGOENGINE_ALIAS,
diff --git a/jobs/mongodb_migration/src/mongodb_migration/main.py b/jobs/mongodb_migration/src/mongodb_migration/main.py
index 58be3983..82b362e6 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/main.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/main.py
@@ -5,0 +6,3 @@ import sys
+from libcommon.log import init_logging
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+
@@ -8,0 +12 @@ from mongodb_migration.plan import Plan
+from mongodb_migration.resources import MigrationsMongoResource
@@ -12,6 +16,17 @@ if __name__ == "__main__":
- collected_migrations = MigrationsCollector().get_migrations()
- try:
- Plan(collected_migrations=collected_migrations).execute()
- sys.exit(0)
- except Exception:
- sys.exit(1)
+
+ init_logging(log_level=job_config.common.log_level)
+ # ^ set first to have logs as soon as possible
+
+ with (
+ CacheMongoResource(database=job_config.cache.mongo_database, host=job_config.cache.mongo_url),
+ QueueMongoResource(database=job_config.queue.mongo_database, host=job_config.queue.mongo_url),
+ MigrationsMongoResource(
+ database=job_config.database_migrations.mongo_database, host=job_config.database_migrations.mongo_url
+ ),
+ ):
+ collected_migrations = MigrationsCollector().get_migrations()
+ try:
+ Plan(collected_migrations=collected_migrations).execute()
+ sys.exit(0)
+ except Exception:
+ sys.exit(1)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/resources.py b/jobs/mongodb_migration/src/mongodb_migration/resources.py
new file mode 100644
index 00000000..0a210baa
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/resources.py
@@ -0,0 +1,21 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from dataclasses import dataclass, field
+
+from libcommon.resources import MongoResource
+
+from mongodb_migration.constants import DATABASE_MIGRATIONS_MONGOENGINE_ALIAS
+
+
+@dataclass
+class MigrationsMongoResource(MongoResource):
+ """
+ A resource that represents a connection to the migrations mongo database.
+
+ Args:
+ database (:obj:`str`): The name of the mongo database.
+ host (:obj:`str`): The host of the mongo database. It must start with ``mongodb://`` or ``mongodb+srv://``.
+ """
+
+ mongoengine_alias: str = field(default=DATABASE_MIGRATIONS_MONGOENGINE_ALIAS, init=False)
diff --git a/jobs/mongodb_migration/tests/conftest.py b/jobs/mongodb_migration/tests/conftest.py
index 6c5cf611..7312e3f5 100644
--- a/jobs/mongodb_migration/tests/conftest.py
+++ b/jobs/mongodb_migration/tests/conftest.py
@@ -4 +4,2 @@
-from pytest import MonkeyPatch, fixture
+from environs import Env
+from pytest import fixture
@@ -6 +6,0 @@ from pytest import MonkeyPatch, fixture
-from mongodb_migration.config import JobConfig
@@ -8,2 +7,0 @@ from mongodb_migration.config import JobConfig
-
-# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200
@@ -11,7 +9,2 @@ from mongodb_migration.config import JobConfig
-def monkeypatch_session():
- monkeypatch_session = MonkeyPatch()
- monkeypatch_session.setenv("CACHE_MONGO_DATABASE", "datasets_server_cache_test")
- monkeypatch_session.setenv("QUEUE_MONGO_DATABASE", "datasets_server_queue_test")
- monkeypatch_session.setenv("MONGODB_MIGRATION_MONGO_DATABASE", "datasets_server_maintenance_test")
- yield monkeypatch_session
- monkeypatch_session.undo()
+def env() -> Env:
+ return Env(expand_vars=True)
@@ -20,10 +13,6 @@ def monkeypatch_session():
-@fixture(scope="session", autouse=True)
-def app_config(monkeypatch_session: MonkeyPatch) -> JobConfig:
- job_config = JobConfig.from_env()
- if (
- "test" not in job_config.cache.mongo_database
- or "test" not in job_config.queue.mongo_database
- or "test" not in job_config.mongodb_migration.mongo_database
- ):
- raise ValueError("Test must be launched on a test mongo database")
- return job_config
+@fixture(scope="session")
+def mongo_host(env: Env) -> str:
+ try:
+ return env.str(name="DATABASE_MIGRATIONS_MONGO_URL")
+ except Exception as e:
+ raise ValueError("DATABASE_MIGRATIONS_MONGO_URL is not set") from e
diff --git a/jobs/mongodb_migration/tests/test_plan.py b/jobs/mongodb_migration/tests/test_plan.py
index f83b5df9..ea66efcb 100644
--- a/jobs/mongodb_migration/tests/test_plan.py
+++ b/jobs/mongodb_migration/tests/test_plan.py
@@ -4 +4 @@
-from typing import List, Optional, Type
+from typing import Iterator, List, Optional, Type
@@ -13,0 +14 @@ from mongodb_migration.plan import Plan, SavedMigrationsError
+from mongodb_migration.resources import MigrationsMongoResource
@@ -17,2 +18,7 @@ from mongodb_migration.plan import Plan, SavedMigrationsError
-def clean_mongo_database() -> None:
- _clean_maintenance_database()
+def migrations_mongo_resource(mongo_host: str) -> Iterator[MigrationsMongoResource]:
+ database = "datasets_server_migrations_test"
+ if "test" not in database:
+ raise ValueError("Test must be launched on a test mongo database")
+ with MigrationsMongoResource(database=database, host=mongo_host) as resource:
+ yield resource
+ _clean_maintenance_database()
diff --git a/jobs/mongodb_migration/tests/test_resources.py b/jobs/mongodb_migration/tests/test_resources.py
new file mode 100644
index 00000000..e8ff0523
--- /dev/null
+++ b/jobs/mongodb_migration/tests/test_resources.py
@@ -0,0 +1,20 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from mongoengine import Document
+from mongoengine.fields import StringField
+
+from mongodb_migration.resources import MigrationsMongoResource
+
+
+def test_cache_database(mongo_host: str) -> None:
+ resource = MigrationsMongoResource(database="test_migrations_database", host=mongo_host)
+
+ class User(Document):
+ name = StringField()
+ meta = {"db_alias": resource.mongoengine_alias}
+
+ assert len(User.objects()) == 0 # type: ignore
+ # clean
+ User.drop_collection() # type: ignore
+ resource.release()
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 89597b52..0a75c74d 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -11,5 +11 @@ from environs import Env
-from libcommon.log import init_logging
-from libcommon.processing_graph import ProcessingGraph, ProcessingGraphSpecification
-from libcommon.queue import connect_to_queue_database
-from libcommon.simple_cache import connect_to_cache_database
-from libcommon.storage import init_dir
+from libcommon.processing_graph import ProcessingGraphSpecification
@@ -24,4 +20 @@ class AssetsConfig:
- _storage_directory: Optional[str] = ASSETS_STORE_DIRECTORY
-
- def __post_init__(self):
- self.storage_directory = init_dir(directory=self._storage_directory, appname="datasets_server_assets")
+ storage_directory: Optional[str] = ASSETS_STORE_DIRECTORY
@@ -35 +28 @@ class AssetsConfig:
- _storage_directory=env.str(name="STORAGE_DIRECTORY", default=ASSETS_STORE_DIRECTORY),
+ storage_directory=env.str(name="STORAGE_DIRECTORY", default=ASSETS_STORE_DIRECTORY),
@@ -50,3 +42,0 @@ class CommonConfig:
- def __post_init__(self):
- init_logging(self.log_level)
-
@@ -73,3 +62,0 @@ class CacheConfig:
- def __post_init__(self):
- connect_to_cache_database(database=self.mongo_database, host=self.mongo_url)
-
@@ -97,3 +83,0 @@ class QueueConfig:
- def __post_init__(self):
- connect_to_queue_database(database=self.mongo_database, host=self.mongo_url)
-
@@ -126,3 +109,0 @@ class ProcessingGraphConfig:
- def __post_init__(self):
- self.graph = ProcessingGraph(self.specification)
-
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
new file mode 100644
index 00000000..7b707218
--- /dev/null
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -0,0 +1,6 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+ASSETS_CACHE_APPNAME = "datasets_server_assets"
+CACHE_MONGOENGINE_ALIAS = "cache"
+QUEUE_MONGOENGINE_ALIAS = "queue"
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index cf808f5b..f86e08bf 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -14 +14 @@ from typing import Dict, Generic, List, Literal, Optional, Type, TypedDict, Type
-from mongoengine import Document, DoesNotExist, connect
+from mongoengine import Document, DoesNotExist
@@ -17,0 +18,2 @@ from mongoengine.queryset.queryset import QuerySet
+from libcommon.constants import QUEUE_MONGOENGINE_ALIAS
+
@@ -99,4 +100,0 @@ def get_datetime() -> datetime:
-def connect_to_queue_database(database: str, host: str) -> None:
- connect(db=database, alias="queue", host=host)
-
-
@@ -130 +128 @@ class Job(Document):
- "db_alias": "queue",
+ "db_alias": QUEUE_MONGOENGINE_ALIAS,
diff --git a/libs/libcommon/src/libcommon/resources.py b/libs/libcommon/src/libcommon/resources.py
new file mode 100644
index 00000000..2eeeece4
--- /dev/null
+++ b/libs/libcommon/src/libcommon/resources.py
@@ -0,0 +1,103 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from dataclasses import dataclass, field
+
+from mongoengine.connection import ConnectionFailure, connect, disconnect
+
+from libcommon.constants import CACHE_MONGOENGINE_ALIAS, QUEUE_MONGOENGINE_ALIAS
+
+
+@dataclass
+class Resource:
+ """
+ A resource that can be allocated and released.
+
+ The method allocate() is called when the resource is created. The method release() allows to free the resource.
+
+ It can be used as a context manager, in which case the resource is released when the context is exited.
+
+ Example:
+ >>> with Resource() as resource:
+ ... pass
+
+ Resources should be inherited from this class and implement the allocate() and release() methods.
+ """
+
+ def __post_init__(self):
+ self.allocate()
+
+ def __enter__(self):
+ return self
+
+ def __exit__(self, exc_type, exc_value, traceback):
+ self.release()
+
+ def allocate(self):
+ pass
+
+ def release(self):
+ pass
+
+
+class MongoConnectionFailure(Exception):
+ pass
+
+
+@dataclass
+class MongoResource(Resource):
+ """
+ A base resource that represents a connection to a database.
+
+ Args:
+ database (:obj:`str`): The name of the mongo database.
+ host (:obj:`str`): The host of the mongo database. It must start with ``mongodb://`` or ``mongodb+srv://``.
+ mongoengine_alias (:obj:`str`): The alias of the connection in mongoengine.
+ server_selection_timeout_ms (:obj:`int`, `optional`, defaults to 30_000): The timeout in milliseconds for
+ server selection.
+ """
+
+ database: str
+ host: str
+ mongoengine_alias: str
+ server_selection_timeout_ms: int = 30_000
+
+ def allocate(self):
+ try:
+ connect(
+ db=self.database,
+ host=self.host,
+ alias=self.mongoengine_alias,
+ serverSelectionTimeoutMS=self.server_selection_timeout_ms,
+ )
+ except ConnectionFailure as e:
+ raise MongoConnectionFailure(f"Failed to connect to MongoDB: {e}") from e
+
+ def release(self):
+ disconnect(alias=self.mongoengine_alias)
+
+
+@dataclass
+class CacheMongoResource(MongoResource):
+ """
+ A resource that represents a connection to the cache mongo database.
+
+ Args:
+ database (:obj:`str`): The name of the mongo database.
+ host (:obj:`str`): The host of the mongo database. It must start with ``mongodb://`` or ``mongodb+srv://``.
+ """
+
+ mongoengine_alias: str = field(default=CACHE_MONGOENGINE_ALIAS, init=False)
+
+
+@dataclass
+class QueueMongoResource(MongoResource):
+ """
+ A resource that represents a connection to the queue mongo database.
+
+ Args:
+ database (:obj:`str`): The name of the mongo database.
+ host (:obj:`str`): The host of the mongo database. It must start with ``mongodb://`` or ``mongodb+srv://``.
+ """
+
+ mongoengine_alias: str = field(default=QUEUE_MONGOENGINE_ALIAS, init=False)
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index f103e286..b9fb1b0f 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -22 +22 @@ from bson.errors import InvalidId
-from mongoengine import Document, DoesNotExist, connect
+from mongoengine import Document, DoesNotExist
@@ -31,0 +32,2 @@ from mongoengine.queryset.queryset import QuerySet
+from libcommon.constants import CACHE_MONGOENGINE_ALIAS
+
@@ -52,4 +53,0 @@ class QuerySetManager(Generic[U]):
-def connect_to_cache_database(database: str, host: str) -> None:
- connect(database, alias="cache", host=host)
-
-
@@ -104 +102 @@ class CachedResponse(Document):
- "db_alias": "cache",
+ "db_alias": CACHE_MONGOENGINE_ALIAS,
diff --git a/libs/libcommon/src/libcommon/storage.py b/libs/libcommon/src/libcommon/storage.py
index 555f9b24..a6dba310 100644
--- a/libs/libcommon/src/libcommon/storage.py
+++ b/libs/libcommon/src/libcommon/storage.py
@@ -10,0 +11,2 @@ from appdirs import user_cache_dir # type:ignore
+from libcommon.constants import ASSETS_CACHE_APPNAME
+
@@ -34,0 +37,14 @@ def init_dir(directory: Optional[StrPath] = None, appname: Optional[str] = None)
+def init_assets_dir(directory: Optional[StrPath] = None) -> StrPath:
+ """Initialize the assets directory.
+
+ If directory is None, it will be set to the default cache location on the machine.
+
+ Args:
+ directory (Optional[Union[str, PathLike[str]]], optional): The directory to initialize. Defaults to None.
+
+ Returns:
+ Union[str, PathLike[str]]: The directory.
+ """
+ return init_dir(directory, appname=ASSETS_CACHE_APPNAME)
+
+
diff --git a/libs/libcommon/tests/conftest.py b/libs/libcommon/tests/conftest.py
index 154c57e8..0226bc1f 100644
--- a/libs/libcommon/tests/conftest.py
+++ b/libs/libcommon/tests/conftest.py
@@ -7 +6,0 @@ from pytest import fixture
-from libcommon.config import CacheConfig, QueueConfig
@@ -9,2 +8,3 @@ from libcommon.config import CacheConfig, QueueConfig
-# CACHE_MONGO_URL and QUEUE_MONGO_URL must be set in the environment, and correspond to a running mongo database
-env = Env(expand_vars=True)
+@fixture(scope="session")
+def env() -> Env:
+ return Env(expand_vars=True)
@@ -13,6 +13,6 @@ env = Env(expand_vars=True)
-@fixture()
-def cache_config() -> CacheConfig:
- cache_config = CacheConfig(mongo_database="datasets_server_cache_test", mongo_url=env.str("CACHE_MONGO_URL"))
- if "test" not in cache_config.mongo_database:
- raise ValueError("Test must be launched on a test mongo database")
- return cache_config
+@fixture(scope="session")
+def cache_mongo_host(env: Env) -> str:
+ try:
+ return env.str(name="CACHE_MONGO_URL")
+ except Exception as e:
+ raise ValueError("CACHE_MONGO_URL is not set") from e
@@ -21,6 +21,6 @@ def cache_config() -> CacheConfig:
-@fixture()
-def queue_config() -> QueueConfig:
- queue_config = QueueConfig(mongo_database="datasets_server_queue_test", mongo_url=env.str("QUEUE_MONGO_URL"))
- if "test" not in queue_config.mongo_database:
- raise ValueError("Test must be launched on a test mongo database")
- return queue_config
+@fixture(scope="session")
+def queue_mongo_host(env: Env) -> str:
+ try:
+ return env.str(name="QUEUE_MONGO_URL")
+ except Exception as e:
+ raise ValueError("QUEUE_MONGO_URL is not set") from e
diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py
index 19b61ebe..5a3d5192 100644
--- a/libs/libcommon/tests/test_processing_steps.py
+++ b/libs/libcommon/tests/test_processing_steps.py
@@ -4,0 +5 @@ from libcommon.config import ProcessingGraphConfig
+from libcommon.processing_graph import ProcessingGraph
@@ -9 +10 @@ def test_default_graph():
- graph = config.graph
+ graph = ProcessingGraph(config.specification)
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index b3ab02d2..285bc086 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -5 +5 @@ import time
-from typing import Optional
+from typing import Iterator, Optional
@@ -9 +8,0 @@ import pytest
-from libcommon.config import QueueConfig
@@ -16,0 +16 @@ from libcommon.queue import (
+from libcommon.resources import QueueMongoResource
@@ -20,2 +20,8 @@ from libcommon.queue import (
-def clean_mongo_database(queue_config: QueueConfig) -> None:
- _clean_queue_database()
+def queue_mongo_resource(queue_mongo_host: str) -> Iterator[QueueMongoResource]:
+ database = "datasets_server_queue_test"
+ host = queue_mongo_host
+ if "test" not in database:
+ raise ValueError("Test must be launched on a test mongo database")
+ with QueueMongoResource(database=database, host=host) as queue_mongo_resource:
+ yield queue_mongo_resource
+ _clean_queue_database()
diff --git a/libs/libcommon/tests/test_resources.py b/libs/libcommon/tests/test_resources.py
new file mode 100644
index 00000000..98679570
--- /dev/null
+++ b/libs/libcommon/tests/test_resources.py
@@ -0,0 +1,111 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from typing import Optional
+
+import pytest
+from mongoengine import Document
+from mongoengine.fields import StringField
+from pymongo.errors import ServerSelectionTimeoutError
+
+from libcommon.resources import (
+ CacheMongoResource,
+ MongoConnectionFailure,
+ MongoResource,
+ QueueMongoResource,
+)
+
+
+def test_database_resource(queue_mongo_host: str) -> None:
+ database_1 = "datasets_server_1"
+ database_2 = "datasets_server_2"
+ host = queue_mongo_host
+ mongoengine_alias = "datasets_server_mongo_alias"
+ server_selection_timeout_ms = 1_000
+ resource_1 = MongoResource(
+ database=database_1,
+ host=host,
+ mongoengine_alias=mongoengine_alias,
+ server_selection_timeout_ms=server_selection_timeout_ms,
+ )
+ with pytest.raises(MongoConnectionFailure):
+ MongoResource(
+ database=database_2,
+ host=host,
+ mongoengine_alias=mongoengine_alias,
+ server_selection_timeout_ms=server_selection_timeout_ms,
+ )
+ resource_1.release()
+ resource_2 = MongoResource(
+ database=database_2,
+ host=host,
+ mongoengine_alias=mongoengine_alias,
+ server_selection_timeout_ms=server_selection_timeout_ms,
+ )
+ resource_2.release()
+
+
[email protected](
+ "host,mongoengine_alias,server_selection_timeout_ms,raises",
+ [
+ (None, "test_timeout_error", 5_000, False),
+ ("mongodb://doesnotexist:123", "test_host_error", 5_000, True),
+ ],
+)
+def test_database_resource_errors(
+ queue_mongo_host: str,
+ host: Optional[str],
+ mongoengine_alias: str,
+ server_selection_timeout_ms: int,
+ raises: bool,
+) -> None:
+ if not host:
+ host = queue_mongo_host
+ database = "datasets_server_test"
+ resource = MongoResource(
+ database=database,
+ host=host,
+ mongoengine_alias=mongoengine_alias,
+ server_selection_timeout_ms=server_selection_timeout_ms,
+ )
+ # ^ this does not raise any issue, as it "only" registers the connection
+
+ class User(Document):
+ name = StringField()
+ meta = {"db_alias": mongoengine_alias}
+
+ if raises:
+ with pytest.raises(ServerSelectionTimeoutError):
+ len(User.objects()) # type: ignore
+ else:
+ assert len(User.objects()) == 0 # type: ignore
+ # clean
+ User.drop_collection() # type: ignore
+
+ resource.release()
+
+
+def test_cache_database(cache_mongo_host: str) -> None:
+ resource = CacheMongoResource(database="test_cache_database", host=cache_mongo_host)
+
+ class User(Document):
+ name = StringField()
+ meta = {"db_alias": resource.mongoengine_alias}
+
+ assert len(User.objects()) == 0 # type: ignore
+ # clean
+ User.drop_collection() # type: ignore
+ resource.release()
+
+
+def test_queue_database(queue_mongo_host: str) -> None:
+ resource = QueueMongoResource(database="test_queue_database", host=queue_mongo_host)
+
+ class User(Document):
+ name = StringField()
+ meta = {"db_alias": resource.mongoengine_alias}
+
+ assert len(User.objects()) == 0 # type: ignore
+ # clean
+ User.drop_collection() # type: ignore
+ resource.release()
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index 71bf609d..2db35861 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -7 +7 @@ from time import process_time
-from typing import Optional
+from typing import Iterator, Optional
@@ -12 +12 @@ from pymongo.errors import DocumentTooLarge
-from libcommon.config import CacheConfig
+from libcommon.resources import CacheMongoResource
@@ -35,2 +35,8 @@ from libcommon.simple_cache import (
-def clean_mongo_database(cache_config: CacheConfig) -> None:
- _clean_cache_database()
+def cache_mongo_resource(cache_mongo_host: str) -> Iterator[CacheMongoResource]:
+ database = "datasets_server_cache_test"
+ host = cache_mongo_host
+ if "test" not in database:
+ raise ValueError("Test must be launched on a test mongo database")
+ with CacheMongoResource(database=database, host=host) as cache_mongo_resource:
+ yield cache_mongo_resource
+ _clean_cache_database()
diff --git a/libs/libcommon/tests/test_storage.py b/libs/libcommon/tests/test_storage.py
index 728a9650..a0307134 100644
--- a/libs/libcommon/tests/test_storage.py
+++ b/libs/libcommon/tests/test_storage.py
@@ -9 +9,2 @@ import pytest
-from libcommon.storage import StrPath, init_dir, remove_dir
+from libcommon.constants import ASSETS_CACHE_APPNAME
+from libcommon.storage import StrPath, init_assets_dir, init_dir, remove_dir
@@ -32 +33,8 @@ def test_init_dir(
- if has_directory is None:
+ if has_directory:
+ directory = str(tmp_path) if is_directory_string else tmp_path
+ result = init_dir(directory=directory, appname=appname)
+ assert result == directory
+ assert subdirectory in str(result), result
+ if appname is not None:
+ assert appname not in str(result), result
+ else:
@@ -38,5 +46,26 @@ def test_init_dir(
- if has_appname:
- assert appname in str(result) is appname, result
- else:
- assert appname is None
- else:
+ if appname:
+ assert appname in str(result), result
+ Path(result).exists()
+ Path(result).is_dir()
+
+
[email protected](
+ "has_directory,is_directory_string",
+ [
+ (False, False),
+ (False, False),
+ (False, True),
+ (False, True),
+ (True, False),
+ (True, False),
+ (True, True),
+ (True, True),
+ ],
+)
+def test_init_assets_dir(
+ tmp_path_factory: pytest.TempPathFactory, has_directory: bool, is_directory_string: bool
+) -> None:
+ subdirectory = "subdirectory"
+ tmp_path = tmp_path_factory.mktemp("test") / subdirectory
+ directory: Optional[StrPath]
+ if has_directory:
@@ -44 +73 @@ def test_init_dir(
- result = init_dir(directory=directory, appname=appname)
+ result = init_assets_dir(directory=directory)
@@ -47,2 +76,9 @@ def test_init_dir(
- if appname is not None:
- assert appname not in str(result), result
+ assert ASSETS_CACHE_APPNAME not in str(result), result
+ else:
+ directory = None
+ result = init_assets_dir(directory=directory)
+ assert result != directory, result
+ assert subdirectory not in str(result), result
+ assert type(result) is str, result
+ assert ASSETS_CACHE_APPNAME in str(result), result
+
diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py
index c9788837..6c7cd4d5 100644
--- a/services/admin/src/admin/app.py
+++ b/services/admin/src/admin/app.py
@@ -4,0 +5,4 @@ import uvicorn # type: ignore
+from libcommon.log import init_logging
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource
+from libcommon.storage import init_assets_dir
@@ -28,4 +32,15 @@ def create_app() -> Starlette:
- processing_steps = list(app_config.processing_graph.graph.steps.values())
- prometheus = Prometheus(
- processing_steps=processing_steps, assets_storage_directory=app_config.assets.storage_directory
- )
+
+ init_logging(log_level=app_config.common.log_level)
+ # ^ set first to have logs as soon as possible
+ assets_directory = init_assets_dir(directory=app_config.assets.storage_directory)
+
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+ processing_steps = list(processing_graph.steps.values())
+ init_processing_steps = processing_graph.get_first_steps()
+
+ resources: list[Resource] = [
+ CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url),
+ QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url),
+ ]
+
+ prometheus = Prometheus(processing_steps=processing_steps, assets_directory=assets_directory)
@@ -50 +65 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.admin.external_auth_url,
@@ -62 +77 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.admin.external_auth_url,
@@ -73 +88 @@ def create_app() -> Starlette:
- init_processing_steps=app_config.processing_graph.graph.get_first_steps(),
+ init_processing_steps=init_processing_steps,
@@ -76 +91 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.admin.external_auth_url,
@@ -89 +104 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.admin.external_auth_url,
@@ -102 +117 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.admin.external_auth_url,
@@ -113 +128 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.admin.external_auth_url,
@@ -126 +141 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.admin.external_auth_url,
@@ -133 +148,2 @@ def create_app() -> Starlette:
- return Starlette(routes=routes, middleware=middleware)
+
+ return Starlette(routes=routes, middleware=middleware, on_shutdown=[resource.release for resource in resources])
diff --git a/services/admin/src/admin/config.py b/services/admin/src/admin/config.py
index 1d4f271a..d1a226e1 100644
--- a/services/admin/src/admin/config.py
+++ b/services/admin/src/admin/config.py
@@ -39,0 +40 @@ ADMIN_CACHE_REPORTS_WITH_CONTENT_NUM_RESULTS = 100
+ADMIN_EXTERNAL_AUTH_URL = None
@@ -48,0 +50 @@ class AdminConfig:
+ external_auth_url: Optional[str] = ADMIN_EXTERNAL_AUTH_URL # not documented
@@ -54 +56 @@ class AdminConfig:
- def from_env() -> "AdminConfig":
+ def from_env(common_config: CommonConfig) -> "AdminConfig":
@@ -56,0 +59,2 @@ class AdminConfig:
+ hf_whoami_path = env.str(name="HF_WHOAMI_PATH", default=ADMIN_HF_WHOAMI_PATH)
+ external_auth_url = None if hf_whoami_path is None else f"{common_config.hf_endpoint}{hf_whoami_path}"
@@ -63,0 +68 @@ class AdminConfig:
+ external_auth_url=external_auth_url,
@@ -65 +70 @@ class AdminConfig:
- hf_whoami_path=env.str(name="HF_WHOAMI_PATH", default=ADMIN_HF_WHOAMI_PATH),
+ hf_whoami_path=hf_whoami_path,
@@ -79,5 +83,0 @@ class AppConfig:
- def __post_init__(self):
- self.external_auth_url = (
- None if self.admin.hf_whoami_path is None else f"{self.common.hf_endpoint}{self.admin.hf_whoami_path}"
- )
-
@@ -86 +86 @@ class AppConfig:
- # First process the common configuration to setup the logging
+ common_config = CommonConfig.from_env()
@@ -88 +88 @@ class AppConfig:
- common=CommonConfig.from_env(),
+ common=common_config,
@@ -93 +93 @@ class AppConfig:
- admin=AdminConfig.from_env(),
+ admin=AdminConfig.from_env(common_config),
diff --git a/services/admin/src/admin/prometheus.py b/services/admin/src/admin/prometheus.py
index b8913016..fc5221a4 100644
--- a/services/admin/src/admin/prometheus.py
+++ b/services/admin/src/admin/prometheus.py
@@ -10,0 +11 @@ from libcommon.simple_cache import get_responses_count_by_kind_status_and_error_
+from libcommon.storage import StrPath
@@ -49 +50 @@ class Prometheus:
- assets_storage_directory: str
+ assets_directory: StrPath
@@ -72 +73 @@ class Prometheus:
- total, used, free, percent = disk_usage(self.assets_storage_directory)
+ total, used, free, percent = disk_usage(str(self.assets_directory))
diff --git a/services/admin/tests/conftest.py b/services/admin/tests/conftest.py
index d515e78d..ccafee55 100644
--- a/services/admin/tests/conftest.py
+++ b/services/admin/tests/conftest.py
@@ -4 +4 @@
-from typing import List
+from typing import Iterator, List
@@ -6 +6,5 @@ from typing import List
-from libcommon.processing_graph import ProcessingStep
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
+from libcommon.queue import _clean_queue_database
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import _clean_cache_database
+from libcommon.storage import StrPath, init_assets_dir
@@ -37 +41,21 @@ def processing_steps(app_config: AppConfig) -> List[ProcessingStep]:
- return list(app_config.processing_graph.graph.steps.values())
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+ return list(processing_graph.steps.values())
+
+
+@fixture(scope="session")
+def assets_directory(app_config: AppConfig) -> StrPath:
+ return init_assets_dir(directory=app_config.assets.storage_directory)
+
+
+@fixture(autouse=True)
+def cache_mongo_resource(app_config: AppConfig) -> Iterator[CacheMongoResource]:
+ with CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url) as resource:
+ yield resource
+ _clean_cache_database()
+
+
+@fixture(autouse=True)
+def queue_mongo_resource(app_config: AppConfig) -> Iterator[QueueMongoResource]:
+ with QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url) as resource:
+ yield resource
+ _clean_queue_database()
diff --git a/services/admin/tests/test_app.py b/services/admin/tests/test_app.py
index 9eb72b76..ca5c7e23 100644
--- a/services/admin/tests/test_app.py
+++ b/services/admin/tests/test_app.py
@@ -8,2 +7,0 @@ from libcommon.processing_graph import ProcessingStep
-from libcommon.queue import _clean_queue_database
-from libcommon.simple_cache import _clean_cache_database
@@ -13 +10,0 @@ from admin.app import create_app
-from admin.config import AppConfig
@@ -21,6 +17,0 @@ def client(monkeypatch_session: pytest.MonkeyPatch) -> TestClient:
[email protected](autouse=True)
-def clean_mongo_databases(app_config: AppConfig) -> None:
- _clean_cache_database()
- _clean_queue_database()
-
-
diff --git a/services/admin/tests/test_app_real.py b/services/admin/tests/test_app_real.py
index 06d860a5..a3183019 100644
--- a/services/admin/tests/test_app_real.py
+++ b/services/admin/tests/test_app_real.py
@@ -4,2 +4 @@
-from libcommon.queue import _clean_queue_database
-from libcommon.simple_cache import _clean_cache_database
+from libcommon.processing_graph import ProcessingGraph
@@ -40,6 +38,0 @@ def real_app_config(real_monkeypatch: MonkeyPatch) -> AppConfig:
-@fixture(autouse=True)
-def real_clean_mongo_databases(real_app_config: AppConfig) -> None:
- _clean_cache_database()
- _clean_queue_database()
-
-
@@ -52 +45,2 @@ def test_force_refresh(
- path = next(iter(real_app_config.processing_graph.graph.steps.values())).endpoint
+ processing_graph = ProcessingGraph(real_app_config.processing_graph.specification)
+ path = next(iter(processing_graph.steps.values())).endpoint
diff --git a/services/admin/tests/test_prometheus.py b/services/admin/tests/test_prometheus.py
index 44252eec..f417fb73 100644
--- a/services/admin/tests/test_prometheus.py
+++ b/services/admin/tests/test_prometheus.py
@@ -0,0 +1,3 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
@@ -4,0 +8 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.storage import StrPath
@@ -6 +9,0 @@ from libcommon.processing_graph import ProcessingStep
-from admin.config import AppConfig
@@ -10,2 +13,4 @@ from admin.prometheus import Prometheus
-def test_prometheus(app_config: AppConfig, processing_steps: List[ProcessingStep]) -> None:
- # we depend on app_config to be sure we already connected to the database
+def test_prometheus(
+ assets_directory: StrPath,
+ processing_steps: List[ProcessingStep],
+) -> None:
@@ -14,3 +19 @@ def test_prometheus(app_config: AppConfig, processing_steps: List[ProcessingStep
- prometheus = Prometheus(
- processing_steps=processing_steps, assets_storage_directory=app_config.assets.storage_directory
- )
+ prometheus = Prometheus(processing_steps=processing_steps, assets_directory=assets_directory)
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 6ada8330..26e66a2b 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -6,0 +7,3 @@ import uvicorn # type: ignore
+from libcommon.log import init_logging
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource
@@ -23,0 +27,4 @@ def create_app() -> Starlette:
+
+ init_logging(log_level=app_config.common.log_level)
+ # ^ set first to have logs as soon as possible
+
@@ -25,0 +33,5 @@ def create_app() -> Starlette:
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+ processing_steps = list(processing_graph.steps.values())
+ processing_steps_required_by_dataset_viewer = processing_graph.get_steps_required_by_dataset_viewer()
+ init_processing_steps = processing_graph.get_first_steps()
+
@@ -32,0 +45,6 @@ def create_app() -> Starlette:
+
+ resources: list[Resource] = [
+ CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url),
+ QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url),
+ ]
+
@@ -37 +55 @@ def create_app() -> Starlette:
- processing_steps_for_valid=app_config.processing_graph.graph.get_steps_required_by_dataset_viewer(),
+ processing_steps_for_valid=processing_steps_required_by_dataset_viewer,
@@ -45,2 +63,2 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
- processing_steps_for_valid=app_config.processing_graph.graph.get_steps_required_by_dataset_viewer(),
+ external_auth_url=app_config.api.external_auth_url,
+ processing_steps_for_valid=processing_steps_required_by_dataset_viewer,
@@ -53 +71 @@ def create_app() -> Starlette:
- processing_steps: List[BaseRoute] = [
+ processing_step_endpoints: List[BaseRoute] = [
@@ -58 +76 @@ def create_app() -> Starlette:
- init_processing_steps=app_config.processing_graph.graph.get_first_steps(),
+ init_processing_steps=init_processing_steps,
@@ -61 +79 @@ def create_app() -> Starlette:
- external_auth_url=app_config.external_auth_url,
+ external_auth_url=app_config.api.external_auth_url,
@@ -66 +84 @@ def create_app() -> Starlette:
- for processing_step in list(app_config.processing_graph.graph.steps.values())
+ for processing_step in processing_steps
@@ -73 +91 @@ def create_app() -> Starlette:
- init_processing_steps=app_config.processing_graph.graph.get_first_steps(),
+ init_processing_steps=init_processing_steps,
@@ -85,2 +103,3 @@ def create_app() -> Starlette:
- routes: List[BaseRoute] = valid + processing_steps + to_protect + protected
- return Starlette(routes=routes, middleware=middleware)
+ routes: List[BaseRoute] = valid + processing_step_endpoints + to_protect + protected
+
+ return Starlette(routes=routes, middleware=middleware, on_shutdown=[resource.release for resource in resources])
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 80ef9a99..f1381415 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -4,0 +5 @@ from dataclasses import dataclass, field
+from typing import Optional
@@ -35,0 +37 @@ class UvicornConfig:
+API_EXTERNAL_AUTH_URL = None
@@ -42,0 +45 @@ class ApiConfig:
+ external_auth_url: Optional[str] = API_EXTERNAL_AUTH_URL # not documented
@@ -48 +51 @@ class ApiConfig:
- def from_env() -> "ApiConfig":
+ def from_env(common_config: CommonConfig) -> "ApiConfig":
@@ -50,0 +54,2 @@ class ApiConfig:
+ hf_auth_path = env.str(name="HF_AUTH_PATH", default=API_HF_AUTH_PATH)
+ external_auth_url = None if hf_auth_path is None else f"{common_config.hf_endpoint}{hf_auth_path}"
@@ -52 +57,2 @@ class ApiConfig:
- hf_auth_path=env.str(name="HF_AUTH_PATH", default=API_HF_AUTH_PATH),
+ external_auth_url=external_auth_url,
+ hf_auth_path=hf_auth_path,
@@ -66,5 +71,0 @@ class AppConfig:
- def __post_init__(self):
- self.external_auth_url = (
- None if self.api.hf_auth_path is None else f"{self.common.hf_endpoint}{self.api.hf_auth_path}"
- )
-
@@ -73 +74 @@ class AppConfig:
- # First process the common configuration to setup the logging
+ common_config = CommonConfig.from_env()
@@ -75 +76 @@ class AppConfig:
- common=CommonConfig.from_env(),
+ common=common_config,
@@ -79 +80 @@ class AppConfig:
- api=ApiConfig.from_env(),
+ api=ApiConfig.from_env(common_config=common_config),
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index 63893ddc..d0afa6f2 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -3,0 +4,6 @@
+from typing import Iterator, List
+
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
+from libcommon.queue import _clean_queue_database
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import _clean_cache_database
@@ -33,2 +39,3 @@ def app_config(monkeypatch_session: MonkeyPatch) -> AppConfig:
-def uvicorn_config(monkeypatch_session: MonkeyPatch):
- return UvicornConfig.from_env()
+def processing_steps(app_config: AppConfig) -> List[ProcessingStep]:
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+ return list(processing_graph.steps.values())
@@ -38,2 +45,2 @@ def uvicorn_config(monkeypatch_session: MonkeyPatch):
-def httpserver_listen_address(uvicorn_config: UvicornConfig):
- return (uvicorn_config.hostname, uvicorn_config.port)
+def first_dataset_processing_step(processing_steps: List[ProcessingStep]):
+ return next(step for step in processing_steps if step.input_type == "dataset")
@@ -43,2 +50,2 @@ def httpserver_listen_address(uvicorn_config: UvicornConfig):
-def hf_endpoint(app_config: AppConfig):
- return app_config.common.hf_endpoint
+def first_config_processing_step(processing_steps: List[ProcessingStep]):
+ return next(step for step in processing_steps if step.input_type == "config")
@@ -48,2 +55,21 @@ def hf_endpoint(app_config: AppConfig):
-def hf_auth_path(app_config: AppConfig):
- return app_config.api.hf_auth_path
+def first_split_processing_step(processing_steps: List[ProcessingStep]):
+ return next(step for step in processing_steps if step.input_type == "split")
+
+
+@fixture(autouse=True)
+def cache_mongo_resource(app_config: AppConfig) -> Iterator[CacheMongoResource]:
+ with CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url) as resource:
+ yield resource
+ _clean_cache_database()
+
+
+@fixture(autouse=True)
+def queue_mongo_resource(app_config: AppConfig) -> Iterator[QueueMongoResource]:
+ with QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url) as resource:
+ yield resource
+ _clean_queue_database()
+
+
+@fixture(scope="session")
+def uvicorn_config(monkeypatch_session: MonkeyPatch):
+ return UvicornConfig.from_env()
@@ -53,2 +79,2 @@ def hf_auth_path(app_config: AppConfig):
-def first_dataset_processing_step(app_config: AppConfig):
- return next(step for step in app_config.processing_graph.graph.steps.values() if step.input_type == "dataset")
+def httpserver_listen_address(uvicorn_config: UvicornConfig):
+ return (uvicorn_config.hostname, uvicorn_config.port)
@@ -58,2 +84,2 @@ def first_dataset_processing_step(app_config: AppConfig):
-def first_config_processing_step(app_config: AppConfig):
- return next(step for step in app_config.processing_graph.graph.steps.values() if step.input_type == "config")
+def hf_endpoint(app_config: AppConfig):
+ return app_config.common.hf_endpoint
@@ -63,2 +89,2 @@ def first_config_processing_step(app_config: AppConfig):
-def first_split_processing_step(app_config: AppConfig):
- return next(step for step in app_config.processing_graph.graph.steps.values() if step.input_type == "split")
+def hf_auth_path(app_config: AppConfig):
+ return app_config.api.hf_auth_path
diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py
index 0ff1f0c1..26feb3c0 100644
--- a/services/api/tests/test_app.py
+++ b/services/api/tests/test_app.py
@@ -8,2 +7,0 @@ from libcommon.processing_graph import ProcessingStep
-from libcommon.queue import _clean_queue_database
-from libcommon.simple_cache import _clean_cache_database
@@ -14 +11,0 @@ from api.app import create_app
-from api.config import AppConfig
@@ -24,6 +20,0 @@ def client(monkeypatch_session: pytest.MonkeyPatch) -> TestClient:
[email protected](autouse=True)
-def clean_mongo_databases(app_config: AppConfig) -> None:
- _clean_cache_database()
- _clean_queue_database()
-
-
diff --git a/services/api/tests/test_app_real.py b/services/api/tests/test_app_real.py
index 721de040..59f90255 100644
--- a/services/api/tests/test_app_real.py
+++ b/services/api/tests/test_app_real.py
@@ -5,2 +4,0 @@
-from libcommon.queue import _clean_queue_database
-from libcommon.simple_cache import _clean_cache_database
@@ -41,6 +38,0 @@ def real_app_config(real_monkeypatch: MonkeyPatch) -> AppConfig:
-@fixture(autouse=True)
-def real_clean_mongo_databases(real_app_config: AppConfig) -> None:
- _clean_cache_database()
- _clean_queue_database()
-
-
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index c668ada2..3bbca9a0 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -78,0 +79 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -95,0 +97 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -112,0 +115 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -156,0 +160 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -180,0 +185,2 @@ services:
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -193,0 +200,2 @@ services:
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -206,0 +215,2 @@ services:
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index fca77917..33384dab 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -78,0 +79 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -95,0 +97 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -112,0 +115 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -144,0 +148 @@ services:
+ FIRST_ROWS_COLUMNS_MAX_NUMBER: ${FIRST_ROWS_COLUMNS_MAX_NUMBER-1_000}
@@ -155,0 +160 @@ services:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -179,0 +185,2 @@ services:
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -192,0 +200,2 @@ services:
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
@@ -205,0 +215,2 @@ services:
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
diff --git a/workers/datasets_based/src/datasets_based/asset.py b/workers/datasets_based/src/datasets_based/asset.py
index dc652da0..b6e65ba2 100644
--- a/workers/datasets_based/src/datasets_based/asset.py
+++ b/workers/datasets_based/src/datasets_based/asset.py
@@ -8,0 +9 @@ import soundfile # type:ignore
+from libcommon.storage import StrPath
@@ -18 +19 @@ def create_asset_dir(
- dataset: str, config: str, split: str, row_idx: int, column: str, assets_directory: str
+ dataset: str, config: str, split: str, row_idx: int, column: str, assets_directory: StrPath
@@ -41 +42 @@ def create_image_file(
- assets_directory: str,
+ assets_directory: StrPath,
@@ -70 +71 @@ def create_audio_files(
- assets_directory: str,
+ assets_directory: StrPath,
diff --git a/workers/datasets_based/src/datasets_based/config.py b/workers/datasets_based/src/datasets_based/config.py
index 61f7b60a..3bbf8e05 100644
--- a/workers/datasets_based/src/datasets_based/config.py
+++ b/workers/datasets_based/src/datasets_based/config.py
@@ -5,2 +5 @@ from dataclasses import dataclass, field
-from pathlib import Path
-from typing import List, Optional, Union
+from typing import List, Optional
@@ -8,2 +6,0 @@ from typing import List, Optional, Union
-import datasets.config
-from datasets.utils.logging import log_levels, set_verbosity
@@ -58 +55 @@ class DatasetsBasedConfig:
- _hf_datasets_cache: Union[str, Path, None] = DATASETS_BASED_HF_DATASETS_CACHE
+ hf_datasets_cache: Optional[str] = DATASETS_BASED_HF_DATASETS_CACHE
@@ -61,5 +57,0 @@ class DatasetsBasedConfig:
- def __post_init__(self) -> None:
- self.hf_datasets_cache = (
- datasets.config.HF_DATASETS_CACHE if self._hf_datasets_cache is None else Path(self._hf_datasets_cache)
- )
-
@@ -72 +64 @@ class DatasetsBasedConfig:
- _hf_datasets_cache=env.str(name="HF_DATASETS_CACHE", default=DATASETS_BASED_HF_DATASETS_CACHE),
+ hf_datasets_cache=env.str(name="HF_DATASETS_CACHE", default=DATASETS_BASED_HF_DATASETS_CACHE),
@@ -86 +77,0 @@ class FirstRowsConfig:
- assets: AssetsConfig = field(default_factory=AssetsConfig)
@@ -98 +88,0 @@ class FirstRowsConfig:
- assets=AssetsConfig.from_env(),
@@ -143,0 +134,14 @@ class ParquetAndDatasetInfoConfig:
+NUMBA_CACHE_DIR: Optional[str] = None
+
+
+@dataclass
+class NumbaConfig:
+ path: Optional[str] = NUMBA_CACHE_DIR # not documented
+
+ @staticmethod
+ def from_env() -> "NumbaConfig":
+ env = Env(expand_vars=True)
+ with env.prefixed("NUMBA_"):
+ return NumbaConfig(path=env.str(name="NUMBA_CACHE_DIR", default=NUMBA_CACHE_DIR))
+
+
@@ -145,0 +150 @@ class AppConfig:
+ assets: AssetsConfig = field(default_factory=AssetsConfig)
@@ -148,0 +154 @@ class AppConfig:
+ numba: NumbaConfig = field(default_factory=NumbaConfig)
@@ -153,21 +158,0 @@ class AppConfig:
- def __post_init__(self):
- # Ensure the datasets library uses the expected HuggingFace endpoint
- datasets.config.HF_ENDPOINT = self.common.hf_endpoint
- # Don't increase the datasets download counts on huggingface.co
- datasets.config.HF_UPDATE_DOWNLOAD_COUNTS = False
- # Set logs from the datasets library to the least verbose
- set_verbosity(log_levels["critical"])
-
- # Note: self.common.hf_endpoint is ignored by the huggingface_hub library for now (see
- # the discussion at https://github.com/huggingface/datasets/pull/5196), and this breaks
- # various of the datasets functions. The fix, for now, is to set the HF_ENDPOINT
- # environment variable to the desired value.
-
- # Add the datasets and numba cache paths to the list of storage paths, to ensure the disk is not full
- env = Env(expand_vars=True)
- numba_path = env.str(name="NUMBA_CACHE_DIR", default=None)
- additional_paths = {str(self.datasets_based.hf_datasets_cache), str(datasets.config.HF_MODULES_CACHE)}
- if numba_path:
- additional_paths.add(numba_path)
- self.worker_loop.storage_paths = list(set(self.worker_loop.storage_paths).union(additional_paths))
-
@@ -177 +162 @@ class AppConfig:
- # First process the common configuration to setup the logging
+ assets=AssetsConfig.from_env(),
@@ -180,0 +166 @@ class AppConfig:
+ numba=NumbaConfig.from_env(),
diff --git a/workers/datasets_based/src/datasets_based/features.py b/workers/datasets_based/src/datasets_based/features.py
index c626ade3..1710b1e2 100644
--- a/workers/datasets_based/src/datasets_based/features.py
+++ b/workers/datasets_based/src/datasets_based/features.py
@@ -20,0 +21 @@ from datasets import (
+from libcommon.storage import StrPath
@@ -52 +53 @@ def image(
- assets_directory: str,
+ assets_directory: StrPath,
@@ -89 +90 @@ def audio(
- assets_directory: str,
+ assets_directory: StrPath,
@@ -127 +128 @@ def get_cell_value(
- assets_directory: str,
+ assets_directory: StrPath,
diff --git a/workers/datasets_based/src/datasets_based/main.py b/workers/datasets_based/src/datasets_based/main.py
index d5608f5d..0c5bb1ff 100644
--- a/workers/datasets_based/src/datasets_based/main.py
+++ b/workers/datasets_based/src/datasets_based/main.py
@@ -3,0 +4,2 @@
+from libcommon.log import init_logging
+from libcommon.processing_graph import ProcessingGraph
@@ -4,0 +7,2 @@ from libcommon.queue import Queue
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.storage import init_assets_dir
@@ -6,0 +11 @@ from datasets_based.config import AppConfig
+from datasets_based.resources import LibrariesResource
@@ -12,9 +17,31 @@ if __name__ == "__main__":
- processing_step = app_config.processing_graph.graph.get_step(app_config.datasets_based.endpoint)
- worker_factory = WorkerFactory(app_config=app_config)
- queue = Queue(type=processing_step.job_type, max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace)
- worker_loop = WorkerLoop(
- queue=queue,
- worker_factory=worker_factory,
- worker_loop_config=app_config.worker_loop,
- )
- worker_loop.loop()
+
+ init_logging(log_level=app_config.common.log_level)
+ # ^ set first to have logs as soon as possible
+ assets_directory = init_assets_dir(directory=app_config.assets.storage_directory)
+
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+ processing_step = processing_graph.get_step(app_config.datasets_based.endpoint)
+
+ with (
+ LibrariesResource(
+ hf_endpoint=app_config.common.hf_endpoint,
+ init_hf_datasets_cache=app_config.datasets_based.hf_datasets_cache,
+ numba_path=app_config.numba.path,
+ ) as libraries_resource,
+ CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url),
+ QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url),
+ ):
+ worker_factory = WorkerFactory(
+ app_config=app_config,
+ processing_graph=processing_graph,
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ assets_directory=assets_directory,
+ )
+ queue = Queue(type=processing_step.job_type, max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace)
+ worker_loop = WorkerLoop(
+ queue=queue,
+ library_cache_paths=libraries_resource.storage_paths,
+ worker_factory=worker_factory,
+ worker_loop_config=app_config.worker_loop,
+ )
+ worker_loop.loop()
diff --git a/workers/datasets_based/src/datasets_based/resources.py b/workers/datasets_based/src/datasets_based/resources.py
new file mode 100644
index 00000000..f3c743d3
--- /dev/null
+++ b/workers/datasets_based/src/datasets_based/resources.py
@@ -0,0 +1,57 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from dataclasses import dataclass, field
+from pathlib import Path
+from typing import Optional
+
+import datasets
+from datasets.utils.logging import get_verbosity, log_levels, set_verbosity
+from libcommon.resources import Resource
+
+
+@dataclass
+class LibrariesResource(Resource):
+ hf_endpoint: str
+ init_hf_datasets_cache: Optional[str] = None
+ numba_path: Optional[str] = None
+
+ previous_hf_endpoint: str = field(init=False)
+ previous_hf_update_download_counts: bool = field(init=False)
+ previous_verbosity: int = field(init=False)
+ hf_datasets_cache: Path = field(init=False)
+ storage_paths: set[str] = field(init=False)
+
+ def allocate(self):
+ self.hf_datasets_cache = (
+ datasets.config.HF_DATASETS_CACHE
+ if self.init_hf_datasets_cache is None
+ else Path(self.init_hf_datasets_cache)
+ )
+
+ # Ensure the datasets library uses the expected HuggingFace endpoint
+ self.previous_hf_endpoint = datasets.config.HF_ENDPOINT
+ datasets.config.HF_ENDPOINT = self.hf_endpoint
+ # Don't increase the datasets download counts on huggingface.co
+ self.previous_hf_update_download_counts = datasets.config.HF_UPDATE_DOWNLOAD_COUNTS
+ datasets.config.HF_UPDATE_DOWNLOAD_COUNTS = False
+ # Set logs from the datasets library to the least verbose
+ self.previous_verbosity = get_verbosity()
+ set_verbosity(log_levels["critical"])
+
+ # Note: self.hf_endpoint is ignored by the huggingface_hub library for now (see
+ # the discussion at https://github.com/huggingface/datasets/pull/5196), and this breaks
+ # various of the datasets functions. The fix, for now, is to set the HF_ENDPOINT
+ # environment variable to the desired value.
+ # TODO: check here if huggingface_hub and datasets use the same endpoint
+
+ # Add the datasets and numba cache paths to the list of storage paths, to ensure the disk is not full
+ storage_paths = {str(self.hf_datasets_cache), str(datasets.config.HF_MODULES_CACHE)}
+ if self.numba_path is not None:
+ storage_paths.add(self.numba_path)
+ self.storage_paths = storage_paths
+
+ def release(self):
+ datasets.config.HF_ENDPOINT = self.previous_hf_endpoint
+ datasets.config.HF_UPDATE_DOWNLOAD_COUNTS = self.previous_hf_update_download_counts
+ set_verbosity(self.previous_verbosity)
diff --git a/workers/datasets_based/src/datasets_based/worker_factory.py b/workers/datasets_based/src/datasets_based/worker_factory.py
index db40eaf5..87323203 100644
--- a/workers/datasets_based/src/datasets_based/worker_factory.py
+++ b/workers/datasets_based/src/datasets_based/worker_factory.py
@@ -4,0 +5,5 @@ from abc import ABC, abstractmethod
+from dataclasses import dataclass
+from pathlib import Path
+
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.storage import StrPath
@@ -36,0 +42 @@ class BaseWorkerFactory(ABC):
+@dataclass
@@ -38,2 +44,4 @@ class WorkerFactory(BaseWorkerFactory):
- def __init__(self, app_config: AppConfig) -> None:
- self.app_config = app_config
+ app_config: AppConfig
+ processing_graph: ProcessingGraph
+ hf_datasets_cache: Path
+ assets_directory: StrPath
@@ -44 +52 @@ class WorkerFactory(BaseWorkerFactory):
- processing_step = self.app_config.processing_graph.graph.get_step_by_job_type(job_type)
+ processing_step = self.processing_graph.get_step_by_job_type(job_type)
@@ -48 +56 @@ class WorkerFactory(BaseWorkerFactory):
- f" {[step.job_type for step in self.app_config.processing_graph.graph.steps.values()]}"
+ f" {[step.job_type for step in self.processing_graph.steps.values()]}"
@@ -51 +59,6 @@ class WorkerFactory(BaseWorkerFactory):
- return ConfigNamesWorker(job_info=job_info, app_config=self.app_config, processing_step=processing_step)
+ return ConfigNamesWorker(
+ job_info=job_info,
+ app_config=self.app_config,
+ processing_step=processing_step,
+ hf_datasets_cache=self.hf_datasets_cache,
+ )
@@ -53 +66,6 @@ class WorkerFactory(BaseWorkerFactory):
- return SplitNamesWorker(job_info=job_info, app_config=self.app_config, processing_step=processing_step)
+ return SplitNamesWorker(
+ job_info=job_info,
+ app_config=self.app_config,
+ processing_step=processing_step,
+ hf_datasets_cache=self.hf_datasets_cache,
+ )
@@ -55 +73,6 @@ class WorkerFactory(BaseWorkerFactory):
- return SplitsWorker(job_info=job_info, app_config=self.app_config, processing_step=processing_step)
+ return SplitsWorker(
+ job_info=job_info,
+ app_config=self.app_config,
+ processing_step=processing_step,
+ hf_datasets_cache=self.hf_datasets_cache,
+ )
@@ -61,0 +85 @@ class WorkerFactory(BaseWorkerFactory):
+ hf_datasets_cache=self.hf_datasets_cache,
@@ -62,0 +87 @@ class WorkerFactory(BaseWorkerFactory):
+ assets_directory=self.assets_directory,
@@ -68,0 +94 @@ class WorkerFactory(BaseWorkerFactory):
+ hf_datasets_cache=self.hf_datasets_cache,
diff --git a/workers/datasets_based/src/datasets_based/worker_loop.py b/workers/datasets_based/src/datasets_based/worker_loop.py
index 11354557..290075b1 100644
--- a/workers/datasets_based/src/datasets_based/worker_loop.py
+++ b/workers/datasets_based/src/datasets_based/worker_loop.py
@@ -7 +7 @@ import time
-from dataclasses import dataclass
+from dataclasses import dataclass, field
@@ -24,0 +25,2 @@ class WorkerLoop:
+ library_cache_paths (`set[str]`):
+ The paths of the library caches. Used to check if the disk is full.
@@ -32,0 +35 @@ class WorkerLoop:
+ library_cache_paths: set[str]
@@ -36,0 +40,5 @@ class WorkerLoop:
+ storage_paths: set[str] = field(init=False)
+
+ def __post_init__(self) -> None:
+ self.storage_paths = set(self.worker_loop_config.storage_paths).union(self.library_cache_paths)
+
@@ -79 +87 @@ class WorkerLoop:
- for path in self.worker_loop_config.storage_paths:
+ for path in self.storage_paths:
diff --git a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py b/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
index 973bc676..300abb8f 100644
--- a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
+++ b/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
@@ -23,0 +24 @@ class DatasetsBasedWorker(Worker):
+ base_datasets_cache: Path
@@ -30 +31,3 @@ class DatasetsBasedWorker(Worker):
- def __init__(self, job_info: JobInfo, app_config: AppConfig, processing_step: ProcessingStep) -> None:
+ def __init__(
+ self, job_info: JobInfo, app_config: AppConfig, processing_step: ProcessingStep, hf_datasets_cache: Path
+ ) -> None:
@@ -37,0 +41 @@ class DatasetsBasedWorker(Worker):
+ self.base_datasets_cache = hf_datasets_cache
@@ -60 +64 @@ class DatasetsBasedWorker(Worker):
- self.set_datasets_cache(self.datasets_based_config.hf_datasets_cache)
+ self.set_datasets_cache(self.base_datasets_cache)
@@ -68 +72 @@ class DatasetsBasedWorker(Worker):
- self.set_datasets_cache(self.datasets_based_config.hf_datasets_cache / cache_subdirectory)
+ self.set_datasets_cache(self.base_datasets_cache / cache_subdirectory)
diff --git a/workers/datasets_based/src/datasets_based/workers/first_rows.py b/workers/datasets_based/src/datasets_based/workers/first_rows.py
index f4aa20eb..3a1eba1b 100644
--- a/workers/datasets_based/src/datasets_based/workers/first_rows.py
+++ b/workers/datasets_based/src/datasets_based/workers/first_rows.py
@@ -9,0 +10 @@ from http import HTTPStatus
+from pathlib import Path
@@ -23,0 +25 @@ from libcommon.simple_cache import SplitFullName as _SplitFullName
+from libcommon.storage import StrPath
@@ -340 +342 @@ def transform_rows(
- assets_directory: str,
+ assets_directory: StrPath,
@@ -406 +408 @@ def compute_first_rows_response(
- assets_directory: str,
+ assets_directory: StrPath,
@@ -439,0 +442,2 @@ def compute_first_rows_response(
+ assets_directory (`str` or `pathlib.Path`):
+ The directory where the assets are stored.
@@ -606,0 +611 @@ class FirstRowsWorker(DatasetsBasedWorker):
+ assets_directory: StrPath
@@ -622,0 +628,2 @@ class FirstRowsWorker(DatasetsBasedWorker):
+ hf_datasets_cache: Path,
+ assets_directory: StrPath,
@@ -624 +631,6 @@ class FirstRowsWorker(DatasetsBasedWorker):
- super().__init__(job_info=job_info, app_config=app_config, processing_step=processing_step)
+ super().__init__(
+ job_info=job_info,
+ app_config=app_config,
+ processing_step=processing_step,
+ hf_datasets_cache=hf_datasets_cache,
+ )
@@ -625,0 +638,2 @@ class FirstRowsWorker(DatasetsBasedWorker):
+ self.assets_directory = assets_directory
+ self.assets_base_url = app_config.assets.base_url
@@ -634,2 +648,2 @@ class FirstRowsWorker(DatasetsBasedWorker):
- assets_base_url=self.first_rows_config.assets.base_url,
- assets_directory=self.first_rows_config.assets.storage_directory,
+ assets_base_url=self.assets_base_url,
+ assets_directory=self.assets_directory,
diff --git a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py b/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
index b3523ecc..74c37021 100644
--- a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
+++ b/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
@@ -619,0 +620 @@ class ParquetAndDatasetInfoWorker(DatasetsBasedWorker):
+ hf_datasets_cache: Path,
@@ -622 +623,6 @@ class ParquetAndDatasetInfoWorker(DatasetsBasedWorker):
- super().__init__(job_info=job_info, app_config=app_config, processing_step=processing_step)
+ super().__init__(
+ job_info=job_info,
+ app_config=app_config,
+ processing_step=processing_step,
+ hf_datasets_cache=hf_datasets_cache,
+ )
diff --git a/workers/datasets_based/tests/conftest.py b/workers/datasets_based/tests/conftest.py
index 669d7e44..2292da13 100644
--- a/workers/datasets_based/tests/conftest.py
+++ b/workers/datasets_based/tests/conftest.py
@@ -7 +6,0 @@ from typing import Iterator
-from libcommon.config import CacheConfig, QueueConfig
@@ -9,0 +9 @@ from libcommon.queue import _clean_queue_database
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -10,0 +11 @@ from libcommon.simple_cache import _clean_cache_database
+from libcommon.storage import StrPath, init_assets_dir
@@ -13,0 +15 @@ from datasets_based.config import AppConfig, FirstRowsConfig
+from datasets_based.resources import LibrariesResource
@@ -65,5 +67,24 @@ def app_config(set_env_vars: MonkeyPatch) -> Iterator[AppConfig]:
- # Clean the database after each test. Must be done in test databases only, ensured by the check above!
- # TODO: use a parameter to pass a reference to the database, instead of relying on the implicit global variable
- # managed by mongoengine
- _clean_cache_database()
- _clean_queue_database()
+
+
+@fixture
+def cache_mongo_resource(app_config: AppConfig) -> Iterator[CacheMongoResource]:
+ with CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url) as resource:
+ yield resource
+ _clean_cache_database()
+
+
+@fixture
+def queue_mongo_resource(app_config: AppConfig) -> Iterator[QueueMongoResource]:
+ with QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url) as resource:
+ yield resource
+ _clean_queue_database()
+
+
+@fixture
+def libraries_resource(app_config: AppConfig) -> Iterator[LibrariesResource]:
+ with LibrariesResource(
+ hf_endpoint=app_config.common.hf_endpoint,
+ init_hf_datasets_cache=app_config.datasets_based.hf_datasets_cache,
+ numba_path=app_config.numba.path,
+ ) as libraries_resource:
+ yield libraries_resource
@@ -77,2 +98,3 @@ def first_rows_config(set_env_vars: MonkeyPatch) -> FirstRowsConfig:
-# Import fixture modules as plugins
-pytest_plugins = ["tests.fixtures.datasets", "tests.fixtures.files", "tests.fixtures.hub"]
+@fixture
+def assets_directory(app_config: AppConfig) -> StrPath:
+ return init_assets_dir(app_config.assets.storage_directory)
@@ -94,14 +116,2 @@ def test_processing_step() -> ProcessingStep:
-@fixture()
-def cache_config(app_config: AppConfig) -> CacheConfig:
- cache_config = app_config.cache
- if "test" not in cache_config.mongo_database:
- raise ValueError("Test must be launched on a test mongo database")
- return cache_config
-
-
-@fixture()
-def queue_config(app_config: AppConfig) -> QueueConfig:
- queue_config = app_config.queue
- if "test" not in queue_config.mongo_database:
- raise ValueError("Test must be launched on a test mongo database")
- return queue_config
+# Import fixture modules as plugins
+pytest_plugins = ["tests.fixtures.datasets", "tests.fixtures.files", "tests.fixtures.hub"]
diff --git a/workers/datasets_based/tests/test_features.py b/workers/datasets_based/tests/test_features.py
index 745b216a..ca2b8057 100644
--- a/workers/datasets_based/tests/test_features.py
+++ b/workers/datasets_based/tests/test_features.py
@@ -10,0 +11 @@ from datasets import Audio, Dataset, Image, Value
+from libcommon.storage import StrPath
@@ -12 +13 @@ from datasets import Audio, Dataset, Image, Value
-from datasets_based.config import AppConfig, FirstRowsConfig
+from datasets_based.config import AppConfig
@@ -60 +61 @@ def test_value(
- first_rows_config: FirstRowsConfig,
+ assets_directory: StrPath,
@@ -74,2 +75,2 @@ def test_value(
- assets_base_url=first_rows_config.assets.base_url,
- assets_directory=first_rows_config.assets.storage_directory,
+ assets_base_url=app_config.assets.base_url,
+ assets_directory=assets_directory,
@@ -302 +303 @@ def test_others(
- first_rows_config: FirstRowsConfig,
+ assets_directory: StrPath,
@@ -318,2 +319,2 @@ def test_others(
- assets_base_url=first_rows_config.assets.base_url,
- assets_directory=first_rows_config.assets.storage_directory,
+ assets_base_url=app_config.assets.base_url,
+ assets_directory=assets_directory,
diff --git a/workers/datasets_based/tests/test_resources.py b/workers/datasets_based/tests/test_resources.py
new file mode 100644
index 00000000..b83582d0
--- /dev/null
+++ b/workers/datasets_based/tests/test_resources.py
@@ -0,0 +1,49 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import datasets
+import pytest
+
+from datasets_based.resources import LibrariesResource
+
+
[email protected](
+ "define_init_hf_datasets_cache,define_numba_path",
+ [(False, False), (False, True), (True, False), (True, True)],
+)
+def test_libraries(tmp_path_factory, define_init_hf_datasets_cache: bool, define_numba_path: bool) -> None:
+ hf_endpoint = "https://another.endpoint"
+ init_hf_datasets_cache: str = (
+ tmp_path_factory.mktemp("hf_datasets_cache") if define_init_hf_datasets_cache else None
+ )
+ numba_path: str = tmp_path_factory.mktemp("numba_path") if define_numba_path else None
+ assert datasets.config.HF_ENDPOINT != hf_endpoint
+ resource = LibrariesResource(
+ hf_endpoint=hf_endpoint,
+ init_hf_datasets_cache=init_hf_datasets_cache,
+ numba_path=numba_path,
+ )
+ assert datasets.config.HF_ENDPOINT == hf_endpoint
+ assert (numba_path in resource.storage_paths) == define_numba_path
+ assert str(resource.hf_datasets_cache) in resource.storage_paths
+ assert str(datasets.config.HF_MODULES_CACHE) in resource.storage_paths
+ assert datasets.config.HF_UPDATE_DOWNLOAD_COUNTS is False
+ assert (resource.hf_datasets_cache == init_hf_datasets_cache) == define_init_hf_datasets_cache
+
+ resource.release()
+
+ assert datasets.config.HF_ENDPOINT != hf_endpoint
+
+
+def test_libraries_context_manager(tmp_path_factory) -> None:
+ hf_endpoint = "https://another.endpoint"
+ init_hf_datasets_cache: str = tmp_path_factory.mktemp("hf_datasets_cache")
+ numba_path: str = tmp_path_factory.mktemp("numba_path")
+ assert datasets.config.HF_ENDPOINT != hf_endpoint
+ with LibrariesResource(
+ hf_endpoint=hf_endpoint,
+ init_hf_datasets_cache=init_hf_datasets_cache,
+ numba_path=numba_path,
+ ):
+ assert datasets.config.HF_ENDPOINT == hf_endpoint
+ assert datasets.config.HF_ENDPOINT != hf_endpoint
diff --git a/workers/datasets_based/tests/test_worker.py b/workers/datasets_based/tests/test_worker.py
index 63eb2f71..1c5e3211 100644
--- a/workers/datasets_based/tests/test_worker.py
+++ b/workers/datasets_based/tests/test_worker.py
@@ -8,0 +9 @@ from libcommon.queue import Priority, Queue, Status
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -11 +12 @@ from libcommon.simple_cache import SplitFullName, upsert_response
-from datasets_based.config import AppConfig, DatasetsBasedConfig
+from datasets_based.config import DatasetsBasedConfig
@@ -16 +17,4 @@ from datasets_based.worker import ERROR_CODES_TO_RETRY, Worker
-def prepare_and_clean_mongo(app_config: AppConfig) -> None:
+def prepare_and_clean_mongo(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> None:
diff --git a/workers/datasets_based/tests/test_worker_factory.py b/workers/datasets_based/tests/test_worker_factory.py
index 07dd5728..279d45d6 100644
--- a/workers/datasets_based/tests/test_worker_factory.py
+++ b/workers/datasets_based/tests/test_worker_factory.py
@@ -6,0 +7 @@ import pytest
+from libcommon.processing_graph import ProcessingGraph
@@ -7,0 +9 @@ from libcommon.queue import Priority
+from libcommon.storage import StrPath
@@ -9,0 +12 @@ from datasets_based.config import AppConfig
+from datasets_based.resources import LibrariesResource
@@ -13,0 +17,5 @@ from datasets_based.worker_factory import WorkerFactory
[email protected]()
+def processing_graph(app_config: AppConfig) -> ProcessingGraph:
+ return ProcessingGraph(app_config.processing_graph.specification)
+
+
@@ -27,2 +35,14 @@ from datasets_based.worker_factory import WorkerFactory
-def test_create_worker(app_config: AppConfig, job_type: str, expected_worker: Optional[str]) -> None:
- worker_factory = WorkerFactory(app_config=app_config)
+def test_create_worker(
+ app_config: AppConfig,
+ processing_graph: ProcessingGraph,
+ libraries_resource: LibrariesResource,
+ assets_directory: StrPath,
+ job_type: str,
+ expected_worker: Optional[str],
+) -> None:
+ worker_factory = WorkerFactory(
+ app_config=app_config,
+ processing_graph=processing_graph,
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ assets_directory=assets_directory,
+ )
diff --git a/workers/datasets_based/tests/test_worker_loop.py b/workers/datasets_based/tests/test_worker_loop.py
index 696db769..e843c976 100644
--- a/workers/datasets_based/tests/test_worker_loop.py
+++ b/workers/datasets_based/tests/test_worker_loop.py
@@ -3,2 +3 @@ from typing import Any, Mapping, Optional
-import pytest
-from libcommon.config import CommonConfig, QueueConfig
+from libcommon.config import CommonConfig
@@ -6,0 +6 @@ from libcommon.queue import Queue
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -8,0 +9 @@ from datasets_based.config import AppConfig, DatasetsBasedConfig, WorkerLoopConf
+from datasets_based.resources import LibrariesResource
@@ -14,6 +14,0 @@ from datasets_based.worker_loop import WorkerLoop
[email protected](autouse=True)
-def prepare_and_clean_mongo(app_config: AppConfig) -> None:
- # prepare the database before each test, and clean it afterwards
- pass
-
-
@@ -54 +49,4 @@ def test_process_next_job(
- queue_config: QueueConfig,
+ app_config: AppConfig,
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
@@ -57 +55 @@ def test_process_next_job(
- queue = Queue(type=test_processing_step.endpoint, max_jobs_per_namespace=queue_config.max_jobs_per_namespace)
+ queue = Queue(type=test_processing_step.endpoint, max_jobs_per_namespace=app_config.queue.max_jobs_per_namespace)
@@ -59 +57 @@ def test_process_next_job(
- worker_factory=worker_factory,
+ library_cache_paths=libraries_resource.storage_paths,
@@ -60,0 +59 @@ def test_process_next_job(
+ worker_factory=worker_factory,
diff --git a/workers/datasets_based/tests/workers/test__datasets_based_worker.py b/workers/datasets_based/tests/workers/test__datasets_based_worker.py
index 1f82113d..5c527daa 100644
--- a/workers/datasets_based/tests/workers/test__datasets_based_worker.py
+++ b/workers/datasets_based/tests/workers/test__datasets_based_worker.py
@@ -13,0 +14 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -16,0 +18 @@ from datasets_based.config import AppConfig
+from datasets_based.resources import LibrariesResource
@@ -39,0 +42 @@ class DummyWorker(DatasetsBasedWorker):
[email protected]
@@ -41,30 +44,38 @@ def get_worker(
- dataset: str,
- config: Optional[str],
- split: Optional[str],
- app_config: AppConfig,
- force: bool = False,
-) -> DummyWorker:
- return DummyWorker(
- job_info={
- "type": DummyWorker.get_job_type(),
- "dataset": dataset,
- "config": config,
- "split": split,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- app_config=app_config,
- processing_step=ProcessingStep(
- endpoint=DummyWorker.get_job_type(),
- input_type="split",
- requires=None,
- required_by_dataset_viewer=False,
- parent=None,
- ancestors=[],
- children=[],
- ),
- )
-
-
-def test_version(app_config: AppConfig) -> None:
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ config: Optional[str],
+ split: Optional[str],
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> DummyWorker:
+ return DummyWorker(
+ job_info={
+ "type": DummyWorker.get_job_type(),
+ "dataset": dataset,
+ "config": config,
+ "split": split,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ endpoint=DummyWorker.get_job_type(),
+ input_type="split",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_worker
+
+
+def test_version(app_config: AppConfig, get_worker) -> None:
@@ -99 +110,7 @@ def test_get_cache_subdirectory(
- app_config: AppConfig, dataset: str, config: Optional[str], split: Optional[str], force: bool, expected: str
+ app_config: AppConfig,
+ get_worker,
+ dataset: str,
+ config: Optional[str],
+ split: Optional[str],
+ force: bool,
+ expected: str,
@@ -106 +123 @@ def test_get_cache_subdirectory(
-def test_set_and_unset_datasets_cache(app_config: AppConfig) -> None:
+def test_set_and_unset_datasets_cache(app_config: AppConfig, get_worker) -> None:
@@ -109 +126 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig) -> None:
- base_path = worker.datasets_based_config.hf_datasets_cache
+ base_path = worker.base_datasets_cache
@@ -117 +134 @@ def test_set_and_unset_datasets_cache(app_config: AppConfig) -> None:
-def test_set_and_unset_cache(app_config: AppConfig) -> None:
+def test_set_and_unset_cache(app_config: AppConfig, get_worker) -> None:
@@ -120 +137 @@ def test_set_and_unset_cache(app_config: AppConfig) -> None:
- datasets_base_path = worker.datasets_based_config.hf_datasets_cache
+ datasets_base_path = worker.base_datasets_cache
@@ -129 +146 @@ def test_set_and_unset_cache(app_config: AppConfig) -> None:
-def test_process(app_config: AppConfig, hub_public_csv: str, config: str) -> None:
+def test_process(app_config: AppConfig, get_worker, hub_public_csv: str, config: str) -> None:
@@ -135 +152 @@ def test_process(app_config: AppConfig, hub_public_csv: str, config: str) -> Non
- datasets_base_path = worker.datasets_based_config.hf_datasets_cache
+ datasets_base_path = worker.base_datasets_cache
@@ -151 +168 @@ def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) ->
-def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig) -> None:
+def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig, get_worker) -> None:
diff --git a/workers/datasets_based/tests/workers/test_config_names.py b/workers/datasets_based/tests/workers/test_config_names.py
index 02a76739..6da4f23c 100644
--- a/workers/datasets_based/tests/workers/test_config_names.py
+++ b/workers/datasets_based/tests/workers/test_config_names.py
@@ -10,0 +11 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -13,0 +15 @@ from datasets_based.config import AppConfig
+from datasets_based.resources import LibrariesResource
@@ -18,0 +21 @@ from ..fixtures.hub import HubDatasets
[email protected]
@@ -20,25 +23,33 @@ def get_worker(
- dataset: str,
- app_config: AppConfig,
- force: bool = False,
-) -> ConfigNamesWorker:
- return ConfigNamesWorker(
- job_info={
- "type": ConfigNamesWorker.get_job_type(),
- "dataset": dataset,
- "config": None,
- "split": None,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- app_config=app_config,
- processing_step=ProcessingStep(
- endpoint=ConfigNamesWorker.get_job_type(),
- input_type="dataset",
- requires=None,
- required_by_dataset_viewer=False,
- parent=None,
- ancestors=[],
- children=[],
- ),
- )
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> ConfigNamesWorker:
+ return ConfigNamesWorker(
+ job_info={
+ "type": ConfigNamesWorker.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ endpoint=ConfigNamesWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_worker
@@ -47 +58 @@ def get_worker(
-def test_should_skip_job(app_config: AppConfig, hub_public_csv: str) -> None:
+def test_should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
@@ -58 +69 @@ def test_should_skip_job(app_config: AppConfig, hub_public_csv: str) -> None:
-def test_process(app_config: AppConfig, hub_public_csv: str) -> None:
+def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
@@ -72 +83 @@ def test_process(app_config: AppConfig, hub_public_csv: str) -> None:
-def test_doesnotexist(app_config: AppConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
@@ -96 +107,7 @@ def test_compute_splits_response_simple_csv(
- hub_datasets: HubDatasets, name: str, use_token: bool, error_code: str, cause: str, app_config: AppConfig
+ hub_datasets: HubDatasets,
+ get_worker,
+ name: str,
+ use_token: bool,
+ error_code: str,
+ cause: str,
+ app_config: AppConfig,
diff --git a/workers/datasets_based/tests/workers/test_dataset_info.py b/workers/datasets_based/tests/workers/test_dataset_info.py
index 34227830..01f1c24a 100644
--- a/workers/datasets_based/tests/workers/test_dataset_info.py
+++ b/workers/datasets_based/tests/workers/test_dataset_info.py
@@ -9,0 +10 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -27,23 +28,34 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> DatasetInfoWorker:
- return DatasetInfoWorker(
- job_info={
- "type": DatasetInfoWorker.get_job_type(),
- "dataset": dataset,
- "config": None,
- "split": None,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- common_config=app_config.common,
- datasets_based_config=app_config.datasets_based,
- processing_step=ProcessingStep(
- endpoint=DatasetInfoWorker.get_job_type(),
- input_type="dataset",
- requires=None,
- required_by_dataset_viewer=False,
- parent=None,
- ancestors=[],
- children=[],
- ),
- )
[email protected]
+def get_worker(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> DatasetInfoWorker:
+ return DatasetInfoWorker(
+ job_info={
+ "type": DatasetInfoWorker.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ common_config=app_config.common,
+ datasets_based_config=app_config.datasets_based,
+ processing_step=ProcessingStep(
+ endpoint=DatasetInfoWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ )
+
+ return _get_worker
@@ -75,0 +88 @@ def test_compute(
+ get_worker,
@@ -95 +108 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
diff --git a/workers/datasets_based/tests/workers/test_first_rows.py b/workers/datasets_based/tests/workers/test_first_rows.py
index 6416f027..ab7ce7a0 100644
--- a/workers/datasets_based/tests/workers/test_first_rows.py
+++ b/workers/datasets_based/tests/workers/test_first_rows.py
@@ -11,0 +12 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -12,0 +14 @@ from libcommon.simple_cache import DoesNotExist, get_response
+from libcommon.storage import StrPath
@@ -14,0 +17 @@ from datasets_based.config import AppConfig, FirstRowsConfig
+from datasets_based.resources import LibrariesResource
@@ -19,0 +23 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
[email protected]
@@ -21,32 +25,44 @@ def get_worker(
- dataset: str,
- config: str,
- split: str,
- app_config: AppConfig,
- first_rows_config: FirstRowsConfig,
- force: bool = False,
-) -> FirstRowsWorker:
- return FirstRowsWorker(
- job_info={
- "type": FirstRowsWorker.get_job_type(),
- "dataset": dataset,
- "config": config,
- "split": split,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- app_config=app_config,
- processing_step=ProcessingStep(
- endpoint=FirstRowsWorker.get_job_type(),
- input_type="split",
- requires=None,
- required_by_dataset_viewer=True,
- parent=None,
- ancestors=[],
- children=[],
- ),
- first_rows_config=first_rows_config,
- )
-
-
-def test_should_skip_job(app_config: AppConfig, first_rows_config: FirstRowsConfig, hub_public_csv: str) -> None:
+ assets_directory: StrPath,
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ config: str,
+ split: str,
+ app_config: AppConfig,
+ first_rows_config: FirstRowsConfig,
+ force: bool = False,
+ ) -> FirstRowsWorker:
+ return FirstRowsWorker(
+ job_info={
+ "type": FirstRowsWorker.get_job_type(),
+ "dataset": dataset,
+ "config": config,
+ "split": split,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ endpoint=FirstRowsWorker.get_job_type(),
+ input_type="split",
+ requires=None,
+ required_by_dataset_viewer=True,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ first_rows_config=first_rows_config,
+ assets_directory=assets_directory,
+ )
+
+ return _get_worker
+
+
+def test_should_skip_job(
+ app_config: AppConfig, get_worker, first_rows_config: FirstRowsConfig, hub_public_csv: str
+) -> None:
@@ -63 +79 @@ def test_should_skip_job(app_config: AppConfig, first_rows_config: FirstRowsConf
-def test_compute(app_config: AppConfig, first_rows_config: FirstRowsConfig, hub_public_csv: str) -> None:
+def test_compute(app_config: AppConfig, get_worker, first_rows_config: FirstRowsConfig, hub_public_csv: str) -> None:
@@ -81 +97 @@ def test_compute(app_config: AppConfig, first_rows_config: FirstRowsConfig, hub_
-def test_doesnotexist(app_config: AppConfig, first_rows_config: FirstRowsConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_worker, first_rows_config: FirstRowsConfig) -> None:
@@ -109,0 +126 @@ def test_number_rows(
+ get_worker,
@@ -169,0 +187 @@ def test_truncation(
+ get_worker,
diff --git a/workers/datasets_based/tests/workers/test_parquet.py b/workers/datasets_based/tests/workers/test_parquet.py
index 7df0715d..c022f651 100644
--- a/workers/datasets_based/tests/workers/test_parquet.py
+++ b/workers/datasets_based/tests/workers/test_parquet.py
@@ -9,0 +10 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -27,23 +28,34 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> ParquetWorker:
- return ParquetWorker(
- job_info={
- "type": ParquetWorker.get_job_type(),
- "dataset": dataset,
- "config": None,
- "split": None,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- common_config=app_config.common,
- datasets_based_config=app_config.datasets_based,
- processing_step=ProcessingStep(
- endpoint=ParquetWorker.get_job_type(),
- input_type="dataset",
- requires=None,
- required_by_dataset_viewer=False,
- parent=None,
- ancestors=[],
- children=[],
- ),
- )
[email protected]
+def get_worker(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> ParquetWorker:
+ return ParquetWorker(
+ job_info={
+ "type": ParquetWorker.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ common_config=app_config.common,
+ datasets_based_config=app_config.datasets_based,
+ processing_step=ProcessingStep(
+ endpoint=ParquetWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ )
+
+ return _get_worker
@@ -75,0 +88 @@ def test_compute(
+ get_worker,
@@ -95 +108 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
diff --git a/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py b/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
index fec1739a..78438dc7 100644
--- a/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
+++ b/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
@@ -13,0 +14 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -16,0 +18 @@ from datasets_based.config import AppConfig, ParquetAndDatasetInfoConfig
+from datasets_based.resources import LibrariesResource
@@ -55,0 +58 @@ def parquet_and_dataset_info_config(
[email protected]
@@ -57,27 +60,35 @@ def get_worker(
- dataset: str,
- app_config: AppConfig,
- parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
- force: bool = False,
-) -> ParquetAndDatasetInfoWorker:
- return ParquetAndDatasetInfoWorker(
- job_info={
- "type": ParquetAndDatasetInfoWorker.get_job_type(),
- "dataset": dataset,
- "config": None,
- "split": None,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- app_config=app_config,
- processing_step=ProcessingStep(
- endpoint=ParquetAndDatasetInfoWorker.get_job_type(),
- input_type="dataset",
- requires=None,
- required_by_dataset_viewer=False,
- parent=None,
- ancestors=[],
- children=[],
- ),
- parquet_and_dataset_info_config=parquet_and_dataset_info_config,
- )
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ app_config: AppConfig,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
+ force: bool = False,
+ ) -> ParquetAndDatasetInfoWorker:
+ return ParquetAndDatasetInfoWorker(
+ job_info={
+ "type": ParquetAndDatasetInfoWorker.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ endpoint=ParquetAndDatasetInfoWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ parquet_and_dataset_info_config=parquet_and_dataset_info_config,
+ )
+
+ return _get_worker
@@ -104 +115,4 @@ def test_compute(
- app_config: AppConfig, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, hub_datasets: HubDatasets
+ app_config: AppConfig,
+ get_worker,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
+ hub_datasets: HubDatasets,
@@ -121 +135,3 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig) -> None:
+def test_doesnotexist(
+ app_config: AppConfig, get_worker, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig
+) -> None:
@@ -248 +264,4 @@ def test_not_supported_if_big(
- app_config: AppConfig, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, hub_public_big: str
+ app_config: AppConfig,
+ get_worker,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
+ hub_public_big: str,
@@ -262 +281 @@ def test_supported_if_gated(
- app_config: AppConfig, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, hub_gated_csv: str
+ app_config: AppConfig, get_worker, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, hub_gated_csv: str
@@ -276,0 +296 @@ def test_not_supported_if_gated_with_extra_fields(
+ get_worker,
@@ -292 +312,4 @@ def test_blocked(
- app_config: AppConfig, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, hub_public_jsonl: str
+ app_config: AppConfig,
+ get_worker,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
+ hub_public_jsonl: str,
@@ -310,0 +334 @@ def test_compute_splits_response_simple_csv_ok(
+ get_worker,
@@ -352,0 +377 @@ def test_compute_splits_response_simple_csv_error(
+ get_worker,
diff --git a/workers/datasets_based/tests/workers/test_sizes.py b/workers/datasets_based/tests/workers/test_sizes.py
index 1d773dba..4d4893ff 100644
--- a/workers/datasets_based/tests/workers/test_sizes.py
+++ b/workers/datasets_based/tests/workers/test_sizes.py
@@ -9,0 +10 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -27,23 +28,34 @@ def prepare_and_clean_mongo(app_config: AppConfig) -> None:
-def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> SizesWorker:
- return SizesWorker(
- job_info={
- "type": SizesWorker.get_job_type(),
- "dataset": dataset,
- "config": None,
- "split": None,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- common_config=app_config.common,
- datasets_based_config=app_config.datasets_based,
- processing_step=ProcessingStep(
- endpoint=SizesWorker.get_job_type(),
- input_type="dataset",
- requires=None,
- required_by_dataset_viewer=False,
- parent=None,
- ancestors=[],
- children=[],
- ),
- )
[email protected]
+def get_worker(
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> SizesWorker:
+ return SizesWorker(
+ job_info={
+ "type": SizesWorker.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ common_config=app_config.common,
+ datasets_based_config=app_config.datasets_based,
+ processing_step=ProcessingStep(
+ endpoint=SizesWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ )
+
+ return _get_worker
@@ -230,0 +243 @@ def test_compute(
+ get_worker,
@@ -250 +263 @@ def test_compute(
-def test_doesnotexist(app_config: AppConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
diff --git a/workers/datasets_based/tests/workers/test_split_names.py b/workers/datasets_based/tests/workers/test_split_names.py
index d61ee180..876a2a63 100644
--- a/workers/datasets_based/tests/workers/test_split_names.py
+++ b/workers/datasets_based/tests/workers/test_split_names.py
@@ -10,0 +11 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -13,0 +15 @@ from datasets_based.config import AppConfig
+from datasets_based.resources import LibrariesResource
@@ -18,0 +21 @@ from ..fixtures.hub import HubDatasets, get_default_config_split
[email protected]
@@ -20,26 +23,34 @@ def get_worker(
- dataset: str,
- config: str,
- app_config: AppConfig,
- force: bool = False,
-) -> SplitNamesWorker:
- return SplitNamesWorker(
- job_info={
- "type": SplitNamesWorker.get_job_type(),
- "dataset": dataset,
- "config": config,
- "split": None,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- app_config=app_config,
- processing_step=ProcessingStep(
- endpoint=SplitNamesWorker.get_job_type(),
- input_type="config",
- requires=None,
- required_by_dataset_viewer=False,
- parent=None,
- ancestors=[],
- children=[],
- ),
- )
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ config: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> SplitNamesWorker:
+ return SplitNamesWorker(
+ job_info={
+ "type": SplitNamesWorker.get_job_type(),
+ "dataset": dataset,
+ "config": config,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ endpoint=SplitNamesWorker.get_job_type(),
+ input_type="config",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_worker
@@ -48 +59 @@ def get_worker(
-def test_process(app_config: AppConfig, hub_public_csv: str) -> None:
+def test_process(app_config: AppConfig, get_worker, hub_public_csv: str) -> None:
@@ -62 +73 @@ def test_process(app_config: AppConfig, hub_public_csv: str) -> None:
-def test_doesnotexist(app_config: AppConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
@@ -87 +98,7 @@ def test_compute_split_names_response(
- hub_datasets: HubDatasets, name: str, use_token: bool, error_code: str, cause: str, app_config: AppConfig
+ hub_datasets: HubDatasets,
+ get_worker,
+ name: str,
+ use_token: bool,
+ error_code: str,
+ cause: str,
+ app_config: AppConfig,
diff --git a/workers/datasets_based/tests/workers/test_splits.py b/workers/datasets_based/tests/workers/test_splits.py
index b253a278..7e85c31a 100644
--- a/workers/datasets_based/tests/workers/test_splits.py
+++ b/workers/datasets_based/tests/workers/test_splits.py
@@ -10,0 +11 @@ from libcommon.queue import Priority
+from libcommon.resources import CacheMongoResource, QueueMongoResource
@@ -13,0 +15 @@ from datasets_based.config import AppConfig
+from datasets_based.resources import LibrariesResource
@@ -18,0 +21 @@ from ..fixtures.hub import HubDatasets
[email protected]
@@ -20,25 +23,33 @@ def get_worker(
- dataset: str,
- app_config: AppConfig,
- force: bool = False,
-) -> SplitsWorker:
- return SplitsWorker(
- job_info={
- "type": SplitsWorker.get_job_type(),
- "dataset": dataset,
- "config": None,
- "split": None,
- "job_id": "job_id",
- "force": force,
- "priority": Priority.NORMAL,
- },
- app_config=app_config,
- processing_step=ProcessingStep(
- endpoint=SplitsWorker.get_job_type(),
- input_type="dataset",
- requires=None,
- required_by_dataset_viewer=True,
- parent=None,
- ancestors=[],
- children=[],
- ),
- )
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+):
+ def _get_worker(
+ dataset: str,
+ app_config: AppConfig,
+ force: bool = False,
+ ) -> SplitsWorker:
+ return SplitsWorker(
+ job_info={
+ "type": SplitsWorker.get_job_type(),
+ "dataset": dataset,
+ "config": None,
+ "split": None,
+ "job_id": "job_id",
+ "force": force,
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ endpoint=SplitsWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=True,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_worker
@@ -47 +58 @@ def get_worker(
-def should_skip_job(app_config: AppConfig, hub_public_csv: str) -> None:
+def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
@@ -58 +69 @@ def should_skip_job(app_config: AppConfig, hub_public_csv: str) -> None:
-def test_process(app_config: AppConfig, hub_public_csv: str) -> None:
+def test_process(app_config: AppConfig, hub_public_csv: str, get_worker) -> None:
@@ -73 +84 @@ def test_process(app_config: AppConfig, hub_public_csv: str) -> None:
-def test_doesnotexist(app_config: AppConfig) -> None:
+def test_doesnotexist(app_config: AppConfig, get_worker) -> None:
@@ -97 +108,7 @@ def test_compute_splits_response_simple_csv(
- hub_datasets: HubDatasets, name: str, use_token: bool, error_code: str, cause: str, app_config: AppConfig
+ hub_datasets: HubDatasets,
+ get_worker,
+ name: str,
+ use_token: bool,
+ error_code: str,
+ cause: str,
+ app_config: AppConfig,
|
|
afd2e49fbc27d000df96b771e753796dc5ade981
|
Andrea Francis Soria Jimenez
| 2023-02-09T17:55:34 |
Dataset info big content error (#780)
|
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index 502ebdfc..7ff95ac6 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -242,0 +243,4 @@ sizes:
+
+# --- datasets_based ---
+datasetsBased:
+ contentMaxBytes: "10_000_000"
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index b5e0f2f2..337db015 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -312,0 +313,4 @@ sizes:
+
+# --- datasets_based ---
+datasetsBased:
+ contentMaxBytes: "10_000_000"
diff --git a/chart/templates/_envDatasetsBased.tpl b/chart/templates/_envDatasetsBased.tpl
index 05d1ed29..ee8fea75 100644
--- a/chart/templates/_envDatasetsBased.tpl
+++ b/chart/templates/_envDatasetsBased.tpl
@@ -10,0 +11,2 @@
+- name: CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes}}
diff --git a/chart/templates/worker/config-names/_container.tpl b/chart/templates/worker/config-names/_container.tpl
index c7e1f76d..bc2f8550 100644
--- a/chart/templates/worker/config-names/_container.tpl
+++ b/chart/templates/worker/config-names/_container.tpl
@@ -18,0 +19,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/dataset-info/_container.tpl b/chart/templates/worker/dataset-info/_container.tpl
index f067ddac..6091e71a 100644
--- a/chart/templates/worker/dataset-info/_container.tpl
+++ b/chart/templates/worker/dataset-info/_container.tpl
@@ -15,0 +16,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/first-rows/_container.tpl b/chart/templates/worker/first-rows/_container.tpl
index 9c2cc318..83e2fd72 100644
--- a/chart/templates/worker/first-rows/_container.tpl
+++ b/chart/templates/worker/first-rows/_container.tpl
@@ -22,0 +23,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/parquet-and-dataset-info/_container.tpl b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
index b35d5be2..b926df09 100644
--- a/chart/templates/worker/parquet-and-dataset-info/_container.tpl
+++ b/chart/templates/worker/parquet-and-dataset-info/_container.tpl
@@ -18,0 +19,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/parquet/_container.tpl b/chart/templates/worker/parquet/_container.tpl
index 6ab0e886..ace15115 100644
--- a/chart/templates/worker/parquet/_container.tpl
+++ b/chart/templates/worker/parquet/_container.tpl
@@ -15,0 +16,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/sizes/_container.tpl b/chart/templates/worker/sizes/_container.tpl
index b67d95c4..b60db74f 100644
--- a/chart/templates/worker/sizes/_container.tpl
+++ b/chart/templates/worker/sizes/_container.tpl
@@ -15,0 +16,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/split-names/_container.tpl b/chart/templates/worker/split-names/_container.tpl
index e66e422c..67879b34 100644
--- a/chart/templates/worker/split-names/_container.tpl
+++ b/chart/templates/worker/split-names/_container.tpl
@@ -18,0 +19,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/templates/worker/splits/_container.tpl b/chart/templates/worker/splits/_container.tpl
index 6d2a4c20..a3715380 100644
--- a/chart/templates/worker/splits/_container.tpl
+++ b/chart/templates/worker/splits/_container.tpl
@@ -18,0 +19,2 @@
+ - name: DATASETS_BASED_CONTENT_MAX_BYTES
+ value: {{ .Values.datasetsBased.contentMaxBytes | quote}}
diff --git a/chart/values.yaml b/chart/values.yaml
index 86d15525..30a1914d 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -379,0 +380,4 @@ sizes:
+
+# --- datasets_based ---
+datasetsBased:
+ contentMaxBytes: "10_000_000"
diff --git a/tools/docker-compose-base.yml b/tools/docker-compose-base.yml
index 8d98fc0e..cb87973b 100644
--- a/tools/docker-compose-base.yml
+++ b/tools/docker-compose-base.yml
@@ -28,0 +29 @@ services:
+ DATASETS_BASED_CONTENT_MAX_BYTES: ${DATASETS_BASED_CONTENT_MAX_BYTES-10_000_000}
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
index 5c33cea9..33a4e81a 100644
--- a/tools/docker-compose-dev-base.yml
+++ b/tools/docker-compose-dev-base.yml
@@ -28,0 +29 @@ services:
+ DATASETS_BASED_CONTENT_MAX_BYTES: ${DATASETS_BASED_CONTENT_MAX_BYTES-10_000_000}
diff --git a/workers/datasets_based/README.md b/workers/datasets_based/README.md
index adf39058..04211f8c 100644
--- a/workers/datasets_based/README.md
+++ b/workers/datasets_based/README.md
@@ -14,0 +15 @@ Set environment variables to configure the datasets-based worker (`DATASETS_BASE
+- `DATASETS_BASED_CONTENT_MAX_BYTES`: the maximum size in bytes of the response content computed by a worker (to prevent returning big responses in the REST API). Defaults to `10_000_000`.
diff --git a/workers/datasets_based/src/datasets_based/config.py b/workers/datasets_based/src/datasets_based/config.py
index cedf201e..61f7b60a 100644
--- a/workers/datasets_based/src/datasets_based/config.py
+++ b/workers/datasets_based/src/datasets_based/config.py
@@ -51,0 +52 @@ DATASETS_BASED_HF_DATASETS_CACHE = None
+DATASETS_BASED_CONTENT_MAX_BYTES = 10_000_000
@@ -57,0 +59 @@ class DatasetsBasedConfig:
+ content_max_bytes: int = DATASETS_BASED_CONTENT_MAX_BYTES
@@ -70,0 +73 @@ class DatasetsBasedConfig:
+ content_max_bytes=env.int(name="CONTENT_MAX_BYTES", default=DATASETS_BASED_CONTENT_MAX_BYTES),
diff --git a/workers/datasets_based/src/datasets_based/worker.py b/workers/datasets_based/src/datasets_based/worker.py
index 45c63764..3268c9b0 100644
--- a/workers/datasets_based/src/datasets_based/worker.py
+++ b/workers/datasets_based/src/datasets_based/worker.py
@@ -22,0 +23 @@ from libcommon.simple_cache import (
+from libcommon.utils import orjson_dumps
@@ -24,0 +26,2 @@ from packaging import version
+from datasets_based.config import DatasetsBasedConfig
+
@@ -29,0 +33 @@ GeneralWorkerErrorCode = Literal[
+ "TooBigContentError",
@@ -106,0 +111,13 @@ class NoGitRevisionError(GeneralWorkerError):
+class TooBigContentError(GeneralWorkerError):
+ """Raised when content size in bytes is bigger than the supported value."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(
+ message=message,
+ status_code=HTTPStatus.NOT_IMPLEMENTED,
+ code="TooBigContentError",
+ cause=cause,
+ disclose_cause=False,
+ )
+
+
@@ -142,0 +160 @@ class Worker(ABC):
+ datasets_based_config: DatasetsBasedConfig
@@ -159,0 +178 @@ class Worker(ABC):
+ datasets_based_config: DatasetsBasedConfig,
@@ -169,0 +189 @@ class Worker(ABC):
+ self.datasets_based_config = datasets_based_config
@@ -299,0 +320,7 @@ class Worker(ABC):
+
+ # Validate content size
+ if len(orjson_dumps(content)) > self.datasets_based_config.content_max_bytes:
+ raise TooBigContentError(
+ "The computed response content exceeds the supported size in bytes"
+ f" ({self.datasets_based_config.content_max_bytes})."
+ )
diff --git a/workers/datasets_based/src/datasets_based/worker_factory.py b/workers/datasets_based/src/datasets_based/worker_factory.py
index 3f2ecef3..db40eaf5 100644
--- a/workers/datasets_based/src/datasets_based/worker_factory.py
+++ b/workers/datasets_based/src/datasets_based/worker_factory.py
@@ -73 +73,4 @@ class WorkerFactory(BaseWorkerFactory):
- job_info=job_info, common_config=self.app_config.common, processing_step=processing_step
+ job_info=job_info,
+ common_config=self.app_config.common,
+ datasets_based_config=self.app_config.datasets_based,
+ processing_step=processing_step,
@@ -77 +80,4 @@ class WorkerFactory(BaseWorkerFactory):
- job_info=job_info, common_config=self.app_config.common, processing_step=processing_step
+ job_info=job_info,
+ common_config=self.app_config.common,
+ datasets_based_config=self.app_config.datasets_based,
+ processing_step=processing_step,
@@ -81 +87,4 @@ class WorkerFactory(BaseWorkerFactory):
- job_info=job_info, common_config=self.app_config.common, processing_step=processing_step
+ job_info=job_info,
+ common_config=self.app_config.common,
+ datasets_based_config=self.app_config.datasets_based,
+ processing_step=processing_step,
diff --git a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py b/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
index fcac9b46..973bc676 100644
--- a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
+++ b/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
@@ -31 +31,6 @@ class DatasetsBasedWorker(Worker):
- super().__init__(job_info=job_info, common_config=app_config.common, processing_step=processing_step)
+ super().__init__(
+ job_info=job_info,
+ common_config=app_config.common,
+ datasets_based_config=app_config.datasets_based,
+ processing_step=processing_step,
+ )
diff --git a/workers/datasets_based/tests/conftest.py b/workers/datasets_based/tests/conftest.py
index 476ecb04..669d7e44 100644
--- a/workers/datasets_based/tests/conftest.py
+++ b/workers/datasets_based/tests/conftest.py
@@ -53,0 +54 @@ def set_env_vars(datasets_cache_directory: Path, modules_cache_directory: Path)
+ mp.setenv("DATASETS_BASED_CONTENT_MAX_BYTES", "10_000_000")
diff --git a/workers/datasets_based/tests/test_worker.py b/workers/datasets_based/tests/test_worker.py
index ceb3913f..63eb2f71 100644
--- a/workers/datasets_based/tests/test_worker.py
+++ b/workers/datasets_based/tests/test_worker.py
@@ -11 +11 @@ from libcommon.simple_cache import SplitFullName, upsert_response
-from datasets_based.config import AppConfig
+from datasets_based.config import AppConfig, DatasetsBasedConfig
@@ -76,0 +77 @@ def test_compare_major_version(
+ datasets_based_config=DatasetsBasedConfig(),
@@ -193,0 +195 @@ def test_should_skip_job(
+ datasets_based_config=DatasetsBasedConfig(),
@@ -233,0 +236 @@ def test_check_type(
+ datasets_based_config=DatasetsBasedConfig(),
@@ -257,0 +261 @@ def test_check_type(
+ datasets_based_config=DatasetsBasedConfig(),
@@ -282,0 +287 @@ def test_create_children_jobs() -> None:
+ datasets_based_config=DatasetsBasedConfig(),
diff --git a/workers/datasets_based/tests/test_worker_loop.py b/workers/datasets_based/tests/test_worker_loop.py
index 8ae1601b..696db769 100644
--- a/workers/datasets_based/tests/test_worker_loop.py
+++ b/workers/datasets_based/tests/test_worker_loop.py
@@ -8 +8 @@ from libcommon.queue import Queue
-from datasets_based.config import AppConfig, WorkerLoopConfig
+from datasets_based.config import AppConfig, DatasetsBasedConfig, WorkerLoopConfig
@@ -39,0 +40 @@ class DummyWorkerFactory(BaseWorkerFactory):
+ self.datasets_based_config = DatasetsBasedConfig()
@@ -43 +44,6 @@ class DummyWorkerFactory(BaseWorkerFactory):
- return DummyWorker(job_info=job_info, common_config=self.common_config, processing_step=self.processing_step)
+ return DummyWorker(
+ job_info=job_info,
+ common_config=self.common_config,
+ datasets_based_config=self.datasets_based_config,
+ processing_step=self.processing_step,
+ )
diff --git a/workers/datasets_based/tests/workers/test__datasets_based_worker.py b/workers/datasets_based/tests/workers/test__datasets_based_worker.py
index 043a248b..1f82113d 100644
--- a/workers/datasets_based/tests/workers/test__datasets_based_worker.py
+++ b/workers/datasets_based/tests/workers/test__datasets_based_worker.py
@@ -3,0 +4 @@
+from dataclasses import replace
@@ -4,0 +6 @@ from datetime import datetime
+from http import HTTPStatus
@@ -11,0 +14 @@ from libcommon.queue import Priority
+from libcommon.simple_cache import get_response
@@ -16 +19 @@ from datasets_based.workers._datasets_based_worker import DatasetsBasedWorker
-from ..fixtures.hub import get_default_config_split
+from ..fixtures.hub import HubDatasets, get_default_config_split
@@ -34 +37 @@ class DummyWorker(DatasetsBasedWorker):
- return {}
+ return {"col1": "a" * 200}
@@ -145,0 +149,16 @@ def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) ->
+
+
+def test_process_big_content(hub_datasets: HubDatasets, app_config: AppConfig) -> None:
+ dataset, config, split = get_default_config_split(hub_datasets["big"]["name"])
+ worker = get_worker(
+ dataset,
+ config,
+ split,
+ app_config=replace(app_config, datasets_based=replace(app_config.datasets_based, content_max_bytes=10)),
+ )
+
+ assert worker.process() is False
+ cached_response = get_response(kind=worker.processing_step.cache_kind, dataset=dataset, config=config, split=split)
+
+ assert cached_response["http_status"] == HTTPStatus.NOT_IMPLEMENTED
+ assert cached_response["error_code"] == "TooBigContentError"
diff --git a/workers/datasets_based/tests/workers/test_dataset_info.py b/workers/datasets_based/tests/workers/test_dataset_info.py
index ba473968..34227830 100644
--- a/workers/datasets_based/tests/workers/test_dataset_info.py
+++ b/workers/datasets_based/tests/workers/test_dataset_info.py
@@ -38,0 +39 @@ def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> Data
+ datasets_based_config=app_config.datasets_based,
diff --git a/workers/datasets_based/tests/workers/test_parquet.py b/workers/datasets_based/tests/workers/test_parquet.py
index 35e6fc69..7df0715d 100644
--- a/workers/datasets_based/tests/workers/test_parquet.py
+++ b/workers/datasets_based/tests/workers/test_parquet.py
@@ -38,0 +39 @@ def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> Parq
+ datasets_based_config=app_config.datasets_based,
diff --git a/workers/datasets_based/tests/workers/test_sizes.py b/workers/datasets_based/tests/workers/test_sizes.py
index bdcf7b77..1d773dba 100644
--- a/workers/datasets_based/tests/workers/test_sizes.py
+++ b/workers/datasets_based/tests/workers/test_sizes.py
@@ -38,0 +39 @@ def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> Size
+ datasets_based_config=app_config.datasets_based,
|
|
7121f536c693bd4cc06f05929efde6872424ad37
|
Rémy
| 2023-02-09T12:47:35 |
ci: use shared action to publish helm chart (#799)
|
diff --git a/.github/workflows/publish-helm.yml b/.github/workflows/publish-helm.yml
index dc643bf1..fe9b8997 100644
--- a/.github/workflows/publish-helm.yml
+++ b/.github/workflows/publish-helm.yml
@@ -15,3 +14,0 @@ jobs:
- env:
- HELM_REPO_USERNAME: ${{ secrets.REGISTRY_USERNAME }}
- HELM_REPO_PASSWORD: ${{ secrets.REGISTRY_PASSWORD }}
@@ -22,7 +19,2 @@ jobs:
- - name: Install Helm
- run: |
- curl https://raw.githubusercontent.com/helm/helm/main/scripts/get-helm-3 | bash
- helm plugin install https://github.com/chartmuseum/helm-push
-
- - name: Tailscale
- uses: tailscale/github-action@v1
+ - name: Helm Publish
+ uses: huggingface/helm-publish-action@latest
@@ -30,11 +22,5 @@ jobs:
- authkey: ${{ secrets.TAILSCALE_AUTHKEY }}
-
- - name: Add repository
- run: |
- helm repo add charts ${{ secrets.REGISTRY_URL }}/chartrepo/charts
-
- - name: Publish
- run: |
- cd chart
- helm dependencies update
- helm cm-push . charts
+ workingDirectory: chart
+ tailscaleKey: ${{ secrets.TAILSCALE_AUTHKEY }}
+ repository: ${{ secrets.REGISTRY_URL }}/chartrepo/charts
+ username: ${{ secrets.REGISTRY_USERNAME }}
+ password: ${{ secrets.REGISTRY_PASSWORD }}
|
|
447a126911ddd94346a5daa8b76a61d1f9949201
|
Rémy
| 2023-02-09T08:33:14 |
feat: allow to use http instead of https (#798)
|
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index 0a7e2486..3efacf03 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -182,0 +183,11 @@ Datasets Server base url
+{{/*
+Return the ingress scheme
+*/}}
+{{- define "datasetsServer.ingress.scheme" -}}
+{{- if .Values.global.huggingface.ingress.ssl -}}
+https://
+{{- else -}}
+http://
+{{- end -}}
+{{- end -}}
+
@@ -187 +198 @@ The assets base URL
-{{- printf "https://%s/assets" (include "datasetsServer.ingress.hostname" .) }}
+{{- printf "%s%s/assets" (include "datasetsServer.ingress.scheme" .) (include "datasetsServer.ingress.hostname" .) }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 0b3b9cf3..86d15525 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -10,0 +11 @@ global:
+ ssl: true
@@ -59 +60 @@ secrets:
- fromSecret: false
+ fromSecret: true
@@ -61 +62 @@ secrets:
- value: hf_app
+ value: ""
|
|
eb78f501a35b7b7f9f5043238ed30198d38d3a14
|
Rémy
| 2023-02-08T15:46:44 |
feat: remove job after 5 minutes (#788)
|
diff --git a/chart/templates/jobs/mongodb-migration/job.yaml b/chart/templates/jobs/mongodb-migration/job.yaml
index 68ea3bd5..a5dc5343 100644
--- a/chart/templates/jobs/mongodb-migration/job.yaml
+++ b/chart/templates/jobs/mongodb-migration/job.yaml
@@ -15,0 +16 @@ spec:
+ ttlSecondsAfterFinished: 300
|
|
3ca72c65f658e66333abf6c1540653a032287864
|
Sylvain Lesage
| 2023-02-08T14:49:31 |
feat: 🎸 add logs when an unexpected error occurs (#789)
|
diff --git a/services/admin/src/admin/routes/backfill.py b/services/admin/src/admin/routes/backfill.py
index 1b449929..95ca237d 100644
--- a/services/admin/src/admin/routes/backfill.py
+++ b/services/admin/src/admin/routes/backfill.py
@@ -57,2 +57,2 @@ def create_backfill_endpoint(
- except Exception:
- return get_json_admin_error_response(UnexpectedError("Unexpected error."), max_age=0)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=0)
diff --git a/services/admin/src/admin/routes/cache_reports.py b/services/admin/src/admin/routes/cache_reports.py
index 4ed5c721..d214383c 100644
--- a/services/admin/src/admin/routes/cache_reports.py
+++ b/services/admin/src/admin/routes/cache_reports.py
@@ -49,2 +49,2 @@ def create_cache_reports_endpoint(
- except Exception:
- return get_json_admin_error_response(UnexpectedError("Unexpected error."), max_age=max_age)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age)
diff --git a/services/admin/src/admin/routes/cache_reports_with_content.py b/services/admin/src/admin/routes/cache_reports_with_content.py
index 45798df8..b38379b1 100644
--- a/services/admin/src/admin/routes/cache_reports_with_content.py
+++ b/services/admin/src/admin/routes/cache_reports_with_content.py
@@ -55,2 +55,2 @@ def create_cache_reports_with_content_endpoint(
- except Exception:
- return get_json_admin_error_response(UnexpectedError("Unexpected error."), max_age=max_age)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age)
diff --git a/services/admin/src/admin/routes/cancel_jobs.py b/services/admin/src/admin/routes/cancel_jobs.py
index 445e5db2..059a9ca7 100644
--- a/services/admin/src/admin/routes/cancel_jobs.py
+++ b/services/admin/src/admin/routes/cancel_jobs.py
@@ -40,2 +40,2 @@ def create_cancel_jobs_endpoint(
- except Exception:
- return get_json_admin_error_response(UnexpectedError("Unexpected error."), max_age=0)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=0)
diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py
index 45ea1279..8f053829 100644
--- a/services/admin/src/admin/routes/force_refresh.py
+++ b/services/admin/src/admin/routes/force_refresh.py
@@ -64,2 +64,2 @@ def create_force_refresh_endpoint(
- except Exception:
- return get_json_admin_error_response(UnexpectedError("Unexpected error."), max_age=0)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=0)
diff --git a/services/admin/src/admin/routes/jobs_duration.py b/services/admin/src/admin/routes/jobs_duration.py
index 548a685e..3dc5b6d4 100644
--- a/services/admin/src/admin/routes/jobs_duration.py
+++ b/services/admin/src/admin/routes/jobs_duration.py
@@ -39,2 +39,2 @@ def create_jobs_duration_per_dataset_endpoint(
- except Exception:
- return get_json_admin_error_response(UnexpectedError("Unexpected error."), max_age=max_age)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age)
diff --git a/services/admin/src/admin/routes/pending_jobs.py b/services/admin/src/admin/routes/pending_jobs.py
index dd154e7b..451e66d9 100644
--- a/services/admin/src/admin/routes/pending_jobs.py
+++ b/services/admin/src/admin/routes/pending_jobs.py
@@ -42,2 +42,2 @@ def create_pending_jobs_endpoint(
- except Exception:
- return get_json_admin_error_response(UnexpectedError("Unexpected error."), max_age=max_age)
+ except Exception as e:
+ return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age)
diff --git a/services/admin/src/admin/utils.py b/services/admin/src/admin/utils.py
index 801fa0ba..5fce2a78 100644
--- a/services/admin/src/admin/utils.py
+++ b/services/admin/src/admin/utils.py
@@ -3,0 +4 @@
+import logging
@@ -60,2 +61,6 @@ class UnexpectedError(AdminCustomError):
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "UnexpectedError")
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "UnexpectedError", cause)
+ if cause:
+ logging.exception(message, exc_info=cause)
+ else:
+ logging.exception(message)
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 1ca2f95f..bd3c80d0 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -59,2 +59,2 @@ def create_valid_endpoint(
- except Exception:
- return get_json_api_error_response(UnexpectedError("Unexpected error."), max_age=max_age_short)
+ except Exception as e:
+ return get_json_api_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age_short)
@@ -86,2 +86,2 @@ def create_is_valid_endpoint(
- except Exception:
- return get_json_api_error_response(error=UnexpectedError("Unexpected error."), max_age=max_age_short)
+ except Exception as e:
+ return get_json_api_error_response(error=UnexpectedError("Unexpected error.", e), max_age=max_age_short)
diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py
index 42715ec5..cc11ee6c 100644
--- a/services/api/src/api/routes/webhook.py
+++ b/services/api/src/api/routes/webhook.py
@@ -109 +109,2 @@ def create_webhook_endpoint(
- except Exception:
+ except Exception as e:
+ logging.exception("Unexpected error", exc_info=e)
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index 293d7278..2800aee1 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -3,0 +4 @@
+import logging
@@ -61,0 +63 @@ class UnexpectedError(ApiCustomError):
+ logging.error(message, exc_info=cause)
diff --git a/workers/datasets_based/src/datasets_based/worker.py b/workers/datasets_based/src/datasets_based/worker.py
index 73f5a765..45c63764 100644
--- a/workers/datasets_based/src/datasets_based/worker.py
+++ b/workers/datasets_based/src/datasets_based/worker.py
@@ -108 +108 @@ class UnexpectedError(GeneralWorkerError):
- """Raised when the response for the split has not been found."""
+ """Raised when the worker raised an unexpected error."""
@@ -117,0 +118 @@ class UnexpectedError(GeneralWorkerError):
+ logging.error(message, exc_info=cause)
|
|
4ce53338431558945d670ee4c51497680ece5b23
|
Sylvain Lesage
| 2023-02-08T11:57:23 |
Fix dockerfiles (#787)
|
diff --git a/jobs/mongodb_migration/Dockerfile b/jobs/mongodb_migration/Dockerfile
index beaa8270..3b83aa7e 100644
--- a/jobs/mongodb_migration/Dockerfile
+++ b/jobs/mongodb_migration/Dockerfile
@@ -8 +8 @@ ENV PYTHONFAULTHANDLER=1 \
- PIP_NO_CACHE_DIR=off \
+ PIP_NO_CACHE_DIR=1 \
@@ -14 +14,2 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VIRTUALENVS_IN_PROJECT=true
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
@@ -20,2 +21,2 @@ RUN apt-get update \
-RUN pip install -U --no-cache-dir pip
-RUN pip install --no-cache-dir "poetry==$POETRY_VERSION"
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
diff --git a/services/admin/Dockerfile b/services/admin/Dockerfile
index 9ee1f079..1ce73069 100644
--- a/services/admin/Dockerfile
+++ b/services/admin/Dockerfile
@@ -8 +8 @@ ENV PYTHONFAULTHANDLER=1 \
- PIP_NO_CACHE_DIR=off \
+ PIP_NO_CACHE_DIR=1 \
@@ -14 +14,2 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VIRTUALENVS_IN_PROJECT=true
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
@@ -20,2 +21,2 @@ RUN apt-get update \
-RUN pip install -U --no-cache-dir pip
-RUN pip install --no-cache-dir "poetry==$POETRY_VERSION"
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
diff --git a/services/api/Dockerfile b/services/api/Dockerfile
index 51dcd060..52483230 100644
--- a/services/api/Dockerfile
+++ b/services/api/Dockerfile
@@ -8 +8 @@ ENV PYTHONFAULTHANDLER=1 \
- PIP_NO_CACHE_DIR=off \
+ PIP_NO_CACHE_DIR=1 \
@@ -14 +14,2 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VIRTUALENVS_IN_PROJECT=true
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
@@ -20,2 +21,2 @@ RUN apt-get update \
-RUN pip install -U --no-cache-dir pip
-RUN pip install --no-cache-dir "poetry==$POETRY_VERSION"
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
diff --git a/tools/Docker.mk b/tools/Docker.mk
index ae004196..b1581f06 100644
--- a/tools/Docker.mk
+++ b/tools/Docker.mk
@@ -7 +7 @@ up:
- docker compose -f $(DOCKER_COMPOSE) up -d --force-recreate --remove-orphans --renew-anon-volumes
+ docker compose -f $(DOCKER_COMPOSE) up -d --build --force-recreate --remove-orphans --renew-anon-volumes
diff --git a/workers/datasets_based/Dockerfile b/workers/datasets_based/Dockerfile
index 0f496ac9..80ac0507 100644
--- a/workers/datasets_based/Dockerfile
+++ b/workers/datasets_based/Dockerfile
@@ -8 +8 @@ ENV PYTHONFAULTHANDLER=1 \
- PIP_NO_CACHE_DIR=off \
+ PIP_NO_CACHE_DIR=1 \
@@ -14 +14,2 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VIRTUALENVS_IN_PROJECT=true
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
@@ -23,2 +24,2 @@ RUN apt-get update \
-RUN pip install -U --no-cache-dir pip
-RUN pip install --no-cache-dir "poetry==$POETRY_VERSION"
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
diff --git a/workers/datasets_based/dev.Dockerfile b/workers/datasets_based/dev.Dockerfile
index 2eea73f5..7b9d54bf 100644
--- a/workers/datasets_based/dev.Dockerfile
+++ b/workers/datasets_based/dev.Dockerfile
@@ -8 +8 @@ ENV PYTHONFAULTHANDLER=1 \
- PIP_NO_CACHE_DIR=off \
+ PIP_NO_CACHE_DIR=1 \
@@ -14 +14,2 @@ ENV PYTHONFAULTHANDLER=1 \
- POETRY_VIRTUALENVS_IN_PROJECT=true
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
@@ -23,2 +24,2 @@ RUN apt-get update \
-RUN pip install -U --no-cache-dir pip
-RUN pip install --no-cache-dir "poetry==$POETRY_VERSION"
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
|
|
d00a6201af457f326b1eaef30fa02ca460842bf7
|
Sylvain Lesage
| 2023-02-08T11:56:14 |
ci: 🎡 run e2e tests only once for a push or pull-request (#786)
|
diff --git a/.github/workflows/e2e.yml b/.github/workflows/e2e.yml
index 78089370..8ba34a94 100644
--- a/.github/workflows/e2e.yml
+++ b/.github/workflows/e2e.yml
@@ -11,0 +12,3 @@ on:
+ - 'libs/**'
+ - 'services/**'
+ - 'workers/**'
@@ -20,0 +24,3 @@ on:
+ - 'libs/**'
+ - 'services/**'
+ - 'workers/**'
diff --git a/.github/workflows/l-libcommon.yml b/.github/workflows/l-libcommon.yml
index 476a5211..3ba6a069 100644
--- a/.github/workflows/l-libcommon.yml
+++ b/.github/workflows/l-libcommon.yml
@@ -32,2 +31,0 @@ jobs:
- e2e-tests:
- uses: ./.github/workflows/_e2e_tests.yml
diff --git a/.github/workflows/s-admin.yml b/.github/workflows/s-admin.yml
index 4974a998..7dd4c411 100644
--- a/.github/workflows/s-admin.yml
+++ b/.github/workflows/s-admin.yml
@@ -34,2 +33,0 @@ jobs:
- e2e-tests:
- uses: ./.github/workflows/_e2e_tests.yml
diff --git a/.github/workflows/s-api.yml b/.github/workflows/s-api.yml
index 5d4e69d9..972d6639 100644
--- a/.github/workflows/s-api.yml
+++ b/.github/workflows/s-api.yml
@@ -34,2 +33,0 @@ jobs:
- e2e-tests:
- uses: ./.github/workflows/_e2e_tests.yml
diff --git a/.github/workflows/w-datasets_based.yml b/.github/workflows/w-datasets_based.yml
index 97b9c30a..2082de4a 100644
--- a/.github/workflows/w-datasets_based.yml
+++ b/.github/workflows/w-datasets_based.yml
@@ -38,2 +37,0 @@ jobs:
- e2e-tests:
- uses: ./.github/workflows/_e2e_tests.yml
|
|
556c2370f44a292dd9018962ae99a9ef946e9098
|
Jatin Kumar
| 2023-02-08T10:41:39 |
Remove first rows fallback variable (#771)
|
diff --git a/chart/templates/worker/first-rows/_container.tpl b/chart/templates/worker/first-rows/_container.tpl
index 7fee8754..9c2cc318 100644
--- a/chart/templates/worker/first-rows/_container.tpl
+++ b/chart/templates/worker/first-rows/_container.tpl
@@ -27,2 +26,0 @@
- - name: FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE
- value: {{ .Values.firstRows.fallbackMaxDatasetSize | quote }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 4c865211..0b3b9cf3 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -285,2 +284,0 @@ firstRows:
- # Max size (in bytes) of the dataset to fallback in normal mode if streaming fails
- fallbackMaxDatasetSize: "100_000_000"
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index 77028267..c668ada2 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -141 +140,0 @@ services:
- FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE: ${FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE-100_000_000}
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index ad95e0ad..fca77917 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -141 +140,0 @@ services:
- FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE: ${FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE-100_000_000}
diff --git a/workers/datasets_based/README.md b/workers/datasets_based/README.md
index 6d8f2fb9..adf39058 100644
--- a/workers/datasets_based/README.md
+++ b/workers/datasets_based/README.md
@@ -42 +41,0 @@ Set environment variables to configure the first rows worker (`FIRST_ROWS_` pref
-- `FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE`: the maximum size in bytes of the dataset to fall back into normal mode if streaming fails. Note that it requires to have the size in the info metadata. Set to `0` to disable the fallback. Defaults to `100_000_000`.
diff --git a/workers/datasets_based/src/datasets_based/config.py b/workers/datasets_based/src/datasets_based/config.py
index 9373befc..cedf201e 100644
--- a/workers/datasets_based/src/datasets_based/config.py
+++ b/workers/datasets_based/src/datasets_based/config.py
@@ -74 +73,0 @@ class DatasetsBasedConfig:
-FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE = 100_000_000
@@ -85 +83,0 @@ class FirstRowsConfig:
- fallback_max_dataset_size: int = FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE
@@ -98,3 +95,0 @@ class FirstRowsConfig:
- fallback_max_dataset_size=env.int(
- name="FALLBACK_MAX_DATASET_SIZE", default=FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE
- ),
diff --git a/workers/datasets_based/src/datasets_based/workers/first_rows.py b/workers/datasets_based/src/datasets_based/workers/first_rows.py
index d95460ac..f4aa20eb 100644
--- a/workers/datasets_based/src/datasets_based/workers/first_rows.py
+++ b/workers/datasets_based/src/datasets_based/workers/first_rows.py
@@ -402 +401,0 @@ def compute_first_rows_response(
- max_size_fallback: Optional[int],
@@ -407,0 +407 @@ def compute_first_rows_response(
+ max_size_fallback: Optional[int] = None,
@@ -638 +637,0 @@ class FirstRowsWorker(DatasetsBasedWorker):
- max_size_fallback=self.first_rows_config.fallback_max_dataset_size,
|
|
5251f9ac2538c9ebe11c42a7731f181c0d02823d
|
Andrea Soria Jimenez
| 2023-02-07T19:56:05 |
Updating docker image hash (#783)
|
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml
index 4d6bbe76..502ebdfc 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/dev.yaml
@@ -29 +29 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -35 +35 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -40 +40 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -46 +46 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 668a0302..b5e0f2f2 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -19 +19 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -25 +25 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -30 +30 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -36 +36 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
diff --git a/chart/values.yaml b/chart/values.yaml
index fab77dc9..4c865211 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -25 +25 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -31 +31 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -36 +36 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
@@ -42 +42 @@ images:
- tag: sha-25ff490
+ tag: sha-27ad2f7
|
|
27ad2f7f555cce350785219e19423e23362c56f6
|
Sylvain Lesage
| 2023-02-07T15:45:39 |
ci: 🎡 the e2e tests must now be run on any code change (#775)
|
diff --git a/.github/workflows/l-libcommon.yml b/.github/workflows/l-libcommon.yml
index 3ba6a069..476a5211 100644
--- a/.github/workflows/l-libcommon.yml
+++ b/.github/workflows/l-libcommon.yml
@@ -31,0 +32,2 @@ jobs:
+ e2e-tests:
+ uses: ./.github/workflows/_e2e_tests.yml
diff --git a/.github/workflows/s-admin.yml b/.github/workflows/s-admin.yml
index 7dd4c411..4974a998 100644
--- a/.github/workflows/s-admin.yml
+++ b/.github/workflows/s-admin.yml
@@ -33,0 +34,2 @@ jobs:
+ e2e-tests:
+ uses: ./.github/workflows/_e2e_tests.yml
diff --git a/.github/workflows/s-api.yml b/.github/workflows/s-api.yml
index 972d6639..5d4e69d9 100644
--- a/.github/workflows/s-api.yml
+++ b/.github/workflows/s-api.yml
@@ -33,0 +34,2 @@ jobs:
+ e2e-tests:
+ uses: ./.github/workflows/_e2e_tests.yml
diff --git a/.github/workflows/w-datasets_based.yml b/.github/workflows/w-datasets_based.yml
index 2082de4a..97b9c30a 100644
--- a/.github/workflows/w-datasets_based.yml
+++ b/.github/workflows/w-datasets_based.yml
@@ -37,0 +38,2 @@ jobs:
+ e2e-tests:
+ uses: ./.github/workflows/_e2e_tests.yml
|
|
763a6a2d727902bf34efcfda2224003b32b8041f
|
Albert Villanova del Moral
| 2023-02-07T14:04:02 |
Make workers' errors derive from WorkerError (#772)
|
diff --git a/workers/datasets_based/src/datasets_based/worker.py b/workers/datasets_based/src/datasets_based/worker.py
index 6c5e968f..73f5a765 100644
--- a/workers/datasets_based/src/datasets_based/worker.py
+++ b/workers/datasets_based/src/datasets_based/worker.py
@@ -25 +25 @@ from packaging import version
-WorkerErrorCode = Literal[
+GeneralWorkerErrorCode = Literal[
@@ -43 +43 @@ class WorkerError(CustomError):
- code: WorkerErrorCode,
+ code: str,
@@ -48 +48 @@ class WorkerError(CustomError):
- message=message, status_code=status_code, code=str(code), cause=cause, disclose_cause=disclose_cause
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
@@ -52 +52,17 @@ class WorkerError(CustomError):
-class ConfigNotFoundError(WorkerError):
+class GeneralWorkerError(WorkerError):
+ """Base class for worker exceptions."""
+
+ def __init__(
+ self,
+ message: str,
+ status_code: HTTPStatus,
+ code: GeneralWorkerErrorCode,
+ cause: Optional[BaseException] = None,
+ disclose_cause: bool = False,
+ ):
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
+
+
+class ConfigNotFoundError(GeneralWorkerError):
@@ -65 +81 @@ class ConfigNotFoundError(WorkerError):
-class SplitNotFoundError(WorkerError):
+class SplitNotFoundError(GeneralWorkerError):
@@ -78 +94 @@ class SplitNotFoundError(WorkerError):
-class NoGitRevisionError(WorkerError):
+class NoGitRevisionError(GeneralWorkerError):
@@ -91 +107 @@ class NoGitRevisionError(WorkerError):
-class UnexpectedError(WorkerError):
+class UnexpectedError(GeneralWorkerError):
diff --git a/workers/datasets_based/src/datasets_based/workers/config_names.py b/workers/datasets_based/src/datasets_based/workers/config_names.py
index 011c8475..8e02ff24 100644
--- a/workers/datasets_based/src/datasets_based/workers/config_names.py
+++ b/workers/datasets_based/src/datasets_based/workers/config_names.py
@@ -10 +9,0 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError
-from libcommon.exceptions import CustomError
@@ -12,0 +12 @@ from libcommon.simple_cache import SplitFullName
+from datasets_based.worker import WorkerError
@@ -18 +18 @@ ConfigNamesWorkerErrorCode = Literal["EmptyDatasetError", "ConfigNamesError"]
-class ConfigNamesWorkerError(CustomError):
+class ConfigNamesWorkerError(WorkerError):
@@ -30 +30 @@ class ConfigNamesWorkerError(CustomError):
- message=message, status_code=status_code, code=str(code), cause=cause, disclose_cause=disclose_cause
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
diff --git a/workers/datasets_based/src/datasets_based/workers/dataset_info.py b/workers/datasets_based/src/datasets_based/workers/dataset_info.py
index a97d81c1..39dcb5fa 100644
--- a/workers/datasets_based/src/datasets_based/workers/dataset_info.py
+++ b/workers/datasets_based/src/datasets_based/workers/dataset_info.py
@@ -9 +8,0 @@ from libcommon.dataset import DatasetNotFoundError
-from libcommon.exceptions import CustomError
@@ -12 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.worker import Worker
+from datasets_based.worker import Worker, WorkerError
@@ -24 +23 @@ class DatasetInfoResponse(TypedDict):
-class DatasetInfoWorkerError(CustomError):
+class DatasetInfoWorkerError(WorkerError):
@@ -35 +34,3 @@ class DatasetInfoWorkerError(CustomError):
- super().__init__(message, status_code, str(code), cause, disclose_cause)
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
diff --git a/workers/datasets_based/src/datasets_based/workers/first_rows.py b/workers/datasets_based/src/datasets_based/workers/first_rows.py
index 4ea3360d..d95460ac 100644
--- a/workers/datasets_based/src/datasets_based/workers/first_rows.py
+++ b/workers/datasets_based/src/datasets_based/workers/first_rows.py
@@ -22 +21,0 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError
-from libcommon.exceptions import CustomError
@@ -29 +28,6 @@ from datasets_based.features import get_cell_value
-from datasets_based.worker import ConfigNotFoundError, JobInfo, SplitNotFoundError
+from datasets_based.worker import (
+ ConfigNotFoundError,
+ JobInfo,
+ SplitNotFoundError,
+ WorkerError,
+)
@@ -45 +49 @@ FirstRowsWorkerErrorCode = Literal[
-class FirstRowsWorkerError(CustomError):
+class FirstRowsWorkerError(WorkerError):
@@ -56 +60,3 @@ class FirstRowsWorkerError(CustomError):
- super().__init__(message, status_code, str(code), cause, disclose_cause)
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
diff --git a/workers/datasets_based/src/datasets_based/workers/parquet.py b/workers/datasets_based/src/datasets_based/workers/parquet.py
index ce9df94a..e156db21 100644
--- a/workers/datasets_based/src/datasets_based/workers/parquet.py
+++ b/workers/datasets_based/src/datasets_based/workers/parquet.py
@@ -9 +8,0 @@ from libcommon.dataset import DatasetNotFoundError
-from libcommon.exceptions import CustomError
@@ -12 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.worker import Worker
+from datasets_based.worker import Worker, WorkerError
@@ -25 +24 @@ class ParquetResponse(TypedDict):
-class ParquetWorkerError(CustomError):
+class ParquetWorkerError(WorkerError):
@@ -36 +35,3 @@ class ParquetWorkerError(CustomError):
- super().__init__(message, status_code, str(code), cause, disclose_cause)
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
diff --git a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py b/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
index 95868bd2..b3523ecc 100644
--- a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
+++ b/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
@@ -28 +27,0 @@ from libcommon.dataset import ask_access
-from libcommon.exceptions import CustomError
@@ -33 +32 @@ from datasets_based.config import AppConfig, ParquetAndDatasetInfoConfig
-from datasets_based.worker import DatasetNotFoundError, JobInfo
+from datasets_based.worker import DatasetNotFoundError, JobInfo, WorkerError
@@ -46 +45 @@ ParquetAndDatasetInfoWorkerErrorCode = Literal[
-class ParquetAndDatasetInfoWorkerError(CustomError):
+class ParquetAndDatasetInfoWorkerError(WorkerError):
@@ -57 +56,3 @@ class ParquetAndDatasetInfoWorkerError(CustomError):
- super().__init__(message, status_code, str(code), cause, disclose_cause)
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
diff --git a/workers/datasets_based/src/datasets_based/workers/sizes.py b/workers/datasets_based/src/datasets_based/workers/sizes.py
index 093afa30..ff450908 100644
--- a/workers/datasets_based/src/datasets_based/workers/sizes.py
+++ b/workers/datasets_based/src/datasets_based/workers/sizes.py
@@ -9 +8,0 @@ from libcommon.dataset import DatasetNotFoundError
-from libcommon.exceptions import CustomError
@@ -12 +11 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.worker import Worker
+from datasets_based.worker import Worker, WorkerError
@@ -58 +57 @@ class SizesResponse(TypedDict):
-class SizesWorkerError(CustomError):
+class SizesWorkerError(WorkerError):
@@ -69 +68,3 @@ class SizesWorkerError(CustomError):
- super().__init__(message, status_code, str(code), cause, disclose_cause)
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
diff --git a/workers/datasets_based/src/datasets_based/workers/split_names.py b/workers/datasets_based/src/datasets_based/workers/split_names.py
index 4a3cfdfd..f196b98c 100644
--- a/workers/datasets_based/src/datasets_based/workers/split_names.py
+++ b/workers/datasets_based/src/datasets_based/workers/split_names.py
@@ -10 +9,0 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError
-from libcommon.exceptions import CustomError
@@ -12,0 +12 @@ from libcommon.simple_cache import SplitFullName
+from datasets_based.worker import WorkerError
@@ -21 +21 @@ SplitNamesWorkerErrorCode = Literal[
-class SplitNamesWorkerError(CustomError):
+class SplitNamesWorkerError(WorkerError):
@@ -33 +33 @@ class SplitNamesWorkerError(CustomError):
- message=message, status_code=status_code, code=str(code), cause=cause, disclose_cause=disclose_cause
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
diff --git a/workers/datasets_based/src/datasets_based/workers/splits.py b/workers/datasets_based/src/datasets_based/workers/splits.py
index 118d1562..fc9a9d71 100644
--- a/workers/datasets_based/src/datasets_based/workers/splits.py
+++ b/workers/datasets_based/src/datasets_based/workers/splits.py
@@ -10 +9,0 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError
-from libcommon.exceptions import CustomError
@@ -12,0 +12 @@ from libcommon.simple_cache import SplitFullName
+from datasets_based.worker import WorkerError
@@ -21 +21 @@ SplitsWorkerErrorCode = Literal[
-class SplitsWorkerError(CustomError):
+class SplitsWorkerError(WorkerError):
@@ -33 +33 @@ class SplitsWorkerError(CustomError):
- message=message, status_code=status_code, code=str(code), cause=cause, disclose_cause=disclose_cause
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
|
|
769bc3aae554ed28d0d49dea314b8a9aad0dd6fa
|
Sylvain Lesage
| 2023-02-07T12:35:50 |
Pass processing step to worker (#779)
|
diff --git a/libs/libcommon/README.md b/libs/libcommon/README.md
index c5996a7d..53027901 100644
--- a/libs/libcommon/README.md
+++ b/libs/libcommon/README.md
@@ -34,10 +33,0 @@ Set environment variables to configure the job queues to precompute API response
-
-## Worker loop configuration
-
-Set environment variables to configure the worker loop that processes the queue.
-
-- `WORKER_LOOP_MAX_DISK_USAGE_PCT`: maximum disk usage of every storage disk in the list (in percentage) to allow a job to start. Set to 0 to disable the test. Defaults to 90.
-- `WORKER_LOOP_MAX_LOAD_PCT`: maximum load of the machine (in percentage: the max between the 1m load and the 5m load divided by the number of CPUs \*100) allowed to start a job. Set to 0 to disable the test. Defaults to 70.
-- `WORKER_LOOP_MAX_MEMORY_PCT`: maximum memory (RAM + SWAP) usage of the machine (in percentage) allowed to start a job. Set to 0 to disable the test. Defaults to 80.
-- `WORKER_LOOP_SLEEP_SECONDS`: wait duration in seconds at each loop iteration before checking if resources are available and processing a job if any is available. Note that the loop doesn't wait just after finishing a job: the next job is immediately processed. Defaults to `15`.
-- `WORKER_LOOP_STORAGE_PATHS`: comma-separated list of paths to check for disk usage. Defaults to empty.
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 68be3939..89597b52 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -7 +7 @@ from dataclasses import dataclass, field
-from typing import List, Optional
+from typing import Optional
@@ -111,31 +110,0 @@ class QueueConfig:
-WORKER_LOOP_MAX_DISK_USAGE_PCT = 90
-WORKER_LOOP_MAX_LOAD_PCT = 70
-WORKER_LOOP_MAX_MEMORY_PCT = 80
-WORKER_LOOP_SLEEP_SECONDS = 15
-
-
-def get_empty_str_list() -> List[str]:
- return []
-
-
-@dataclass
-class WorkerLoopConfig:
- max_disk_usage_pct: int = WORKER_LOOP_MAX_DISK_USAGE_PCT
- max_load_pct: int = WORKER_LOOP_MAX_LOAD_PCT
- max_memory_pct: int = WORKER_LOOP_MAX_MEMORY_PCT
- sleep_seconds: int = WORKER_LOOP_SLEEP_SECONDS
- storage_paths: List[str] = field(default_factory=get_empty_str_list)
-
- @staticmethod
- def from_env() -> "WorkerLoopConfig":
- env = Env(expand_vars=True)
- with env.prefixed("WORKER_LOOP_"):
- return WorkerLoopConfig(
- max_disk_usage_pct=env.int(name="MAX_DISK_USAGE_PCT", default=WORKER_LOOP_MAX_DISK_USAGE_PCT),
- max_load_pct=env.int(name="MAX_LOAD_PCT", default=WORKER_LOOP_MAX_LOAD_PCT),
- max_memory_pct=env.int(name="MAX_MEMORY_PCT", default=WORKER_LOOP_MAX_MEMORY_PCT),
- sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_LOOP_SLEEP_SECONDS),
- storage_paths=env.list(name="STORAGE_PATHS", default=get_empty_str_list()),
- )
-
-
diff --git a/workers/datasets_based/README.md b/workers/datasets_based/README.md
index 786b6ced..6d8f2fb9 100644
--- a/workers/datasets_based/README.md
+++ b/workers/datasets_based/README.md
@@ -72,0 +73,10 @@ See [../../libs/libcommon/README.md](../../libs/libcommon/README.md) for more in
+
+## Worker loop configuration
+
+Set environment variables to configure the worker loop that processes the queue.
+
+- `WORKER_LOOP_MAX_DISK_USAGE_PCT`: maximum disk usage of every storage disk in the list (in percentage) to allow a job to start. Set to 0 to disable the test. Defaults to 90.
+- `WORKER_LOOP_MAX_LOAD_PCT`: maximum load of the machine (in percentage: the max between the 1m load and the 5m load divided by the number of CPUs \*100) allowed to start a job. Set to 0 to disable the test. Defaults to 70.
+- `WORKER_LOOP_MAX_MEMORY_PCT`: maximum memory (RAM + SWAP) usage of the machine (in percentage) allowed to start a job. Set to 0 to disable the test. Defaults to 80.
+- `WORKER_LOOP_SLEEP_SECONDS`: wait duration in seconds at each loop iteration before checking if resources are available and processing a job if any is available. Note that the loop doesn't wait just after finishing a job: the next job is immediately processed. Defaults to `15`.
+- `WORKER_LOOP_STORAGE_PATHS`: comma-separated list of paths to check for disk usage. Defaults to empty.
diff --git a/workers/datasets_based/src/datasets_based/config.py b/workers/datasets_based/src/datasets_based/config.py
index a69b6105..9373befc 100644
--- a/workers/datasets_based/src/datasets_based/config.py
+++ b/workers/datasets_based/src/datasets_based/config.py
@@ -17 +16,0 @@ from libcommon.config import (
- WorkerLoopConfig,
@@ -19,0 +19,31 @@ from libcommon.config import (
+WORKER_LOOP_MAX_DISK_USAGE_PCT = 90
+WORKER_LOOP_MAX_LOAD_PCT = 70
+WORKER_LOOP_MAX_MEMORY_PCT = 80
+WORKER_LOOP_SLEEP_SECONDS = 15
+
+
+def get_empty_str_list() -> List[str]:
+ return []
+
+
+@dataclass
+class WorkerLoopConfig:
+ max_disk_usage_pct: int = WORKER_LOOP_MAX_DISK_USAGE_PCT
+ max_load_pct: int = WORKER_LOOP_MAX_LOAD_PCT
+ max_memory_pct: int = WORKER_LOOP_MAX_MEMORY_PCT
+ sleep_seconds: int = WORKER_LOOP_SLEEP_SECONDS
+ storage_paths: List[str] = field(default_factory=get_empty_str_list)
+
+ @staticmethod
+ def from_env() -> "WorkerLoopConfig":
+ env = Env(expand_vars=True)
+ with env.prefixed("WORKER_LOOP_"):
+ return WorkerLoopConfig(
+ max_disk_usage_pct=env.int(name="MAX_DISK_USAGE_PCT", default=WORKER_LOOP_MAX_DISK_USAGE_PCT),
+ max_load_pct=env.int(name="MAX_LOAD_PCT", default=WORKER_LOOP_MAX_LOAD_PCT),
+ max_memory_pct=env.int(name="MAX_MEMORY_PCT", default=WORKER_LOOP_MAX_MEMORY_PCT),
+ sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_LOOP_SLEEP_SECONDS),
+ storage_paths=env.list(name="STORAGE_PATHS", default=get_empty_str_list()),
+ )
+
+
@@ -87,4 +116,0 @@ PARQUET_AND_DATASET_INFO_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s"
-def get_empty_str_list() -> List[str]:
- return []
-
-
diff --git a/workers/datasets_based/src/datasets_based/main.py b/workers/datasets_based/src/datasets_based/main.py
index 7c91470f..d5608f5d 100644
--- a/workers/datasets_based/src/datasets_based/main.py
+++ b/workers/datasets_based/src/datasets_based/main.py
@@ -7 +7 @@ from datasets_based.config import AppConfig
-from datasets_based.worker_factory import DatasetBasedWorkerFactory
+from datasets_based.worker_factory import WorkerFactory
@@ -13 +13 @@ if __name__ == "__main__":
- worker_factory = DatasetBasedWorkerFactory(app_config=app_config)
+ worker_factory = WorkerFactory(app_config=app_config)
diff --git a/workers/datasets_based/src/datasets_based/worker.py b/workers/datasets_based/src/datasets_based/worker.py
index 19f65a34..6c5e968f 100644
--- a/workers/datasets_based/src/datasets_based/worker.py
+++ b/workers/datasets_based/src/datasets_based/worker.py
@@ -413,15 +412,0 @@ class Worker(ABC):
-
-
-class WorkerFactory(ABC):
- """
- Base class for worker factories. A worker factory is a class that creates a worker.
-
- It cannot be instantiated directly, but must be subclassed.
- """
-
- def create_worker(self, job_info: JobInfo) -> Worker:
- return self._create_worker(job_info=job_info)
-
- @abstractmethod
- def _create_worker(self, job_info: JobInfo) -> Worker:
- pass
diff --git a/workers/datasets_based/src/datasets_based/worker_factory.py b/workers/datasets_based/src/datasets_based/worker_factory.py
index ca505b9c..3f2ecef3 100644
--- a/workers/datasets_based/src/datasets_based/worker_factory.py
+++ b/workers/datasets_based/src/datasets_based/worker_factory.py
@@ -3,0 +4,2 @@
+from abc import ABC, abstractmethod
+
@@ -9 +11 @@ from datasets_based.config import (
-from datasets_based.worker import JobInfo, Worker, WorkerFactory
+from datasets_based.worker import JobInfo, Worker
@@ -20 +22,16 @@ from datasets_based.workers.splits import SplitsWorker
-class DatasetBasedWorkerFactory(WorkerFactory):
+class BaseWorkerFactory(ABC):
+ """
+ Base class for worker factories. A worker factory is a class that creates a worker.
+
+ It cannot be instantiated directly, but must be subclassed.
+ """
+
+ def create_worker(self, job_info: JobInfo) -> Worker:
+ return self._create_worker(job_info=job_info)
+
+ @abstractmethod
+ def _create_worker(self, job_info: JobInfo) -> Worker:
+ pass
+
+
+class WorkerFactory(BaseWorkerFactory):
@@ -25,0 +43,7 @@ class DatasetBasedWorkerFactory(WorkerFactory):
+ try:
+ processing_step = self.app_config.processing_graph.graph.get_step_by_job_type(job_type)
+ except ValueError as e:
+ raise ValueError(
+ f"Unsupported job type: '{job_type}'. The job types declared in the processing graph are:"
+ f" {[step.job_type for step in self.app_config.processing_graph.graph.steps.values()]}"
+ ) from e
@@ -27 +51 @@ class DatasetBasedWorkerFactory(WorkerFactory):
- return ConfigNamesWorker(job_info=job_info, app_config=self.app_config)
+ return ConfigNamesWorker(job_info=job_info, app_config=self.app_config, processing_step=processing_step)
@@ -29 +53 @@ class DatasetBasedWorkerFactory(WorkerFactory):
- return SplitNamesWorker(job_info=job_info, app_config=self.app_config)
+ return SplitNamesWorker(job_info=job_info, app_config=self.app_config, processing_step=processing_step)
@@ -31 +55 @@ class DatasetBasedWorkerFactory(WorkerFactory):
- return SplitsWorker(job_info=job_info, app_config=self.app_config)
+ return SplitsWorker(job_info=job_info, app_config=self.app_config, processing_step=processing_step)
@@ -32,0 +57 @@ class DatasetBasedWorkerFactory(WorkerFactory):
+ first_rows_config = FirstRowsConfig.from_env()
@@ -34 +59,4 @@ class DatasetBasedWorkerFactory(WorkerFactory):
- job_info=job_info, app_config=self.app_config, first_rows_config=FirstRowsConfig.from_env()
+ job_info=job_info,
+ app_config=self.app_config,
+ processing_step=processing_step,
+ first_rows_config=first_rows_config,
@@ -39,0 +68 @@ class DatasetBasedWorkerFactory(WorkerFactory):
+ processing_step=processing_step,
@@ -43 +72,3 @@ class DatasetBasedWorkerFactory(WorkerFactory):
- return ParquetWorker(job_info=job_info, app_config=self.app_config)
+ return ParquetWorker(
+ job_info=job_info, common_config=self.app_config.common, processing_step=processing_step
+ )
@@ -45 +76,3 @@ class DatasetBasedWorkerFactory(WorkerFactory):
- return DatasetInfoWorker(job_info=job_info, app_config=self.app_config)
+ return DatasetInfoWorker(
+ job_info=job_info, common_config=self.app_config.common, processing_step=processing_step
+ )
@@ -47 +80,3 @@ class DatasetBasedWorkerFactory(WorkerFactory):
- return SizesWorker(job_info=job_info, app_config=self.app_config)
+ return SizesWorker(
+ job_info=job_info, common_config=self.app_config.common, processing_step=processing_step
+ )
diff --git a/workers/datasets_based/src/datasets_based/worker_loop.py b/workers/datasets_based/src/datasets_based/worker_loop.py
index 0c27d5c3..11354557 100644
--- a/workers/datasets_based/src/datasets_based/worker_loop.py
+++ b/workers/datasets_based/src/datasets_based/worker_loop.py
@@ -9 +8,0 @@ from dataclasses import dataclass
-from libcommon.config import WorkerLoopConfig
@@ -13 +12,2 @@ from psutil import cpu_count, disk_usage, getloadavg, swap_memory, virtual_memor
-from datasets_based.worker import WorkerFactory
+from datasets_based.config import WorkerLoopConfig
+from datasets_based.worker_factory import BaseWorkerFactory
@@ -34 +34 @@ class WorkerLoop:
- worker_factory: WorkerFactory
+ worker_factory: BaseWorkerFactory
diff --git a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py b/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
index 8e060272..fcac9b46 100644
--- a/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
+++ b/workers/datasets_based/src/datasets_based/workers/_datasets_based_worker.py
@@ -12,0 +13 @@ import datasets.config
+from libcommon.processing_graph import ProcessingStep
@@ -29,9 +30 @@ class DatasetsBasedWorker(Worker):
- def __init__(self, job_info: JobInfo, app_config: AppConfig) -> None:
- job_type = job_info["type"]
- try:
- processing_step = app_config.processing_graph.graph.get_step_by_job_type(job_type)
- except ValueError as e:
- raise ValueError(
- f"Unsupported job type: '{job_type}'. The job types declared in the processing graph are:"
- f" {[step.job_type for step in app_config.processing_graph.graph.steps.values()]}"
- ) from e
+ def __init__(self, job_info: JobInfo, app_config: AppConfig, processing_step: ProcessingStep) -> None:
diff --git a/workers/datasets_based/src/datasets_based/workers/dataset_info.py b/workers/datasets_based/src/datasets_based/workers/dataset_info.py
index 62555120..a97d81c1 100644
--- a/workers/datasets_based/src/datasets_based/workers/dataset_info.py
+++ b/workers/datasets_based/src/datasets_based/workers/dataset_info.py
@@ -12,2 +12 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.config import AppConfig
-from datasets_based.worker import JobInfo, Worker
+from datasets_based.worker import Worker
@@ -97,11 +95,0 @@ class DatasetInfoWorker(Worker):
- def __init__(self, job_info: JobInfo, app_config: AppConfig) -> None:
- job_type = job_info["type"]
- try:
- processing_step = app_config.processing_graph.graph.get_step_by_job_type(job_type)
- except ValueError as e:
- raise ValueError(
- f"Unsupported job type: '{job_type}'. The job types declared in the processing graph are:"
- f" {[step.job_type for step in app_config.processing_graph.graph.steps.values()]}"
- ) from e
- super().__init__(job_info=job_info, common_config=app_config.common, processing_step=processing_step)
-
diff --git a/workers/datasets_based/src/datasets_based/workers/first_rows.py b/workers/datasets_based/src/datasets_based/workers/first_rows.py
index 6c657913..4ea3360d 100644
--- a/workers/datasets_based/src/datasets_based/workers/first_rows.py
+++ b/workers/datasets_based/src/datasets_based/workers/first_rows.py
@@ -22,0 +23 @@ from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
@@ -610,2 +611,8 @@ class FirstRowsWorker(DatasetsBasedWorker):
- def __init__(self, job_info: JobInfo, app_config: AppConfig, first_rows_config: FirstRowsConfig) -> None:
- super().__init__(job_info=job_info, app_config=app_config)
+ def __init__(
+ self,
+ job_info: JobInfo,
+ app_config: AppConfig,
+ processing_step: ProcessingStep,
+ first_rows_config: FirstRowsConfig,
+ ) -> None:
+ super().__init__(job_info=job_info, app_config=app_config, processing_step=processing_step)
diff --git a/workers/datasets_based/src/datasets_based/workers/parquet.py b/workers/datasets_based/src/datasets_based/workers/parquet.py
index 9c66ebe8..ce9df94a 100644
--- a/workers/datasets_based/src/datasets_based/workers/parquet.py
+++ b/workers/datasets_based/src/datasets_based/workers/parquet.py
@@ -12,2 +12 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.config import AppConfig
-from datasets_based.worker import JobInfo, Worker
+from datasets_based.worker import Worker
@@ -100,11 +98,0 @@ class ParquetWorker(Worker):
- def __init__(self, job_info: JobInfo, app_config: AppConfig) -> None:
- job_type = job_info["type"]
- try:
- processing_step = app_config.processing_graph.graph.get_step_by_job_type(job_type)
- except ValueError as e:
- raise ValueError(
- f"Unsupported job type: '{job_type}'. The job types declared in the processing graph are:"
- f" {[step.job_type for step in app_config.processing_graph.graph.steps.values()]}"
- ) from e
- super().__init__(job_info=job_info, common_config=app_config.common, processing_step=processing_step)
-
diff --git a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py b/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
index 2cccb3d8..95868bd2 100644
--- a/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
+++ b/workers/datasets_based/src/datasets_based/workers/parquet_and_dataset_info.py
@@ -28,0 +29 @@ from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
@@ -614 +615,5 @@ class ParquetAndDatasetInfoWorker(DatasetsBasedWorker):
- self, job_info: JobInfo, app_config: AppConfig, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig
+ self,
+ job_info: JobInfo,
+ app_config: AppConfig,
+ processing_step: ProcessingStep,
+ parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig,
@@ -616 +621 @@ class ParquetAndDatasetInfoWorker(DatasetsBasedWorker):
- super().__init__(job_info=job_info, app_config=app_config)
+ super().__init__(job_info=job_info, app_config=app_config, processing_step=processing_step)
diff --git a/workers/datasets_based/src/datasets_based/workers/sizes.py b/workers/datasets_based/src/datasets_based/workers/sizes.py
index 79ade862..093afa30 100644
--- a/workers/datasets_based/src/datasets_based/workers/sizes.py
+++ b/workers/datasets_based/src/datasets_based/workers/sizes.py
@@ -12,2 +12 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response
-from datasets_based.config import AppConfig
-from datasets_based.worker import JobInfo, Worker
+from datasets_based.worker import Worker
@@ -181,11 +179,0 @@ class SizesWorker(Worker):
- def __init__(self, job_info: JobInfo, app_config: AppConfig) -> None:
- job_type = job_info["type"]
- try:
- processing_step = app_config.processing_graph.graph.get_step_by_job_type(job_type)
- except ValueError as e:
- raise ValueError(
- f"Unsupported job type: '{job_type}'. The job types declared in the processing graph are:"
- f" {[step.job_type for step in app_config.processing_graph.graph.steps.values()]}"
- ) from e
- super().__init__(job_info=job_info, common_config=app_config.common, processing_step=processing_step)
-
diff --git a/workers/datasets_based/tests/test_worker_factory.py b/workers/datasets_based/tests/test_worker_factory.py
index bb8d56e4..07dd5728 100644
--- a/workers/datasets_based/tests/test_worker_factory.py
+++ b/workers/datasets_based/tests/test_worker_factory.py
@@ -11 +11 @@ from datasets_based.worker import JobInfo
-from datasets_based.worker_factory import DatasetBasedWorkerFactory
+from datasets_based.worker_factory import WorkerFactory
@@ -28 +28 @@ def test_create_worker(app_config: AppConfig, job_type: str, expected_worker: Op
- worker_factory = DatasetBasedWorkerFactory(app_config=app_config)
+ worker_factory = WorkerFactory(app_config=app_config)
diff --git a/workers/datasets_based/tests/test_worker_loop.py b/workers/datasets_based/tests/test_worker_loop.py
index f233bac1..8ae1601b 100644
--- a/workers/datasets_based/tests/test_worker_loop.py
+++ b/workers/datasets_based/tests/test_worker_loop.py
@@ -4 +4 @@ import pytest
-from libcommon.config import CommonConfig, QueueConfig, WorkerLoopConfig
+from libcommon.config import CommonConfig, QueueConfig
@@ -8,2 +8,3 @@ from libcommon.queue import Queue
-from datasets_based.config import AppConfig
-from datasets_based.worker import JobInfo, Worker, WorkerFactory
+from datasets_based.config import AppConfig, WorkerLoopConfig
+from datasets_based.worker import JobInfo, Worker
+from datasets_based.worker_factory import BaseWorkerFactory
@@ -36 +37 @@ class DummyWorker(Worker):
-class DummyWorkerFactory(WorkerFactory):
+class DummyWorkerFactory(BaseWorkerFactory):
diff --git a/workers/datasets_based/tests/workers/test__datasets_based_worker.py b/workers/datasets_based/tests/workers/test__datasets_based_worker.py
index 3995803e..043a248b 100644
--- a/workers/datasets_based/tests/workers/test__datasets_based_worker.py
+++ b/workers/datasets_based/tests/workers/test__datasets_based_worker.py
@@ -9,0 +10 @@ import pytest
+from libcommon.processing_graph import ProcessingStep
@@ -53,0 +55,9 @@ def get_worker(
+ processing_step=ProcessingStep(
+ endpoint=DummyWorker.get_job_type(),
+ input_type="split",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_config_names.py b/workers/datasets_based/tests/workers/test_config_names.py
index 3a58469b..02a76739 100644
--- a/workers/datasets_based/tests/workers/test_config_names.py
+++ b/workers/datasets_based/tests/workers/test_config_names.py
@@ -8,0 +9 @@ from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
@@ -33,0 +35,9 @@ def get_worker(
+ processing_step=ProcessingStep(
+ endpoint=ConfigNamesWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_dataset_info.py b/workers/datasets_based/tests/workers/test_dataset_info.py
index d2546ffd..ba473968 100644
--- a/workers/datasets_based/tests/workers/test_dataset_info.py
+++ b/workers/datasets_based/tests/workers/test_dataset_info.py
@@ -7,0 +8 @@ import pytest
+from libcommon.processing_graph import ProcessingStep
@@ -37 +38,10 @@ def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> Data
- app_config=app_config,
+ common_config=app_config.common,
+ processing_step=ProcessingStep(
+ endpoint=DatasetInfoWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_first_rows.py b/workers/datasets_based/tests/workers/test_first_rows.py
index d3c1e283..6416f027 100644
--- a/workers/datasets_based/tests/workers/test_first_rows.py
+++ b/workers/datasets_based/tests/workers/test_first_rows.py
@@ -9,0 +10 @@ from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
@@ -37,0 +39,9 @@ def get_worker(
+ processing_step=ProcessingStep(
+ endpoint=FirstRowsWorker.get_job_type(),
+ input_type="split",
+ requires=None,
+ required_by_dataset_viewer=True,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_parquet.py b/workers/datasets_based/tests/workers/test_parquet.py
index d62f678b..35e6fc69 100644
--- a/workers/datasets_based/tests/workers/test_parquet.py
+++ b/workers/datasets_based/tests/workers/test_parquet.py
@@ -7,0 +8 @@ import pytest
+from libcommon.processing_graph import ProcessingStep
@@ -37 +38,10 @@ def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> Parq
- app_config=app_config,
+ common_config=app_config.common,
+ processing_step=ProcessingStep(
+ endpoint=ParquetWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py b/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
index 696300a2..fec1739a 100644
--- a/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
+++ b/workers/datasets_based/tests/workers/test_parquet_and_dataset_info.py
@@ -11,0 +12 @@ from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
@@ -71,0 +73,9 @@ def get_worker(
+ processing_step=ProcessingStep(
+ endpoint=ParquetAndDatasetInfoWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_sizes.py b/workers/datasets_based/tests/workers/test_sizes.py
index 712a31f1..bdcf7b77 100644
--- a/workers/datasets_based/tests/workers/test_sizes.py
+++ b/workers/datasets_based/tests/workers/test_sizes.py
@@ -7,0 +8 @@ import pytest
+from libcommon.processing_graph import ProcessingStep
@@ -37 +38,10 @@ def get_worker(dataset: str, app_config: AppConfig, force: bool = False) -> Size
- app_config=app_config,
+ common_config=app_config.common,
+ processing_step=ProcessingStep(
+ endpoint=SizesWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_split_names.py b/workers/datasets_based/tests/workers/test_split_names.py
index 0c030ff3..d61ee180 100644
--- a/workers/datasets_based/tests/workers/test_split_names.py
+++ b/workers/datasets_based/tests/workers/test_split_names.py
@@ -8,0 +9 @@ from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
@@ -34,0 +36,9 @@ def get_worker(
+ processing_step=ProcessingStep(
+ endpoint=SplitNamesWorker.get_job_type(),
+ input_type="config",
+ requires=None,
+ required_by_dataset_viewer=False,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
diff --git a/workers/datasets_based/tests/workers/test_splits.py b/workers/datasets_based/tests/workers/test_splits.py
index 66bb9523..b253a278 100644
--- a/workers/datasets_based/tests/workers/test_splits.py
+++ b/workers/datasets_based/tests/workers/test_splits.py
@@ -8,0 +9 @@ from libcommon.exceptions import CustomError
+from libcommon.processing_graph import ProcessingStep
@@ -33,0 +35,9 @@ def get_worker(
+ processing_step=ProcessingStep(
+ endpoint=SplitsWorker.get_job_type(),
+ input_type="dataset",
+ requires=None,
+ required_by_dataset_viewer=True,
+ parent=None,
+ ancestors=[],
+ children=[],
+ ),
|
|
91837d71befc13458b90df29588e614ed2b5665f
|
Albert Villanova del Moral
| 2023-02-07T12:25:40 |
Fix CI mypy error: "WorkerFactory" has no attribute "app_config" (#778)
|
diff --git a/workers/datasets_based/src/datasets_based/worker_loop.py b/workers/datasets_based/src/datasets_based/worker_loop.py
index 3d03d207..0c27d5c3 100644
--- a/workers/datasets_based/src/datasets_based/worker_loop.py
+++ b/workers/datasets_based/src/datasets_based/worker_loop.py
@@ -100 +99,0 @@ class WorkerLoop:
- self.info(f"Using endpoint {self.worker_factory.app_config.common.hf_endpoint}")
|
|
2e920f6f2241aedda244c2802fb53a6b7b0dafb2
|
Quentin Lhoest
| 2023-02-06T13:56:37 |
Use hub-ci locally (#774)
|
diff --git a/Makefile b/Makefile
index b375e873..bc174bb0 100644
--- a/Makefile
+++ b/Makefile
@@ -11 +11 @@ dev-start: export COMPOSE_PROJECT_NAME := dev-datasets-server
-dev-stop: export COMPOSE_PROJECT_NAME = dev-datasets-server
+dev-stop: export COMPOSE_PROJECT_NAME := dev-datasets-server
diff --git a/front/admin_ui/README.md b/front/admin_ui/README.md
index d655de97..2d88a546 100644
--- a/front/admin_ui/README.md
+++ b/front/admin_ui/README.md
@@ -20 +20 @@ To connect to your local DEV endpoint:
-DEV=1 poetry run python app.py
+DEV=1 HF_TOKEN=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 poetry run python app.py
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
index 3650bcfb..5c33cea9 100644
--- a/tools/docker-compose-dev-base.yml
+++ b/tools/docker-compose-dev-base.yml
@@ -6,2 +6,2 @@ services:
- COMMON_HF_ENDPOINT: ${COMMON_HF_ENDPOINT-https://huggingface.co}
- COMMON_HF_TOKEN: ${COMMON_HF_TOKEN-}
+ COMMON_HF_ENDPOINT: ${COMMON_HF_ENDPOINT-https://hub-ci.huggingface.co}
+ COMMON_HF_TOKEN: ${COMMON_HF_TOKEN-hf_app_datasets-server_token}
@@ -10 +10 @@ services:
- HF_ENDPOINT: ${COMMON_HF_ENDPOINT-https://huggingface.co} # see https://github.com/huggingface/datasets/pull/5196#issuecomment-1322191411
+ HF_ENDPOINT: ${COMMON_HF_ENDPOINT-https://hub-ci.huggingface.co} # see https://github.com/huggingface/datasets/pull/5196#issuecomment-1322191411
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 8cbc66ed..ad95e0ad 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -27 +27 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -54 +54 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -83 +83 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -100 +100 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -117 +117 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -135 +135 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -161 +161 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -167 +167 @@ services:
- PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-}
+ PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-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}
@@ -182 +182 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -195 +195 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
@@ -208 +208 @@ services:
- file: docker-compose-base.yml
+ file: docker-compose-dev-base.yml
diff --git a/workers/datasets_based/src/datasets_based/worker.py b/workers/datasets_based/src/datasets_based/worker.py
index d55f8d2e..19f65a34 100644
--- a/workers/datasets_based/src/datasets_based/worker.py
+++ b/workers/datasets_based/src/datasets_based/worker.py
@@ -183,2 +183,2 @@ class Worker(ABC):
- def critical(self, msg: str) -> None:
- self.log(level=logging.CRITICAL, msg=msg)
+ def warning(self, msg: str) -> None:
+ self.log(level=logging.WARNING, msg=msg)
@@ -188,0 +189,3 @@ class Worker(ABC):
+ def critical(self, msg: str) -> None:
+ self.log(level=logging.CRITICAL, msg=msg)
+
diff --git a/workers/datasets_based/src/datasets_based/worker_loop.py b/workers/datasets_based/src/datasets_based/worker_loop.py
index 0c27d5c3..3d03d207 100644
--- a/workers/datasets_based/src/datasets_based/worker_loop.py
+++ b/workers/datasets_based/src/datasets_based/worker_loop.py
@@ -99,0 +100 @@ class WorkerLoop:
+ self.info(f"Using endpoint {self.worker_factory.app_config.common.hf_endpoint}")
|
|
7188c0a4b84b10b88e39ff9f3ad88ae04843402e
|
Sylvain Lesage
| 2023-02-06T11:24:57 |
refactor: 💡 hard-code the value of the fallback (#773)
|
diff --git a/workers/datasets_based/src/datasets_based/workers/first_rows.py b/workers/datasets_based/src/datasets_based/workers/first_rows.py
index 5426a2b3..6c657913 100644
--- a/workers/datasets_based/src/datasets_based/workers/first_rows.py
+++ b/workers/datasets_based/src/datasets_based/workers/first_rows.py
@@ -7,0 +8 @@ import time
+import warnings
@@ -394 +395 @@ def compute_first_rows_response(
- max_size_fallback: int,
+ max_size_fallback: Optional[int],
@@ -421 +422 @@ def compute_first_rows_response(
- max_size_fallback (`int`):
+ max_size_fallback (`int` or `None`): **DEPRECATED**
@@ -422,0 +424 @@ def compute_first_rows_response(
+ This argument is now hard-coded to 100MB, and will be removed in a future version.
@@ -537 +539,10 @@ def compute_first_rows_response(
- if info.size_in_bytes is None or info.size_in_bytes > max_size_fallback:
+ MAX_SIZE_FALLBACK = 100_000_000
+ if max_size_fallback:
+ warnings.warn(
+ (
+ f"The parameter 'max_size_fallback' is deprecated. The hard-coded value `{MAX_SIZE_FALLBACK}`"
+ " will be used instead."
+ ),
+ category=DeprecationWarning,
+ )
+ if info.size_in_bytes is None or info.size_in_bytes > MAX_SIZE_FALLBACK:
|
|
3bb80b725c361bb5072e14ffc82ab4838ad12405
|
Quentin Lhoest
| 2023-02-03T11:09:14 |
Locally use volumes for workers code (#766)
|
diff --git a/DEVELOPER_GUIDE.md b/DEVELOPER_GUIDE.md
index 1073d9cd..dc0be457 100644
--- a/DEVELOPER_GUIDE.md
+++ b/DEVELOPER_GUIDE.md
@@ -16 +16,3 @@ Install docker (see https://docs.docker.com/engine/install/ubuntu/#install-using
-```
+Run the project locally:
+
+```bash
@@ -19,0 +22,9 @@ make start
+Run the project in development mode:
+
+```bash
+make dev-start
+```
+
+In development mode, you don't need to rebuild the docker images to apply a change in a worker.
+You can just restart the worker's docker container and it will apply your changes.
+
@@ -33 +44 @@ then:
-```
+```bash
diff --git a/Makefile b/Makefile
index 153acb7b..b375e873 100644
--- a/Makefile
+++ b/Makefile
@@ -6 +5,0 @@ export PORT_REVERSE_PROXY := 8100
-export COMPOSE_PROJECT_NAME := datasets-server
@@ -8,2 +7,11 @@ export COMPOSE_PROJECT_NAME := datasets-server
-# makefile variables
-DOCKER_COMPOSE := ./tools/docker-compose-datasets-server.yml
+# environment variables per target
+start: export COMPOSE_PROJECT_NAME := datasets-server
+stop: export COMPOSE_PROJECT_NAME := datasets-server
+dev-start: export COMPOSE_PROJECT_NAME := dev-datasets-server
+dev-stop: export COMPOSE_PROJECT_NAME = dev-datasets-server
+
+# makefile variables per target
+start: DOCKER_COMPOSE := ./tools/docker-compose-datasets-server.yml
+stop: DOCKER_COMPOSE := ./tools/docker-compose-datasets-server.yml
+dev-start: DOCKER_COMPOSE := ./tools/docker-compose-dev-datasets-server.yml
+dev-stop: DOCKER_COMPOSE := ./tools/docker-compose-dev-datasets-server.yml
@@ -20,0 +29,8 @@ stop:
+.PHONY: dev-start
+dev-start:
+ MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up
+
+.PHONY: dev-stop
+dev-stop:
+ MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down
+
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
new file mode 100644
index 00000000..3650bcfb
--- /dev/null
+++ b/tools/docker-compose-dev-base.yml
@@ -0,0 +1,33 @@
+version: "3.9"
+services:
+ common:
+ environment:
+ # common
+ COMMON_HF_ENDPOINT: ${COMMON_HF_ENDPOINT-https://huggingface.co}
+ COMMON_HF_TOKEN: ${COMMON_HF_TOKEN-}
+ COMMON_LOG_LEVEL: ${COMMON_LOG_LEVEL-INFO}
+ # huggingface_hub
+ HF_ENDPOINT: ${COMMON_HF_ENDPOINT-https://huggingface.co} # see https://github.com/huggingface/datasets/pull/5196#issuecomment-1322191411
+ # cache
+ CACHE_MONGO_URL: ${CACHE_MONGO_URL-mongodb://mongodb} # use mongo container by default
+ CACHE_MONGO_DATABASE: ${CACHE_MONGO_DATABASE-datasets_server_cache}
+ # queue
+ QUEUE_MAX_JOBS_PER_NAMESPACE: ${QUEUE_MAX_JOBS_PER_NAMESPACE-1}
+ QUEUE_MONGO_URL: ${QUEUE_MONGO_URL-mongodb://mongodb} # use mongo container by default
+ QUEUE_MONGO_DATABASE: ${QUEUE_MONGO_DATABASE-datasets_server_queue}
+ # worker
+ WORKER_LOOP_MAX_DISK_USAGE_PCT: ${WORKER_LOOP_MAX_DISK_USAGE_PCT-90}
+ WORKER_LOOP_MAX_LOAD_PCT: ${WORKER_LOOP_MAX_LOAD_PCT-70}
+ WORKER_LOOP_MAX_MEMORY_PCT: ${WORKER_LOOP_MAX_MEMORY_PCT-80}
+ WORKER_LOOP_SLEEP_SECONDS: ${WORKER_LOOP_SLEEP_SECONDS-15}
+ datasets-worker:
+ extends:
+ service: common
+ environment:
+ # datasets
+ DATASETS_BASED_HF_DATASETS_CACHE: ${HF_DATASETS_CACHE-/datasets-cache}
+ HF_MODULES_CACHE: ${HF_DATASETS_CACHE-/modules-cache}
+ NUMBA_CACHE_DIR: ${NUMBA_CACHE_DIR-/numba-cache}
+ # volumes to local source directory for development
+ volumes:
+ - ../workers/datasets_based/src:/src/workers/datasets_based/src
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
new file mode 100644
index 00000000..8cbc66ed
--- /dev/null
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -0,0 +1,233 @@
+version: "3.9"
+services:
+ reverse-proxy:
+ image: docker.io/nginx:1.20
+ # image: ${IMAGE_REVERSE_PROXY?IMAGE_REVERSE_PROXY env var must be provided}
+ volumes:
+ - ../chart/nginx-templates/:/etc/nginx/templates:ro
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:ro
+ - ../chart/static-files/openapi.json:/static-files/openapi.json:ro
+ ports:
+ - "${PORT_REVERSE_PROXY-8000}:80"
+ environment:
+ ASSETS_DIRECTORY: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ HOST: localhost
+ PORT: 80
+ URL_ADMIN: http://admin:${ADMIN_UVICORN_PORT-8081}
+ URL_API: http://api:${API_UVICORN_PORT-8080}
+ depends_on:
+ - api
+ - admin
+ admin:
+ build:
+ context: ..
+ dockerfile: services/admin/Dockerfile
+ # image: ${IMAGE_SERVICE_ADMIN?IMAGE_SERVICE_ADMIN env var must be provided}
+ extends:
+ file: docker-compose-base.yml
+ service: common
+ environment:
+ # service
+ ADMIN_HF_ORGANIZATION: ${ADMIN_HF_ORGANIZATION-huggingface}
+ ADMIN_CACHE_REPORTS_NUM_RESULTS: ${ADMIN_CACHE_REPORTS_NUM_RESULTS-100}
+ ADMIN_CACHE_REPORTS_WITH_CONTENT_NUM_RESULTS: ${ADMIN_CACHE_REPORTS_WITH_CONTENT_NUM_RESULTS-100}
+ ADMIN_HF_WHOAMI_PATH: ${ADMIN_HF_WHOAMI_PATH-/api/whoami-v2}
+ ADMIN_MAX_AGE: ${ADMIN_MAX_AGE-10}
+ # prometheus
+ PROMETHEUS_MULTIPROC_DIR: ${PROMETHEUS_MULTIPROC_DIR-}
+ # uvicorn
+ ADMIN_UVICORN_HOSTNAME: 0.0.0.0 # required for docker compose
+ ADMIN_UVICORN_NUM_WORKERS: ${ADMIN_UVICORN_NUM_WORKERS-2}
+ ADMIN_UVICORN_PORT: ${ADMIN_UVICORN_PORT-8081}
+ depends_on:
+ - mongodb
+ restart: always
+ ports:
+ # for debug
+ - ${ADMIN_UVICORN_PORT-8081}:${ADMIN_UVICORN_PORT-8081}
+ api:
+ build:
+ context: ..
+ dockerfile: services/api/Dockerfile
+ # image: ${IMAGE_SERVICE_API?IMAGE_SERVICE_API env var must be provided}
+ extends:
+ file: docker-compose-base.yml
+ service: common
+ environment:
+ # service
+ API_HF_AUTH_PATH: ${API_HF_AUTH_PATH-/api/datasets/%s/auth-check}
+ API_MAX_AGE_LONG: ${API_MAX_AGE_LONG-120}
+ API_MAX_AGE_SHORT: ${API_MAX_AGE_SHORT-10}
+ # prometheus
+ PROMETHEUS_MULTIPROC_DIR: ${PROMETHEUS_MULTIPROC_DIR-}
+ # uvicorn
+ API_UVICORN_HOSTNAME: 0.0.0.0 # required for docker compose
+ API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2}
+ API_UVICORN_PORT: ${API_UVICORN_PORT-8080}
+ ports:
+ # for debug
+ - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080}
+ depends_on:
+ - mongodb
+ restart: unless-stopped
+ worker-config-names:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ volumes:
+ - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
+ - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
+ - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ DATASETS_BASED_ENDPOINT: "/config-names" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ worker-split-names:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ volumes:
+ - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
+ - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
+ - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ DATASETS_BASED_ENDPOINT: "/split-names" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ worker-splits:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ volumes:
+ - splits-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
+ - splits-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
+ - splits-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ DATASETS_BASED_ENDPOINT: "/splits" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ worker-first-rows:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ volumes:
+ - assets:${ASSETS_STORAGE_DIRECTORY-/assets}:rw
+ - first-rows-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
+ - first-rows-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
+ - first-rows-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ ASSETS_BASE_URL: "http://localhost:${PORT_REVERSE_PROXY-8000}/assets" # hard-coded to work with the reverse-proxy
+ ASSETS_STORAGE_DIRECTORY: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ DATASETS_BASED_ENDPOINT: "/first-rows" # hard-coded
+ FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE: ${FIRST_ROWS_FALLBACK_MAX_DATASET_SIZE-100_000_000}
+ FIRST_ROWS_MAX_BYTES: ${FIRST_ROWS_MAX_BYTES-1_000_000}
+ FIRST_ROWS_MAX_NUMBER: ${FIRST_ROWS_MAX_NUMBER-100}
+ FIRST_ROWS_MIN_CELL_BYTES: ${FIRST_ROWS_MIN_CELL_BYTES-100}
+ FIRST_ROWS_MIN_NUMBER: ${FIRST_ROWS_MIN_NUMBER-10}
+ WORKER_LOOP_STORAGE_PATHS: ${ASSETS_STORAGE_DIRECTORY-/assets}
+ # ^ note: the datasets cache is automatically added, so no need to add it here
+ depends_on:
+ - mongodb
+ restart: always
+ worker-parquet-and-dataset-info:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ volumes:
+ - parquet-datasets-cache:${HF_DATASETS_CACHE-/datasets-cache}:rw
+ - parquet-modules-cache:${HF_DATASETS_CACHE-/modules-cache}:rw
+ - parquet-numba-cache:${NUMBA_CACHE_DIR-/numba-cache}:rw
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ DATASETS_BASED_ENDPOINT: "/parquet-and-dataset-info" # hard-coded
+ PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS: ${PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS-}
+ PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE: ${PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE-Update parquet files}
+ PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-}
+ PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE-100_000_000}
+ PARQUET_AND_DATASET_INFO_SOURCE_REVISION: ${PARQUET_AND_DATASET_INFO_SOURCE_REVISION-main}
+ PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS-}
+ PARQUET_AND_DATASET_INFO_TARGET_REVISION: ${PARQUET_AND_DATASET_INFO_TARGET_REVISION-refs/convert/parquet}
+ PARQUET_AND_DATASET_INFO_URL_TEMPLATE: ${PARQUET_AND_DATASET_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
+ depends_on:
+ - mongodb
+ restart: always
+ worker-parquet:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ DATASETS_BASED_ENDPOINT: "/parquet" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ worker-dataset-info:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ DATASETS_BASED_ENDPOINT: "/dataset-info" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ worker-sizes:
+ build:
+ context: ..
+ dockerfile: workers/datasets_based/dev.Dockerfile
+ # image: ${IMAGE_WORKER_DATASETS_BASED?IMAGE_WORKER_DATASETS_BASED env var must be provided}
+ extends:
+ file: docker-compose-base.yml
+ service: datasets-worker
+ environment:
+ DATASETS_BASED_ENDPOINT: "/sizes" # hard-coded
+ depends_on:
+ - mongodb
+ restart: always
+ mongodb:
+ image: docker.io/mongo
+ volumes:
+ - mongo:/data/db:rw
+ ports:
+ # for debug
+ - "${MONGO_PORT-27017}:27017"
+volumes:
+ assets:
+ mongo:
+ splits-datasets-cache:
+ splits-modules-cache:
+ splits-numba-cache:
+ first-rows-datasets-cache:
+ first-rows-modules-cache:
+ first-rows-numba-cache:
+ parquet-datasets-cache:
+ parquet-modules-cache:
+ parquet-numba-cache:
diff --git a/workers/datasets_based/dev.Dockerfile b/workers/datasets_based/dev.Dockerfile
new file mode 100644
index 00000000..2eea73f5
--- /dev/null
+++ b/workers/datasets_based/dev.Dockerfile
@@ -0,0 +1,40 @@
+# build with
+# docker build -t some_tag_worker -f Dockerfile ../..
+FROM python:3.9.15-slim
+
+ENV PYTHONFAULTHANDLER=1 \
+ PYTHONUNBUFFERED=1 \
+ PYTHONHASHSEED=random \
+ PIP_NO_CACHE_DIR=off \
+ PIP_DISABLE_PIP_VERSION_CHECK=on \
+ PIP_DEFAULT_TIMEOUT=100 \
+ POETRY_NO_INTERACTION=1 \
+ # Versions:
+ POETRY_VERSION=1.3.2 \
+ POETRY_VIRTUALENVS_IN_PROJECT=true
+
+# System deps:
+RUN apt-get update \
+ && apt-get install -y build-essential unzip wget python3-dev make \
+ libicu-dev ffmpeg libavcodec-extra libsndfile1 llvm pkg-config \
+ poppler-utils \
+ && rm -rf /var/lib/apt/lists/*
+
+RUN pip install -U --no-cache-dir pip
+RUN pip install --no-cache-dir "poetry==$POETRY_VERSION"
+
+WORKDIR /src
+COPY workers/datasets_based/vendors ./workers/datasets_based/vendors/
+COPY workers/datasets_based/poetry.lock ./workers/datasets_based/poetry.lock
+COPY workers/datasets_based/pyproject.toml ./workers/datasets_based/pyproject.toml
+COPY libs/libcommon ./libs/libcommon
+WORKDIR /src/workers/datasets_based/
+RUN poetry install --no-cache
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# No need to copy the source code since we map a volume in docker-compose-base.yaml
+# Removed: COPY workers/datasets_based/src ./src
+# Removed: RUN poetry install --no-cache
+# However we need to install the package when the container starts
+# Added: poetry install
+ENTRYPOINT ["/bin/sh", "-c" , "poetry install && poetry run python src/datasets_based/main.py"]
diff --git a/workers/datasets_based/src/datasets_based/worker_loop.py b/workers/datasets_based/src/datasets_based/worker_loop.py
index 0a225a73..0c27d5c3 100644
--- a/workers/datasets_based/src/datasets_based/worker_loop.py
+++ b/workers/datasets_based/src/datasets_based/worker_loop.py
@@ -99,0 +100 @@ class WorkerLoop:
+ self.info("Worker started")
|
|
7b2ef4d3ea0a1facb7789c3a959602279c8e932a
|
Quentin Lhoest
| 2023-02-03T11:04:53 |
create doc for every pr (#768)
|
diff --git a/.github/workflows/doc-pr-build.yml b/.github/workflows/doc-pr-build.yml
index f962ada5..44de5eca 100644
--- a/.github/workflows/doc-pr-build.yml
+++ b/.github/workflows/doc-pr-build.yml
@@ -8,3 +7,0 @@ on:
- paths:
- - 'docs/**'
- - '.github/workflows/doc-pr-build.yml'
|
|
c199a3f19a5866d2540790e57510ee5f6cfd48a6
|
Quentin Lhoest
| 2023-02-02T19:41:17 |
Add refresh dataset ui (#760)
|
diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py
index 571d717b..e5724f67 100644
--- a/front/admin_ui/app.py
+++ b/front/admin_ui/app.py
@@ -1,0 +2,2 @@ import os
+import urllib.parse
+from itertools import product
@@ -25 +27,4 @@ def healthcheck():
- response = requests.head(f"{DSS_ENDPOINT}/admin/pending-jobs", timeout=10)
+ try:
+ response = requests.head(f"{DSS_ENDPOINT}/admin/pending-jobs", timeout=10)
+ except requests.ConnectionError as error:
+ return f"❌ Failed to connect to {DSS_ENDPOINT} (error {error})"
@@ -40 +45 @@ with gr.Blocks() as demo:
- token_box = gr.Textbox(label="token", placeholder="hf_xxx", type="password")
+ token_box = gr.Textbox(HF_TOKEN or "", label="token", placeholder="hf_xxx", type="password")
@@ -60,0 +66,8 @@ with gr.Blocks() as demo:
+ with gr.Tab("Refresh dataset"):
+ refresh_type = gr.Textbox(label="Processing step type", placeholder="first-rows")
+ refresh_dataset_name = gr.Textbox(label="dataset", placeholder="c4")
+ refresh_config_name = gr.Textbox(label="config (optional)", placeholder="en")
+ refresh_split_name = gr.Textbox(label="split (optional)", placeholder="train, test")
+ gr.Markdown("*you can select multiple values by separating them with commas, e.g. split='train, test'*")
+ refresh_dataset_button = gr.Button("Force resfresh dataset")
+ refresh_dataset_output = gr.Markdown("")
@@ -82 +94,0 @@ with gr.Blocks() as demo:
- token = token or HF_TOKEN
@@ -94 +106,5 @@ with gr.Blocks() as demo:
- pending_jobs_df["created_at"] = pd.to_datetime(pending_jobs_df["created_at"], errors="coerce")
+ if "created_at" in pending_jobs_df.columns:
+ pending_jobs_df["created_at"] = pd.to_datetime(pending_jobs_df["created_at"], errors="coerce")
+ most_recent = pending_jobs_df.nlargest(5, "created_at")
+ else:
+ most_recent = pd.DataFrame()
@@ -101 +117 @@ with gr.Blocks() as demo:
- recent_pending_jobs_table: gr.update(value=pending_jobs_df.nlargest(5, "created_at"))
+ recent_pending_jobs_table: gr.update(value=most_recent)
@@ -116,0 +133,33 @@ with gr.Blocks() as demo:
+ def refresh_dataset(token, refresh_types, refresh_dataset_names, refresh_config_names, refresh_split_names):
+ headers = {"Authorization": f"Bearer {token}"}
+ all_results = ""
+ for refresh_type, refresh_dataset_name, refresh_config_name, refresh_split_name in product(
+ refresh_types.split(","), refresh_dataset_names.split(","), refresh_config_names.split(","), refresh_split_names.split(",")
+ ):
+ refresh_types = refresh_types.strip()
+ refresh_dataset_name = refresh_dataset_name.strip()
+ params = {"dataset": refresh_dataset_name}
+ if refresh_config_name:
+ refresh_config_name = refresh_config_name.strip()
+ params["config"] = refresh_config_name
+ if refresh_split_name:
+ refresh_split_name = refresh_split_name.strip()
+ params["split"] = refresh_split_name
+ params = urllib.parse.urlencode(params)
+ response = requests.post(f"{DSS_ENDPOINT}/admin/force-refresh/{refresh_type}?{params}", headers=headers, timeout=60)
+ if response.status_code == 200:
+ result = f"[{refresh_dataset_name}] ✅ Added processing step to the queue: '{refresh_type}'"
+ if refresh_config_name:
+ result += f", for config '{refresh_config_name}'"
+ if refresh_split_name:
+ result += f", for split '{refresh_split_name}'"
+ else:
+ result = f"[{refresh_dataset_name}] ❌ Failed to add processing step to the queue. Error {response.status_code}"
+ try:
+ if response.json().get("error"):
+ result += f": {response.json()['error']}"
+ except requests.JSONDecodeError:
+ result += f": {response.content}"
+ all_results += result.strip("\n") + "\n"
+ return "```\n" + all_results + "\n```"
+
@@ -117,0 +167 @@ with gr.Blocks() as demo:
+
@@ -120,0 +171 @@ with gr.Blocks() as demo:
+ refresh_dataset_button.click(refresh_dataset, inputs=[token_box, refresh_type, refresh_dataset_name, refresh_config_name, refresh_split_name], outputs=refresh_dataset_output)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.