hash
stringlengths
40
40
authorName
stringclasses
42 values
authorEmail
stringclasses
41 values
date
timestamp[ms]date
2021-07-26 09:52:55
2025-07-18 10:19:56
subject
stringlengths
11
116
diff
stringlengths
0
987k
a8846dcceb2a7a6f2e38fa2370da633f9f3e117a
Quentin Lhoest
2023-05-05T14:20:43
unsupport image and audio (#1144)
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 03cb5aa5..d809d6c5 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -85 +85 @@ def get_parquet_fs(dataset: str, hf_token: Optional[str]) -> HfFileSystem: -UNSUPPORTED_FEATURES_MAGIC_STRINGS = ["'binary'"] +UNSUPPORTED_FEATURES_MAGIC_STRINGS = ["Image(", "Audio(", "'binary'"]
af91bf788c18b0f98989d0e148f85427b06b7678
Quentin Lhoest
2023-05-04T12:04:29
increase api replicas + skip cacheMaintenance (#1140)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 500ef0ab..b450c431 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -115 +115 @@ cacheMaintenance: - action: "backfill" + action: "skip" @@ -245 +245 @@ api: - replicas: 4 + replicas: 12
37ed2cd0ee7bde4804b632097e4ee6f90cac5288
Andrea Francis Soria Jimenez
2023-05-03T21:32:47
Fix parent class for split-names-from-dataset-info and first-rows-from-parquet (#1136)
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index a030b649..2c548a81 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -160 +160,2 @@ class JobRunnerFactory(BaseJobRunnerFactory): - app_config=self.app_config, + common_config=self.app_config.common, + worker_config=self.app_config.worker, @@ -162 +162,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - hf_datasets_cache=self.hf_datasets_cache, @@ -183 +182,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - hf_datasets_cache=self.hf_datasets_cache, diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index 5a042943..5963b50e 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -15,0 +16 @@ from worker.job_runner import ( + JobRunner, @@ -20 +20,0 @@ from worker.job_runner import ( -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner @@ -98 +98 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> -class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): +class SplitNamesFromDatasetInfoJobRunner(JobRunner): diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index 7ecf9185..02c11c15 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -7 +6,0 @@ from http import HTTPStatus -from pathlib import Path @@ -29,0 +29 @@ from worker.job_runner import ( + JobRunner, @@ -33 +32,0 @@ from worker.job_runner import ( -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner @@ -283 +282 @@ def compute_first_rows_response( -class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): +class SplitFirstRowsFromParquetJobRunner(JobRunner): @@ -300 +298,0 @@ class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): - hf_datasets_cache: Path, @@ -305 +303,2 @@ class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): - app_config=app_config, + common_config=app_config.common, + worker_config=app_config.worker, @@ -307 +305,0 @@ class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): - hf_datasets_cache=hf_datasets_cache, diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index e8ab3e7e..8c400486 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -22 +21,0 @@ from worker.job_runners.config.split_names_from_dataset_info import ( -from worker.resources import LibrariesResource @@ -29 +27,0 @@ def get_job_runner( - libraries_resource: LibrariesResource, @@ -49 +47,2 @@ def get_job_runner( - app_config=app_config, + common_config=app_config.common, + worker_config=app_config.worker, @@ -60 +58,0 @@ def get_job_runner( - hf_datasets_cache=libraries_resource.hf_datasets_cache, diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index 5fd059fc..7fc27d51 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -24 +23,0 @@ from worker.job_runners.split.first_rows_from_parquet import ( -from worker.resources import LibrariesResource @@ -35 +33,0 @@ def get_job_runner( - libraries_resource: LibrariesResource, @@ -67 +64,0 @@ def get_job_runner( - hf_datasets_cache=libraries_resource.hf_datasets_cache,
d303acb8c67bddc643b336dd8dee1ecf06abf68c
Sylvain Lesage
2023-05-03T19:09:22
fix: 🐛 don't refresh the UnexpectedError entries in next sync (#1135)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index cc0ab550..500ef0ab 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -120 +120 @@ cacheMaintenance: - error_codes_to_retry: "UnexpectedError" + error_codes_to_retry: ""
0558564fd356578e0c02eecc9fcda8a813b98ef2
Quentin Lhoest
2023-05-03T16:28:36
Re-lower row group (#1134)
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 3334c625..b70c6239 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -28 +28 @@ PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION = 2 -PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 2 +PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 3 @@ -36,2 +36,3 @@ PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION = 1 -PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS = 100 -PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS = 100 +PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS = 100 +PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS = 100 +PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_BINARY_DATASETS = 100 diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 6db5a292..96c90580 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -16,0 +17 @@ import datasets.config +import datasets.info @@ -37 +38,6 @@ from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFo -from libcommon.constants import PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION +from libcommon.constants import ( + PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS, + PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_BINARY_DATASETS, + PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS, + PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, +) @@ -665,0 +672,24 @@ def raise_if_too_big_from_external_data_files( +def get_writer_batch_size(ds_config_info: datasets.info.DatasetInfo) -> Optional[int]: + """ + Get the writer_batch_size that defines the maximum row group size in the parquet files. + The default in `datasets` is 1,000 but we lower it to 100 for image datasets. + This allows to optimize random access to parquet file, since accessing 1 row requires + to read its entire row group. + Args: + ds_config_info (`datasets.info.DatasetInfo`): + Dataset info from `datasets`. + Returns: + writer_batch_size (`Optional[int]`): + Writer batch size to pass to a dataset builder. + If `None`, then it will use the `datasets` default. + """ + if "Audio(" in str(ds_config_info.features): + return PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS + elif "Image(" in str(ds_config_info.features): + return PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS + elif "'binary'" in str(ds_config_info.features): + return PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_ROW_GROUP_SIZE_FOR_BINARY_DATASETS + else: + return None + + @@ -814,0 +845,5 @@ def compute_config_parquet_and_info_response( + writer_batch_size = get_writer_batch_size(builder.info) + if writer_batch_size is not None and ( + builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size + ): + builder._writer_batch_size = writer_batch_size diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index d13c2c91..6ee535da 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -9,0 +10 @@ import datasets.builder +import datasets.info @@ -12,0 +14 @@ import requests +from datasets import Audio, Features, Image, Value @@ -28,0 +31 @@ from worker.job_runners.config.parquet_and_info import ( + get_writer_batch_size, @@ -619,0 +623,15 @@ def test_parse_repo_filename(filename: str, split: str, config: str, raises: boo + + [email protected]( + "ds_info, has_big_chunks", + [ + (datasets.info.DatasetInfo(), False), + (datasets.info.DatasetInfo(features=Features({"text": Value("string")})), False), + (datasets.info.DatasetInfo(features=Features({"image": Image()})), True), + (datasets.info.DatasetInfo(features=Features({"audio": Audio()})), True), + (datasets.info.DatasetInfo(features=Features({"nested": [{"image": Image()}]})), True), + (datasets.info.DatasetInfo(features=Features({"blob": Value("binary")})), True), + ], +) +def test_get_writer_batch_size(ds_info: datasets.info.DatasetInfo, has_big_chunks: bool) -> None: + assert get_writer_batch_size(ds_info) == (100 if has_big_chunks else None)
9e8c65bd8cebeec0e695ada39ee1a43ae11654f9
Andrea Francis Soria Jimenez
2023-05-03T16:19:11
Aggregated config and dataset level for opt-in-out-urls-scan (#1108)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 7c0655d8..9982228d 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -13,0 +14 @@ from libcommon.constants import ( + PROCESSING_STEP_CONFIG_OPT_IN_OUT_URLS_COUNT_VERSION, @@ -18,0 +20 @@ from libcommon.constants import ( + PROCESSING_STEP_DATASET_OPT_IN_OUT_URLS_COUNT_VERSION, @@ -272,0 +275,10 @@ class ProcessingGraphConfig: + "config-opt-in-out-urls-count": { + "input_type": "config", + "requires": ["/split-names-from-streaming", "split-opt-in-out-urls-count"], + "job_runner_version": PROCESSING_STEP_CONFIG_OPT_IN_OUT_URLS_COUNT_VERSION, + }, + "dataset-opt-in-out-urls-count": { + "input_type": "dataset", + "requires": ["/config-names", "config-opt-in-out-urls-count"], + "job_runner_version": PROCESSING_STEP_DATASET_OPT_IN_OUT_URLS_COUNT_VERSION, + }, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index b36f6211..3334c625 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -18,0 +19 @@ PROCESSING_STEP_CONFIG_INFO_VERSION = 2 +PROCESSING_STEP_CONFIG_OPT_IN_OUT_URLS_COUNT_VERSION = 1 @@ -20,0 +22 @@ PROCESSING_STEP_DATASET_IS_VALID_VERSION = 2 +PROCESSING_STEP_DATASET_OPT_IN_OUT_URLS_COUNT_VERSION = 1 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index fed3cdd6..1f4d9dcb 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -74,0 +75 @@ def graph() -> ProcessingGraph: + "dataset-opt-in-out-urls-count", @@ -106 +107 @@ def graph() -> ProcessingGraph: - ["split-first-rows-from-streaming", "dataset-split-names"], + ["split-first-rows-from-streaming", "dataset-split-names", "config-opt-in-out-urls-count"], @@ -219 +220 @@ def graph() -> ProcessingGraph: - [], + ["config-opt-in-out-urls-count"], @@ -230,0 +232,31 @@ def graph() -> ProcessingGraph: + ( + "config-opt-in-out-urls-count", + ["dataset-opt-in-out-urls-count"], + ["split-opt-in-out-urls-count", "/split-names-from-streaming"], + [ + "/config-names", + "/split-names-from-streaming", + "split-first-rows-from-streaming", + "/split-names-from-dataset-info", + "config-info", + "config-parquet-and-info", + "split-opt-in-out-urls-count", + "split-opt-in-out-urls-scan", + ], + ), + ( + "dataset-opt-in-out-urls-count", + [], + ["config-opt-in-out-urls-count", "/config-names"], + [ + "/config-names", + "/split-names-from-streaming", + "split-first-rows-from-streaming", + "/split-names-from-dataset-info", + "config-info", + "config-parquet-and-info", + "config-opt-in-out-urls-count", + "split-opt-in-out-urls-count", + "split-opt-in-out-urls-scan", + ], + ), diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 5f6acbb0..5f4b63fd 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -348,0 +349,5 @@ def get_CONFIG_STATE_DICT(dataset: str, config: str, split_states: List[Any], ca + { + "id": f"config-opt-in-out-urls-count,{dataset},{config}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, @@ -459,0 +465,5 @@ def test_dataset_state_as_dict() -> None: + { + "id": f"dataset-opt-in-out-urls-count,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, @@ -518,0 +529 @@ def test_plan() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -534,0 +546 @@ def test_plan() -> None: + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -549,0 +562 @@ def test_plan_job_creation_and_termination() -> None: + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -568,0 +582 @@ def test_plan_job_creation_and_termination() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -583,0 +598 @@ def test_plan_job_creation_and_termination() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -625,0 +641,2 @@ def test_plan_job_creation_and_termination() -> None: + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -633,0 +651 @@ def test_plan_job_creation_and_termination() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -647,0 +666 @@ def test_plan_job_creation_and_termination() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -660,0 +680,2 @@ def test_plan_job_creation_and_termination() -> None: + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -703,0 +725,2 @@ def test_plan_retry_error() -> None: + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -711,0 +735 @@ def test_plan_retry_error() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -730,0 +755,2 @@ def test_plan_retry_error() -> None: + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -738,0 +765 @@ def test_plan_retry_error() -> None: + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -787,0 +815,2 @@ def test_plan_incoherent_state() -> None: + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -795,0 +825 @@ def test_plan_incoherent_state() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -811,4 +841 @@ def test_plan_incoherent_state() -> None: - "up_to_date": [ - "/config-names,dataset", - "/split-names-from-dataset-info,dataset,config1", - ], + "up_to_date": ["/config-names,dataset", "/split-names-from-dataset-info,dataset,config1"], @@ -822,0 +850,2 @@ def test_plan_incoherent_state() -> None: + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -830,0 +860 @@ def test_plan_incoherent_state() -> None: + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -898,0 +929,2 @@ def test_plan_updated_at() -> None: + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -905,0 +938 @@ def test_plan_updated_at() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -923,0 +957,2 @@ def test_plan_updated_at() -> None: + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -931,0 +967 @@ def test_plan_updated_at() -> None: + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -967,0 +1004,2 @@ def test_plan_job_runner_version() -> None: + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -975,0 +1014 @@ def test_plan_job_runner_version() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -994,0 +1034,2 @@ def test_plan_job_runner_version() -> None: + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -1002,0 +1044 @@ def test_plan_job_runner_version() -> None: + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -1053,0 +1096,2 @@ def test_plan_git_revision( + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -1061,0 +1106 @@ def test_plan_git_revision( + "dataset-opt-in-out-urls-count,dataset", @@ -1079,0 +1125,2 @@ def test_plan_git_revision( + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -1087,0 +1135 @@ def test_plan_git_revision( + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -1110,0 +1159,2 @@ def test_plan_git_revision( + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -1118,0 +1169 @@ def test_plan_git_revision( + "dataset-opt-in-out-urls-count,dataset", @@ -1135,0 +1187,2 @@ def test_plan_git_revision( + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -1143,0 +1197 @@ def test_plan_git_revision( + "CreateJob[dataset-opt-in-out-urls-count,dataset]", @@ -1215,0 +1270,2 @@ def test_plan_update_fan_in_parent() -> None: + "config-opt-in-out-urls-count,dataset,config1", + "config-opt-in-out-urls-count,dataset,config2", @@ -1221,0 +1278 @@ def test_plan_update_fan_in_parent() -> None: + "dataset-opt-in-out-urls-count,dataset", @@ -1242,0 +1300,2 @@ def test_plan_update_fan_in_parent() -> None: + "CreateJob[config-opt-in-out-urls-count,dataset,config1]", + "CreateJob[config-opt-in-out-urls-count,dataset,config2]", @@ -1248,0 +1308 @@ def test_plan_update_fan_in_parent() -> None: + "CreateJob[dataset-opt-in-out-urls-count,dataset]", diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 47b080b7..3313fd35 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -141,0 +142,2 @@ class EndpointConfig: + "dataset": ["dataset-opt-in-out-urls-count"], + "config": ["config-opt-in-out-urls-count"], diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index e2f61053..26706d8d 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -80 +80 @@ def test_endpoints_definition() -> None: - assert sorted(list(opt_in_out_urls)) == ["split"] + assert sorted(list(opt_in_out_urls)) == ["config", "dataset", "split"] @@ -81,0 +82,2 @@ def test_endpoints_definition() -> None: + assert opt_in_out_urls["config"] is not None + assert opt_in_out_urls["dataset"] is not None @@ -82,0 +85,2 @@ def test_endpoints_definition() -> None: + assert len(opt_in_out_urls["config"]) == 1 # Only has one processing step + assert len(opt_in_out_urls["dataset"]) == 1 # Only has one processing step diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 1f06598e..a030b649 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -14,0 +15,3 @@ from worker.job_runners.config.info import ConfigInfoJobRunner +from worker.job_runners.config.opt_in_out_urls_count import ( + ConfigOptInOutUrlsCountJobRunner, +) @@ -26,0 +30,3 @@ from worker.job_runners.dataset.is_valid import DatasetIsValidJobRunner +from worker.job_runners.dataset.opt_in_out_urls_count import ( + DatasetOptInOutUrlsCountJobRunner, +) @@ -194,0 +201,14 @@ class JobRunnerFactory(BaseJobRunnerFactory): + if job_type == ConfigOptInOutUrlsCountJobRunner.get_job_type(): + return ConfigOptInOutUrlsCountJobRunner( + job_info=job_info, + common_config=self.app_config.common, + worker_config=self.app_config.worker, + processing_step=processing_step, + ) + if job_type == DatasetOptInOutUrlsCountJobRunner.get_job_type(): + return DatasetOptInOutUrlsCountJobRunner( + job_info=job_info, + common_config=self.app_config.common, + worker_config=self.app_config.worker, + processing_step=processing_step, + ) @@ -220,0 +241,2 @@ class JobRunnerFactory(BaseJobRunnerFactory): + ConfigOptInOutUrlsCountJobRunner.get_job_type(), + DatasetOptInOutUrlsCountJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py new file mode 100644 index 00000000..eabb7cab --- /dev/null +++ b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py @@ -0,0 +1,123 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from http import HTTPStatus +from typing import Any, Literal, Mapping, Optional, Tuple + +from libcommon.constants import PROCESSING_STEP_CONFIG_OPT_IN_OUT_URLS_COUNT_VERSION +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response + +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, + get_previous_step_or_raise, +) +from worker.utils import OptInOutUrlsCountResponse + +ConfigOptInOutUrlsCountJobRunnerErrorCode = Literal["PreviousStepFormatError"] + + +class ConfigOptInOutUrlsCountJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: ConfigOptInOutUrlsCountJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class PreviousStepFormatError(ConfigOptInOutUrlsCountJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +def compute_opt_in_out_urls_scan_response(dataset: str, config: str) -> Tuple[OptInOutUrlsCountResponse, float]: + logging.info(f"get config-opt-in-out-urls-count for dataset={dataset} config={config}") + + split_names_response = get_previous_step_or_raise( + kinds=["/split-names-from-streaming"], dataset=dataset, config=config + ) + content = split_names_response.response["content"] + if "splits" not in content: + raise PreviousStepFormatError("Previous step did not return the expected content: 'splits'.") + + urls_columns = [] + num_opt_in_urls = 0 + num_opt_out_urls = 0 + num_urls = 0 + num_scanned_rows = 0 + try: + total = 0 + pending = 0 + for split_item in content["splits"]: + split = split_item["split"] + total += 1 + try: + response = get_response( + kind="split-opt-in-out-urls-count", dataset=dataset, config=config, split=split + ) + except DoesNotExist: + logging.debug("No response found in previous step for this dataset: 'split-opt-in-out-urls-count'.") + pending += 1 + continue + if response["http_status"] != HTTPStatus.OK: + logging.debug(f"Previous step gave an error: {response['http_status']}.") + continue + split_opt_in_out_content = response["content"] + urls_columns.extend(split_opt_in_out_content["urls_columns"]) + num_opt_in_urls += split_opt_in_out_content["num_opt_in_urls"] + num_opt_out_urls += split_opt_in_out_content["num_opt_out_urls"] + num_urls += split_opt_in_out_content["num_urls"] + num_scanned_rows += split_opt_in_out_content["num_scanned_rows"] + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + unique_urls_columns = sorted(list(set(urls_columns))) + has_urls_columns = len(unique_urls_columns) > 0 + progress = (total - pending) / total if total else 1.0 + + return ( + OptInOutUrlsCountResponse( + urls_columns=unique_urls_columns, + has_urls_columns=has_urls_columns, + num_opt_in_urls=num_opt_in_urls, + num_opt_out_urls=num_opt_out_urls, + num_scanned_rows=num_scanned_rows, + num_urls=num_urls, + ), + progress, + ) + + +class ConfigOptInOutUrlsCountJobRunner(JobRunner): + @staticmethod + def get_job_type() -> str: + return "config-opt-in-out-urls-count" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_CONFIG_OPT_IN_OUT_URLS_COUNT_VERSION + + def compute(self) -> JobResult: + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") + if self.config is None: + raise ParameterMissingError("'config' parameter is required") + response_content, progress = compute_opt_in_out_urls_scan_response(dataset=self.dataset, config=self.config) + return JobResult(response_content, progress=progress) + + def get_new_splits(self, _: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by the compute.""" + return {SplitFullName(dataset=self.dataset, config=self.config, split=None)} diff --git a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py new file mode 100644 index 00000000..717fdb9d --- /dev/null +++ b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py @@ -0,0 +1,122 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from http import HTTPStatus +from typing import Any, Literal, Mapping, Optional, Tuple + +from libcommon.constants import PROCESSING_STEP_DATASET_OPT_IN_OUT_URLS_COUNT_VERSION +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response + +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, + get_previous_step_or_raise, +) +from worker.utils import OptInOutUrlsCountResponse + +DatasetOptInOutUrlsCountJobRunnerErrorCode = Literal["PreviousStepFormatError"] + + +class DatasetOptInOutUrlsCountJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: DatasetOptInOutUrlsCountJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class PreviousStepFormatError(DatasetOptInOutUrlsCountJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +def compute_opt_in_out_urls_count_response(dataset: str) -> Tuple[OptInOutUrlsCountResponse, float]: + logging.info(f"get opt-in-out-urls-count for dataset={dataset}") + + config_names_response = get_previous_step_or_raise(kinds=["/config-names"], dataset=dataset) + content = config_names_response.response["content"] + if "config_names" not in content: + raise PreviousStepFormatError("Previous step did not return the expected content: 'config_names'.") + + urls_columns = [] + num_opt_in_urls = 0 + num_opt_out_urls = 0 + num_urls = 0 + num_scanned_rows = 0 + try: + total = 0 + pending = 0 + for config_item in content["config_names"]: + config = config_item["config"] + total += 1 + try: + response = get_response(kind="config-opt-in-out-urls-count", dataset=dataset, config=config) + except DoesNotExist: + logging.debug("No response found in previous step for this dataset: 'config-opt-in-out-urls-count'.") + pending += 1 + continue + if response["http_status"] != HTTPStatus.OK: + logging.debug(f"Previous step gave an error: {response['http_status']}.") + continue + else: + if response["progress"] and response["progress"] < 1.0: + logging.debug(f"Previous step is still in progress: {response['progress']}.") + pending += 1 + continue + split_opt_in_out_content = response["content"] + urls_columns.extend(split_opt_in_out_content["urls_columns"]) + num_opt_in_urls += split_opt_in_out_content["num_opt_in_urls"] + num_opt_out_urls += split_opt_in_out_content["num_opt_out_urls"] + num_urls += split_opt_in_out_content["num_urls"] + num_scanned_rows += split_opt_in_out_content["num_scanned_rows"] + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + unique_urls_columns = sorted(list(set(urls_columns))) + has_urls_columns = len(unique_urls_columns) > 0 + progress = (total - pending) / total if total else 1.0 + + return ( + OptInOutUrlsCountResponse( + urls_columns=unique_urls_columns, + has_urls_columns=has_urls_columns, + num_opt_in_urls=num_opt_in_urls, + num_opt_out_urls=num_opt_out_urls, + num_scanned_rows=num_scanned_rows, + num_urls=num_urls, + ), + progress, + ) + + +class DatasetOptInOutUrlsCountJobRunner(JobRunner): + @staticmethod + def get_job_type() -> str: + return "dataset-opt-in-out-urls-count" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_DATASET_OPT_IN_OUT_URLS_COUNT_VERSION + + def compute(self) -> JobResult: + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") + response_content, progress = compute_opt_in_out_urls_count_response(dataset=self.dataset) + return JobResult(response_content, progress=progress) + + def get_new_splits(self, _: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by the compute.""" + return {SplitFullName(dataset=self.dataset, config=None, split=None)} diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py index 29276d99..38ca9d0c 100644 --- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py @@ -6 +6 @@ from http import HTTPStatus -from typing import Any, List, Literal, Mapping, Optional, TypedDict +from typing import Any, Literal, Mapping, Optional @@ -16,0 +17 @@ from worker.job_runner import ( +from worker.utils import OptInOutUrlsCountResponse @@ -44,9 +44,0 @@ class PreviousStepFormatError(SplitOptInOutUrlsCountJobRunnerError): -class OptInOutUrlsCountResponse(TypedDict): - urls_columns: List[str] - num_opt_in_urls: int - num_opt_out_urls: int - num_urls: int - num_scanned_rows: int - has_urls_columns: bool - - diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py index 6cd4ad55..f0c6965d 100644 --- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py @@ -8 +8 @@ from pathlib import Path -from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict, Union +from typing import Any, List, Literal, Mapping, Optional, Tuple, Union @@ -25 +25,6 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner -from worker.utils import SplitFirstRowsResponse, get_rows_or_raise +from worker.utils import ( + OptInOutUrlsScanResponse, + OptUrl, + SplitFirstRowsResponse, + get_rows_or_raise, +) @@ -95,17 +99,0 @@ class ExternalServerError(SplitOptInOutUrlsScanJobRunnerError): -class OptUrl(TypedDict): - url: str - row_idx: int - column_name: str - - -class OptInOutUrlsScanResponse(TypedDict): - urls_columns: List[str] - opt_in_urls: List[OptUrl] - opt_out_urls: List[OptUrl] - num_opt_in_urls: int - num_opt_out_urls: int - num_urls: int - num_scanned_rows: int - has_urls_columns: bool - - diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 85966d05..dafe7ac1 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -86,0 +87,20 @@ class SplitFirstRowsResponse(TypedDict): +class OptUrl(TypedDict): + url: str + row_idx: int + column_name: str + + +class OptInOutUrlsCountResponse(TypedDict): + urls_columns: List[str] + num_opt_in_urls: int + num_opt_out_urls: int + num_urls: int + num_scanned_rows: int + has_urls_columns: bool + + +class OptInOutUrlsScanResponse(OptInOutUrlsCountResponse): + opt_in_urls: List[OptUrl] + opt_out_urls: List[OptUrl] + + diff --git a/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py new file mode 100644 index 00000000..883f8d2f --- /dev/null +++ b/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py @@ -0,0 +1,185 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Callable, List + +import pytest +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response + +from worker.config import AppConfig +from worker.job_runner import PreviousStepError +from worker.job_runners.config.opt_in_out_urls_count import ( + ConfigOptInOutUrlsCountJobRunner, +) + + [email protected](autouse=True) +def prepare_and_clean_mongo(app_config: AppConfig) -> None: + # prepare the database before each test, and clean it afterwards + pass + + +GetJobRunner = Callable[[str, str, AppConfig, bool], ConfigOptInOutUrlsCountJobRunner] + + [email protected] +def get_job_runner( + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + app_config: AppConfig, + force: bool = False, + ) -> ConfigOptInOutUrlsCountJobRunner: + return ConfigOptInOutUrlsCountJobRunner( + job_info={ + "type": ConfigOptInOutUrlsCountJobRunner.get_job_type(), + "dataset": dataset, + "config": config, + "split": None, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + common_config=app_config.common, + worker_config=app_config.worker, + processing_step=ProcessingStep( + name=ConfigOptInOutUrlsCountJobRunner.get_job_type(), + input_type="config", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + parents=[], + job_runner_version=ConfigOptInOutUrlsCountJobRunner.get_job_runner_version(), + ), + ) + + return _get_job_runner + + [email protected]( + "dataset,config,split_names_status,split_names_content,spawning_status" + + ",spawning_content,expected_error_code,expected_content,should_raise", + [ + ( + "dataset_ok", + "config", + HTTPStatus.OK, + { + "splits": [ + {"dataset": "dataset_ok", "config": "config", "split": "split"}, + {"dataset": "dataset_ok", "config": "config", "split": "split2"}, + ] + }, + [HTTPStatus.OK, HTTPStatus.OK], + [ + { + "urls_columns": ["url"], + "num_opt_in_urls": 1, + "num_opt_out_urls": 2, + "num_urls": 10, + "num_scanned_rows": 100, + "has_urls_columns": True, + }, + { + "urls_columns": [], + "num_opt_in_urls": 0, + "num_opt_out_urls": 0, + "num_urls": 0, + "num_scanned_rows": 30, + "has_urls_columns": False, + }, + ], + None, + { + "urls_columns": ["url"], + "num_opt_in_urls": 1, + "num_opt_out_urls": 2, + "num_urls": 10, + "num_scanned_rows": 130, + "has_urls_columns": True, + }, + False, + ), + ( + "previos_step_error", + "config", + HTTPStatus.INTERNAL_SERVER_ERROR, + {}, + [], + [], + "PreviousStepError", + None, + True, + ), + ( + "previous_step_format_error", + "config", + HTTPStatus.OK, + { + "splits": [ + {"dataset": "dataset_ok", "config": "config", "split": "split"}, + {"dataset": "dataset_ok", "config": "config", "split": "split2"}, + ] + }, + [HTTPStatus.OK], + [{"wrong_format": None}], + "PreviousStepFormatError", + None, + True, + ), + ], +) +def test_compute( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset: str, + config: str, + split_names_status: HTTPStatus, + split_names_content: Any, + spawning_status: List[HTTPStatus], + spawning_content: List[Any], + expected_error_code: str, + expected_content: Any, + should_raise: bool, +) -> None: + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content=split_names_content, + http_status=split_names_status, + ) + + if split_names_status == HTTPStatus.OK: + for split_item, status, content in zip(split_names_content["splits"], spawning_status, spawning_content): + upsert_response( + kind="split-opt-in-out-urls-count", + dataset=dataset, + config=split_item["config"], + split=split_item["split"], + content=content, + http_status=status, + ) + + job_runner = get_job_runner(dataset, config, app_config, False) + if should_raise: + with pytest.raises(Exception) as e: + job_runner.compute() + assert e.type.__name__ == expected_error_code + else: + assert job_runner.compute().content == expected_content + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = config = "doesnotexist" + job_runner = get_job_runner(dataset, config, app_config, False) + with pytest.raises(PreviousStepError): + job_runner.compute() diff --git a/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py new file mode 100644 index 00000000..e979466f --- /dev/null +++ b/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py @@ -0,0 +1,180 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Callable, List + +import pytest +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response + +from worker.config import AppConfig +from worker.job_runner import PreviousStepError +from worker.job_runners.dataset.opt_in_out_urls_count import ( + DatasetOptInOutUrlsCountJobRunner, +) + + [email protected](autouse=True) +def prepare_and_clean_mongo(app_config: AppConfig) -> None: + # prepare the database before each test, and clean it afterwards + pass + + +GetJobRunner = Callable[[str, AppConfig, bool], DatasetOptInOutUrlsCountJobRunner] + + [email protected] +def get_job_runner( + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + app_config: AppConfig, + force: bool = False, + ) -> DatasetOptInOutUrlsCountJobRunner: + return DatasetOptInOutUrlsCountJobRunner( + job_info={ + "type": DatasetOptInOutUrlsCountJobRunner.get_job_type(), + "dataset": dataset, + "config": None, + "split": None, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + common_config=app_config.common, + worker_config=app_config.worker, + processing_step=ProcessingStep( + name=DatasetOptInOutUrlsCountJobRunner.get_job_type(), + input_type="config", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + parents=[], + job_runner_version=DatasetOptInOutUrlsCountJobRunner.get_job_runner_version(), + ), + ) + + return _get_job_runner + + [email protected]( + "dataset,config_names_status,config_names_content,config_upstream_status" + + ",config_upstream_content,expected_error_code,expected_content,should_raise", + [ + ( + "dataset_ok", + HTTPStatus.OK, + { + "config_names": [ + {"dataset": "dataset_ok", "config": "config1"}, + {"dataset": "dataset_ok", "config": "config2"}, + ] + }, + [HTTPStatus.OK, HTTPStatus.OK], + [ + { + "urls_columns": ["image_url", "url"], + "num_opt_in_urls": 10, + "num_opt_out_urls": 20, + "num_urls": 100, + "num_scanned_rows": 100, + "has_urls_columns": True, + }, + { + "urls_columns": ["image_url", "label", "url"], + "num_opt_in_urls": 10, + "num_opt_out_urls": 0, + "num_urls": 50, + "num_scanned_rows": 300, + "has_urls_columns": True, + }, + ], + None, + { + "urls_columns": ["image_url", "label", "url"], + "num_opt_in_urls": 20, + "num_opt_out_urls": 20, + "num_urls": 150, + "num_scanned_rows": 400, + "has_urls_columns": True, + }, + False, + ), + ( + "previos_step_error", + HTTPStatus.INTERNAL_SERVER_ERROR, + {}, + [], + [], + "PreviousStepError", + None, + True, + ), + ( + "previous_step_format_error", + HTTPStatus.OK, + { + "config_names": [ + {"dataset": "dataset_ok", "config": "config1"}, + {"dataset": "dataset_ok", "config": "config2"}, + ] + }, + [HTTPStatus.OK], + [{"wrong_format": None}], + "PreviousStepFormatError", + None, + True, + ), + ], +) +def test_compute( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset: str, + config_names_status: HTTPStatus, + config_names_content: Any, + config_upstream_status: List[HTTPStatus], + config_upstream_content: List[Any], + expected_error_code: str, + expected_content: Any, + should_raise: bool, +) -> None: + upsert_response( + kind="/config-names", + dataset=dataset, + content=config_names_content, + http_status=config_names_status, + ) + + if config_names_status == HTTPStatus.OK: + for split_item, status, content in zip( + config_names_content["config_names"], config_upstream_status, config_upstream_content + ): + upsert_response( + kind="config-opt-in-out-urls-count", + dataset=dataset, + config=split_item["config"], + content=content, + http_status=status, + ) + + job_runner = get_job_runner(dataset, app_config, False) + if should_raise: + with pytest.raises(Exception) as e: + job_runner.compute() + assert e.type.__name__ == expected_error_code + else: + assert job_runner.compute().content == expected_content + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = "doesnotexist" + job_runner = get_job_runner(dataset, app_config, False) + with pytest.raises(PreviousStepError): + job_runner.compute()
847ca2b8e117040f7e16051775c2229d1939f79c
Sylvain Lesage
2023-05-03T15:26:05
Fix dataset split names (#1125)
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 54600ec2..b36f6211 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -30 +30 @@ PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION = 3 -PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 2 +PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 3 diff --git a/services/worker/src/worker/job_runners/dataset/split_names.py b/services/worker/src/worker/job_runners/dataset/split_names.py index bf0dfcfb..ed2996d8 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names.py +++ b/services/worker/src/worker/job_runners/dataset/split_names.py @@ -71,2 +71 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - - config_names_best_response = get_previous_step_or_raise(kinds=["/config-names", "dataset-info"], dataset=dataset) + config_names_best_response = get_previous_step_or_raise(kinds=["/config-names"], dataset=dataset) @@ -74,12 +73,3 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - if config_names_best_response.kind == "/config-names": - if "config_names" not in content: - raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") - config_names = [config_name_item["config"] for config_name_item in content["config_names"]] - elif config_names_best_response.kind == "dataset-info": - if "dataset_info" not in content: - raise PreviousStepFormatError("'dataset-info' did not return the expected content: 'dataset_info'.") - config_names = list(content["dataset-info"].keys()) - else: - raise PreviousStepFormatError( - "Previous step '/config-names' or 'dataset-info' did not return the expected content." - ) + if "config_names" not in content: + raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + config_names = [config_name_item["config"] for config_name_item in content["config_names"]] @@ -87,3 +77 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - raise PreviousStepFormatError( - "Previous steps '/config-names' or 'dataset-info' did not return a list of config names." - ) + raise PreviousStepFormatError("Previous step '/config-names' did not return a list of config names.")
3db2a069019389aa331791fca7f508d4eb3be99e
Polina Kazakova
2023-05-03T12:10:03
Delete `dataset-split-names-from-streaming` job runner (#1106)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 0daba420..aebdd394 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -61,0 +62,9 @@ from mongodb_migration.migrations._20230428145000_queue_delete_ttl_index import +from mongodb_migration.migrations._20230428175100_cache_delete_dataset_split_names_from_streaming import ( + MigrationCacheDeleteDatasetSplitNamesFromStreaming, +) +from mongodb_migration.migrations._20230428181800_queue_delete_dataset_split_names_from_streaming import ( + MigrationQueueDeleteDatasetSplitNamesFromStreaming, +) +from mongodb_migration.migrations._20230428193100_metrics_delete_dataset_split_names_from_streaming import ( + MigrationMetricsDeleteDatasetSplitNamesFromStreaming, +) @@ -140,0 +150,12 @@ class MigrationsCollector: + MigrationCacheDeleteDatasetSplitNamesFromStreaming( + version="20230428175100", + description="delete the cache entries of kind 'dataset-split-names-from-streaming'", + ), + MigrationQueueDeleteDatasetSplitNamesFromStreaming( + version="20230428181800", + description="delete the jobs of type 'dataset-split-names-from-streaming'", + ), + MigrationMetricsDeleteDatasetSplitNamesFromStreaming( + version="20230428193100", + description="delete the queue and cache metrics for step 'dataset-split-names-from-streaming'", + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428175100_cache_delete_dataset_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428175100_cache_delete_dataset_split_names_from_streaming.py new file mode 100644 index 00000000..bd9b5c15 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428175100_cache_delete_dataset_split_names_from_streaming.py @@ -0,0 +1,31 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + + +import logging + +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +cache_kind = "dataset-split-names-from-streaming" + + +class MigrationCacheDeleteDatasetSplitNamesFromStreaming(Migration): + def up(self) -> None: + logging.info(f"Delete cache entries of kind {cache_kind}") + db = get_db(CACHE_MONGOENGINE_ALIAS) + + # delete existing documents + db[CACHE_COLLECTION_RESPONSES].delete_many({"kind": cache_kind}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {cache_kind} kind") + + db = get_db(CACHE_MONGOENGINE_ALIAS) + if db[CACHE_COLLECTION_RESPONSES].count_documents({"kind": cache_kind}): + raise ValueError(f"Found documents with kind {cache_kind}") diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428181800_queue_delete_dataset_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428181800_queue_delete_dataset_split_names_from_streaming.py new file mode 100644 index 00000000..b15505c0 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428181800_queue_delete_dataset_split_names_from_streaming.py @@ -0,0 +1,30 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + + +import logging + +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +job_type = "dataset-split-names-from-streaming" + + +class MigrationQueueDeleteDatasetSplitNamesFromStreaming(Migration): + def up(self) -> None: + logging.info(f"Delete jobs of type {job_type}") + + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].delete_many({"type": job_type}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {job_type} type") + + db = get_db(QUEUE_MONGOENGINE_ALIAS) + if db[QUEUE_COLLECTION_JOBS].count_documents({"type": job_type}): + raise ValueError(f"Found documents with type {job_type}") diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428193100_metrics_delete_dataset_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428193100_metrics_delete_dataset_split_names_from_streaming.py new file mode 100644 index 00000000..21472b0e --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428193100_metrics_delete_dataset_split_names_from_streaming.py @@ -0,0 +1,37 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + + +import logging + +from libcommon.constants import ( + METRICS_COLLECTION_CACHE_TOTAL_METRIC, + METRICS_COLLECTION_JOB_TOTAL_METRIC, + METRICS_MONGOENGINE_ALIAS, +) +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +job_type = cache_kind = "dataset-split-names-from-streaming" + + +class MigrationMetricsDeleteDatasetSplitNamesFromStreaming(Migration): + def up(self) -> None: + logging.info(f"Delete job metrics of type {job_type}") + + db = get_db(METRICS_MONGOENGINE_ALIAS) + db[METRICS_COLLECTION_JOB_TOTAL_METRIC].delete_many({"queue": job_type}) + db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].delete_many({"kind": cache_kind}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {job_type} type or {cache_kind} kind") + + db = get_db(METRICS_MONGOENGINE_ALIAS) + if db[METRICS_COLLECTION_JOB_TOTAL_METRIC].count_documents({"queue": job_type}): + raise ValueError(f"Found documents with type {job_type}") + if db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].count_documents({"kind": cache_kind}): + raise ValueError(f"Found documents with kind {cache_kind}") diff --git a/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230424174000_cache_delete_parquet_and_dataset_info.py similarity index 100% rename from jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py rename to jobs/mongodb_migration/tests/migrations/test_20230424174000_cache_delete_parquet_and_dataset_info.py diff --git a/jobs/mongodb_migration/tests/migrations/test_20230428175100_cache_delete_dataset_split_names_from_streaming.py b/jobs/mongodb_migration/tests/migrations/test_20230428175100_cache_delete_dataset_split_names_from_streaming.py new file mode 100644 index 00000000..3c7cfcea --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230428175100_cache_delete_dataset_split_names_from_streaming.py @@ -0,0 +1,27 @@ +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230428175100_cache_delete_dataset_split_names_from_streaming import ( + MigrationCacheDeleteDatasetSplitNamesFromStreaming, +) + + +def test_cache_delete_dataset_split_names_from_streaming(mongo_host: str) -> None: + kind = "dataset-split-names-from-streaming" + with MongoResource( + database="test_cache_delete_dataset_split_names_from_streaming", host=mongo_host, mongoengine_alias="cache" + ): + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many([{"kind": kind, "dataset": "dataset", "http_status": 200}]) + assert db[CACHE_COLLECTION_RESPONSES].find_one({"kind": kind}) # Ensure there is at least one record to delete + + migration = MigrationCacheDeleteDatasetSplitNamesFromStreaming( + version="20230428180400", + description=f"remove cache for kind {kind}", + ) + migration.up() + + assert not db[CACHE_COLLECTION_RESPONSES].find_one({"kind": kind}) # Ensure 0 records with old kind + + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230428181800_queue_delete_dataset_split_names_from_streaming.py b/jobs/mongodb_migration/tests/migrations/test_20230428181800_queue_delete_dataset_split_names_from_streaming.py new file mode 100644 index 00000000..e4cb761e --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230428181800_queue_delete_dataset_split_names_from_streaming.py @@ -0,0 +1,39 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230428181800_queue_delete_dataset_split_names_from_streaming import ( + MigrationQueueDeleteDatasetSplitNamesFromStreaming, +) + + +def test_queue_delete_dataset_split_names_from_streaming(mongo_host: str) -> None: + job_type = "dataset-split-names-from-streaming" + with MongoResource( + database="test_queue_dataset_split_names_from_streaming", host=mongo_host, mongoengine_alias="queue" + ): + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].insert_many( + [ + { + "type": job_type, + "unicity_id": f"Job[{job_type}][dataset][config][split]", + "dataset": "dataset", + "http_status": 200, + } + ] + ) + assert db[QUEUE_COLLECTION_JOBS].find_one({"type": job_type}) # Ensure there is at least one record to delete + + migration = MigrationQueueDeleteDatasetSplitNamesFromStreaming( + version="20230428190400", + description=f"remove jobs of type '{job_type}'", + ) + migration.up() + + assert not db[QUEUE_COLLECTION_JOBS].find_one({"type": job_type}) # Ensure 0 records with old type + + db[QUEUE_COLLECTION_JOBS].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230428193100_metrics_delete_dataset_split_names_from_streaming.py b/jobs/mongodb_migration/tests/migrations/test_20230428193100_metrics_delete_dataset_split_names_from_streaming.py new file mode 100644 index 00000000..be8433a4 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230428193100_metrics_delete_dataset_split_names_from_streaming.py @@ -0,0 +1,48 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from libcommon.constants import ( + METRICS_COLLECTION_CACHE_TOTAL_METRIC, + METRICS_COLLECTION_JOB_TOTAL_METRIC, + METRICS_MONGOENGINE_ALIAS, +) +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230428193100_metrics_delete_dataset_split_names_from_streaming import ( + MigrationMetricsDeleteDatasetSplitNamesFromStreaming, +) + + +def test_metrics_delete_dataset_split_names_from_streaming(mongo_host: str) -> None: + step_name = job_type = cache_kind = "dataset-split-names-from-streaming" + with MongoResource( + database="test_metrics_delete_dataset_split_names_from_streaming", + host=mongo_host, + mongoengine_alias=METRICS_MONGOENGINE_ALIAS, + ): + db = get_db(METRICS_MONGOENGINE_ALIAS) + db[METRICS_COLLECTION_JOB_TOTAL_METRIC].insert_many([{"queue": job_type, "status": "waiting", "total": 0}]) + db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].insert_many([{"kind": cache_kind, "http_status": 400, "total": 0}]) + assert db[METRICS_COLLECTION_JOB_TOTAL_METRIC].find_one( + {"queue": job_type} + ) # Ensure there is at least one record to delete + assert db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].find_one( + {"kind": cache_kind} + ) # Ensure there is at least one record to delete + + migration = MigrationMetricsDeleteDatasetSplitNamesFromStreaming( + version="20230428193700", + description=f"delete the queue and cache metrics for step '{step_name}'", + ) + migration.up() + + assert not db[METRICS_COLLECTION_JOB_TOTAL_METRIC].find_one( + {"queue": job_type} + ) # Ensure 0 records after deletion + assert not db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].find_one( + {"kind": cache_kind} + ) # Ensure 0 records after deletion + + db[METRICS_COLLECTION_JOB_TOTAL_METRIC].drop() + db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].drop() diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 5d009a23..7c0655d8 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -22 +21,0 @@ from libcommon.constants import ( - PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION, @@ -245,5 +243,0 @@ class ProcessingGraphConfig: - "dataset-split-names-from-streaming": { - "input_type": "dataset", - "requires": ["/split-names-from-streaming", "/config-names"], - "job_runner_version": PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION, - }, # to be deprecated diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index ee8d82b4..54600ec2 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -24 +23,0 @@ PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION = 2 -PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION = 2 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 5414854b..fed3cdd6 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -76 +75,0 @@ def graph() -> ProcessingGraph: - "dataset-split-names-from-streaming", @@ -107 +106 @@ def graph() -> ProcessingGraph: - ["split-first-rows-from-streaming", "dataset-split-names-from-streaming", "dataset-split-names"], + ["split-first-rows-from-streaming", "dataset-split-names"], @@ -117,6 +115,0 @@ def graph() -> ProcessingGraph: - ( - "dataset-split-names-from-streaming", - [], - ["/config-names", "/split-names-from-streaming"], - ["/config-names", "/split-names-from-streaming"], - ), diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 3a663188..5f6acbb0 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -445,5 +444,0 @@ def test_dataset_state_as_dict() -> None: - { - "id": f"dataset-split-names-from-streaming,{DATASET_NAME}", - "job_state": {"is_in_process": False}, - "cache_state": {"exists": False, "is_success": False}, - }, @@ -528 +522,0 @@ def test_plan() -> None: - "dataset-split-names-from-streaming,dataset", @@ -545 +538,0 @@ def test_plan() -> None: - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -561 +553,0 @@ def test_plan_job_creation_and_termination() -> None: - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -581 +572,0 @@ def test_plan_job_creation_and_termination() -> None: - "dataset-split-names-from-streaming,dataset", @@ -597 +587,0 @@ def test_plan_job_creation_and_termination() -> None: - "dataset-split-names-from-streaming,dataset", @@ -648 +637,0 @@ def test_plan_job_creation_and_termination() -> None: - "dataset-split-names-from-streaming,dataset", @@ -663 +651,0 @@ def test_plan_job_creation_and_termination() -> None: - "dataset-split-names-from-streaming,dataset", @@ -728 +715,0 @@ def test_plan_retry_error() -> None: - "dataset-split-names-from-streaming,dataset", @@ -756 +742,0 @@ def test_plan_retry_error() -> None: - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -814 +799,0 @@ def test_plan_incoherent_state() -> None: - "dataset-split-names-from-streaming,dataset", @@ -850 +834,0 @@ def test_plan_incoherent_state() -> None: - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -926 +909,0 @@ def test_plan_updated_at() -> None: - "dataset-split-names-from-streaming,dataset", @@ -953 +935,0 @@ def test_plan_updated_at() -> None: - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -998 +979,0 @@ def test_plan_job_runner_version() -> None: - "dataset-split-names-from-streaming,dataset", @@ -1026 +1006,0 @@ def test_plan_job_runner_version() -> None: - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -1086 +1065,0 @@ def test_plan_git_revision( - "dataset-split-names-from-streaming,dataset", @@ -1113 +1091,0 @@ def test_plan_git_revision( - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -1145 +1122,0 @@ def test_plan_git_revision( - "dataset-split-names-from-streaming,dataset", @@ -1171 +1147,0 @@ def test_plan_git_revision( - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -1249 +1224,0 @@ def test_plan_update_fan_in_parent() -> None: - "dataset-split-names-from-streaming,dataset", @@ -1278 +1252,0 @@ def test_plan_update_fan_in_parent() -> None: - "CreateJob[dataset-split-names-from-streaming,dataset]", diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index db98d6a5..47b080b7 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -124 +123,0 @@ class EndpointConfig: - "dataset-split-names-from-streaming", diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index 039ab75b..e2f61053 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -39 +39 @@ def test_endpoints_definition() -> None: - assert len(splits["dataset"]) == 3 # Has four processing steps + assert len(splits["dataset"]) == 2 # Has two processing steps diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index c1d2f069..1f06598e 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -33,3 +32,0 @@ from worker.job_runners.dataset.split_names_from_dataset_info import ( -from worker.job_runners.dataset.split_names_from_streaming import ( - DatasetSplitNamesFromStreamingJobRunner, -) @@ -168,7 +164,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - if job_type == DatasetSplitNamesFromStreamingJobRunner.get_job_type(): - return DatasetSplitNamesFromStreamingJobRunner( - job_info=job_info, - processing_step=processing_step, - common_config=self.app_config.common, - worker_config=self.app_config.worker, - ) @@ -226 +215,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - DatasetSplitNamesFromStreamingJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py deleted file mode 100644 index 8db0d3ff..00000000 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py +++ /dev/null @@ -1,160 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import logging -from http import HTTPStatus -from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict - -from libcommon.constants import ( - PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION, -) -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response - -from worker.job_runner import ( - JobResult, - JobRunner, - JobRunnerError, - ParameterMissingError, - get_previous_step_or_raise, -) -from worker.utils import ConfigItem, SplitItem - -DatasetSplitNamesFromStreamingJobRunnerErrorCode = Literal["PreviousStepFormatError", "ResponseNotReady"] - - -class DatasetSplitNamesFromStreamingJobRunnerError(JobRunnerError): - """Base class for dataset split from streaming names job runner exceptions.""" - - def __init__( - self, - message: str, - status_code: HTTPStatus, - code: DatasetSplitNamesFromStreamingJobRunnerErrorCode, - cause: Optional[BaseException] = None, - disclose_cause: bool = False, - ): - super().__init__( - message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause - ) - - -class PreviousStepFormatError(DatasetSplitNamesFromStreamingJobRunnerError): - """Raised when the content of the previous step has not the expected format.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) - - -class ResponseNotReadyError(DatasetSplitNamesFromStreamingJobRunnerError): - """Raised when the response has not been processed yet from any source.""" - - def __init__(self, message: str): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseNotReady") - - -class FailedConfigItem(ConfigItem): - error: Mapping[str, Any] - - -class DatasetSplitNamesFromStreamingResponse(TypedDict): - splits: List[SplitItem] - pending: List[ConfigItem] - failed: List[FailedConfigItem] - - -def compute_dataset_split_names_from_streaming_response( - dataset: str, -) -> Tuple[DatasetSplitNamesFromStreamingResponse, float]: - """ - Get the response of /splits for one specific dataset on huggingface.co - computed from responses cached in /split-names-from-streaming step. - - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated by a `/`. - Returns: - `DatasetSplitNamesFromStreamingResponse`: An object with a list of split names for the dataset [splits], - a list of pending configs to be processed [pending] and the list of errors [failed] by config. - <Tip> - Raises the following errors: - - [`~job_runner.PreviousStepError`] - If the the previous step gave an error. - - [`~job_runners.dataset.split_names_from_streaming.PreviousStepFormatError`] - If the content of the previous step has not the expected format - </Tip> - """ - logging.info(f"get dataset split names from dataset info for dataset={dataset}") - - config_names_best_response = get_previous_step_or_raise(kinds=["/config-names"], dataset=dataset) - content = config_names_best_response.response["content"] - if "config_names" not in content: - raise PreviousStepFormatError("Previous step did not return the expected content: 'config_names'.") - config_content = content["config_names"] - - try: - splits: List[SplitItem] = [] - pending: List[ConfigItem] = [] - failed: List[FailedConfigItem] = [] - total = 0 - for config_item in config_content: - config = config_item["config"] - total += 1 - try: - response = get_response(kind="/split-names-from-streaming", dataset=dataset, config=config) - except DoesNotExist: - logging.debug("No response found in previous step '/split-names-from-streaming' for this dataset.") - pending.append(ConfigItem({"dataset": dataset, "config": config})) - continue - if response["http_status"] != HTTPStatus.OK: - logging.debug(f"Previous step gave an error: {response['http_status']}.") - failed.append( - FailedConfigItem( - { - "dataset": dataset, - "config": config, - "error": response["content"], - } - ) - ) - continue - splits.extend( - [ - SplitItem({"dataset": dataset, "config": config, "split": split_content["split"]}) - for split_content in response["content"]["splits"] - ] - ) - except Exception as e: - raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e - - progress = (total - len(pending)) / total if total else 1.0 - - return ( - DatasetSplitNamesFromStreamingResponse( - { - "splits": splits, - "pending": pending, - "failed": failed, - } - ), - progress, - ) - - -class DatasetSplitNamesFromStreamingJobRunner(JobRunner): - @staticmethod - def get_job_type() -> str: - return "dataset-split-names-from-streaming" - - @staticmethod - def get_job_runner_version() -> int: - return PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION - - def compute(self) -> JobResult: - if self.dataset is None: - raise ParameterMissingError("'dataset' parameter is required") - response_content, progress = compute_dataset_split_names_from_streaming_response(dataset=self.dataset) - return JobResult(response_content, progress=progress) - - def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]: - """Get the set of new splits, from the content created by the compute.""" - return {SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["splits"]} diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py deleted file mode 100644 index ec18d17c..00000000 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py +++ /dev/null @@ -1,269 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from http import HTTPStatus -from typing import Any, Callable - -import pytest -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Priority -from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import SplitFullName, upsert_response - -from worker.config import AppConfig -from worker.job_runner import PreviousStepError -from worker.job_runners.dataset.split_names_from_streaming import ( - DatasetSplitNamesFromStreamingJobRunner, - PreviousStepFormatError, -) - -GetJobRunner = Callable[[str, AppConfig, bool], DatasetSplitNamesFromStreamingJobRunner] - - [email protected] -def get_job_runner( - cache_mongo_resource: CacheMongoResource, - queue_mongo_resource: QueueMongoResource, -) -> GetJobRunner: - def _get_job_runner( - dataset: str, - app_config: AppConfig, - force: bool = False, - ) -> DatasetSplitNamesFromStreamingJobRunner: - return DatasetSplitNamesFromStreamingJobRunner( - job_info={ - "type": DatasetSplitNamesFromStreamingJobRunner.get_job_type(), - "dataset": dataset, - "config": None, - "split": None, - "job_id": "job_id", - "force": force, - "priority": Priority.NORMAL, - }, - common_config=app_config.common, - worker_config=app_config.worker, - processing_step=ProcessingStep( - name=DatasetSplitNamesFromStreamingJobRunner.get_job_type(), - input_type="dataset", - requires=[], - required_by_dataset_viewer=False, - ancestors=[], - children=[], - parents=[], - job_runner_version=DatasetSplitNamesFromStreamingJobRunner.get_job_runner_version(), - ), - ) - - return _get_job_runner - - [email protected]( - "dataset,split_names_from_streaming,expected_content,progress", - [ - ( - "pending_response", - [ - { - "config": "config_a", - "response": { - "splits": [ - { - "dataset": "pending_response", - "config": "config_a", - "split": "split_a", - } - ] - }, - } - ], - { - "splits": [ - { - "dataset": "pending_response", - "config": "config_a", - "split": "split_a", - }, - ], - "pending": [{"dataset": "pending_response", "config": "config_b"}], - "failed": [], - }, - 0.5, - ), - ( - "complete", - [ - { - "config": "config_a", - "response": { - "splits": [ - { - "dataset": "complete", - "config": "config_a", - "split": "split_a", - } - ] - }, - }, - { - "config": "config_b", - "response": { - "splits": [ - { - "dataset": "complete", - "config": "config_b", - "split": "split_b", - } - ] - }, - }, - ], - { - "splits": [ - { - "dataset": "complete", - "config": "config_a", - "split": "split_a", - }, - { - "dataset": "complete", - "config": "config_b", - "split": "split_b", - }, - ], - "pending": [], - "failed": [], - }, - 1, - ), - ], -) -def test_compute_progress( - app_config: AppConfig, - get_job_runner: GetJobRunner, - dataset: str, - split_names_from_streaming: Any, - expected_content: Any, - progress: float, -) -> None: - upsert_response( - kind="/config-names", - dataset=dataset, - content={ - "config_names": [ - { - "dataset": dataset, - "config": "config_a", - }, - {"dataset": dataset, "config": "config_b"}, - ] - }, - http_status=HTTPStatus.OK, - ) - for config in split_names_from_streaming: - upsert_response( - kind="/split-names-from-streaming", - dataset=dataset, - config=config["config"], - content=config["response"], - http_status=HTTPStatus.OK, - ) - job_runner = get_job_runner(dataset, app_config, False) - response = job_runner.compute() - assert response.content == expected_content - assert response.progress == progress - - -def test_compute_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset = "error" - config = "error" - upsert_response( - kind="/config-names", - dataset=dataset, - content={ - "config_names": [ - { - "dataset": dataset, - "config": config, - } - ] - }, - http_status=HTTPStatus.OK, - ) - upsert_response( - kind="/split-names-from-streaming", - dataset=dataset, - config=config, - content={}, - http_status=HTTPStatus.INTERNAL_SERVER_ERROR, - ) - job_runner = get_job_runner(dataset, app_config, False) - response = job_runner.compute() - assert response.content == { - "splits": [], - "failed": [{"dataset": dataset, "config": config, "error": {}}], - "pending": [], - } - assert response.progress == 1.0 - - -def test_compute_format_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset = "error" - config = "error" - upsert_response( - kind="/config-names", - dataset=dataset, - content={ - "config_names": [ - { - "dataset": dataset, - "config": config, - } - ] - }, - http_status=HTTPStatus.OK, - ) - upsert_response( - kind="/split-names-from-streaming", - dataset=dataset, - config=config, - content={"wrong_format": []}, - http_status=HTTPStatus.OK, - ) - job_runner = get_job_runner(dataset, app_config, False) - with pytest.raises(PreviousStepFormatError): - job_runner.compute() - - -def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset = "doesnotexist" - job_runner = get_job_runner(dataset, app_config, False) - with pytest.raises(PreviousStepError): - job_runner.compute() - - -def test_get_new_splits(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset = "dataset" - job_runner = get_job_runner(dataset, app_config, False) - content = { - "splits": [ - { - "dataset": dataset, - "config": "config_a", - "split": "split_a", - }, - { - "dataset": dataset, - "config": "config_b", - "split": "split_b", - }, - ], - "pending": [], - "failed": [], - } - expected = { - SplitFullName(dataset=dataset, config="config_a", split="split_a"), - SplitFullName(dataset=dataset, config="config_b", split="split_b"), - } - new_splits = job_runner.get_new_splits(content=content) - assert new_splits - assert new_splits == expected
d0d46288ffd662d9f2be7a43a26075f21662f3aa
Sylvain Lesage
2023-05-03T11:58:38
fix: 🐛 fix the URL for /admin/force-refresh (#1131)
diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py index aef2d748..47c9fef6 100644 --- a/front/admin_ui/app.py +++ b/front/admin_ui/app.py @@ -72 +72 @@ with gr.Blocks() as demo: - refresh_dataset_button = gr.Button("Force resfresh dataset") + refresh_dataset_button = gr.Button("Force refresh dataset") @@ -249 +249 @@ The cache is outdated or in an incoherent state. Here is the plan to backfill th - response = requests.post(f"{DSS_ENDPOINT}/admin/force-refresh/{refresh_type}?{params}", headers=headers, timeout=60) + response = requests.post(f"{DSS_ENDPOINT}/admin/force-refresh{refresh_type}?{params}", headers=headers, timeout=60)
2bf624a0d60a351c30248880378f30583758f08e
Andrea Francis Soria Jimenez
2023-05-02T20:48:07
Remove spawning url content from API /opt-in-out-urls (#1124)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 4de84c4d..5d009a23 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -27,0 +28 @@ from libcommon.constants import ( + PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_COUNT_VERSION, @@ -272,0 +274,5 @@ class ProcessingGraphConfig: + "split-opt-in-out-urls-count": { + "input_type": "split", + "requires": ["split-opt-in-out-urls-scan"], + "job_runner_version": PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_COUNT_VERSION, + }, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index e0d55214..ee8d82b4 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -31,0 +32 @@ PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 2 +PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_COUNT_VERSION = 1 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index ec81fa88..5414854b 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -213 +213 @@ def graph() -> ProcessingGraph: - [], + ["split-opt-in-out-urls-count"], @@ -223,0 +224,14 @@ def graph() -> ProcessingGraph: + ( + "split-opt-in-out-urls-count", + [], + ["split-opt-in-out-urls-scan"], + [ + "/config-names", + "/split-names-from-streaming", + "split-first-rows-from-streaming", + "/split-names-from-dataset-info", + "config-info", + "config-parquet-and-info", + "split-opt-in-out-urls-scan", + ], + ), diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index c2a30cc4..3a663188 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -285,0 +286,5 @@ def get_SPLIT_STATE_DICT(dataset: str, config: str, split: str) -> Any: + { + "id": f"split-opt-in-out-urls-count,{dataset},{config},{split}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, @@ -813,0 +819,2 @@ def test_plan_incoherent_state() -> None: + "split-opt-in-out-urls-count,dataset,config1,split1", + "split-opt-in-out-urls-count,dataset,config1,split2", @@ -847,0 +855,2 @@ def test_plan_incoherent_state() -> None: + "CreateJob[split-opt-in-out-urls-count,dataset,config1,split1]", + "CreateJob[split-opt-in-out-urls-count,dataset,config1,split2]", diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 1357f6d2..db98d6a5 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -142,2 +142,2 @@ class EndpointConfig: - "/opt-in-out-urls-scan": { - "split": ["split-opt-in-out-urls-scan"], + "/opt-in-out-urls": { + "split": ["split-opt-in-out-urls-count"], diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index 238b685e..039ab75b 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -77,0 +78,6 @@ def test_endpoints_definition() -> None: + opt_in_out_urls = definition["/opt-in-out-urls"] + assert opt_in_out_urls is not None + assert sorted(list(opt_in_out_urls)) == ["split"] + assert opt_in_out_urls["split"] is not None + assert len(opt_in_out_urls["split"]) == 1 # Only has one processing step + diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index b5ac1d2e..c1d2f069 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -41,0 +42,3 @@ from worker.job_runners.split.first_rows_from_streaming import ( +from worker.job_runners.split.opt_in_out_urls_count import ( + SplitOptInOutUrlsCountJobRunner, +) @@ -202,0 +206,8 @@ class JobRunnerFactory(BaseJobRunnerFactory): + if job_type == SplitOptInOutUrlsCountJobRunner.get_job_type(): + return SplitOptInOutUrlsCountJobRunner( + job_info=job_info, + processing_step=processing_step, + common_config=self.app_config.common, + worker_config=self.app_config.worker, + ) + @@ -219,0 +231 @@ class JobRunnerFactory(BaseJobRunnerFactory): + SplitOptInOutUrlsCountJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py new file mode 100644 index 00000000..29276d99 --- /dev/null +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py @@ -0,0 +1,104 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from http import HTTPStatus +from typing import Any, List, Literal, Mapping, Optional, TypedDict + +from libcommon.constants import PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_COUNT_VERSION +from libcommon.simple_cache import SplitFullName + +from worker.job_runner import ( + CompleteJobResult, + JobRunner, + JobRunnerError, + get_previous_step_or_raise, +) + +SplitOptInOutUrlsCountJobRunnerErrorCode = Literal["PreviousStepFormatError"] + + +class SplitOptInOutUrlsCountJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: SplitOptInOutUrlsCountJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class PreviousStepFormatError(SplitOptInOutUrlsCountJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +class OptInOutUrlsCountResponse(TypedDict): + urls_columns: List[str] + num_opt_in_urls: int + num_opt_out_urls: int + num_urls: int + num_scanned_rows: int + has_urls_columns: bool + + +def compute_opt_in_out_urls_count_response( + dataset: str, + config: str, + split: str, +) -> OptInOutUrlsCountResponse: + logging.info(f"get opt-in-out-urls-count for dataset={dataset} config={config} split={split}") + + opt_in_out_urls_scan = get_previous_step_or_raise( + kinds=["split-opt-in-out-urls-scan"], dataset=dataset, config=config, split=split + ) + + try: + content = opt_in_out_urls_scan.response["content"] + opt_in_out_urls_count = OptInOutUrlsCountResponse( + has_urls_columns=content["has_urls_columns"], + num_opt_in_urls=content["num_opt_in_urls"], + num_opt_out_urls=content["num_opt_out_urls"], + num_scanned_rows=content["num_scanned_rows"], + num_urls=content["num_urls"], + urls_columns=content["urls_columns"], + ) + except KeyError as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + return opt_in_out_urls_count + + +class SplitOptInOutUrlsCountJobRunner(JobRunner): + @staticmethod + def get_job_type() -> str: + return "split-opt-in-out-urls-count" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_COUNT_VERSION + + def compute(self) -> CompleteJobResult: + if self.config is None or self.split is None: + raise ValueError("config and split are required") + return CompleteJobResult( + compute_opt_in_out_urls_count_response( + dataset=self.dataset, + config=self.config, + split=self.split, + ) + ) + + def get_new_splits(self, _: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by compute.""" + if self.config is None or self.split is None: + raise ValueError("config and split are required") + return {SplitFullName(dataset=self.dataset, config=self.config, split=self.split)} diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py new file mode 100644 index 00000000..92921827 --- /dev/null +++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py @@ -0,0 +1,156 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Callable + +import pytest +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response + +from worker.config import AppConfig +from worker.job_runner import PreviousStepError +from worker.job_runners.split.opt_in_out_urls_count import ( + SplitOptInOutUrlsCountJobRunner, +) + + [email protected](autouse=True) +def prepare_and_clean_mongo(app_config: AppConfig) -> None: + # prepare the database before each test, and clean it afterwards + pass + + +GetJobRunner = Callable[[str, str, str, AppConfig, bool], SplitOptInOutUrlsCountJobRunner] + + [email protected] +def get_job_runner( + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + split: str, + app_config: AppConfig, + force: bool = False, + ) -> SplitOptInOutUrlsCountJobRunner: + return SplitOptInOutUrlsCountJobRunner( + job_info={ + "type": SplitOptInOutUrlsCountJobRunner.get_job_type(), + "dataset": dataset, + "config": config, + "split": split, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + common_config=app_config.common, + worker_config=app_config.worker, + processing_step=ProcessingStep( + name=SplitOptInOutUrlsCountJobRunner.get_job_type(), + input_type="split", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + parents=[], + job_runner_version=SplitOptInOutUrlsCountJobRunner.get_job_runner_version(), + ), + ) + + return _get_job_runner + + [email protected]( + "dataset,config,split,upstream_status,upstream_content,expected_error_code,expected_content,should_raise", + [ + ( + "dataset_ok", + "config_ok", + "split_ok", + HTTPStatus.OK, + { + "has_urls_columns": True, + "num_scanned_rows": 4, + "opt_in_urls": [ + {"url": "http://testurl.test/test_image3-optIn.jpg", "row_idx": 3, "column_name": "col"} + ], + "opt_out_urls": [ + {"url": "http://testurl.test/test_image-optOut.jpg", "row_idx": 0, "column_name": "col"} + ], + "urls_columns": ["col"], + "num_opt_out_urls": 1, + "num_opt_in_urls": 1, + "num_urls": 4, + }, + None, + { + "has_urls_columns": True, + "num_scanned_rows": 4, + "urls_columns": ["col"], + "num_opt_out_urls": 1, + "num_opt_in_urls": 1, + "num_urls": 4, + }, + False, + ), + ( + "dataset_previous_step_error", + "config_previous_step_error", + "split_previous_step_error", + HTTPStatus.INTERNAL_SERVER_ERROR, + {}, + "PreviousStepError", + None, + True, + ), + ( + "dataset_format_error", + "config_format_error", + "split_format_error", + HTTPStatus.OK, + {"wrong_format": None}, + "PreviousStepFormatError", + None, + True, + ), + ], +) +def test_compute( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset: str, + config: str, + split: str, + upstream_status: HTTPStatus, + upstream_content: Any, + expected_error_code: str, + expected_content: Any, + should_raise: bool, +) -> None: + upsert_response( + kind="split-opt-in-out-urls-scan", + dataset=dataset, + config=config, + split=split, + content=upstream_content, + http_status=upstream_status, + ) + job_runner = get_job_runner(dataset, config, split, app_config, False) + if should_raise: + with pytest.raises(Exception) as e: + job_runner.compute() + assert e.type.__name__ == expected_error_code + else: + assert job_runner.compute().content == expected_content + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = config = split = "doesnotexist" + job_runner = get_job_runner(dataset, config, split, app_config, False) + with pytest.raises(PreviousStepError): + job_runner.compute()
810a549ba374df6379a95067851b927b78a1978b
Sylvain Lesage
2023-05-02T13:38:04
Recompute all the cache entries with error code "UnexpectedError" (#1117)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index f322447c..a315e649 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.12.3 +version: 1.12.4 diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 692d1e31..cc0ab550 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -115 +115 @@ cacheMaintenance: - action: "skip" + action: "backfill" @@ -118,0 +119,2 @@ cacheMaintenance: + backfill: + error_codes_to_retry: "UnexpectedError" @@ -132,0 +135 @@ backfill: + error_codes_to_retry: "" diff --git a/chart/templates/cron-jobs/backfill/_container.tpl b/chart/templates/cron-jobs/backfill/_container.tpl index e2e1afb1..12899c8c 100644 --- a/chart/templates/cron-jobs/backfill/_container.tpl +++ b/chart/templates/cron-jobs/backfill/_container.tpl @@ -17,0 +18,2 @@ + - name: CACHE_MAINTENANCE_BACKFILL_ERROR_CODES_TO_RETRY + value: {{ .Values.backfill.error_codes_to_retry | quote }} diff --git a/chart/templates/jobs/cache-maintenance/_container.tpl b/chart/templates/jobs/cache-maintenance/_container.tpl index 08fad153..4b79a2cf 100644 --- a/chart/templates/jobs/cache-maintenance/_container.tpl +++ b/chart/templates/jobs/cache-maintenance/_container.tpl @@ -14,0 +15,2 @@ + - name: CACHE_MAINTENANCE_BACKFILL_ERROR_CODES_TO_RETRY + value: {{ .Values.cacheMaintenance.backfill.error_codes_to_retry | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index 7d40148a..ad96b39a 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -227,0 +228,2 @@ cacheMaintenance: + backfill: + error_codes_to_retry: "" @@ -241,0 +244 @@ backfill: + error_codes_to_retry: "" diff --git a/jobs/cache_maintenance/README.md b/jobs/cache_maintenance/README.md index 504df6e0..be7e5a7f 100644 --- a/jobs/cache_maintenance/README.md +++ b/jobs/cache_maintenance/README.md @@ -20,0 +21,4 @@ Set environment variables to configure the job (`CACHE_MAINTENANCE_` prefix): +Specific to the backfill action: + +- `CACHE_MAINTENANCE_BACKFILL_ERROR_CODES_TO_RETRY`: the list of error codes to retry. Defaults to None. + diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index a0045ae4..09517cb7 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -5 +5 @@ import logging -from typing import Optional +from typing import List, Optional @@ -15,0 +16 @@ def backfill_cache( + error_codes_to_retry: Optional[List[str]] = None, @@ -27 +28 @@ def backfill_cache( - f"{analyzed_datasets} analyzed datasets ({len(supported_dataset_infos)} datasets left):" + f"{analyzed_datasets} analyzed datasets (total: {len(supported_dataset_infos)} datasets):" @@ -39 +40,6 @@ def backfill_cache( - dataset_state = DatasetState(dataset=dataset, processing_graph=processing_graph, revision=dataset_info.sha) + dataset_state = DatasetState( + dataset=dataset, + processing_graph=processing_graph, + revision=dataset_info.sha, + error_codes_to_retry=error_codes_to_retry, + ) diff --git a/jobs/cache_maintenance/src/cache_maintenance/config.py b/jobs/cache_maintenance/src/cache_maintenance/config.py index 60858020..79a75ec5 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/config.py +++ b/jobs/cache_maintenance/src/cache_maintenance/config.py @@ -5 +5 @@ from dataclasses import dataclass, field -from typing import Optional +from typing import List, Optional @@ -16,0 +17,16 @@ from libcommon.config import ( +CACHE_MAINTENANCE_BACKFILL_ERROR_CODES_TO_RETRY = None + + +@dataclass(frozen=True) +class BackfillConfig: + error_codes_to_retry: Optional[List[str]] = CACHE_MAINTENANCE_BACKFILL_ERROR_CODES_TO_RETRY + + @classmethod + def from_env(cls) -> "BackfillConfig": + env = Env(expand_vars=True) + + return cls( + error_codes_to_retry=env.list(name="CACHE_MAINTENANCE_BACKFILL_ERROR_CODES_TO_RETRY", default=""), + ) + + @@ -27,0 +44 @@ class JobConfig: + backfill: BackfillConfig = field(default_factory=BackfillConfig) @@ -40,0 +58 @@ class JobConfig: + backfill=BackfillConfig.from_env(), diff --git a/jobs/cache_maintenance/src/cache_maintenance/main.py b/jobs/cache_maintenance/src/cache_maintenance/main.py index 1db6f002..e9acb982 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/main.py +++ b/jobs/cache_maintenance/src/cache_maintenance/main.py @@ -63,0 +64 @@ def run_job() -> None: + error_codes_to_retry=job_config.backfill.error_codes_to_retry, @@ -76 +77,2 @@ if __name__ == "__main__": - except Exception: + except Exception as e: + logging.exception(e) diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index 4fb9ff0d..08724648 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -86,3 +85,0 @@ class JobState: -ERROR_CODES_TO_RETRY: List[str] = [] - - @@ -96,0 +94 @@ class CacheState: + error_codes_to_retry: Optional[List[str]] = None @@ -121,3 +119,7 @@ class CacheState: - return self.cache_entry_metadata is not None and ( - self.cache_entry_metadata["http_status"] >= 400 - and self.cache_entry_metadata["error_code"] in ERROR_CODES_TO_RETRY + return ( + self.error_codes_to_retry is not None + and self.cache_entry_metadata is not None + and ( + self.cache_entry_metadata["http_status"] >= 400 + and self.cache_entry_metadata["error_code"] in self.error_codes_to_retry + ) @@ -142,0 +145 @@ class ArtifactState: + error_codes_to_retry: Optional[List[str]] = None @@ -165 +168,5 @@ class ArtifactState: - cache_kind=self.step.cache_kind, dataset=self.dataset, config=self.config, split=self.split + cache_kind=self.step.cache_kind, + dataset=self.dataset, + config=self.config, + split=self.split, + error_codes_to_retry=self.error_codes_to_retry, @@ -191,0 +199 @@ class SplitState: + error_codes_to_retry: Optional[List[str]] = None @@ -197 +205,7 @@ class SplitState: - step.name: ArtifactState(step=step, dataset=self.dataset, config=self.config, split=self.split) + step.name: ArtifactState( + step=step, + dataset=self.dataset, + config=self.config, + split=self.split, + error_codes_to_retry=self.error_codes_to_retry, + ) @@ -215,0 +230 @@ class ConfigState: + error_codes_to_retry: Optional[List[str]] = None @@ -223 +238,7 @@ class ConfigState: - step.name: ArtifactState(step=step, dataset=self.dataset, config=self.config, split=None) + step.name: ArtifactState( + step=step, + dataset=self.dataset, + config=self.config, + split=None, + error_codes_to_retry=self.error_codes_to_retry, + ) @@ -234 +255,7 @@ class ConfigState: - SplitState(self.dataset, self.config, split_name, processing_graph=self.processing_graph) + SplitState( + self.dataset, + self.config, + split_name, + processing_graph=self.processing_graph, + error_codes_to_retry=self.error_codes_to_retry, + ) @@ -349,0 +377 @@ class DatasetState: + error_codes_to_retry: Optional[List[str]] = None @@ -361 +389,7 @@ class DatasetState: - step.name: ArtifactState(step=step, dataset=self.dataset, config=None, split=None) + step.name: ArtifactState( + step=step, + dataset=self.dataset, + config=None, + split=None, + error_codes_to_retry=self.error_codes_to_retry, + ) @@ -370 +404,6 @@ class DatasetState: - ConfigState(dataset=self.dataset, config=config_name, processing_graph=self.processing_graph) + ConfigState( + dataset=self.dataset, + config=config_name, + processing_graph=self.processing_graph, + error_codes_to_retry=self.error_codes_to_retry, + ) diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 1899df4f..c2a30cc4 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -6 +5,0 @@ from typing import Any, Dict, List, Mapping, Optional, TypedDict -from unittest.mock import patch @@ -469,2 +468,10 @@ CONFIG_INFO_OK = {"config": CONFIG_NAME_1, "content": "not important"} -def get_dataset_state(git_revision: Optional[str] = CURRENT_GIT_REVISION) -> DatasetState: - return DatasetState(dataset=DATASET_NAME, processing_graph=PROCESSING_GRAPH, revision=git_revision) +def get_dataset_state( + git_revision: Optional[str] = CURRENT_GIT_REVISION, + error_codes_to_retry: Optional[List[str]] = None, +) -> DatasetState: + return DatasetState( + dataset=DATASET_NAME, + processing_graph=PROCESSING_GRAPH, + revision=git_revision, + error_codes_to_retry=error_codes_to_retry, + ) @@ -479,0 +487 @@ def assert_dataset_state( + error_codes_to_retry: Optional[List[str]] = None, @@ -481 +489 @@ def assert_dataset_state( - dataset_state = get_dataset_state(git_revision=git_revision) + dataset_state = get_dataset_state(git_revision=git_revision, error_codes_to_retry=error_codes_to_retry) @@ -685,59 +693,31 @@ def test_plan_retry_error() -> None: - with patch("libcommon.state.ERROR_CODES_TO_RETRY", [ERROR_CODE_TO_RETRY]): - assert_dataset_state( - # The config names are known - config_names=TWO_CONFIG_NAMES, - # The split names are not yet known - split_names_in_first_config=[], - # "/config-names,dataset" is in the cache, but it's not categorized in up to date, - # but in "cache_is_error_to_retry" due to the error code - cache_status={ - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/split-names-from-dataset-info,dataset,config1", - "/split-names-from-dataset-info,dataset,config2", - "/split-names-from-streaming,dataset,config1", - "/split-names-from-streaming,dataset,config2", - "config-info,dataset,config1", - "config-info,dataset,config2", - "config-parquet,dataset,config1", - "config-parquet,dataset,config2", - "config-parquet-and-info,dataset,config1", - "config-parquet-and-info,dataset,config2", - "config-size,dataset,config1", - "config-size,dataset,config2", - "dataset-info,dataset", - "dataset-is-valid,dataset", - "dataset-parquet,dataset", - "dataset-size,dataset", - "dataset-split-names,dataset", - "dataset-split-names-from-dataset-info,dataset", - "dataset-split-names-from-streaming,dataset", - ], - "cache_is_error_to_retry": ["/config-names,dataset"], - "cache_is_job_runner_obsolete": [], - "up_to_date": [], - }, - queue_status={"in_process": []}, - # The "/config-names,dataset" artifact will be retried - tasks=[ - "CreateJob[/config-names,dataset]", - "CreateJob[/split-names-from-dataset-info,dataset,config1]", - "CreateJob[/split-names-from-dataset-info,dataset,config2]", - "CreateJob[/split-names-from-streaming,dataset,config1]", - "CreateJob[/split-names-from-streaming,dataset,config2]", - "CreateJob[config-info,dataset,config1]", - "CreateJob[config-info,dataset,config2]", - "CreateJob[config-parquet,dataset,config1]", - "CreateJob[config-parquet,dataset,config2]", - "CreateJob[config-parquet-and-info,dataset,config1]", - "CreateJob[config-parquet-and-info,dataset,config2]", - "CreateJob[config-size,dataset,config1]", - "CreateJob[config-size,dataset,config2]", - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-is-valid,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]", + assert_dataset_state( + error_codes_to_retry=[ERROR_CODE_TO_RETRY], + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are not yet known + split_names_in_first_config=[], + # "/config-names,dataset" is in the cache, but it's not categorized in up to date, + # but in "cache_is_error_to_retry" due to the error code + cache_status={ + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-parquet-and-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "config-size,dataset,config1", + "config-size,dataset,config2", + "dataset-info,dataset", + "dataset-is-valid,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", @@ -745 +725,29 @@ def test_plan_retry_error() -> None: - ) + "cache_is_error_to_retry": ["/config-names,dataset"], + "cache_is_job_runner_obsolete": [], + "up_to_date": [], + }, + queue_status={"in_process": []}, + # The "/config-names,dataset" artifact will be retried + tasks=[ + "CreateJob[/config-names,dataset]", + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-info,dataset,config2]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[config-size,dataset,config1]", + "CreateJob[config-size,dataset,config2]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-is-valid,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + )
6e9d1bede9432a7b6acac280618d73def8bbffff
Sylvain Lesage
2023-05-02T13:36:54
fix: 🐛 extend takes one argument (a list), not one arg per elem (#1116)
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py index bada9c29..ae155743 100644 --- a/libs/libcommon/src/libcommon/exceptions.py +++ b/libs/libcommon/src/libcommon/exceptions.py @@ -40 +40 @@ class CustomError(LoggedError): - disclose_cause: bool = False, + disclose_cause: Optional[bool] = None, @@ -46,0 +47 @@ class CustomError(LoggedError): + self.disclose_cause = disclose_cause if disclose_cause is not None else cause is not None @@ -52 +52,0 @@ class CustomError(LoggedError): - self.disclose_cause = disclose_cause @@ -57 +56,0 @@ class CustomError(LoggedError): - self.disclose_cause = False diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index 54a8f26d..0f61ffcb 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -240 +240 @@ class PreviousStepError(JobRunnerError): - disclose_cause = response["details"] == response["content"] + disclose_cause = orjson_dumps(response["details"]) == orjson_dumps(response["content"]) @@ -260 +260 @@ class PreviousStepError(JobRunnerError): - error_with_cause["cause_traceback"].extend(*response["details"]["cause_traceback"]) + error_with_cause["cause_traceback"].extend(response["details"]["cause_traceback"]) diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py index 2ad532cb..e9c4ee3d 100644 --- a/services/worker/tests/test_job_runner.py +++ b/services/worker/tests/test_job_runner.py @@ -12 +12,6 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import CachedResponse, SplitFullName, upsert_response +from libcommon.simple_cache import ( + CachedResponse, + SplitFullName, + get_response_with_details, + upsert_response, +) @@ -15 +20,6 @@ from worker.config import WorkerConfig -from worker.job_runner import ERROR_CODES_TO_RETRY, CompleteJobResult, JobRunner +from worker.job_runner import ( + ERROR_CODES_TO_RETRY, + CompleteJobResult, + JobRunner, + PreviousStepError, +) @@ -376,0 +387,59 @@ def test_raise_if_parallel_response_exists( + + [email protected]("disclose_cause", [False, True]) +def test_previous_step_error(disclose_cause: bool) -> None: + dataset = "dataset" + config = "config" + split = "split" + kind = "cache_kind" + error_code = "ErrorCode" + error_message = "error message" + cause_exception = "CauseException" + cause_message = "cause message" + cause_traceback = ["traceback1", "traceback2"] + details = { + "error": error_message, + "cause_exception": cause_exception, + "cause_message": cause_message, + "cause_traceback": cause_traceback, + } + content = details if disclose_cause else {"error": error_message} + job_runner_version = 1 + dataset_git_revision = "dataset_git_revision" + progress = 1.0 + upsert_response( + kind=kind, + dataset=dataset, + config=config, + split=split, + content=content, + http_status=HTTPStatus.INTERNAL_SERVER_ERROR, + error_code=error_code, + details=details, + job_runner_version=job_runner_version, + dataset_git_revision=dataset_git_revision, + progress=progress, + ) + response = get_response_with_details(kind=kind, dataset=dataset, config=config, split=split) + error = PreviousStepError.from_response(response=response, kind=kind, dataset=dataset, config=config, split=split) + assert error.disclose_cause == disclose_cause + assert error.status_code == HTTPStatus.INTERNAL_SERVER_ERROR + assert error.code == error_code + assert error.as_response_without_cause() == { + "error": error_message, + } + assert error.as_response_with_cause() == { + "error": error_message, + "cause_exception": cause_exception, + "cause_message": cause_message, + "cause_traceback": [ + "The previous step failed, the error is copied to this step:", + f" {kind=} {dataset=} {config=} {split=}", + "---", + *cause_traceback, + ], + } + if disclose_cause: + assert error.as_response() == error.as_response_with_cause() + else: + assert error.as_response() == error.as_response_without_cause()
f2ed26e96da297680720a0868db1b242ab421430
Sylvain Lesage
2023-05-02T09:41:28
Add error detail in admin UI (#1113)
diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py index c3868c7c..aef2d748 100644 --- a/front/admin_ui/app.py +++ b/front/admin_ui/app.py @@ -8 +7,0 @@ import gradio as gr -import matplotlib.pyplot as plt @@ -11,0 +11 @@ import duckdb +import json @@ -151 +151,3 @@ with gr.Blocks() as demo: - "progress": cached_response["progress"] + "progress": cached_response["progress"], + "updated_at": cached_response["updated_at"], + "details": json.dumps(cached_response["details"]), diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index ced94f68..f5f1f0b1 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -136,0 +137 @@ def upsert_response( + updated_at: Optional[datetime] = None, @@ -145 +146 @@ def upsert_response( - updated_at=get_datetime(), + updated_at=updated_at or get_datetime(), @@ -387,0 +389,2 @@ class CacheReport(TypedDict): + details: Mapping[str, Any] + updated_at: datetime @@ -441,16 +444 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo - objects = list( - queryset.order_by("+id") - .only( - "id", - "kind", - "dataset", - "config", - "split", - "http_status", - "error_code", - "job_runner_version", - "dataset_git_revision", - "progress", - ) - .limit(limit) - ) + objects = list(queryset.order_by("+id").exclude("content").limit(limit)) @@ -465,0 +454,2 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo + "details": object.details, + "updated_at": object.updated_at, @@ -486,11 +476 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L - responses = CachedResponse.objects(kind=kind, dataset=dataset).only( - "kind", - "dataset", - "config", - "split", - "http_status", - "error_code", - "job_runner_version", - "dataset_git_revision", - "progress", - ) + responses = CachedResponse.objects(kind=kind, dataset=dataset).exclude("content") @@ -504,0 +485,2 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L + "details": response.details, + "updated_at": response.updated_at, @@ -515,2 +496,0 @@ class CacheReportWithContent(CacheReport): - details: Mapping[str, Any] - updated_at: datetime @@ -559,19 +539 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int) - objects = list( - queryset.order_by("+id") - .only( - "id", - "kind", - "dataset", - "config", - "split", - "http_status", - "error_code", - "content", - "job_runner_version", - "dataset_git_revision", - "details", - "updated_at", - "progress", - ) - .limit(limit) - ) + objects = list(queryset.order_by("+id").limit(limit)) diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index 78f2a0a4..90743b34 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -363,0 +364 @@ def test_get_cache_reports() -> None: + updated_at_a = datetime(2020, 1, 1, 0, 0, 0) @@ -368,0 +370 @@ def test_get_cache_reports() -> None: + updated_at=updated_at_a, @@ -380,0 +383 @@ def test_get_cache_reports() -> None: + updated_at_b = datetime(2020, 1, 1, 0, 0, 1) @@ -390,0 +394 @@ def test_get_cache_reports() -> None: + updated_at=updated_at_b, @@ -401,0 +406 @@ def test_get_cache_reports() -> None: + updated_at_c = datetime(2020, 1, 1, 0, 0, 2) @@ -410,0 +416 @@ def test_get_cache_reports() -> None: + updated_at=updated_at_c, @@ -418,0 +425 @@ def test_get_cache_reports() -> None: + updated_at=updated_at_c, @@ -420 +426,0 @@ def test_get_cache_reports() -> None: - @@ -429,0 +436 @@ def test_get_cache_reports() -> None: + updated_at=updated_at_c, @@ -440,0 +448,2 @@ def test_get_cache_reports() -> None: + "details": {}, + "updated_at": updated_at_a, @@ -451,0 +461,2 @@ def test_get_cache_reports() -> None: + "details": details_b, + "updated_at": updated_at_b, @@ -468,0 +480,2 @@ def test_get_cache_reports() -> None: + "details": details_c, + "updated_at": updated_at_c, @@ -479,3 +491,0 @@ def test_get_cache_reports() -> None: - REDACTED_DATE = datetime(2020, 1, 1, 0, 0, 0) - for c in response_with_content["cache_reports_with_content"]: - c["updated_at"] = REDACTED_DATE @@ -494 +504 @@ def test_get_cache_reports() -> None: - "updated_at": REDACTED_DATE, + "updated_at": updated_at_a, @@ -508 +518 @@ def test_get_cache_reports() -> None: - "updated_at": REDACTED_DATE, + "updated_at": updated_at_b, @@ -515,2 +524,0 @@ def test_get_cache_reports() -> None: - for c in response_with_content["cache_reports_with_content"]: - c["updated_at"] = REDACTED_DATE @@ -530 +538 @@ def test_get_cache_reports() -> None: - "updated_at": REDACTED_DATE, + "updated_at": updated_at_c, @@ -547,0 +556 @@ def test_get_cache_reports() -> None: + assert result_a[0]["details"] == {} @@ -555,0 +565,2 @@ def test_get_cache_reports() -> None: + assert result["details"] == details_c + assert result["updated_at"] == updated_at_c
d5ab5fa87371e7b20ba8c064bc200ff7aabd29fe
Sylvain Lesage
2023-04-30T09:52:26
feat: 🎸 restore backfill job as a cron job at 12:00 every day (#1109)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index b56d0178..f322447c 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.12.2 +version: 1.12.3 diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 9a5c43f6..692d1e31 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -115 +115 @@ cacheMaintenance: - action: "backfill" + action: "skip" @@ -129 +129,3 @@ backfill: - enabled: false + enabled: true + log: + level: "debug" @@ -131,2 +133,2 @@ backfill: - schedule: "40 */6 * * *" - # every six hour + schedule: "00 12 * * *" + # every day at 12:00 diff --git a/chart/templates/cron-jobs/backfill/_container.tpl b/chart/templates/cron-jobs/backfill/_container.tpl index e8de2a64..e2e1afb1 100644 --- a/chart/templates/cron-jobs/backfill/_container.tpl +++ b/chart/templates/cron-jobs/backfill/_container.tpl @@ -12 +11,0 @@ - {{ include "envLog" . | nindent 2 }} @@ -18,0 +18,2 @@ + - name: LOG_LEVEL + value: {{ .Values.backfill.log.level | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index 570df9c9..7d40148a 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -238,0 +239,2 @@ backfill: + log: + level: "info"
2829133093d213b1ccfedca2d5b0b0f926204fa6
Julien Chaumond
2023-04-28T18:19:44
Some Doc tweaks (#1107)
diff --git a/docs/source/first_rows.mdx b/docs/source/first_rows.mdx index df387525..96c8107b 100644 --- a/docs/source/first_rows.mdx +++ b/docs/source/first_rows.mdx @@ -30 +30 @@ def query(): - response = requests.request("GET", API_URL, headers=headers) + response = requests.get(API_URL, headers=headers) diff --git a/docs/source/parquet.mdx b/docs/source/parquet.mdx index 94c13805..a70712fb 100644 --- a/docs/source/parquet.mdx +++ b/docs/source/parquet.mdx @@ -3 +3 @@ -Datasets can be published in any format (CSV, JSONL, directories of images, etc.) to the Hub, and they are easily accessed with the 🤗 [Datasets](https://huggingface.co/docs/datasets/) library. For a more performant experience (especially when it comes to large datasets), Datasets Server automatically converts every dataset to the [Parquet](https://parquet.apache.org/) format. The Parquet files are published to the Hub on a specific `ref/convert/parquet` branch (like this `amazon_polarity` [branch](https://huggingface.co/datasets/amazon_polarity/tree/refs%2Fconvert%2Fparquet) for example). +Datasets can be published in any format (CSV, JSONL, directories of images, etc.) to the Hub, and they are easily accessed with the 🤗 [Datasets](https://huggingface.co/docs/datasets/) library. For a more performant experience (especially when it comes to large datasets), Datasets Server automatically converts every public dataset to the [Parquet](https://parquet.apache.org/) format. The Parquet files are published to the Hub on a specific `ref/convert/parquet` branch (like this `amazon_polarity` [branch](https://huggingface.co/datasets/amazon_polarity/tree/refs%2Fconvert%2Fparquet) for example). @@ -16 +16 @@ def query(): - response = requests.request("GET", API_URL, headers=headers) + response = requests.get(API_URL, headers=headers) diff --git a/docs/source/quick_start.mdx b/docs/source/quick_start.mdx index 4d41ce04..ea8b9428 100644 --- a/docs/source/quick_start.mdx +++ b/docs/source/quick_start.mdx @@ -18 +18 @@ Each feature is served through an endpoint summarized in the table below: -| [/parquet](./parquet) | GET | Get the list of parquet files of a dataset. | `dataset`: name of the datasett | +| [/parquet](./parquet) | GET | Get the list of parquet files of a dataset. | `dataset`: name of the dataset | @@ -55 +55 @@ def query(): - response = requests.request("GET", API_URL, headers=headers) + response = requests.get(API_URL, headers=headers) @@ -98 +98 @@ def query(): - response = requests.request("GET", API_URL) + response = requests.get(API_URL) @@ -137 +137 @@ def query(): - response = requests.request("GET", API_URL) + response = requests.get(API_URL) @@ -178 +178 @@ def query(): - response = requests.request("GET", API_URL) + response = requests.get(API_URL) @@ -211 +211 @@ curl https://datasets-server.huggingface.co/splits?dataset=rotten_tomatoes \ -The `/first-rows` endpoint returns a JSON list of the first 100 rows of a dataset. You should specify the dataset name, configuration name (you can find out the configuration name from the `/splits` endpoint), and split name of the dataset you'd like to preview: +The `/first-rows` endpoint returns a JSON list of the first 100 rows of a dataset. It also returns the types of data features ("columns" data types). You should specify the dataset name, configuration name (you can find out the configuration name from the `/splits` endpoint), and split name of the dataset you'd like to preview: @@ -219 +219 @@ def query(): - response = requests.request("GET", API_URL) + response = requests.get(API_URL) @@ -252 +252 @@ curl https://datasets-server.huggingface.co/first-rows?dataset=rotten_tomatoes&c -The datasets-server converts every dataset on the Hub to the [parquet](https://parquet.apache.org/) format. The `/parquet` endpoint returns a JSON list of the parquet URLs for a dataset: +The datasets-server converts every public dataset on the Hub to the [parquet](https://parquet.apache.org/) format. The `/parquet` endpoint returns a JSON list of the parquet URLs for a dataset: @@ -260 +260 @@ def query(): - response = requests.request("GET", API_URL) + response = requests.get(API_URL) diff --git a/docs/source/splits.mdx b/docs/source/splits.mdx index 8d72ade9..c9018e7d 100644 --- a/docs/source/splits.mdx +++ b/docs/source/splits.mdx @@ -24 +24 @@ def query(): - response = requests.request("GET", API_URL, headers=headers) + response = requests.get(API_URL, headers=headers) diff --git a/docs/source/valid.mdx b/docs/source/valid.mdx index af818e49..3acd4e3d 100644 --- a/docs/source/valid.mdx +++ b/docs/source/valid.mdx @@ -30 +30 @@ def query(): - response = requests.request("GET", API_URL) + response = requests.get(API_URL) @@ -86 +86 @@ def query(): - response = requests.request("GET", API_URL, headers=headers) + response = requests.get(API_URL, headers=headers)
e0983833c3940ab839404c26974c8e405aacfef0
Sylvain Lesage
2023-04-28T13:31:15
feat: 🎸 change TTL for finished jobs from 7 days to 1 day (#1105)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 22eee82c..0daba420 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -58,0 +59,3 @@ from mongodb_migration.migrations._20230427121500_metrics_delete_parquet_and_dat +from mongodb_migration.migrations._20230428145000_queue_delete_ttl_index import ( + MigrationQueueDeleteTTLIndexOnFinishedAt, +) @@ -133,0 +137,4 @@ class MigrationsCollector: + MigrationQueueDeleteTTLIndexOnFinishedAt( + version="20230428145000", + description="delete the TTL index on the 'finished_at' field in the queue database", + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428145000_queue_delete_ttl_index.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428145000_queue_delete_ttl_index.py new file mode 100644 index 00000000..fd3508b0 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230428145000_queue_delete_ttl_index.py @@ -0,0 +1,49 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +import logging +from typing import Any, List, Mapping + +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +field_name = "finished_at" + + +def get_index_names(index_information: Mapping[str, Any], field_name: str) -> List[str]: + return [ + name + for name, value in index_information.items() + if isinstance(value, dict) + and "expireAfterSeconds" in value + and "key" in value + and value["key"] == [(field_name, 1)] + ] + + +class MigrationQueueDeleteTTLIndexOnFinishedAt(Migration): + def up(self) -> None: + logging.info( + f"Delete ttl index on field {field_name}. Mongoengine will create it again with a different TTL parameter" + ) + + db = get_db(QUEUE_MONGOENGINE_ALIAS) + collection = db[QUEUE_COLLECTION_JOBS] + ttl_index_names = get_index_names(index_information=collection.index_information(), field_name=field_name) + if len(ttl_index_names) != 1: + raise ValueError(f"Expected 1 ttl index on field {field_name}, found {len(ttl_index_names)}") + collection.drop_index(ttl_index_names[0]) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info("Check that the index does not exists anymore") + + db = get_db(QUEUE_MONGOENGINE_ALIAS) + collection = db[QUEUE_COLLECTION_JOBS] + ttl_index_names = get_index_names(index_information=collection.index_information(), field_name=field_name) + if len(ttl_index_names) > 0: + raise ValueError(f"Found TTL index for field {field_name}") diff --git a/jobs/mongodb_migration/tests/migrations/test_20230428145000_queue_delete_ttl_index.py b/jobs/mongodb_migration/tests/migrations/test_20230428145000_queue_delete_ttl_index.py new file mode 100644 index 00000000..58ffa7a0 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230428145000_queue_delete_ttl_index.py @@ -0,0 +1,34 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS +from libcommon.queue import Job +from libcommon.resources import MongoResource +from libcommon.utils import get_datetime +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230428145000_queue_delete_ttl_index import ( + MigrationQueueDeleteTTLIndexOnFinishedAt, + get_index_names, +) + + +def test_queue_delete_ttl_index(mongo_host: str) -> None: + with MongoResource(database="test_queue_delete_ttl_index", host=mongo_host, mongoengine_alias="queue"): + Job(type="test", dataset="test", unicity_id="test", namespace="test", created_at=get_datetime()).save() + db = get_db(QUEUE_MONGOENGINE_ALIAS) + assert ( + len(get_index_names(db[QUEUE_COLLECTION_JOBS].index_information(), "finished_at")) == 1 + ) # Ensure the TTL index exists + + migration = MigrationQueueDeleteTTLIndexOnFinishedAt( + version="20230428145000", + description="remove ttl index on field 'finished_at'", + ) + migration.up() + + assert ( + len(get_index_names(db[QUEUE_COLLECTION_JOBS].index_information(), "finished_at")) == 0 + ) # Ensure the TTL index exists # Ensure 0 records with old type + + db[QUEUE_COLLECTION_JOBS].drop() diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index d33fc93c..e0d55214 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -13 +13 @@ QUEUE_MONGOENGINE_ALIAS = "queue" -QUEUE_TTL_SECONDS = 604_800 # 7 days +QUEUE_TTL_SECONDS = 86_400 # 1 day diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 21220785..12c47435 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -12 +12,2 @@ import pytz -from libcommon.queue import EmptyQueueError, Priority, Queue, Status +from libcommon.constants import QUEUE_TTL_SECONDS +from libcommon.queue import EmptyQueueError, Job, Priority, Queue, Status @@ -364,0 +366,12 @@ def test_queue_kill_zombies() -> None: + + +def test_has_ttl_index_on_finished_at_field() -> None: + ttl_index_names = [ + name + for name, value in Job._get_collection().index_information().items() + if "expireAfterSeconds" in value and "key" in value and value["key"] == [("finished_at", 1)] + ] + assert len(ttl_index_names) == 1 + ttl_index_name = ttl_index_names[0] + assert ttl_index_name == "finished_at_1" + assert Job._get_collection().index_information()[ttl_index_name]["expireAfterSeconds"] == QUEUE_TTL_SECONDS
352ac00745dc6ecc89c1a1b5a37e0d58cab6ade6
Sylvain Lesage
2023-04-28T11:50:04
fix: 🐛 revert to the previous value (#1103)
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 44668f5e..d33fc93c 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -13 +13 @@ QUEUE_MONGOENGINE_ALIAS = "queue" -QUEUE_TTL_SECONDS = 172_800 # 2 days +QUEUE_TTL_SECONDS = 604_800 # 7 days
259fd092c12d240d9b8d733c965c4b9362e90684
Sylvain Lesage
2023-04-28T11:31:58
feat: 🎸 add logs to backill job (#1102)
diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index 585db391..a0045ae4 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -27,2 +27,3 @@ def backfill_cache( - f"{analyzed_datasets} analyzed datasets: {backfilled_datasets} backfilled datasets" - f" ({100 * backfilled_datasets / analyzed_datasets:.2f}%), with {total_created_jobs} created jobs." + f"{analyzed_datasets} analyzed datasets ({len(supported_dataset_infos)} datasets left):" + f" {backfilled_datasets} backfilled datasets ({100 * backfilled_datasets / analyzed_datasets:.2f}%), with" + f" {total_created_jobs} created jobs."
377270df0fa64891e5af0c06745ed714795b1046
Sylvain Lesage
2023-04-28T11:26:09
feat: 🎸 add debug logs (#1101)
diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index 588d847a..585db391 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -25,2 +25,2 @@ def backfill_cache( - def log() -> None: - logging.info( + def get_log() -> str: + return ( @@ -43,0 +44 @@ def backfill_cache( + logging.debug(get_log()) @@ -45 +46 @@ def backfill_cache( - log() + logging.info(get_log()) @@ -47 +48 @@ def backfill_cache( - log() + logging.info(get_log())
3593f57d880307663220ff188e6126ecda7f8e8e
Sylvain Lesage
2023-04-28T11:24:20
feat: 🎸 add an index to the jobs collection (#1100)
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index d33fc93c..44668f5e 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -13 +13 @@ QUEUE_MONGOENGINE_ALIAS = "queue" -QUEUE_TTL_SECONDS = 604_800 # 7 days +QUEUE_TTL_SECONDS = 172_800 # 2 days diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 83e7b868..35aa36d6 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -141,0 +142 @@ class Job(Document): + ("priority", "status", "created_at", "namespace", "type", "unicity_id"),
e057599aca88679aa7172d75bb7d0fec0fb8ae13
Sylvain Lesage
2023-04-28T10:46:41
Add logs and move backfill job to one shot job (#1098)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 600399ec..b56d0178 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.12.1 +version: 1.12.2 diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 35d50a55..9a5c43f6 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -115 +115 @@ cacheMaintenance: - action: "skip" + action: "backfill" @@ -116,0 +117,2 @@ cacheMaintenance: + log: + level: "debug" @@ -127 +129 @@ backfill: - enabled: true + enabled: false diff --git a/chart/templates/jobs/cache-maintenance/_container.tpl b/chart/templates/jobs/cache-maintenance/_container.tpl index 5394c7e1..08fad153 100644 --- a/chart/templates/jobs/cache-maintenance/_container.tpl +++ b/chart/templates/jobs/cache-maintenance/_container.tpl @@ -9 +8,0 @@ - {{ include "envLog" . | nindent 2 }} @@ -15,0 +15,2 @@ + - name: LOG_LEVEL + value: {{ .Values.cacheMaintenance.log.level | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index 395a718b..570df9c9 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -225,0 +226,2 @@ cacheMaintenance: + log: + level: "info" diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index 2e65cfed..588d847a 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -22 +22 @@ def backfill_cache( - created_jobs = 0 + total_created_jobs = 0 @@ -27,2 +27,2 @@ def backfill_cache( - f" {analyzed_datasets} analyzed datasets: {backfilled_datasets} backfilled datasets" - f" ({100 * backfilled_datasets / analyzed_datasets:.2f}%), with {created_jobs} created jobs." + f"{analyzed_datasets} analyzed datasets: {backfilled_datasets} backfilled datasets" + f" ({100 * backfilled_datasets / analyzed_datasets:.2f}%), with {total_created_jobs} created jobs." @@ -39 +39,2 @@ def backfill_cache( - if dataset_state.should_be_backfilled: + created_jobs = dataset_state.backfill() + if created_jobs > 0: @@ -41 +42 @@ def backfill_cache( - created_jobs += dataset_state.backfill() + total_created_jobs += created_jobs diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index 5ead27ad..4fb9ff0d 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -6,0 +7 @@ import contextlib +import logging @@ -333,0 +335 @@ class Plan: + logging.debug(f"Running task {task.id}")
51fb7d42b6b552e4dadae10c3165d1322589ea9e
Sylvain Lesage
2023-04-28T10:24:57
fix: 🐛 report backfill every 100 anaylized datasets (#1097)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 1cecc3bf..35d50a55 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -129,2 +129,2 @@ backfill: - schedule: "0 * * * *" - # every hour + schedule: "40 */6 * * *" + # every six hour diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index 9c309e7a..2e65cfed 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -43 +43 @@ def backfill_cache( - if backfilled_datasets % log_batch == 0: + if analyzed_datasets % log_batch == 0:
cf92023ea6a89fdefecc2e21e7cb700e5a133583
Sylvain Lesage
2023-04-28T09:18:35
fix: 🐛 add missing environment variables for migration job (#1096)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index b7622deb..600399ec 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.12.0 +version: 1.12.1 diff --git a/chart/templates/jobs/mongodb-migration/_container.tpl b/chart/templates/jobs/mongodb-migration/_container.tpl index 4a78df16..b9cba938 100644 --- a/chart/templates/jobs/mongodb-migration/_container.tpl +++ b/chart/templates/jobs/mongodb-migration/_container.tpl @@ -9,0 +10 @@ + {{ include "envMetrics" . | nindent 2 }}
d380761996d62918f64819eacb198ba74515d8f7
Sylvain Lesage
2023-04-28T08:58:11
Update backfill job, and setup a cronjob in prod (#1077)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 05776ee9..b7622deb 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.11.0 +version: 1.12.0 diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index b996b418..67a25b69 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -111,0 +112,3 @@ cacheMaintenance: +backfill: + enabled: false + diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index c63fdc31..1cecc3bf 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -125,0 +126,14 @@ cacheMaintenance: +backfill: + enabled: true + action: "backfill" + schedule: "0 * * * *" + # every hour + nodeSelector: {} + resources: + requests: + cpu: 1 + limits: + cpu: 1 + memory: "512Mi" + tolerations: [] + diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index bfc62fbb..549664ce 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -85,0 +86,5 @@ app.kubernetes.io/component: "{{ include "name" . }}-metrics-collector" +{{- define "labels.backfill" -}} +{{ include "hf.labels.commons" . }} +app.kubernetes.io/component: "{{ include "name" . }}-backfill" +{{- end -}} + diff --git a/chart/templates/cron-jobs/backfill/_container.tpl b/chart/templates/cron-jobs/backfill/_container.tpl new file mode 100644 index 00000000..e8de2a64 --- /dev/null +++ b/chart/templates/cron-jobs/backfill/_container.tpl @@ -0,0 +1,19 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "containerBackfill" -}} +- name: "{{ include "name" . }}-backfill" + image: {{ include "jobs.cacheMaintenance.image" . }} + imagePullPolicy: {{ .Values.images.pullPolicy }} + securityContext: + allowPrivilegeEscalation: false + resources: {{ toYaml .Values.backfill.resources | nindent 4 }} + env: + {{ include "envLog" . | nindent 2 }} + {{ include "envCache" . | nindent 2 }} + {{ include "envQueue" . | nindent 2 }} + {{ include "envCommon" . | nindent 2 }} + {{ include "envMetrics" . | nindent 2 }} + - name: CACHE_MAINTENANCE_ACTION + value: {{ .Values.backfill.action | quote }} +{{- end -}} diff --git a/chart/templates/cron-jobs/backfill/job.yaml b/chart/templates/cron-jobs/backfill/job.yaml new file mode 100644 index 00000000..1fe18605 --- /dev/null +++ b/chart/templates/cron-jobs/backfill/job.yaml @@ -0,0 +1,24 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- if and .Values.images.jobs.cacheMaintenance .Values.backfill.enabled }} +apiVersion: batch/v1 +kind: CronJob +metadata: + labels: {{ include "labels.backfill" . | nindent 4 }} + name: "{{ include "name" . }}-job-backfill" + namespace: {{ .Release.Namespace }} +spec: + schedule: {{ .Values.backfill.schedule | quote }} + jobTemplate: + spec: + ttlSecondsAfterFinished: 3600 + template: + spec: + restartPolicy: OnFailure + {{- include "image.imagePullSecrets" . | nindent 6 }} + nodeSelector: {{ toYaml .Values.backfill.nodeSelector | nindent 12 }} + tolerations: {{ toYaml .Values.backfill.tolerations | nindent 12 }} + containers: {{ include "containerBackfill" . | nindent 12 }} + securityContext: {{ include "securityContext" . | nindent 12 }} +{{- end}} diff --git a/chart/values.yaml b/chart/values.yaml index b7176319..395a718b 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -234,0 +235,14 @@ cacheMaintenance: +backfill: + enabled: false + action: "backfill" + schedule: "0 */3 * * *" + # every 3 hours + nodeSelector: {} + resources: + requests: + cpu: 0 + limits: + cpu: 0 + tolerations: [] + + diff --git a/jobs/cache_maintenance/README.md b/jobs/cache_maintenance/README.md new file mode 100644 index 00000000..504df6e0 --- /dev/null +++ b/jobs/cache_maintenance/README.md @@ -0,0 +1,29 @@ +# Datasets server maintenance job + +> Job to run maintenance actions on the datasets-server + +Available actions: + +- `backfill`: backfill the cache (i.e. create jobs to add the missing entries or update the outdated entries) +- `metrics`: compute and store the cache and queue metrics +- `skip`: do nothing + +## Configuration + +The script can be configured using environment variables. They are grouped by scope. + +### Actions + +Set environment variables to configure the job (`CACHE_MAINTENANCE_` prefix): + +- `CACHE_MAINTENANCE_ACTION`: the action to launch, among `backfill`, `metrics`, `skip`. Defaults to `skip`. + +### Common + +See [../../libs/libcommon/README.md](../../libs/libcommon/README.md) for more information about the common configuration. + +## Launch + +```shell +make run +``` diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index a304a7df..9c309e7a 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -7,4 +7,3 @@ from typing import Optional -import libcommon -from libcommon.operations import update_dataset -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Priority +from libcommon.dataset import get_supported_dataset_infos +from libcommon.processing_graph import ProcessingGraph +from libcommon.state import DatasetState @@ -14 +13 @@ def backfill_cache( - init_processing_steps: list[ProcessingStep], + processing_graph: ProcessingGraph, @@ -18,3 +17,4 @@ def backfill_cache( - logging.info("backfill init processing steps for supported datasets") - supported_datasets = libcommon.dataset.get_supported_datasets(hf_endpoint=hf_endpoint, hf_token=hf_token) - logging.info(f"about to backfill {len(supported_datasets)} datasets") + logging.info("backfill supported datasets") + supported_dataset_infos = get_supported_dataset_infos(hf_endpoint=hf_endpoint, hf_token=hf_token) + logging.info(f"analyzing {len(supported_dataset_infos)} supported datasets") + analyzed_datasets = 0 @@ -21,0 +22 @@ def backfill_cache( + created_jobs = 0 @@ -23,9 +24,5 @@ def backfill_cache( - for dataset in libcommon.dataset.get_supported_datasets(hf_endpoint=hf_endpoint, hf_token=hf_token): - update_dataset( - dataset=dataset, - init_processing_steps=init_processing_steps, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - force=False, - priority=Priority.LOW, - do_check_support=False, + + def log() -> None: + logging.info( + f" {analyzed_datasets} analyzed datasets: {backfilled_datasets} backfilled datasets" + f" ({100 * backfilled_datasets / analyzed_datasets:.2f}%), with {created_jobs} created jobs." @@ -33 +30,13 @@ def backfill_cache( - backfilled_datasets += 1 + + for dataset_info in supported_dataset_infos: + analyzed_datasets += 1 + + dataset = dataset_info.id + if not dataset: + # should not occur + continue + dataset_state = DatasetState(dataset=dataset, processing_graph=processing_graph, revision=dataset_info.sha) + if dataset_state.should_be_backfilled: + backfilled_datasets += 1 + created_jobs += dataset_state.backfill() + @@ -35 +44,3 @@ def backfill_cache( - logging.info(f"{backfilled_datasets} datasets have been backfilled") + log() + + log() diff --git a/jobs/cache_maintenance/src/cache_maintenance/main.py b/jobs/cache_maintenance/src/cache_maintenance/main.py index 41513a40..1db6f002 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/main.py +++ b/jobs/cache_maintenance/src/cache_maintenance/main.py @@ -19 +18,0 @@ from cache_maintenance.metrics import collect_metrics -from cache_maintenance.upgrade import upgrade_cache @@ -25 +24 @@ def run_job() -> None: - supported_actions = ["backfill", "upgrade", "collect-metrics"] + supported_actions = ["backfill", "collect-metrics", "skip"] @@ -58,2 +56,0 @@ def run_job() -> None: - init_processing_steps = processing_graph.get_first_steps() - processing_steps = list(processing_graph.steps.values()) @@ -64 +61 @@ def run_job() -> None: - init_processing_steps=init_processing_steps, + processing_graph=processing_graph, @@ -68,5 +65,2 @@ def run_job() -> None: - if action == "upgrade": - upgrade_cache(processing_steps) - - if action == "collect-metrics": - collect_metrics(processing_steps) + elif action == "collect-metrics": + collect_metrics(processing_graph=processing_graph) diff --git a/jobs/cache_maintenance/src/cache_maintenance/metrics.py b/jobs/cache_maintenance/src/cache_maintenance/metrics.py index e5f8a097..45ecff05 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/metrics.py +++ b/jobs/cache_maintenance/src/cache_maintenance/metrics.py @@ -7 +7 @@ from libcommon.metrics import CacheTotalMetric, JobTotalMetric -from libcommon.processing_graph import ProcessingStep +from libcommon.processing_graph import ProcessingGraph @@ -12 +12 @@ from libcommon.simple_cache import get_responses_count_by_kind_status_and_error_ -def collect_metrics(processing_steps: list[ProcessingStep]) -> None: +def collect_metrics(processing_graph: ProcessingGraph) -> None: @@ -15 +15 @@ def collect_metrics(processing_steps: list[ProcessingStep]) -> None: - for processing_step in processing_steps: + for processing_step in processing_graph.steps.values(): diff --git a/jobs/cache_maintenance/src/cache_maintenance/upgrade.py b/jobs/cache_maintenance/src/cache_maintenance/upgrade.py deleted file mode 100644 index 52764614..00000000 --- a/jobs/cache_maintenance/src/cache_maintenance/upgrade.py +++ /dev/null @@ -1,43 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import logging -from typing import List - -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Priority, Queue -from libcommon.simple_cache import get_outdated_split_full_names_for_step - - -def upgrade_cache(processing_steps: List[ProcessingStep]) -> None: - logging.info("upgrade cache") - queue = Queue() - for processing_step in processing_steps: - current_version = processing_step.job_runner_version - cache_records = get_outdated_split_full_names_for_step(processing_step.cache_kind, current_version) - logging.info( - ( - f"processing_step={processing_step.cache_kind} ", - ( - f"{len(cache_records)} cache entries are outdated for" - f" processing_step={processing_step.cache_kind} (version {current_version}). Creating jobs to" - " update them." - ), - ) - ) - - for cache_info in cache_records: - logging.debug( - ( - f"upsert_job for processing_step={processing_step.job_type}", - f" dataset={cache_info.dataset} config={cache_info.config} split={cache_info.split}", - ), - ) - queue.upsert_job( - job_type=processing_step.job_type, - dataset=cache_info.dataset, - config=cache_info.config, - split=cache_info.split, - force=False, - priority=Priority.LOW, - ) diff --git a/jobs/cache_maintenance/tests/test_backfill_cache.py b/jobs/cache_maintenance/tests/test_backfill_cache.py deleted file mode 100644 index 8e0d6626..00000000 --- a/jobs/cache_maintenance/tests/test_backfill_cache.py +++ /dev/null @@ -1,32 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from unittest.mock import patch - -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Queue, Status - -from cache_maintenance.backfill import backfill_cache - - -def test_backfill_cache() -> None: - some_datasets = ["dummy", "dummy_2", "dummy_3"] - datasets_number = len(some_datasets) - step_name = "dummy_step" - with patch("libcommon.dataset.get_supported_datasets", return_value=some_datasets): - init_processing_steps = [ - ProcessingStep( - name=step_name, - input_type="dataset", - requires=[], - required_by_dataset_viewer=False, - ancestors=[], - children=[], - parents=[], - job_runner_version=1, - ) - ] - backfill_cache(init_processing_steps=init_processing_steps, hf_endpoint="hf_endpoint", hf_token="hf_token") - queue = Queue() - assert queue.count_jobs(Status.WAITING, step_name) == datasets_number - assert queue.count_jobs(Status.STARTED, step_name) == 0 diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py index 1e29e4bd..ca864e22 100644 --- a/jobs/cache_maintenance/tests/test_collect_metrics.py +++ b/jobs/cache_maintenance/tests/test_collect_metrics.py @@ -7 +7 @@ from libcommon.metrics import CacheTotalMetric, JobTotalMetric -from libcommon.processing_graph import ProcessingStep +from libcommon.processing_graph import ProcessingGraph @@ -20 +20,4 @@ def test_collect_metrics() -> None: - test_type = "test_type" + step_name = "test_type" + processing_graph = ProcessingGraph( + processing_graph_specification={step_name: {"input_type": "dataset", "job_runner_version": 1}} + ) @@ -22,2 +25,3 @@ def test_collect_metrics() -> None: - queue.upsert_job(job_type=test_type, dataset="dataset", config="config", split="split") - + queue.upsert_job( + job_type=processing_graph.get_step(step_name).job_type, dataset="dataset", config="config", split="split" + ) @@ -25 +29,6 @@ def test_collect_metrics() -> None: - kind=test_type, dataset=dataset, config=config, split=split, content=content, http_status=HTTPStatus.OK + kind=processing_graph.get_step(step_name).cache_kind, + dataset=dataset, + config=config, + split=split, + content=content, + http_status=HTTPStatus.OK, @@ -28,14 +37 @@ def test_collect_metrics() -> None: - processing_steps = [ - ProcessingStep( - name=test_type, - input_type="dataset", - requires=[], - required_by_dataset_viewer=False, - ancestors=[], - children=[], - parents=[], - job_runner_version=1, - ) - ] - - collect_metrics(processing_steps=processing_steps) + collect_metrics(processing_graph=processing_graph) diff --git a/jobs/cache_maintenance/tests/test_upgrade_cache.py b/jobs/cache_maintenance/tests/test_upgrade_cache.py deleted file mode 100644 index 34c8caf1..00000000 --- a/jobs/cache_maintenance/tests/test_upgrade_cache.py +++ /dev/null @@ -1,48 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from http import HTTPStatus - -import pytest -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Queue -from libcommon.simple_cache import upsert_response - -from cache_maintenance.upgrade import upgrade_cache - - [email protected]( - "job_runner_version, upserts_job", - [ - (0, True), - (1, False), - (2, False), - ], -) -def test_upgrade_cache( - job_runner_version: int, - upserts_job: bool, -) -> None: - processing_steps = [ - ProcessingStep( - name="/config-names", - input_type="dataset", - requires=[], - required_by_dataset_viewer=False, - ancestors=[], - children=[], - parents=[], - job_runner_version=1, - ) - ] - queue = Queue() - assert not queue.is_job_in_process(job_type="/config-names", dataset="dataset") - upsert_response( - kind="/config-names", - dataset="dataset", - content={}, - http_status=HTTPStatus.OK, - job_runner_version=job_runner_version, - ) - upgrade_cache(processing_steps=processing_steps) - assert queue.is_job_in_process(job_type="/config-names", dataset="dataset") == upserts_job diff --git a/jobs/mongodb_migration/README.md b/jobs/mongodb_migration/README.md index 2e0539fe..4194579c 100644 --- a/jobs/mongodb_migration/README.md +++ b/jobs/mongodb_migration/README.md @@ -9 +9 @@ The script can be configured using environment variables. They are grouped by sc -### Admin service +### Migration script @@ -20 +20 @@ See [../../libs/libcommon/README.md](../../libs/libcommon/README.md) for more in -## Script +## Launch @@ -22,17 +22,3 @@ See [../../libs/libcommon/README.md](../../libs/libcommon/README.md) for more in -The script: - -- `run`: run all the migrations. First look at the previously executed migrations. Run the new ones, and revert them in case of error. - -To launch the scripts: - -- if the image runs in a docker container: - - ```shell - docker exec -it datasets-server_mongodb_migration_1 make <SCRIPT> - ``` - -- if the image runs in a kube pod: - - ```shell - kubectl exec datasets-server-prod-mongodb_migration-5cc8f8fcd7-k7jfc -- make <SCRIPT> - ``` +```shell +make run +``` diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py index 706c0f10..810f8aae 100644 --- a/libs/libcommon/src/libcommon/dataset.py +++ b/libs/libcommon/src/libcommon/dataset.py @@ -192,0 +193,36 @@ def ask_access( +def raise_if_not_supported(dataset_info: DatasetInfo) -> None: + """ + Raise an error if the dataset is not supported by the datasets-server. + Args: + dataset_info (`DatasetInfo`): + The dataset info. + Returns: + `None` + <Tip> + Raises the following errors: + - [`~libcommon.dataset.DisabledViewerError`]: if the dataset viewer is disabled. + - [`~libcommon.dataset.DatasetNotFoundError`]: if the dataset id does not exist, or if the dataset is private + </Tip> + """ + if not dataset_info.id or dataset_info.private: + raise DatasetNotFoundError(DOES_NOT_EXIST_OR_PRIVATE_DATASET_ERROR_MESSAGE) + if dataset_info.cardData and not dataset_info.cardData.get("viewer", True): + raise DisabledViewerError("The dataset viewer has been disabled on this dataset.") + + +def is_supported(dataset_info: DatasetInfo) -> bool: + """ + Check if the dataset is supported by the datasets-server. + Args: + dataset_info (`DatasetInfo`): + The dataset info. + Returns: + `bool`: True if the dataset is supported, False otherwise. + """ + try: + raise_if_not_supported(dataset_info) + except DatasetError: + return False + return True + + @@ -260,4 +296 @@ def get_dataset_info_for_supported_datasets( - if dataset_info.private: - raise DatasetNotFoundError(DOES_NOT_EXIST_OR_PRIVATE_DATASET_ERROR_MESSAGE) - if dataset_info.cardData and not dataset_info.cardData.get("viewer", True): - raise DisabledViewerError("The dataset viewer has been disabled on this dataset.") + raise_if_not_supported(dataset_info) @@ -349,3 +382,2 @@ def check_support( -def get_supported_datasets(hf_endpoint: str, hf_token: Optional[str] = None) -> list[str]: - return [d.id for d in HfApi(endpoint=hf_endpoint, token=hf_token).list_datasets() if d.id and not d.private] - # no timeout on this function. It's used only in the admin service +def get_supported_dataset_infos(hf_endpoint: str, hf_token: Optional[str] = None) -> list[DatasetInfo]: + return [d for d in HfApi(endpoint=hf_endpoint, token=hf_token).list_datasets() if is_supported(d)] diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index f4557dc5..5ead27ad 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -327 +327,6 @@ class Plan: - def run(self) -> None: + def run(self) -> int: + """Run all the tasks in the plan. + + Returns: + The number of tasks that were run. + """ @@ -329,0 +335 @@ class Plan: + return len(self.tasks) @@ -348,0 +355 @@ class DatasetState: + should_be_backfilled: bool = field(init=False) @@ -366,0 +374 @@ class DatasetState: + self.should_be_backfilled = len(self.plan.tasks) > 0 @@ -500,2 +508,7 @@ class DatasetState: - def backfill(self) -> None: - self.plan.run() + def backfill(self) -> int: + """Backfill the cache. + + Returns: + The number of jobs created. + """ + return self.plan.run()
8e56fef3fa81e2b452283e90c5377eeaa46eef0f
Sylvain Lesage
2023-04-28T08:58:01
feat: 🎸 create children jobs even in case of error (#1069)
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index f9ca6d28..ced94f68 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -332,7 +331,0 @@ def get_best_response( -def get_split_full_names_for_dataset_and_kind(dataset: str, kind: str) -> set[SplitFullName]: - return { - SplitFullName(dataset=response.dataset, config=response.config, split=response.split) - for response in CachedResponse.objects(dataset=dataset, kind=kind).only("dataset", "config", "split") - } - - diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index 1c6b699c..78f2a0a4 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -20 +19,0 @@ from libcommon.simple_cache import ( - SplitFullName, @@ -31 +29,0 @@ from libcommon.simple_cache import ( - get_split_full_names_for_dataset_and_kind, @@ -217,39 +214,0 @@ def test_big_row() -> None: -def test_get_split_full_names_for_dataset_and_kind() -> None: - kind_a = "test_kind_a" - kind_b = "test_kind_b" - dataset_a = "test_dataset_a" - dataset_b = "test_dataset_b" - dataset_c = "test_dataset_c" - config_a = "test_config_a" - config_b = "test_config_b" - split_a = "test_split_a" - split_b = None - upsert_response(kind=kind_a, dataset=dataset_a, content={}, http_status=HTTPStatus.OK) - upsert_response( - kind=kind_b, dataset=dataset_a, config=config_a, split=split_a, content={}, http_status=HTTPStatus.OK - ) - upsert_response( - kind=kind_b, dataset=dataset_a, config=config_b, split=split_a, content={}, http_status=HTTPStatus.OK - ) - upsert_response( - kind=kind_b, dataset=dataset_a, config=config_b, split=split_b, content={}, http_status=HTTPStatus.OK - ) - upsert_response(kind=kind_a, dataset=dataset_b, content={}, http_status=HTTPStatus.OK) - result = get_split_full_names_for_dataset_and_kind(dataset=dataset_a, kind=kind_a) - expected = {SplitFullName(dataset_a, None, None)} - assert len(result) == len(expected) and all(x in expected for x in result) - # ^ compare the contents of the lists without caring about the order - result = get_split_full_names_for_dataset_and_kind(dataset=dataset_a, kind=kind_b) - expected = { - SplitFullName(dataset_a, config_a, split_a), - SplitFullName(dataset_a, config_b, split_b), - SplitFullName(dataset_a, config_b, split_a), - } - assert len(result) == len(expected) and all(x in expected for x in result) - # ^ compare the contents of the lists without caring about the order - assert get_split_full_names_for_dataset_and_kind(dataset=dataset_b, kind=kind_a) == { - SplitFullName(dataset_b, None, None) - } - assert get_split_full_names_for_dataset_and_kind(dataset=dataset_c, kind=kind_a) == set() - - diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index 0f4669e4..54a8f26d 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -24 +23,0 @@ from libcommon.simple_cache import ( - delete_response, @@ -28 +26,0 @@ from libcommon.simple_cache import ( - get_split_full_names_for_dataset_and_kind, @@ -547,10 +545,19 @@ class JobRunner(ABC): - if response_in_cache["http_status"] != HTTPStatus.OK: - # if the response is not valid, we don't create the children jobs - return - new_split_full_names_for_split: set[SplitFullName] = self.get_new_splits(response_in_cache["content"]) - new_split_full_names_for_config: set[SplitFullName] = { - SplitFullName(dataset=s.dataset, config=s.config, split=None) for s in new_split_full_names_for_split - } - new_split_full_names_for_dataset: set[SplitFullName] = { - SplitFullName(dataset=s.dataset, config=None, split=None) for s in new_split_full_names_for_config - } # should be self.dataset + if response_in_cache["http_status"] == HTTPStatus.OK: + new_split_full_names_for_split: set[SplitFullName] = self.get_new_splits(response_in_cache["content"]) + new_split_full_names_for_config: set[SplitFullName] = { + SplitFullName(dataset=s.dataset, config=s.config, split=None) for s in new_split_full_names_for_split + } + elif self.processing_step.input_type == "split": + new_split_full_names_for_split = { + SplitFullName(dataset=self.dataset, config=self.config, split=self.split) + } + new_split_full_names_for_config = {SplitFullName(dataset=self.dataset, config=self.config, split=None)} + elif self.processing_step.input_type == "config": + new_split_full_names_for_split = set() + new_split_full_names_for_config = {SplitFullName(dataset=self.dataset, config=self.config, split=None)} + + else: + new_split_full_names_for_split = set() + new_split_full_names_for_config = set() + new_split_full_names_for_dataset = {SplitFullName(dataset=self.dataset, config=None, split=None)} + @@ -565,16 +571,0 @@ class JobRunner(ABC): - # remove obsolete responses from the cache - split_full_names_in_cache = get_split_full_names_for_dataset_and_kind( - dataset=self.dataset, kind=processing_step.cache_kind - ) - split_full_names_to_delete = split_full_names_in_cache.difference(new_split_full_names) - for split_full_name in split_full_names_to_delete: - delete_response( - kind=processing_step.cache_kind, - dataset=split_full_name.dataset, - config=split_full_name.config, - split=split_full_name.split, - ) - logging.debug( - f"{len(split_full_names_to_delete)} obsolete responses" - f"of kind {processing_step.cache_kind} deleted from cache for dataset={self.dataset}" - )
c57e4a7618db38aeebeecc627e008f48a27c3ec1
Sylvain Lesage
2023-04-27T13:18:44
Add migration for metrics (#1093)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index bc2d2014..22eee82c 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -55,0 +56,3 @@ from mongodb_migration.migrations._20230424174000_cache_delete_parquet_and_datas +from mongodb_migration.migrations._20230427121500_metrics_delete_parquet_and_dataset_info import ( + MigrationMetricsDeleteParquetAndDatasetInfo, +) @@ -126,0 +130,4 @@ class MigrationsCollector: + MigrationMetricsDeleteParquetAndDatasetInfo( + version="20230427121500", + description="delete the queue and cache metrics for step '/parquet-and-dataset-info'", + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/config.py b/jobs/mongodb_migration/src/mongodb_migration/config.py index 26d60a22..770b4352 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/config.py +++ b/jobs/mongodb_migration/src/mongodb_migration/config.py @@ -7 +7 @@ from environs import Env -from libcommon.config import CacheConfig, LogConfig, QueueConfig +from libcommon.config import CacheConfig, LogConfig, MetricsConfig, QueueConfig @@ -32,0 +33 @@ class JobConfig: + metrics: MetricsConfig = field(default_factory=MetricsConfig) @@ -40,0 +42 @@ class JobConfig: + metrics=MetricsConfig.from_env(), diff --git a/jobs/mongodb_migration/src/mongodb_migration/constants.py b/jobs/mongodb_migration/src/mongodb_migration/constants.py index bf3b9f23..b1e86d08 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/constants.py +++ b/jobs/mongodb_migration/src/mongodb_migration/constants.py @@ -3,0 +4 @@ +DATABASE_MIGRATIONS_COLLECTION_MIGRATIONS = "databaseMigrations" diff --git a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py index 3ea56734..9be31dc5 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py +++ b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py @@ -11 +11,4 @@ from mongoengine.queryset.queryset import QuerySet -from mongodb_migration.constants import DATABASE_MIGRATIONS_MONGOENGINE_ALIAS +from mongodb_migration.constants import ( + DATABASE_MIGRATIONS_COLLECTION_MIGRATIONS, + DATABASE_MIGRATIONS_MONGOENGINE_ALIAS, +) @@ -42 +45 @@ class DatabaseMigration(Document): - "collection": "databaseMigrations", + "collection": DATABASE_MIGRATIONS_COLLECTION_MIGRATIONS, diff --git a/jobs/mongodb_migration/src/mongodb_migration/main.py b/jobs/mongodb_migration/src/mongodb_migration/main.py index 6a747aab..af285d96 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/main.py +++ b/jobs/mongodb_migration/src/mongodb_migration/main.py @@ -8 +8,5 @@ from libcommon.log import init_logging -from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.resources import ( + CacheMongoResource, + MetricsMongoResource, + QueueMongoResource, +) @@ -25,0 +30,3 @@ def run_job() -> None: + MetricsMongoResource( + database=job_config.metrics.mongo_database, host=job_config.metrics.mongo_url + ) as metrics_resource, @@ -37,0 +45,5 @@ def run_job() -> None: + if not metrics_resource.is_available(): + logging.warning( + "The connection to the metrics database could not be established. The migration job is skipped." + ) + return diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221116133500_queue_job_add_force.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221116133500_queue_job_add_force.py index 549b61d4..875b80a9 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221116133500_queue_job_add_force.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221116133500_queue_job_add_force.py @@ -5,0 +6 @@ import logging +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -18,2 +19,2 @@ class MigrationAddForceToJob(Migration): - db = get_db("queue") - db["jobsBlue"].update_many({"force": {"$exists": False}}, {"$set": {"force": False}}) + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many({"force": {"$exists": False}}, {"$set": {"force": False}}) @@ -23,2 +24,2 @@ class MigrationAddForceToJob(Migration): - db = get_db("queue") - db["jobsBlue"].update_many({}, {"$unset": {"force": ""}}) + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many({}, {"$unset": {"force": ""}}) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py index 5f3ab06b..28e4ddf5 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py @@ -5,0 +6 @@ import logging +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -18,2 +19,2 @@ class MigrationAddPriorityToJob(Migration): - db = get_db("queue") - db["jobsBlue"].update_many({"priority": {"$exists": False}}, {"$set": {"priority": "normal"}}) + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many({"priority": {"$exists": False}}, {"$set": {"priority": "normal"}}) @@ -23,2 +24,2 @@ class MigrationAddPriorityToJob(Migration): - db = get_db("queue") - db["jobsBlue"].update_many({}, {"$unset": {"priority": ""}}) + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many({}, {"$unset": {"priority": ""}}) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py index e10637ff..1c45c5dd 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216112500_cache_split_names_from_streaming.py @@ -5,0 +6 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -15 +15,0 @@ split_names_tmp = "/split-names-TMP" -db_name = "cache" @@ -22 +22 @@ class MigrationCacheUpdateSplitNames(Migration): - db = get_db(db_name) + db = get_db(CACHE_MONGOENGINE_ALIAS) @@ -24 +24 @@ class MigrationCacheUpdateSplitNames(Migration): - db["cachedResponsesBlue"].update_many( + db[CACHE_COLLECTION_RESPONSES].update_many( @@ -28 +28,3 @@ class MigrationCacheUpdateSplitNames(Migration): - db["cachedResponsesBlue"].update_many({"kind": split_names}, {"$set": {"kind": split_names_from_streaming}}) + db[CACHE_COLLECTION_RESPONSES].update_many( + {"kind": split_names}, {"$set": {"kind": split_names_from_streaming}} + ) @@ -30 +32 @@ class MigrationCacheUpdateSplitNames(Migration): - db["cachedResponsesBlue"].delete_many({"kind": split_names_tmp}) + db[CACHE_COLLECTION_RESPONSES].delete_many({"kind": split_names_tmp}) @@ -34,2 +36,4 @@ class MigrationCacheUpdateSplitNames(Migration): - db = get_db(db_name) - db["cachedResponsesBlue"].update_many({"kind": split_names_from_streaming}, {"$set": {"kind": split_names}}) + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many( + {"kind": split_names_from_streaming}, {"$set": {"kind": split_names}} + ) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py index 3684b9a0..74ec791a 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230216141000_queue_split_names_from_streaming.py @@ -5,0 +6 @@ import logging +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -26,2 +27,2 @@ class MigrationQueueUpdateSplitNames(Migration): - db = get_db("queue") - db["jobsBlue"].update_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many( @@ -52,2 +53,2 @@ class MigrationQueueUpdateSplitNames(Migration): - db = get_db("queue") - db["jobsBlue"].update_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many( diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py index a56f1185..a30d5236 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py @@ -5,0 +6 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -18,2 +19,2 @@ class MigrationAddProgressToCacheResponse(Migration): - db = get_db("cache") - db["cachedResponsesBlue"].update_many({"progress": {"$exists": False}}, {"$set": {"progress": 1.0}}) + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many({"progress": {"$exists": False}}, {"$set": {"progress": 1.0}}) @@ -23,2 +24,2 @@ class MigrationAddProgressToCacheResponse(Migration): - db = get_db("cache") - db["cachedResponsesBlue"].update_many({}, {"$unset": {"progress": ""}}) + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many({}, {"$unset": {"progress": ""}}) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py index bcbf4326..e052a46c 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py @@ -5,0 +6 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -18,2 +19,2 @@ class MigrationAddJobRunnerVersionToCacheResponse(Migration): - db = get_db("cache") - db["cachedResponsesBlue"].update_many( + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many( @@ -39,2 +40,2 @@ class MigrationAddJobRunnerVersionToCacheResponse(Migration): - db = get_db("cache") - db["cachedResponsesBlue"].update_many({}, {"$unset": {"job_runner_version": ""}}) + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many({}, {"$unset": {"job_runner_version": ""}}) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py index 2f261957..976b3e7d 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py @@ -5,0 +6 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -17,2 +18,2 @@ class MigrationRemoveWorkerVersionFromCachedResponse(Migration): - db = get_db("cache") - db["cachedResponsesBlue"].update_many({}, {"$unset": {"worker_version": ""}}) + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many({}, {"$unset": {"worker_version": ""}}) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320163700_cache_first_rows_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320163700_cache_first_rows_from_streaming.py index 37a9ecb5..013b7d4b 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320163700_cache_first_rows_from_streaming.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320163700_cache_first_rows_from_streaming.py @@ -5,0 +6 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -14 +14,0 @@ split_first_rows_from_streaming = "split-first-rows-from-streaming" -db_name = "cache" @@ -21 +21 @@ class MigrationCacheUpdateFirstRows(Migration): - db = get_db(db_name) + db = get_db(CACHE_MONGOENGINE_ALIAS) @@ -24 +24 @@ class MigrationCacheUpdateFirstRows(Migration): - db["cachedResponsesBlue"].update_many( + db[CACHE_COLLECTION_RESPONSES].update_many( @@ -30,2 +30,2 @@ class MigrationCacheUpdateFirstRows(Migration): - db = get_db(db_name) - db["cachedResponsesBlue"].update_many( + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many( diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320165700_queue_first_rows_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320165700_queue_first_rows_from_streaming.py index 5eeaf1f7..414c10b1 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320165700_queue_first_rows_from_streaming.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320165700_queue_first_rows_from_streaming.py @@ -5,0 +6 @@ import logging +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -14 +14,0 @@ split_first_rows_from_streaming = "split-first-rows-from-streaming" -db_name = "queue" @@ -26,2 +26,2 @@ class MigrationQueueUpdateFirstRows(Migration): - db = get_db("queue") - db["jobsBlue"].update_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many( @@ -52,2 +52,2 @@ class MigrationQueueUpdateFirstRows(Migration): - db = get_db("queue") - db["jobsBlue"].update_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many( diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323155000_cache_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323155000_cache_dataset_info.py index 53d3929a..e6bdc93c 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323155000_cache_dataset_info.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323155000_cache_dataset_info.py @@ -5,0 +6 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -14 +14,0 @@ dataset_info_updated = "dataset-info" -db_name = "cache" @@ -20 +20 @@ class MigrationCacheUpdateDatasetInfo(Migration): - db = get_db(db_name) + db = get_db(CACHE_MONGOENGINE_ALIAS) @@ -23 +23 @@ class MigrationCacheUpdateDatasetInfo(Migration): - db["cachedResponsesBlue"].update_many({"kind": dataset_info}, {"$set": {"kind": dataset_info_updated}}) + db[CACHE_COLLECTION_RESPONSES].update_many({"kind": dataset_info}, {"$set": {"kind": dataset_info_updated}}) @@ -27,2 +27,2 @@ class MigrationCacheUpdateDatasetInfo(Migration): - db = get_db(db_name) - db["cachedResponsesBlue"].update_many({"kind": dataset_info_updated}, {"$set": {"kind": dataset_info}}) + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many({"kind": dataset_info_updated}, {"$set": {"kind": dataset_info}}) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323160000_queue_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323160000_queue_dataset_info.py index 19e73670..7cc842e3 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323160000_queue_dataset_info.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323160000_queue_dataset_info.py @@ -2,0 +3 @@ import logging +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -11 +11,0 @@ dataset_info_updated = "dataset-info" -db_name = "queue" @@ -23,2 +23,2 @@ class MigrationQueueUpdateDatasetInfo(Migration): - db = get_db(db_name) - db["jobsBlue"].update_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many( @@ -49,2 +49,2 @@ class MigrationQueueUpdateDatasetInfo(Migration): - db = get_db(db_name) - db["jobsBlue"].update_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many( diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091400_queue_delete_splits.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091400_queue_delete_splits.py index e56ac03c..f83aec2c 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091400_queue_delete_splits.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091400_queue_delete_splits.py @@ -2,0 +3 @@ import logging +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -8 +8,0 @@ job_type = "/splits" -db_name = "queue" @@ -16,2 +16,2 @@ class MigrationQueueDeleteSplits(Migration): - db = get_db(db_name) - db["jobsBlue"].delete_many({"type": job_type}) + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].delete_many({"type": job_type}) @@ -25,2 +25,2 @@ class MigrationQueueDeleteSplits(Migration): - db = get_db(db_name) - if db[db_name].count_documents({"type": job_type}): + db = get_db(QUEUE_MONGOENGINE_ALIAS) + if db[QUEUE_COLLECTION_JOBS].count_documents({"type": job_type}): diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091500_cache_delete_splits.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091500_cache_delete_splits.py index 19aaa468..b2034ee3 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091500_cache_delete_splits.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091500_cache_delete_splits.py @@ -5,0 +6 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -11 +11,0 @@ cache_kind = "/splits" -db_name = "cache" @@ -17 +17 @@ class MigrationCacheDeleteSplits(Migration): - db = get_db(db_name) + db = get_db(CACHE_MONGOENGINE_ALIAS) @@ -20 +20 @@ class MigrationCacheDeleteSplits(Migration): - db["cachedResponsesBlue"].delete_many({"kind": cache_kind}) + db[CACHE_COLLECTION_RESPONSES].delete_many({"kind": cache_kind}) @@ -28,2 +28,2 @@ class MigrationCacheDeleteSplits(Migration): - db = get_db(db_name) - if db[db_name].count_documents({"kind": cache_kind}): + db = get_db(CACHE_MONGOENGINE_ALIAS) + if db[CACHE_COLLECTION_RESPONSES].count_documents({"kind": cache_kind}): diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424173000_queue_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424173000_queue_delete_parquet_and_dataset_info.py index 0c9dd3a9..375feb02 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424173000_queue_delete_parquet_and_dataset_info.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424173000_queue_delete_parquet_and_dataset_info.py @@ -6,0 +7 @@ import logging +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -12 +12,0 @@ job_type = "/parquet-and-dataset-info" -db_name = "queue" @@ -19,2 +19,2 @@ class MigrationQueueDeleteParquetAndDatasetInfo(Migration): - db = get_db(db_name) - db["jobsBlue"].delete_many({"type": job_type}) + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].delete_many({"type": job_type}) @@ -28,2 +28,2 @@ class MigrationQueueDeleteParquetAndDatasetInfo(Migration): - db = get_db(db_name) - if db[db_name].count_documents({"type": job_type}): + db = get_db(QUEUE_MONGOENGINE_ALIAS) + if db[QUEUE_COLLECTION_JOBS].count_documents({"type": job_type}): diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424174000_cache_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424174000_cache_delete_parquet_and_dataset_info.py index 3fd8edd6..a1b9a45f 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424174000_cache_delete_parquet_and_dataset_info.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424174000_cache_delete_parquet_and_dataset_info.py @@ -6,0 +7 @@ import logging +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -12 +12,0 @@ cache_kind = "/parquet-and-dataset-info" -db_name = "cache" @@ -18 +18 @@ class MigrationCacheDeleteParquetAndDatasetInfo(Migration): - db = get_db(db_name) + db = get_db(CACHE_MONGOENGINE_ALIAS) @@ -21 +21 @@ class MigrationCacheDeleteParquetAndDatasetInfo(Migration): - db["cachedResponsesBlue"].delete_many({"kind": cache_kind}) + db[CACHE_COLLECTION_RESPONSES].delete_many({"kind": cache_kind}) @@ -29,2 +29,2 @@ class MigrationCacheDeleteParquetAndDatasetInfo(Migration): - db = get_db(db_name) - if db[db_name].count_documents({"kind": cache_kind}): + db = get_db(CACHE_MONGOENGINE_ALIAS) + if db[CACHE_COLLECTION_RESPONSES].count_documents({"kind": cache_kind}): diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230427121500_metrics_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230427121500_metrics_delete_parquet_and_dataset_info.py new file mode 100644 index 00000000..afaef85d --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230427121500_metrics_delete_parquet_and_dataset_info.py @@ -0,0 +1,38 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + + +import logging + +from libcommon.constants import ( + METRICS_COLLECTION_CACHE_TOTAL_METRIC, + METRICS_COLLECTION_JOB_TOTAL_METRIC, + METRICS_MONGOENGINE_ALIAS, +) +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +job_type = "/parquet-and-dataset-info" +cache_kind = "/parquet-and-dataset-info" + + +class MigrationMetricsDeleteParquetAndDatasetInfo(Migration): + def up(self) -> None: + logging.info(f"Delete job metrics of type {job_type}") + + db = get_db(METRICS_MONGOENGINE_ALIAS) + db[METRICS_COLLECTION_JOB_TOTAL_METRIC].delete_many({"queue": job_type}) + db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].delete_many({"kind": cache_kind}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {job_type} type or {cache_kind} kind") + + db = get_db(METRICS_MONGOENGINE_ALIAS) + if db[METRICS_COLLECTION_JOB_TOTAL_METRIC].count_documents({"queue": job_type}): + raise ValueError(f"Found documents with type {job_type}") + if db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].count_documents({"kind": cache_kind}): + raise ValueError(f"Found documents with kind {cache_kind}") diff --git a/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py b/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py index 5d3563bb..09333d62 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230309141600_cache_add_job_runner_version.py @@ -5,0 +6 @@ import pytest +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -16,2 +17,2 @@ def test_cache_add_job_runner_version_without_worker_version(mongo_host: str) -> - db = get_db("cache") - db["cachedResponsesBlue"].insert_many( + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many( @@ -24 +25 @@ def test_cache_add_job_runner_version_without_worker_version(mongo_host: str) -> - result = db["cachedResponsesBlue"].find_one({"dataset": "dataset_without_worker_version"}) + result = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": "dataset_without_worker_version"}) @@ -27 +28 @@ def test_cache_add_job_runner_version_without_worker_version(mongo_host: str) -> - db["cachedResponsesBlue"].drop() + db[CACHE_COLLECTION_RESPONSES].drop() @@ -41,2 +42,2 @@ def test_cache_add_job_runner_version(mongo_host: str, worker_version: str, expe - db = get_db("cache") - db["cachedResponsesBlue"].insert_many( + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many( @@ -49 +50 @@ def test_cache_add_job_runner_version(mongo_host: str, worker_version: str, expe - result = db["cachedResponsesBlue"].find_one({"dataset": "dataset"}) + result = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": "dataset"}) @@ -52 +53 @@ def test_cache_add_job_runner_version(mongo_host: str, worker_version: str, expe - db["cachedResponsesBlue"].drop() + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py index 8a94eec7..1eff4c01 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py @@ -3,0 +4 @@ +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -16,3 +17,3 @@ def test_cache_remove_worker_version(mongo_host: str) -> None: - db = get_db("cache") - db["cachedResponsesBlue"].delete_many({}) - db["cachedResponsesBlue"].insert_many( + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].delete_many({}) + db[CACHE_COLLECTION_RESPONSES].insert_many( @@ -32 +33 @@ def test_cache_remove_worker_version(mongo_host: str) -> None: - result = db["cachedResponsesBlue"].find_one({"dataset": "dataset_without_worker_version"}) + result = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": "dataset_without_worker_version"}) @@ -38 +39 @@ def test_cache_remove_worker_version(mongo_host: str) -> None: - db["cachedResponsesBlue"].drop() + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230320163700_cache_first_rows_from_streaming.py b/jobs/mongodb_migration/tests/migrations/test_20230320163700_cache_first_rows_from_streaming.py index 7c3b5bca..023f969d 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230320163700_cache_first_rows_from_streaming.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230320163700_cache_first_rows_from_streaming.py @@ -2,0 +3 @@ +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -13,3 +14,3 @@ def test_cache_update_first_rows_kind(mongo_host: str) -> None: - db = get_db("cache") - db["cachedResponsesBlue"].insert_many([{"kind": "/first-rows", "dataset": "dataset", "http_status": 200}]) - assert db["cachedResponsesBlue"].find_one( + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many([{"kind": "/first-rows", "dataset": "dataset", "http_status": 200}]) + assert db[CACHE_COLLECTION_RESPONSES].find_one( @@ -25 +26 @@ def test_cache_update_first_rows_kind(mongo_host: str) -> None: - assert not db["cachedResponsesBlue"].find_one({"kind": "/first-rows"}) # Ensure 0 records with old kind + assert not db[CACHE_COLLECTION_RESPONSES].find_one({"kind": "/first-rows"}) # Ensure 0 records with old kind @@ -27 +28 @@ def test_cache_update_first_rows_kind(mongo_host: str) -> None: - assert db["cachedResponsesBlue"].find_one({"kind": "split-first-rows-from-streaming"}) + assert db[CACHE_COLLECTION_RESPONSES].find_one({"kind": "split-first-rows-from-streaming"}) @@ -29 +30 @@ def test_cache_update_first_rows_kind(mongo_host: str) -> None: - db["cachedResponsesBlue"].drop() + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230320165700_queue_first_rows_from_streaming.py b/jobs/mongodb_migration/tests/migrations/test_20230320165700_queue_first_rows_from_streaming.py index 4ee385da..79e21656 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230320165700_queue_first_rows_from_streaming.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230320165700_queue_first_rows_from_streaming.py @@ -2,0 +3 @@ +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -15,2 +16,2 @@ def test_queue_update_first_rows_type_and_unicity_id(mongo_host: str) -> None: - db = get_db("queue") - db["jobsBlue"].insert_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].insert_many( @@ -26 +27,3 @@ def test_queue_update_first_rows_type_and_unicity_id(mongo_host: str) -> None: - assert db["jobsBlue"].find_one({"type": "/first-rows"}) # Ensure there is at least one record to update + assert db[QUEUE_COLLECTION_JOBS].find_one( + {"type": "/first-rows"} + ) # Ensure there is at least one record to update @@ -36 +39 @@ def test_queue_update_first_rows_type_and_unicity_id(mongo_host: str) -> None: - assert not db["jobsBlue"].find_one({"type": "/first-rows"}) # Ensure 0 records with old type + assert not db[QUEUE_COLLECTION_JOBS].find_one({"type": "/first-rows"}) # Ensure 0 records with old type @@ -38 +41 @@ def test_queue_update_first_rows_type_and_unicity_id(mongo_host: str) -> None: - result = db["jobsBlue"].find_one({"type": "split-first-rows-from-streaming"}) + result = db[QUEUE_COLLECTION_JOBS].find_one({"type": "split-first-rows-from-streaming"}) @@ -41 +44 @@ def test_queue_update_first_rows_type_and_unicity_id(mongo_host: str) -> None: - db["jobsBlue"].drop() + db[QUEUE_COLLECTION_JOBS].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230323155000_cache_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230323155000_cache_dataset_info.py index 5d1f3b1b..680b2879 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230323155000_cache_dataset_info.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230323155000_cache_dataset_info.py @@ -3,0 +4 @@ +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -15,3 +16,5 @@ def test_cache_update_dataset_info_kind(mongo_host: str) -> None: - db = get_db("cache") - db["cachedResponsesBlue"].insert_many([{"kind": old_kind, "dataset": "dataset", "http_status": 200}]) - assert db["cachedResponsesBlue"].find_one({"kind": old_kind}) # Ensure there is at least one record to update + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many([{"kind": old_kind, "dataset": "dataset", "http_status": 200}]) + assert db[CACHE_COLLECTION_RESPONSES].find_one( + {"kind": old_kind} + ) # Ensure there is at least one record to update @@ -25 +28 @@ def test_cache_update_dataset_info_kind(mongo_host: str) -> None: - assert not db["cachedResponsesBlue"].find_one({"kind": old_kind}) # Ensure 0 records with old kind + assert not db[CACHE_COLLECTION_RESPONSES].find_one({"kind": old_kind}) # Ensure 0 records with old kind @@ -27 +30 @@ def test_cache_update_dataset_info_kind(mongo_host: str) -> None: - assert db["cachedResponsesBlue"].find_one({"kind": new_kind}) + assert db[CACHE_COLLECTION_RESPONSES].find_one({"kind": new_kind}) @@ -29 +32 @@ def test_cache_update_dataset_info_kind(mongo_host: str) -> None: - db["cachedResponsesBlue"].drop() + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230323160000_queue_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230323160000_queue_dataset_info.py index 423bef04..4005b47a 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230323160000_queue_dataset_info.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230323160000_queue_dataset_info.py @@ -3,0 +4 @@ +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -17,2 +18,2 @@ def test_queue_update_dataset_info_type_and_unicity_id(mongo_host: str) -> None: - db = get_db("queue") - db["jobsBlue"].insert_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].insert_many( @@ -28 +29 @@ def test_queue_update_dataset_info_type_and_unicity_id(mongo_host: str) -> None: - assert db["jobsBlue"].find_one({"type": old_kind}) # Ensure there is at least one record to update + assert db[QUEUE_COLLECTION_JOBS].find_one({"type": old_kind}) # Ensure there is at least one record to update @@ -36 +37 @@ def test_queue_update_dataset_info_type_and_unicity_id(mongo_host: str) -> None: - assert not db["jobsBlue"].find_one({"type": old_kind}) # Ensure 0 records with old type + assert not db[QUEUE_COLLECTION_JOBS].find_one({"type": old_kind}) # Ensure 0 records with old type @@ -38 +39 @@ def test_queue_update_dataset_info_type_and_unicity_id(mongo_host: str) -> None: - result = db["jobsBlue"].find_one({"type": new_kind}) + result = db[QUEUE_COLLECTION_JOBS].find_one({"type": new_kind}) @@ -41 +42 @@ def test_queue_update_dataset_info_type_and_unicity_id(mongo_host: str) -> None: - db["jobsBlue"].drop() + db[QUEUE_COLLECTION_JOBS].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230407091400_queue_delete_splits.py b/jobs/mongodb_migration/tests/migrations/test_20230407091400_queue_delete_splits.py index 96b722a6..1e96842d 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230407091400_queue_delete_splits.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230407091400_queue_delete_splits.py @@ -3,0 +4 @@ +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -15,2 +16,2 @@ def test_queue_remove_splits(mongo_host: str) -> None: - db = get_db("queue") - db["jobsBlue"].insert_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].insert_many( @@ -26 +27 @@ def test_queue_remove_splits(mongo_host: str) -> None: - assert db["jobsBlue"].find_one({"type": job_type}) # Ensure there is at least one record to delete + assert db[QUEUE_COLLECTION_JOBS].find_one({"type": job_type}) # Ensure there is at least one record to delete @@ -34 +35 @@ def test_queue_remove_splits(mongo_host: str) -> None: - assert not db["jobsBlue"].find_one({"type": job_type}) # Ensure 0 records with old type + assert not db[QUEUE_COLLECTION_JOBS].find_one({"type": job_type}) # Ensure 0 records with old type @@ -36 +37 @@ def test_queue_remove_splits(mongo_host: str) -> None: - db["jobsBlue"].drop() + db[QUEUE_COLLECTION_JOBS].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230407091500_cache_remove_splits.py b/jobs/mongodb_migration/tests/migrations/test_20230407091500_cache_remove_splits.py index 5be77a69..8738a4be 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230407091500_cache_remove_splits.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230407091500_cache_remove_splits.py @@ -3,0 +4 @@ +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -15,3 +16,3 @@ def test_cache_remove_splits(mongo_host: str) -> None: - db = get_db("cache") - db["cachedResponsesBlue"].insert_many([{"kind": kind, "dataset": "dataset", "http_status": 200}]) - assert db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure there is at least one record to update + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many([{"kind": kind, "dataset": "dataset", "http_status": 200}]) + assert db[CACHE_COLLECTION_RESPONSES].find_one({"kind": kind}) # Ensure there is at least one record to update @@ -25 +26 @@ def test_cache_remove_splits(mongo_host: str) -> None: - assert not db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure 0 records with old kind + assert not db[CACHE_COLLECTION_RESPONSES].find_one({"kind": kind}) # Ensure 0 records with old kind @@ -27 +28 @@ def test_cache_remove_splits(mongo_host: str) -> None: - db["cachedResponsesBlue"].drop() + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py index 1e37b92d..e3a6d039 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py @@ -3,0 +4 @@ +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -17,3 +18,3 @@ def test_cache_delete_parquet_and_dataset_info(mongo_host: str) -> None: - db = get_db("cache") - db["cachedResponsesBlue"].insert_many([{"kind": kind, "dataset": "dataset", "http_status": 200}]) - assert db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure there is at least one record to delete + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many([{"kind": kind, "dataset": "dataset", "http_status": 200}]) + assert db[CACHE_COLLECTION_RESPONSES].find_one({"kind": kind}) # Ensure there is at least one record to delete @@ -27 +28 @@ def test_cache_delete_parquet_and_dataset_info(mongo_host: str) -> None: - assert not db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure 0 records with old kind + assert not db[CACHE_COLLECTION_RESPONSES].find_one({"kind": kind}) # Ensure 0 records with old kind @@ -29 +30 @@ def test_cache_delete_parquet_and_dataset_info(mongo_host: str) -> None: - db["cachedResponsesBlue"].drop() + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230424173000_queue_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230424173000_queue_delete_parquet_and_dataset_info.py index 4219f2d1..91c6434e 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230424173000_queue_delete_parquet_and_dataset_info.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230424173000_queue_delete_parquet_and_dataset_info.py @@ -3,0 +4 @@ +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS @@ -17,2 +18,2 @@ def test_queue_delete_parquet_and_dataset_info(mongo_host: str) -> None: - db = get_db("queue") - db["jobsBlue"].insert_many( + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].insert_many( @@ -28 +29 @@ def test_queue_delete_parquet_and_dataset_info(mongo_host: str) -> None: - assert db["jobsBlue"].find_one({"type": job_type}) # Ensure there is at least one record to delete + assert db[QUEUE_COLLECTION_JOBS].find_one({"type": job_type}) # Ensure there is at least one record to delete @@ -36 +37 @@ def test_queue_delete_parquet_and_dataset_info(mongo_host: str) -> None: - assert not db["jobsBlue"].find_one({"type": job_type}) # Ensure 0 records with old type + assert not db[QUEUE_COLLECTION_JOBS].find_one({"type": job_type}) # Ensure 0 records with old type @@ -38 +39 @@ def test_queue_delete_parquet_and_dataset_info(mongo_host: str) -> None: - db["jobsBlue"].drop() + db[QUEUE_COLLECTION_JOBS].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230427121500_metrics_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230427121500_metrics_delete_parquet_and_dataset_info.py new file mode 100644 index 00000000..9494234c --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230427121500_metrics_delete_parquet_and_dataset_info.py @@ -0,0 +1,50 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from libcommon.constants import ( + METRICS_COLLECTION_CACHE_TOTAL_METRIC, + METRICS_COLLECTION_JOB_TOTAL_METRIC, + METRICS_MONGOENGINE_ALIAS, +) +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230427121500_metrics_delete_parquet_and_dataset_info import ( + MigrationMetricsDeleteParquetAndDatasetInfo, +) + + +def test_metrics_delete_parquet_and_dataset_info(mongo_host: str) -> None: + step_name = "/parquet-and-dataset-info" + job_type = step_name + cache_kind = step_name + with MongoResource( + database="test_metrics_delete_parquet_and_dataset_info", + host=mongo_host, + mongoengine_alias=METRICS_MONGOENGINE_ALIAS, + ): + db = get_db(METRICS_MONGOENGINE_ALIAS) + db[METRICS_COLLECTION_JOB_TOTAL_METRIC].insert_many([{"queue": job_type, "status": "waiting", "total": 0}]) + db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].insert_many([{"kind": cache_kind, "http_status": 400, "total": 0}]) + assert db[METRICS_COLLECTION_JOB_TOTAL_METRIC].find_one( + {"queue": job_type} + ) # Ensure there is at least one record to delete + assert db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].find_one( + {"kind": cache_kind} + ) # Ensure there is at least one record to delete + + migration = MigrationMetricsDeleteParquetAndDatasetInfo( + version="20230427121500", + description=f"delete the queue and cache metrics for step '{step_name}'", + ) + migration.up() + + assert not db[METRICS_COLLECTION_JOB_TOTAL_METRIC].find_one( + {"queue": job_type} + ) # Ensure 0 records after deletion + assert not db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].find_one( + {"kind": cache_kind} + ) # Ensure 0 records after deletion + + db[METRICS_COLLECTION_JOB_TOTAL_METRIC].drop() + db[METRICS_COLLECTION_CACHE_TOTAL_METRIC].drop() diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 840569ff..d33fc93c 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -5 +5 @@ ASSETS_CACHE_APPNAME = "datasets_server_assets" -CACHED_ASSETS_CACHE_APPNAME = "datasets_server_cached_assets" +CACHE_COLLECTION_RESPONSES = "cachedResponsesBlue" @@ -7 +7,3 @@ CACHE_MONGOENGINE_ALIAS = "cache" -QUEUE_MONGOENGINE_ALIAS = "queue" +CACHED_ASSETS_CACHE_APPNAME = "datasets_server_cached_assets" +METRICS_COLLECTION_CACHE_TOTAL_METRIC = "cacheTotalMetric" +METRICS_COLLECTION_JOB_TOTAL_METRIC = "jobTotalMetric" @@ -8,0 +11,2 @@ METRICS_MONGOENGINE_ALIAS = "metrics" +QUEUE_COLLECTION_JOBS = "jobsBlue" +QUEUE_MONGOENGINE_ALIAS = "queue" diff --git a/libs/libcommon/src/libcommon/metrics.py b/libs/libcommon/src/libcommon/metrics.py index 6d60a107..8c4343ce 100644 --- a/libs/libcommon/src/libcommon/metrics.py +++ b/libs/libcommon/src/libcommon/metrics.py @@ -12 +12,5 @@ from mongoengine.queryset.queryset import QuerySet -from libcommon.constants import METRICS_MONGOENGINE_ALIAS +from libcommon.constants import ( + METRICS_COLLECTION_CACHE_TOTAL_METRIC, + METRICS_COLLECTION_JOB_TOTAL_METRIC, + METRICS_MONGOENGINE_ALIAS, +) @@ -52 +56 @@ class JobTotalMetric(Document): - "collection": "jobTotalMetric", + "collection": METRICS_COLLECTION_JOB_TOTAL_METRIC, @@ -78 +82 @@ class CacheTotalMetric(Document): - "collection": "cacheTotalMetric", + "collection": METRICS_COLLECTION_CACHE_TOTAL_METRIC, diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 02938948..83e7b868 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -19 +19,5 @@ from mongoengine.queryset.queryset import QuerySet -from libcommon.constants import QUEUE_MONGOENGINE_ALIAS, QUEUE_TTL_SECONDS +from libcommon.constants import ( + QUEUE_COLLECTION_JOBS, + QUEUE_MONGOENGINE_ALIAS, + QUEUE_TTL_SECONDS, +) @@ -127 +131 @@ class Job(Document): - "collection": "jobsBlue", + "collection": QUEUE_COLLECTION_JOBS, diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index 24141b27..f9ca6d28 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -35 +35 @@ from mongoengine.queryset.queryset import QuerySet -from libcommon.constants import CACHE_MONGOENGINE_ALIAS +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS @@ -103 +103 @@ class CachedResponse(Document): - "collection": "cachedResponsesBlue", + "collection": CACHE_COLLECTION_RESPONSES,
847da8d7a11eba19dee40f57b00ec85df977065a
Sylvain Lesage
2023-04-27T12:19:11
Add WORKER_JOB_TYPES_BLOCKED (#1092)
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index e413d595..b996b418 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -202 +202,2 @@ workers: - workerOnlyJobTypes: "" + workerJobTypesBlocked: "" + workerJobTypesOnly: "" @@ -216 +217,2 @@ workers: - workerOnlyJobTypes: "config-parquet,dataset-parquet,config-info,dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" + workerJobTypesBlocked: "/config-names,/split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan" + workerJobTypesOnly: "" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index d610b212..c63fdc31 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -239 +239,2 @@ workers: - workerOnlyJobTypes: "" + workerJobTypesBlocked: "" + workerJobTypesOnly: "" @@ -254 +255,2 @@ workers: - workerOnlyJobTypes: "/config-names,config-parquet,dataset-parquet,config-info,dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" + workerJobTypesBlocked: "/config-names,/split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan" + workerJobTypesOnly: "" diff --git a/chart/templates/worker/_container.tpl b/chart/templates/worker/_container.tpl index bcd65a1e..44f17fa3 100644 --- a/chart/templates/worker/_container.tpl +++ b/chart/templates/worker/_container.tpl @@ -20,2 +20,4 @@ - - name: WORKER_ONLY_JOB_TYPES - value: {{ .workerValues.workerOnlyJobTypes | quote }} + - name: WORKER_JOB_TYPES_BLOCKED + value: {{ .workerValues.workerJobTypesBlocked | quote }} + - name: WORKER_JOB_TYPES_ONLY + value: {{ .workerValues.workerJobTypesOnly | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index c37fca78..b7176319 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -367,0 +368,2 @@ workers: + # job types that this worker will not process + workerJobTypesBlocked: "" @@ -369,16 +371 @@ workers: - workerOnlyJobTypes: "" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - - - # name of the deployment - deployName: "light" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "config-parquet,dataset-parquet,config-info,dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" + workerJobTypesOnly: "" diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 80ea87de..02938948 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -328 +328,4 @@ class Queue: - self, priority: Priority, only_job_types: Optional[list[str]] = None + self, + priority: Priority, + job_types_blocked: Optional[list[str]] = None, + job_types_only: Optional[list[str]] = None, @@ -340 +343,2 @@ class Queue: - only_job_types: if not None, only jobs of the given types are considered. + job_types_blocked: if not None, jobs of the given types are not considered. + job_types_only: if not None, only jobs of the given types are considered. @@ -347,5 +351,3 @@ class Queue: - logging.debug("Getting next waiting job for priority %s, among types %s", priority, only_job_types or "all") - started_jobs = ( - Job.objects(type__in=only_job_types, status=Status.STARTED) - if only_job_types - else Job.objects(status=Status.STARTED) + logging.debug( + f"Getting next waiting job for priority {priority}, blocked types: {job_types_blocked}, only types:" + f" {job_types_only}" @@ -352,0 +355,6 @@ class Queue: + filters = {} + if job_types_blocked: + filters["type__nin"] = job_types_blocked + if job_types_only: + filters["type__in"] = job_types_only + started_jobs = Job.objects(status=Status.STARTED, **filters) @@ -358,13 +366,2 @@ class Queue: - ( - Job.objects( - type__in=only_job_types, - status=Status.WAITING, - namespace__nin=set(started_job_namespaces), - priority=priority, - ) - if only_job_types - else Job.objects( - status=Status.WAITING, - namespace__nin=set(started_job_namespaces), - priority=priority, - ) + Job.objects( + status=Status.WAITING, namespace__nin=set(started_job_namespaces), priority=priority, **filters @@ -376,0 +374 @@ class Queue: + @@ -407,15 +405,6 @@ class Queue: - ( - Job.objects( - type__in=only_job_types, - status=Status.WAITING, - namespace__in=least_common_namespaces_group, - unicity_id__nin=started_unicity_ids, - priority=priority, - ) - if only_job_types - else Job.objects( - status=Status.WAITING, - namespace__in=least_common_namespaces_group, - unicity_id__nin=started_unicity_ids, - priority=priority, - ) + Job.objects( + status=Status.WAITING, + namespace__in=least_common_namespaces_group, + unicity_id__nin=started_unicity_ids, + priority=priority, + **filters, @@ -435 +424,3 @@ class Queue: - def get_next_waiting_job(self, only_job_types: Optional[list[str]] = None) -> Job: + def get_next_waiting_job( + self, job_types_blocked: Optional[list[str]] = None, job_types_only: Optional[list[str]] = None + ) -> Job: @@ -446 +437,2 @@ class Queue: - only_job_types: if not None, only jobs of the given types are considered. + job_types_blocked: if not None, jobs of the given types are not considered. + job_types_only: if not None, only jobs of the given types are considered. @@ -455 +447,3 @@ class Queue: - return self._get_next_waiting_job_for_priority(priority=priority, only_job_types=only_job_types) + return self._get_next_waiting_job_for_priority( + priority=priority, job_types_blocked=job_types_blocked, job_types_only=job_types_only + ) @@ -460 +454,3 @@ class Queue: - def start_job(self, only_job_types: Optional[list[str]] = None) -> JobInfo: + def start_job( + self, job_types_blocked: Optional[list[str]] = None, job_types_only: Optional[list[str]] = None + ) -> JobInfo: @@ -466 +462,2 @@ class Queue: - only_job_types: if not None, only jobs of the given types are considered. + job_types_blocked: if not None, jobs of the given types are not considered. + job_types_only: if not None, only jobs of the given types are considered. @@ -474,2 +471,4 @@ class Queue: - logging.debug("looking for a job to start, among the following types: %s", only_job_types or "all") - next_waiting_job = self.get_next_waiting_job(only_job_types=only_job_types) + logging.debug(f"looking for a job to start, blocked types: {job_types_blocked}, only types: {job_types_only}") + next_waiting_job = self.get_next_waiting_job( + job_types_blocked=job_types_blocked, job_types_only=job_types_only + ) @@ -479 +478,5 @@ class Queue: - if only_job_types and next_waiting_job.type not in only_job_types: + if job_types_blocked and next_waiting_job.type in job_types_blocked: + raise RuntimeError( + f"The job type {next_waiting_job.type} is in the list of blocked job types {job_types_only}" + ) + if job_types_only and next_waiting_job.type not in job_types_only: @@ -481 +484 @@ class Queue: - f"The job type {next_waiting_job.type} is not in the list of allowed job types {only_job_types}" + f"The job type {next_waiting_job.type} is not in the list of allowed job types {job_types_only}" @@ -711 +714 @@ class Queue: - logging.info(f"Killing {len(zombies)} zombies. Job ids = " + zombies_examples_str) + logging.info(f"Killing {len(zombies)} zombies. Job ids = {zombies_examples_str}") @@ -724 +727 @@ class Queue: - logging.info("Killing a long job. Job id = " + long_job_id) + logging.info(f"Killing a long job. Job id = {long_job_id}") diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 67a0f6a2..21220785 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -211 +211 @@ def test_max_jobs_per_namespace(max_jobs_per_namespace: Optional[int]) -> None: - "job_type,only_job_types", + "job_type,job_types_blocked,job_types_only,should_raise", @@ -213,4 +213,10 @@ def test_max_jobs_per_namespace(max_jobs_per_namespace: Optional[int]) -> None: - ("test_type", None), - ("test_type", ["test_type"]), - ("test_type", ["other_type", "test_type"]), - ("test_type", ["other_type"]), + ("test_type", None, None, False), + ("test_type", None, ["test_type"], False), + ("test_type", ["other_type"], None, False), + ("test_type", ["other_type"], ["test_type"], False), + ("test_type", None, ["other_type"], True), + ("test_type", ["test_type"], None, True), + ("test_type", ["test_type"], ["test_type"], True), + ("test_type", ["other_type", "test_type"], None, True), + ("test_type", ["other_type"], ["other_type"], True), + ("test_type", ["other_type", "test_type"], ["other_type", "test_type"], True), @@ -219 +225,3 @@ def test_max_jobs_per_namespace(max_jobs_per_namespace: Optional[int]) -> None: -def test_only_job_types(job_type: str, only_job_types: Optional[list[str]]) -> None: +def test_job_types_only( + job_type: str, job_types_blocked: Optional[list[str]], job_types_only: Optional[list[str]], should_raise: bool +) -> None: @@ -224 +232 @@ def test_only_job_types(job_type: str, only_job_types: Optional[list[str]]) -> N - if only_job_types and job_type not in only_job_types: + if should_raise: @@ -226 +234 @@ def test_only_job_types(job_type: str, only_job_types: Optional[list[str]]) -> N - queue.start_job(only_job_types=only_job_types) + queue.start_job(job_types_blocked=job_types_blocked, job_types_only=job_types_only) @@ -228 +236 @@ def test_only_job_types(job_type: str, only_job_types: Optional[list[str]]) -> N - job_info = queue.start_job(only_job_types=only_job_types) + job_info = queue.start_job(job_types_blocked=job_types_blocked, job_types_only=job_types_only) @@ -318 +326 @@ def test_queue_heartbeat() -> None: - queue.start_job([job_type]) + queue.start_job(job_types_only=[job_type]) @@ -332 +340 @@ def test_queue_get_zombies() -> None: - queue.start_job([job_type]) + queue.start_job(job_types_only=[job_type]) @@ -334 +342 @@ def test_queue_get_zombies() -> None: - queue.start_job([job_type]) + queue.start_job(job_types_only=[job_type]) @@ -346 +354 @@ def test_queue_kill_zombies() -> None: - queue.start_job([job_type]) + queue.start_job(job_types_only=[job_type]) @@ -348 +356 @@ def test_queue_kill_zombies() -> None: - queue.start_job([job_type]) + queue.start_job(job_types_only=[job_type]) diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index ea8ce32c..1899df4f 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -592 +592 @@ def test_plan_job_creation_and_termination() -> None: - job_info = Queue().start_job(only_job_types=["/config-names"]) + job_info = Queue().start_job(job_types_only=["/config-names"]) diff --git a/services/worker/README.md b/services/worker/README.md index a2436173..2ce0d474 100644 --- a/services/worker/README.md +++ b/services/worker/README.md @@ -14,0 +15,2 @@ Set environment variables to configure the worker. +- `WORKER_JOB_TYPES_BLOCKED`: comma-separated list of job types that will not be processed, e.g. "/config-names,dataset-split-names". If empty, no job type is blocked. Defaults to empty. +- `WORKER_JOB_TYPES_ONLY`: comma-separated list of the non-blocked job types to process, e.g. "/config-names,dataset-split-names". If empty, the worker processes all the non-blocked jobs. Defaults to empty. @@ -20 +21,0 @@ Set environment variables to configure the worker. -- `WORKER_ONLY_JOB_TYPES`: comma-separated list of the job types to process, e.g. "/config-names,dataset-split-names". If empty, the worker processes all the jobs. Defaults to empty. diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index 1f869141..a61eb872 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -91,0 +92 @@ module = [ + "aiolimiter.*", diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index 20d3e7c6..eef8bb35 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -37,0 +38,5 @@ class WorkerConfig: + heartbeat_interval_seconds: int = WORKER_HEARTBEAT_INTERVAL_SECONDS + job_types_blocked: list[str] = field(default_factory=get_empty_str_list) + job_types_only: list[str] = field(default_factory=get_empty_str_list) + kill_long_job_interval_seconds: int = WORKER_KILL_LONG_JOB_INTERVAL_SECONDS + kill_zombies_interval_seconds: int = WORKER_KILL_ZOMBIES_INTERVAL_SECONDS @@ -38,0 +44 @@ class WorkerConfig: + max_job_duration_seconds: int = WORKER_MAX_JOB_DURATION_SECONDS @@ -41 +47 @@ class WorkerConfig: - only_job_types: list[str] = field(default_factory=get_empty_str_list) + max_missing_heartbeats: int = WORKER_MAX_MISSING_HEARTBEATS @@ -43 +48,0 @@ class WorkerConfig: - storage_paths: List[str] = field(default_factory=get_empty_str_list) @@ -45,5 +50 @@ class WorkerConfig: - heartbeat_interval_seconds: int = WORKER_HEARTBEAT_INTERVAL_SECONDS - max_missing_heartbeats: int = WORKER_MAX_MISSING_HEARTBEATS - kill_zombies_interval_seconds: int = WORKER_KILL_ZOMBIES_INTERVAL_SECONDS - max_job_duration_seconds: int = WORKER_MAX_JOB_DURATION_SECONDS - kill_long_job_interval_seconds: int = WORKER_KILL_LONG_JOB_INTERVAL_SECONDS + storage_paths: List[str] = field(default_factory=get_empty_str_list) @@ -57,9 +57,0 @@ class WorkerConfig: - max_disk_usage_pct=env.int(name="MAX_DISK_USAGE_PCT", default=WORKER_MAX_DISK_USAGE_PCT), - max_load_pct=env.int(name="MAX_LOAD_PCT", default=WORKER_MAX_LOAD_PCT), - max_memory_pct=env.int(name="MAX_MEMORY_PCT", default=WORKER_MAX_MEMORY_PCT), - sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_SLEEP_SECONDS), - only_job_types=env.list(name="ONLY_JOB_TYPES", default=get_empty_str_list()), - storage_paths=env.list(name="STORAGE_PATHS", default=get_empty_str_list()), - state_file_path=env.str( - name="STATE_FILE_PATH", default=WORKER_STATE_FILE_PATH - ), # this environment variable is not expected to be set explicitly, it's set by the worker executor @@ -69 +61,5 @@ class WorkerConfig: - max_missing_heartbeats=env.int(name="MAX_MISSING_HEARTBEATS", default=WORKER_MAX_MISSING_HEARTBEATS), + job_types_blocked=env.list(name="JOB_TYPES_BLOCKED", default=get_empty_str_list()), + job_types_only=env.list(name="JOB_TYPES_ONLY", default=get_empty_str_list()), + kill_long_job_interval_seconds=env.int( + name="KILL_LONG_JOB_INTERVAL_SECONDS", default=WORKER_KILL_LONG_JOB_INTERVAL_SECONDS + ), @@ -72,0 +69 @@ class WorkerConfig: + max_disk_usage_pct=env.int(name="MAX_DISK_USAGE_PCT", default=WORKER_MAX_DISK_USAGE_PCT), @@ -76,3 +73,8 @@ class WorkerConfig: - kill_long_job_interval_seconds=env.int( - name="KILL_LONG_JOB_INTERVAL_SECONDS", default=WORKER_KILL_LONG_JOB_INTERVAL_SECONDS - ), + max_load_pct=env.int(name="MAX_LOAD_PCT", default=WORKER_MAX_LOAD_PCT), + max_memory_pct=env.int(name="MAX_MEMORY_PCT", default=WORKER_MAX_MEMORY_PCT), + max_missing_heartbeats=env.int(name="MAX_MISSING_HEARTBEATS", default=WORKER_MAX_MISSING_HEARTBEATS), + sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_SLEEP_SECONDS), + state_file_path=env.str( + name="STATE_FILE_PATH", default=WORKER_STATE_FILE_PATH + ), # this environment variable is not expected to be set explicitly, it's set by the worker executor + storage_paths=env.list(name="STORAGE_PATHS", default=get_empty_str_list()), @@ -97,0 +100,2 @@ class DatasetsBasedConfig: +FIRST_ROWS_CELL_MIN_BYTES = 100 +FIRST_ROWS_COLUMNS_MAX_NUMBER = 1_000 @@ -100 +103,0 @@ FIRST_ROWS_MAX_NUMBER = 100 -FIRST_ROWS_CELL_MIN_BYTES = 100 @@ -102 +104,0 @@ FIRST_ROWS_MIN_NUMBER = 10 -FIRST_ROWS_COLUMNS_MAX_NUMBER = 1_000 @@ -106,0 +109 @@ class FirstRowsConfig: + columns_max_number: int = FIRST_ROWS_COLUMNS_MAX_NUMBER @@ -111 +113,0 @@ class FirstRowsConfig: - columns_max_number: int = FIRST_ROWS_COLUMNS_MAX_NUMBER @@ -117,0 +120 @@ class FirstRowsConfig: + columns_max_number=env.int(name="COLUMNS_MAX_NUMBER", default=FIRST_ROWS_COLUMNS_MAX_NUMBER), @@ -122 +124,0 @@ class FirstRowsConfig: - columns_max_number=env.int(name="COLUMNS_MAX_NUMBER", default=FIRST_ROWS_COLUMNS_MAX_NUMBER), @@ -131 +132,0 @@ OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN = None -OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH = 1000 @@ -132,0 +134 @@ OPT_IN_OUT_URLS_SCAN_SPAWNING_URL = "https://opts-api.spawningaiapi.com/api/v2/q +OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH = 1000 @@ -137,2 +138,0 @@ class OptInOutUrlsScanConfig: - spawning_url: str = OPT_IN_OUT_URLS_SCAN_SPAWNING_URL - rows_max_number: int = OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER @@ -140,2 +139,0 @@ class OptInOutUrlsScanConfig: - urls_number_per_batch: int = OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH - spawning_token: Optional[str] = OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN @@ -143,0 +142,4 @@ class OptInOutUrlsScanConfig: + rows_max_number: int = OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER + spawning_token: Optional[str] = OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN + spawning_url: str = OPT_IN_OUT_URLS_SCAN_SPAWNING_URL + urls_number_per_batch: int = OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH @@ -150 +151,0 @@ class OptInOutUrlsScanConfig: - rows_max_number=env.int(name="ROWS_MAX_NUMBER", default=OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER), @@ -152,4 +152,0 @@ class OptInOutUrlsScanConfig: - urls_number_per_batch=env.int( - name="URLS_NUMBER_PER_BATCH", default=OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH - ), - spawning_token=env.str(name="SPAWNING_TOKEN", default=OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN), @@ -161,0 +159,2 @@ class OptInOutUrlsScanConfig: + rows_max_number=env.int(name="ROWS_MAX_NUMBER", default=OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER), + spawning_token=env.str(name="SPAWNING_TOKEN", default=OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN), @@ -162,0 +162,3 @@ class OptInOutUrlsScanConfig: + urls_number_per_batch=env.int( + name="URLS_NUMBER_PER_BATCH", default=OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH + ), @@ -178 +179,0 @@ class ParquetAndInfoConfig: - supported_datasets: List[str] = field(default_factory=get_empty_str_list) @@ -183,0 +185 @@ class ParquetAndInfoConfig: + supported_datasets: List[str] = field(default_factory=get_empty_str_list) @@ -193 +194,0 @@ class ParquetAndInfoConfig: - supported_datasets=env.list(name="SUPPORTED_DATASETS", default=get_empty_str_list()), @@ -197,0 +199 @@ class ParquetAndInfoConfig: + supported_datasets=env.list(name="SUPPORTED_DATASETS", default=get_empty_str_list()), diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py index 6b7dade9..863f609e 100644 --- a/services/worker/src/worker/loop.py +++ b/services/worker/src/worker/loop.py @@ -21,4 +20,0 @@ from worker.job_runner_factory import BaseJobRunnerFactory -class UnknownJobTypeError(Exception): - pass - - @@ -128,7 +124,4 @@ class Loop: - job_info = self.queue.start_job(only_job_types=self.worker_config.only_job_types) - if self.worker_config.only_job_types and job_info["type"] not in self.worker_config.only_job_types: - raise UnknownJobTypeError( - f"Job of type {job_info['type']} is not supported (only" - f" ${', '.join(self.worker_config.only_job_types)}). The queue should not have provided this" - " job. It is in an inconsistent state. Please report this issue to the datasets team." - ) + job_info = self.queue.start_job( + job_types_blocked=self.worker_config.job_types_blocked, + job_types_only=self.worker_config.job_types_only, + )
8f3056af4ac67a167d52ddf20d53cd19bcc510eb
Sylvain Lesage
2023-04-27T07:22:16
feat: 🎸 use a common function to forge unique ids (#1090)
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 0bb98dce..80ea87de 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -20 +20 @@ from libcommon.constants import QUEUE_MONGOENGINE_ALIAS, QUEUE_TTL_SECONDS -from libcommon.utils import get_datetime +from libcommon.utils import get_datetime, inputs_to_string @@ -248 +248 @@ class Queue: - unicity_id=f"Job[{job_type}][{dataset}][{config}][{split}]", + unicity_id=inputs_to_string(dataset=dataset, config=config, split=split, prefix=job_type), diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index a63891b8..f4557dc5 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -19,0 +20 @@ from libcommon.simple_cache import ( +from libcommon.utils import inputs_to_string @@ -157 +158 @@ class ArtifactState: - self.id = ",".join([p for p in (self.step.name, self.dataset, self.config, self.split) if p]) + self.id = inputs_to_string(dataset=self.dataset, config=self.config, split=self.split, prefix=self.step.name) diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py index 39db0ed7..6e9b483c 100644 --- a/libs/libcommon/src/libcommon/utils.py +++ b/libs/libcommon/src/libcommon/utils.py @@ -6 +6 @@ from datetime import datetime, timezone -from typing import Any +from typing import Any, Optional @@ -26,0 +27,16 @@ def get_datetime() -> datetime: + + +def inputs_to_string( + dataset: str, + config: Optional[str] = None, + split: Optional[str] = None, + prefix: Optional[str] = None, +) -> str: + result = dataset + if config is not None: + result = f"{result},{config}" + if split is not None: + result = f"{result},{split}" + if prefix is not None: + result = f"{prefix},{result}" + return result diff --git a/libs/libcommon/tests/test_utils.py b/libs/libcommon/tests/test_utils.py new file mode 100644 index 00000000..ab60a202 --- /dev/null +++ b/libs/libcommon/tests/test_utils.py @@ -0,0 +1,22 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import pytest + +from libcommon.utils import inputs_to_string + + [email protected]( + "dataset,config,split,prefix,expected", + [ + ("dataset", None, None, None, "dataset"), + ("dataset", "config", None, None, "dataset,config"), + ("dataset", None, "split", None, "dataset"), + ("dataset", "config", "split", None, "dataset,config,split"), + ("dataset", None, None, "prefix", "prefix,dataset"), + ("dataset", "config", "split", "prefix", "prefix,dataset,config,split"), + ], +) +def test_inputs_to_string(dataset: str, config: str, split: str, prefix: str, expected: str) -> None: + result = inputs_to_string(dataset=dataset, config=config, split=split, prefix=prefix) + assert result == expected
c72264aa625d9036677e8cfd9f4db18549762d48
Albert Villanova del Moral
2023-04-26T13:33:52
Raise informative error when importing non-installed module (#1084)
diff --git a/services/worker/src/worker/job_runners/config_names.py b/services/worker/src/worker/job_runners/config_names.py index 8043c227..38f7f43f 100644 --- a/services/worker/src/worker/job_runners/config_names.py +++ b/services/worker/src/worker/job_runners/config_names.py @@ -16 +16 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner -ConfigNamesJobRunnerErrorCode = Literal["EmptyDatasetError", "ConfigNamesError"] +ConfigNamesJobRunnerErrorCode = Literal["EmptyDatasetError", "DatasetModuleNotInstalledError", "ConfigNamesError"] @@ -41,0 +42,7 @@ class EmptyDatasetError(ConfigNamesJobRunnerError): +class DatasetModuleNotInstalledError(ConfigNamesJobRunnerError): + """Raised when the dataset tries to import a module that is not installed.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "DatasetModuleNotInstalledError", cause, True) + + @@ -79,0 +87,2 @@ def compute_config_names_response( + - [`~job_runners.config_names.DatasetModuleNotInstalledError`] + The dataset tries to import a module that is not installed. @@ -93,0 +103,4 @@ def compute_config_names_response( + except ImportError as err: + raise DatasetModuleNotInstalledError( + "The dataset tries to import a module that is not installed.", cause=err + ) from err
bed57e5cb706c25fe29205f14e79ba07ff880636
Polina Kazakova
2023-04-26T11:24:09
Remove /parquet-and-dataset-info job runner (#1043)
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index 7275ea30..3fa823f7 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -39 +39 @@ -# specific to the /parquet-and-dataset-info and config-parquet-and-info job runners +# specific to 'config-parquet-and-info' job runner diff --git a/e2e/tests/test_11_api.py b/e2e/tests/test_11_api.py index fbeae374..2be6433b 100644 --- a/e2e/tests/test_11_api.py +++ b/e2e/tests/test_11_api.py @@ -54 +54 @@ def test_auth_e2e( - ("/parquet-and-dataset-info", "dataset"), + ("/parquet-and-dataset-info", "config"), diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 02cf9a88..bc2d2014 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -49,0 +50,6 @@ from mongodb_migration.migrations._20230407091500_cache_delete_splits import ( +from mongodb_migration.migrations._20230424173000_queue_delete_parquet_and_dataset_info import ( + MigrationQueueDeleteParquetAndDatasetInfo, +) +from mongodb_migration.migrations._20230424174000_cache_delete_parquet_and_dataset_info import ( + MigrationCacheDeleteParquetAndDatasetInfo, +) @@ -113,0 +120,7 @@ class MigrationsCollector: + MigrationQueueDeleteParquetAndDatasetInfo( + version="20230424173000", + description="delete the jobs of type '/parquet-and-dataset-info'", + ), + MigrationCacheDeleteParquetAndDatasetInfo( + version="20230424174000", description="delete the cache entries of kind '/parquet-and-dataset-info'" + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424173000_queue_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424173000_queue_delete_parquet_and_dataset_info.py new file mode 100644 index 00000000..0c9dd3a9 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424173000_queue_delete_parquet_and_dataset_info.py @@ -0,0 +1,30 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + + +import logging + +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +job_type = "/parquet-and-dataset-info" +db_name = "queue" + + +class MigrationQueueDeleteParquetAndDatasetInfo(Migration): + def up(self) -> None: + logging.info(f"Delete jobs of type {job_type}") + + db = get_db(db_name) + db["jobsBlue"].delete_many({"type": job_type}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {job_type} type") + + db = get_db(db_name) + if db[db_name].count_documents({"type": job_type}): + raise ValueError(f"Found documents with type {job_type}") diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424174000_cache_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424174000_cache_delete_parquet_and_dataset_info.py new file mode 100644 index 00000000..3fd8edd6 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230424174000_cache_delete_parquet_and_dataset_info.py @@ -0,0 +1,31 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + + +import logging + +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +cache_kind = "/parquet-and-dataset-info" +db_name = "cache" + + +class MigrationCacheDeleteParquetAndDatasetInfo(Migration): + def up(self) -> None: + logging.info(f"Delete cache entries of kind {cache_kind}") + db = get_db(db_name) + + # delete existing documents + db["cachedResponsesBlue"].delete_many({"kind": cache_kind}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {cache_kind} kind") + + db = get_db(db_name) + if db[db_name].count_documents({"kind": cache_kind}): + raise ValueError(f"Found documents with kind {cache_kind}") diff --git a/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py new file mode 100644 index 00000000..1e37b92d --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230424173000_cache_delete_parquet_and_dataset_info.py @@ -0,0 +1,29 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230424174000_cache_delete_parquet_and_dataset_info import ( + MigrationCacheDeleteParquetAndDatasetInfo, +) + + +def test_cache_delete_parquet_and_dataset_info(mongo_host: str) -> None: + kind = "/parquet-and-dataset-info" + with MongoResource( + database="test_cache_delete_parquet_and_dataset_info", host=mongo_host, mongoengine_alias="cache" + ): + db = get_db("cache") + db["cachedResponsesBlue"].insert_many([{"kind": kind, "dataset": "dataset", "http_status": 200}]) + assert db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure there is at least one record to delete + + migration = MigrationCacheDeleteParquetAndDatasetInfo( + version="20230424173000", + description=f"remove cache for kind {kind}", + ) + migration.up() + + assert not db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure 0 records with old kind + + db["cachedResponsesBlue"].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230424173000_queue_delete_parquet_and_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230424173000_queue_delete_parquet_and_dataset_info.py new file mode 100644 index 00000000..4219f2d1 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230424173000_queue_delete_parquet_and_dataset_info.py @@ -0,0 +1,38 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230424173000_queue_delete_parquet_and_dataset_info import ( + MigrationQueueDeleteParquetAndDatasetInfo, +) + + +def test_queue_delete_parquet_and_dataset_info(mongo_host: str) -> None: + job_type = "/parquet-and-dataset-info" + with MongoResource( + database="test_queue_delete_parquet_and_dataset_info", host=mongo_host, mongoengine_alias="queue" + ): + db = get_db("queue") + db["jobsBlue"].insert_many( + [ + { + "type": job_type, + "unicity_id": f"Job[{job_type}][dataset][config][split]", + "dataset": "dataset", + "http_status": 200, + } + ] + ) + assert db["jobsBlue"].find_one({"type": job_type}) # Ensure there is at least one record to delete + + migration = MigrationQueueDeleteParquetAndDatasetInfo( + version="20230424173000", + description=f"remove jobs of type '{job_type}'", + ) + migration.up() + + assert not db["jobsBlue"].find_one({"type": job_type}) # Ensure 0 records with old type + + db["jobsBlue"].drop() diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index a95be214..4de84c4d 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -24 +23,0 @@ from libcommon.constants import ( - PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION, @@ -205,4 +203,0 @@ class ProcessingGraphConfig: - "/parquet-and-dataset-info": { - "input_type": "dataset", - "job_runner_version": PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION, - }, diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 9c302ba9..ec81fa88 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -160 +159,0 @@ def graph() -> ProcessingGraph: - ("/parquet-and-dataset-info", [], [], []), @@ -241 +240 @@ def test_default_graph_first_steps(graph: ProcessingGraph) -> None: - [graph.get_step(step_name) for step_name in {"/config-names", "/parquet-and-dataset-info"}], + [graph.get_step(step_name) for step_name in {"/config-names"}], diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 8f203fad..ea8ce32c 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -426,5 +425,0 @@ def test_dataset_state_as_dict() -> None: - { - "id": f"/parquet-and-dataset-info,{DATASET_NAME}", - "job_state": {"is_in_process": False}, - "cache_state": {"exists": False, "is_success": False}, - }, @@ -514 +508,0 @@ def test_plan() -> None: - "/parquet-and-dataset-info,dataset", @@ -532 +525,0 @@ def test_plan() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -549 +541,0 @@ def test_plan_job_creation_and_termination() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -570 +561,0 @@ def test_plan_job_creation_and_termination() -> None: - "/parquet-and-dataset-info,dataset", @@ -587 +577,0 @@ def test_plan_job_creation_and_termination() -> None: - "/parquet-and-dataset-info,dataset", @@ -627 +616,0 @@ def test_plan_job_creation_and_termination() -> None: - "/parquet-and-dataset-info,dataset", @@ -655 +643,0 @@ def test_plan_job_creation_and_termination() -> None: - "/parquet-and-dataset-info,dataset", @@ -709 +696,0 @@ def test_plan_retry_error() -> None: - "/parquet-and-dataset-info,dataset", @@ -738 +724,0 @@ def test_plan_retry_error() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -798 +783,0 @@ def test_plan_incoherent_state() -> None: - "/parquet-and-dataset-info,dataset", @@ -833 +817,0 @@ def test_plan_incoherent_state() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -908 +891,0 @@ def test_plan_updated_at() -> None: - "/parquet-and-dataset-info,dataset", @@ -935 +917,0 @@ def test_plan_updated_at() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -981 +962,0 @@ def test_plan_job_runner_version() -> None: - "/parquet-and-dataset-info,dataset", @@ -1010 +990,0 @@ def test_plan_job_runner_version() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -1071 +1050,0 @@ def test_plan_git_revision( - "/parquet-and-dataset-info,dataset", @@ -1099 +1077,0 @@ def test_plan_git_revision( - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -1132 +1109,0 @@ def test_plan_git_revision( - "/parquet-and-dataset-info,dataset", @@ -1159 +1135,0 @@ def test_plan_git_revision( - "CreateJob[/parquet-and-dataset-info,dataset]", @@ -1241 +1216,0 @@ def test_plan_update_fan_in_parent() -> None: - "/parquet-and-dataset-info,dataset", @@ -1270 +1244,0 @@ def test_plan_update_fan_in_parent() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 0e7b2a76..1357f6d2 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -131 +130,0 @@ class EndpointConfig: - "dataset": ["/parquet-and-dataset-info"], diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index ec72f741..238b685e 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -50 +50 @@ def test_endpoints_definition() -> None: - assert sorted(list(parquet_and_info)) == ["config", "dataset"] + assert sorted(list(parquet_and_info)) == ["config"] diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 33113bca..b5ac1d2e 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -36 +35,0 @@ from worker.job_runners.dataset.split_names_from_streaming import ( -from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoJobRunner @@ -103,7 +101,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - if job_type == ParquetAndDatasetInfoJobRunner.get_job_type(): - return ParquetAndDatasetInfoJobRunner( - job_info=job_info, - app_config=self.app_config, - processing_step=processing_step, - hf_datasets_cache=self.hf_datasets_cache, - ) @@ -215 +206,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - ParquetAndDatasetInfoJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 28e879ac..6db5a292 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -682 +682 @@ def compute_config_parquet_and_info_response( - Get the response of /parquet-and-dataset-info for one specific dataset on huggingface.co. + Get the response of config-parquet-and-info for one specific dataset and config on huggingface.co. @@ -781 +781 @@ def compute_config_parquet_and_info_response( - logging.info(f"get config names for {dataset=}") + logging.info(f"getting config names for {dataset=}") diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py deleted file mode 100644 index a122c73a..00000000 --- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py +++ /dev/null @@ -1,930 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import contextlib -import glob -import logging -import re -from functools import partial -from http import HTTPStatus -from multiprocessing.pool import ThreadPool -from pathlib import Path -from typing import Any, Dict, List, Literal, Mapping, Optional, Tuple, TypedDict -from urllib.parse import quote - -import datasets -import datasets.config -import datasets.info -import numpy as np -import requests -from datasets import ( - DownloadConfig, - get_dataset_config_names, - get_dataset_infos, - load_dataset_builder, -) -from datasets.builder import DatasetBuilder -from datasets.data_files import EmptyDatasetError as _EmptyDatasetError -from datasets.download import StreamingDownloadManager -from datasets.utils.file_utils import ( - get_authentication_headers_for_url, - http_head, - is_relative_path, - url_or_path_join, -) -from datasets.utils.py_utils import asdict, map_nested -from huggingface_hub._commit_api import ( - CommitOperation, - CommitOperationAdd, - CommitOperationDelete, -) -from huggingface_hub.hf_api import DatasetInfo, HfApi, RepoFile -from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFoundError -from libcommon.constants import ( - PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS, - PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS, - PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION, -) -from libcommon.dataset import DatasetNotFoundError, ask_access -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import JobInfo -from libcommon.simple_cache import SplitFullName - -from worker.config import AppConfig, ParquetAndInfoConfig -from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner -from worker.job_runners.config.parquet_and_info import ParquetFile, ParquetFileItem - -ParquetAndDatasetInfoJobRunnerErrorCode = Literal[ - "DatasetRevisionNotFoundError", - "EmptyDatasetError", - "ConfigNamesError", - "DatasetInBlockListError", - "DatasetTooBigFromHubError", - "DatasetTooBigFromDatasetsError", - "UnsupportedExternalFilesError", - "DatasetWithTooManyExternalFilesError", - "DatasetWithTooBigExternalFilesError", - "ExternalFilesSizeRequestHTTPError", - "ExternalFilesSizeRequestConnectionError", - "ExternalFilesSizeRequestTimeoutError", - "ExternalFilesSizeRequestError", -] - - -class ParquetAndDatasetInfoJobRunnerError(JobRunnerError): - """Base class for exceptions in this module.""" - - def __init__( - self, - message: str, - status_code: HTTPStatus, - code: ParquetAndDatasetInfoJobRunnerErrorCode, - cause: Optional[BaseException] = None, - disclose_cause: bool = False, - ): - super().__init__( - message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause - ) - - -class DatasetRevisionNotFoundError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the revision of a dataset repository does not exist.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_FOUND, "DatasetRevisionNotFoundError", cause, False) - - -class ConfigNamesError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the configuration names could not be fetched.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ConfigNamesError", cause, True) - - -class EmptyDatasetError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the dataset has no data.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "EmptyDatasetError", cause, True) - - -class DatasetInBlockListError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the dataset is in the list of blocked datasets.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetInBlockListError", cause, False) - - -class DatasetTooBigFromHubError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the dataset size (sum of files on the Hub) is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromHubError", cause, False) - - -class DatasetTooBigFromDatasetsError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromDatasetsError", cause, False) - - -class ParquetAndDatasetInfoResponse(TypedDict): - parquet_files: List[ParquetFileItem] - dataset_info: Dict[str, Any] - - -DATASET_TYPE = "dataset" - - -# TODO: use huggingface_hub's hf_hub_url after -# https://github.com/huggingface/huggingface_hub/issues/1082 -def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url_template: str) -> str: - return (hf_endpoint + url_template) % (repo_id, quote(revision, safe=""), filename) - - -p = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>\w+(\.\w+)*?)(-[0-9]{5}-of-[0-9]{5})?.parquet") - - -def parse_repo_filename(filename: str) -> Tuple[str, str]: - parts = filename.split("/") - if len(parts) != 2: - raise ValueError(f"Invalid filename: {filename}") - config, fname = parts - m = p.match(fname) - if not m: - raise ValueError(f"Cannot parse {filename}") - split = m.group("split") - return config, split - - -def create_parquet_file_item( - repo_file: RepoFile, - dataset: str, - hf_endpoint: str, - target_revision: str, - url_template: str, -) -> ParquetFileItem: - if repo_file.size is None: - raise ValueError(f"Cannot get size of {repo_file.rfilename}") - config, split = parse_repo_filename(repo_file.rfilename) - return { - "dataset": dataset, - "config": config, - "split": split, - "url": hf_hub_url( - repo_id=dataset, - filename=repo_file.rfilename, - hf_endpoint=hf_endpoint, - revision=target_revision, - url_template=url_template, - ), - "filename": Path(repo_file.rfilename).name, - "size": repo_file.size, - } - - -def raise_if_blocked( - dataset: str, - blocked_datasets: List[str], -) -> None: - """ - Raise an error if the dataset is in the list of blocked datasets - - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - blocked_datasets (`List[str]`): - The list of blocked datasets. If empty, no dataset is blocked. - Returns: - `None` - <Tip> - Raises the following errors: - - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`] - If the dataset is in the list of blocked datasets. - </Tip> - """ - if dataset in blocked_datasets: - raise DatasetInBlockListError( - "The parquet conversion has been disabled for this dataset for now. Please open an issue in" - " https://github.com/huggingface/datasets-server if you want this dataset to be supported." - ) - - -def get_dataset_info_or_raise( - dataset: str, - hf_endpoint: str, - hf_token: Optional[str], - revision: str, -) -> DatasetInfo: - """ - Return the dataset info if possible. - Raise an error if the dataset cannot be accessed (does not exist, gated with extra fields, private) - - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - hf_endpoint (`str`): - The Hub endpoint (for example: "https://huggingface.co") - hf_token (`str`, `optional`): - An app authentication token with read access to all the datasets. - revision (`str`): - The git revision (e.g. "main" or sha) of the dataset - Returns: - `DatasetInfo`: The dataset info - <Tip> - Raises the following errors: - - [`~.job_runner.DatasetNotFoundError`] - If the repository to download from cannot be found. This may be because it doesn't exist, - or because it is set to `private` and you do not have access. - - [`~job_runners.parquet_and_dataset_info.DatasetRevisionNotFoundError`] - If the revision does not exist or cannot be accessed using the token. - </Tip> - """ - try: - dataset_info = HfApi(endpoint=hf_endpoint, token=hf_token).dataset_info( - repo_id=dataset, revision=revision, files_metadata=True - ) - except RepositoryNotFoundError as err: - raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err - except RevisionNotFoundError as err: - raise DatasetRevisionNotFoundError("The dataset revision does not exist on the Hub.") from err - if dataset_info.private: - raise DatasetNotFoundError("The dataset does not exist on the Hub.") - return dataset_info - - -def raise_if_too_big_from_hub( - dataset_info: DatasetInfo, - max_dataset_size: int, -) -> None: - """ - Raise an error if the dataset is too big to be converted to parquet - - Args: - dataset_info (`DatasetInfo`): - The dataset info - max_dataset_size (`int`): - The maximum size of the dataset in bytes - Returns: - `None` - <Tip> - Raises the following errors: - - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromHubError`] - If the dataset is too big to be converted to parquet - </Tip> - """ - dataset_size: int = sum(sibling.size for sibling in dataset_info.siblings if sibling.size is not None) - if dataset_size > max_dataset_size: - raise DatasetTooBigFromHubError( - f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes. " - f"Current size of files on the hub is {dataset_size} bytes." - ) - - -def raise_if_too_big_from_datasets( - dataset: str, - hf_endpoint: str, - hf_token: Optional[str], - revision: str, - max_dataset_size: int, -) -> None: - """ - Raise an error if the dataset is too big to be converted to parquet, as measured by the sum of the configs - sizes given by the datasets library - - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - hf_endpoint (`str`): - The Hub endpoint (for example: "https://huggingface.co") - hf_token (`str`, `optional`): - An app authentication token with read access to all the datasets. - revision (`str`): - The git revision (e.g. "main" or sha) of the dataset - max_dataset_size (`int`): - The maximum size of the dataset in bytes - Returns: - `None` - <Tip> - Raises the following errors: - - [`ValueError`] - If the datasets.config.HF_ENDPOINT is not set to the expected value - - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromDatasetsError`] - If the dataset is too big to be converted to parquet - </Tip> - """ - if datasets.config.HF_ENDPOINT != hf_endpoint: - raise ValueError( - f"Invalid datasets.config.HF_ENDPOINT value: '{datasets.config.HF_ENDPOINT}'. Please set it to:" - f" '{hf_endpoint}'." - ) - dataset_size = 0 - with contextlib.suppress(Exception): - infos = get_dataset_infos(path=dataset, revision=revision, use_auth_token=hf_token) - dataset_size = sum(value.dataset_size for value in infos.values() if value.dataset_size is not None) - if dataset_size > max_dataset_size: - raise DatasetTooBigFromDatasetsError( - f"The dataset is too big to be converted to Parquet. The size of the dataset ({dataset_size} B, as given" - f" per the datasets library) exceeds the maximum supported size ({max_dataset_size} B). Please report the" - " issue." - ) - - -def raise_if_not_supported( - dataset: str, - hf_endpoint: str, - hf_token: Optional[str], - committer_hf_token: Optional[str], - revision: str, - supported_datasets: List[str], - blocked_datasets: List[str], - max_dataset_size: int, -) -> None: - """ - Raise an error if the dataset is not supported: - - if the dataset is in the list of blocked datasets - - if the dataset cannot be accessed (does not exist, gated with extra fields, private) - - if the dataset is too big, and not in the list of supported datasets - - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - hf_endpoint (`str`): - The Hub endpoint (for example: "https://huggingface.co") - hf_token (`str`, `optional`): - An app authentication token with read access to all the datasets. - committer_hf_token (`str`, `optional`): - A user authentication token (See https://huggingface.co/settings/token) with write access. It must: - - be part of the `huggingface` organization (to create the ref/convert/parquet "branch") - - be part of the `datasets-maintainers` organization (to push to the ref/convert/parquet "branch") - revision (`str`): - The git revision (e.g. "main" or sha) of the dataset - supported_datasets (`List[str]`): - The list of supported datasets, saving the blocked datasets. If empty, all datasets are supported - (saving the blocked datasets). - blocked_datasets (`List[str]`): - The list of blocked datasets. If empty, no dataset is blocked. - max_dataset_size (`int`): - The maximum size of a dataset in bytes. If the dataset is under the limit (which means that the size - can be fetched), it will be allowed. - Returns: - `ParquetResponseResult`: An object with the parquet_response - (dataset and list of parquet files) and the dataset_git_revision (sha) if any. - <Tip> - Raises the following errors: - - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`] - If the dataset is in the list of blocked datasets. - - [`~libcommon.dataset.GatedExtraFieldsError`]: if the dataset is gated, with extra fields. - Programmatic access is not implemented for this type of dataset because there is no easy - way to get the list of extra fields. - - [`~libcommon.dataset.GatedDisabledError`]: if the dataset is gated, but disabled. - - [`~libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the - token does not give the sufficient access to the dataset, or if the dataset is private - (private datasets are not supported by the datasets server) - - ['~requests.exceptions.HTTPError']: any other error when asking access - - [`~job_runners.parquet_and_dataset_info.DatasetRevisionNotFoundError`] - If the revision does not exist or cannot be accessed using the token. - - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromHubError`] - If the dataset is too big to be converted to parquet - - [`ValueError`] - If the datasets.config.HF_ENDPOINT is not set to the expected value - - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromDatasetsError`] - If the dataset is too big to be converted to parquet - </Tip> - """ - raise_if_blocked(dataset=dataset, blocked_datasets=blocked_datasets) - ask_access(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=committer_hf_token) - dataset_info = get_dataset_info_or_raise( - dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, revision=revision - ) - if dataset in supported_datasets: - return - raise_if_too_big_from_datasets( - dataset=dataset, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - revision=revision, - max_dataset_size=max_dataset_size, - ) - raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size) - - -class EmptySplitsError(Exception): - pass - - -class SplitInfoFormatError(Exception): - pass - - -class EmptyConfigNameError(Exception): - pass - - -class EmptyDownloadSizeError(Exception): - pass - - -class EmptyFeaturesError(Exception): - pass - - -class DatasetWithTooManyExternalFilesError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyExternalFilesError", cause, True) - - -class DatasetWithTooBigExternalFilesError(ParquetAndDatasetInfoJobRunnerError): - """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooBigExternalFilesError", cause, True) - - -class UnsupportedExternalFilesError(ParquetAndDatasetInfoJobRunnerError): - """Raised when we failed to get the size of the external files.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "UnsupportedExternalFilesError", cause, True) - - -class ExternalFilesSizeRequestHTTPError(ParquetAndDatasetInfoJobRunnerError): - """Raised when we failed to get the size of the external files.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestHTTPError", cause, True) - - -class ExternalFilesSizeRequestConnectionError(ParquetAndDatasetInfoJobRunnerError): - """Raised when we failed to get the size of the external files.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestConnectionError", cause, True) - - -class ExternalFilesSizeRequestTimeoutError(ParquetAndDatasetInfoJobRunnerError): - """Raised when we failed to get the size of the external files.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestTimeoutError", cause, True) - - -class ExternalFilesSizeRequestError(ParquetAndDatasetInfoJobRunnerError): - """Raised when we failed to get the size of the external files.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestError", cause, True) - - -def _request_size(url: str, hf_token: Optional[str] = None) -> Optional[int]: - headers = get_authentication_headers_for_url(url, use_auth_token=hf_token) - response = http_head(url, headers=headers, max_retries=3) - response.raise_for_status() - size = response.headers.get("Content-Length") if response.ok else None - return int(size) if size is not None else size - - -class _MockStreamingDownloadManager(StreamingDownloadManager): # type: ignore - def __init__(self, *args: Any, **kwargs: Any) -> None: - super().__init__(*args, **kwargs) - self.ext_data_files: List[str] = [] - - def download(self, url_or_urls: Any) -> Any: - url_or_urls = map_nested( - self._download, - url_or_urls, - map_tuple=True, - parallel_min_length=np.inf, - # ^ parallel_min_length has int type, but is currently used in datasets for a comparison only - # and it works with np.inf. No conversion is involved - # (would raise: OverflowError: cannot convert float infinity to integer) - ) - return url_or_urls - - def _download(self, urlpath: Any) -> str: - urlpath_str = str(urlpath) - if is_relative_path(urlpath_str): - # append the relative path to the base_path - urlpath_str = url_or_path_join(self._base_path, urlpath_str) - elif not urlpath_str.startswith(self._base_path): - # it's an external file - self.ext_data_files.append(urlpath_str) - return urlpath_str - - -def raise_if_too_big_from_external_data_files( - builder: DatasetBuilder, max_dataset_size: int, max_external_data_files: int, hf_token: Optional[str] -) -> None: - # Packaged dataset modules only load data files that are inside the dataset repository. - # No need to check them since they're already caught by `raise_if_too_big_from_hub` - if type(builder).__module__.startswith("datasets."): - return - # For datasets with a loading script however, we need to check the downloaded files - mock_dl_manager = _MockStreamingDownloadManager( - base_path=builder.base_path, download_config=DownloadConfig(use_auth_token=hf_token) - ) - try: - builder._split_generators(mock_dl_manager) - except (requests.exceptions.RequestException, NotImplementedError) as error: - if isinstance(error, NotImplementedError): - # we can ignore the errors from functions not implemented in streaming mode like `.extract()` on TAR files - if "is not implemented in streaming mode." not in str(error): - raise UnsupportedExternalFilesError( - ( - "Couldn't get the list of external files in `_split_generators` because it doesn't support" - f" streaming:\n{error}" - ), - error, - ) from error - elif isinstance(error, requests.exceptions.HTTPError): - raise ExternalFilesSizeRequestHTTPError( - ( - "Couldn't get the list of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - elif isinstance(error, requests.exceptions.ConnectionError): - raise ExternalFilesSizeRequestConnectionError( - ( - "Couldn't get the list of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - elif isinstance(error, requests.exceptions.Timeout): - raise ExternalFilesSizeRequestTimeoutError( - ( - "Couldn't get the list of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - else: - raise ExternalFilesSizeRequestError( - ( - "Couldn't get the list of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - ext_data_files = mock_dl_manager.ext_data_files - if len(ext_data_files) > max_external_data_files: - raise DatasetWithTooManyExternalFilesError( - f"The conversion to parquet is limited to datasets with less than {max_external_data_files} files. " - f"However it uses {len(ext_data_files)} data files." - ) - elif ext_data_files: - try: - with ThreadPool(16) as pool: - total_size = 0 - get_size = partial(_request_size, hf_token=hf_token) - for i, size in enumerate(pool.imap_unordered(get_size, ext_data_files)): - if size is not None: - total_size += size - if total_size > max_dataset_size: - raise DatasetWithTooBigExternalFilesError( - f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes." - f" However {i + 1} data files of {len(ext_data_files)} are already bigger than" - f" {total_size} bytes." - ) - except requests.exceptions.RequestException as error: - if isinstance(error, requests.exceptions.HTTPError): - raise ExternalFilesSizeRequestHTTPError( - ( - "Couldn't get the size of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - elif isinstance(error, requests.exceptions.ConnectionError): - raise ExternalFilesSizeRequestConnectionError( - ( - "Couldn't get the size of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - elif isinstance(error, requests.exceptions.Timeout): - raise ExternalFilesSizeRequestTimeoutError( - ( - "Couldn't get the size of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - else: - raise ExternalFilesSizeRequestError( - ( - "Couldn't get the size of external files in `_split_generators` because a request" - f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" - " (e.g. inside a data/ folder)." - ), - error, - ) from error - - -def get_writer_batch_size(ds_config_info: datasets.info.DatasetInfo) -> Optional[int]: - """ - Get the writer_batch_size that defines the maximum row group size in the parquet files. - The default in `datasets` is 1,000 but we lower it to 100 for image datasets. - This allows to optimize random access to parquet file, since accessing 1 row requires - to read its entire row group. - - Args: - ds_config_info (`datasets.info.DatasetInfo`): - Dataset info from `datasets`. - - Returns: - writer_batch_size (`Optional[int]`): - Writer batch size to pass to a dataset builder. - If `None`, then it will use the `datasets` default. - """ - if "Audio(" in str(ds_config_info.features): - return PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS - elif "Image(" in str(ds_config_info.features): - return PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS - else: - return None - - -def compute_parquet_and_dataset_info_response( - dataset: str, - hf_endpoint: str, - hf_token: Optional[str], - committer_hf_token: Optional[str], - source_revision: str, - target_revision: str, - commit_message: str, - url_template: str, - supported_datasets: List[str], - blocked_datasets: List[str], - max_dataset_size: int, - max_external_data_files: int, -) -> ParquetAndDatasetInfoResponse: - """ - Get the response of /parquet-and-dataset-info for one specific dataset on huggingface.co. - It is assumed that the dataset can be accessed with the token. - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - hf_endpoint (`str`): - The Hub endpoint (for example: "https://huggingface.co") - hf_token (`str`, `optional`): - An app authentication token with read access to all the datasets. - committer_hf_token (`str`, `optional`): - A user authentication token (See https://huggingface.co/settings/token) with write access. It must: - - be part of the `huggingface` organization (to create the ref/convert/parquet "branch") - - be part of the `datasets-maintainers` organization (to push to the ref/convert/parquet "branch") - source_revision (`str`): - The git revision (e.g. "main" or sha) of the dataset used to prepare the parquet files - target_revision (`str`): - The target git revision (e.g. "ref/convert/parquet") of the dataset where to store the parquet files - commit_message (`str`): - The commit message to use when storing the parquet files - url_template (`str`): - The template to use to build the parquet file url - supported_datasets (`List[str]`): - The list of supported datasets, saving the blocked datasets. If empty, all datasets are supported - (saving the blocked datasets). - blocked_datasets (`List[str]`): - The list of blocked datasets. If empty, no dataset is blocked. - max_dataset_size (`int`): - The maximum size of a dataset in bytes. If the dataset is under the limit (which means that the size - can be fetched), it will be allowed. - max_external_data_files (`int`): - The maximum number of external data files of a dataset. This is for datasets with loading scripts only. - Returns: - `ParquetAndDatasetInfoResponse`: An object with the parquet_and_dataset_info_response - (dataset info and list of parquet files). - <Tip> - Raises the following errors: - - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`] - If the dataset is in the list of blocked datasets. - - [`libcommon.dataset.GatedExtraFieldsError`]: if the dataset is gated, with extra fields. - Programmatic access is not implemented for this type of dataset because there is no easy - way to get the list of extra fields. - - [`libcommon.dataset.GatedDisabledError`]: if the dataset is gated, but disabled. - - [`libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the - token does not give the sufficient access to the dataset, or if the dataset is private - (private datasets are not supported by the datasets server) - - ['HTTPError'](https://requests.readthedocs.io/en/latest/api/#requests.HTTPError): any other error when - asking access - - [`~job_runners.parquet_and_dataset_info.DatasetRevisionNotFoundError`] - If the revision does not exist or cannot be accessed using the token. - - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromHubError`] - If the dataset is too big to be converted to parquet - - [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) - If the datasets.config.HF_ENDPOINT is not set to the expected value - - [`~job_runners.parquet_and_dataset_info.DatasetTooBigFromDatasetsError`] - If the dataset is too big to be converted to parquet - - [`~job_runners.parquet_and_dataset_info.EmptyDatasetError`] - The dataset is empty. - - [`~job_runners.parquet_and_dataset_info.ConfigNamesError`] - If the list of configurations could not be obtained using the datasets library. - - [`~job_runners.parquet_and_dataset_info.DatasetInBlockListError`] - If the dataset is in the list of blocked datasets. - - [`~job_runners.parquet_and_dataset_info.DatasetWithTooManyExternalFilesError`] - If the dataset has too many external files to be converted to parquet - - [`~job_runners.parquet_and_dataset_info.DatasetWithTooBigExternalFilesError`] - If the dataset is too big external files be converted to parquet - - [`~job_runners.parquet_and_dataset_info.UnsupportedExternalFilesError`] - If we failed to get the external files sizes to make sure we can convert the dataset to parquet - - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestHTTPError`] - If we failed to get the external files sizes to make sure we can convert the dataset to parquet - - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestConnectionError`] - If we failed to get the external files sizes to make sure we can convert the dataset to parquet - - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestTimeoutError`] - If we failed to get the external files sizes to make sure we can convert the dataset to parquet - - [`~job_runners.parquet_and_dataset_info.ExternalFilesSizeRequestError`] - If we failed to get the external files sizes to make sure we can convert the dataset to parquet - </Tip> - """ - logging.info(f"get parquet files and dataset info for dataset={dataset}") - - raise_if_not_supported( - dataset=dataset, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - committer_hf_token=committer_hf_token, - revision=source_revision, - supported_datasets=supported_datasets, - blocked_datasets=blocked_datasets, - max_dataset_size=max_dataset_size, - ) - - hf_api = HfApi(endpoint=hf_endpoint, token=hf_token) - committer_hf_api = HfApi(endpoint=hf_endpoint, token=committer_hf_token) - - # get the sorted list of configurations - try: - config_names = sorted( - str(config) - for config in get_dataset_config_names(path=dataset, revision=source_revision, use_auth_token=hf_token) - ) - except _EmptyDatasetError as err: - raise EmptyDatasetError("The dataset is empty.", cause=err) from err - except Exception as err: - raise ConfigNamesError("Cannot get the config names for the dataset.", cause=err) from err - - # prepare the parquet files locally - parquet_files: List[ParquetFile] = [] - dataset_info: Dict[str, Any] = {} - download_config = DownloadConfig(delete_extracted=True) - for config in config_names: - builder = load_dataset_builder( - path=dataset, - name=config, - revision=source_revision, - use_auth_token=hf_token, - download_config=download_config, - ) - writer_batch_size = get_writer_batch_size(builder.info) - if writer_batch_size is not None and ( - builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size - ): - builder._writer_batch_size = writer_batch_size - raise_if_too_big_from_external_data_files( - builder=builder, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - hf_token=hf_token, - ) - builder.download_and_prepare(file_format="parquet") # the parquet files are stored in the cache dir - dataset_info[config] = asdict(builder.info) - # ^ see - # https://github.dev/huggingface/datasets/blob/e183a269067575db8765ee979bd8523d14a1adae/src/datasets/info.py#L244-L245 - # note that asdict() is not typed in the datasets library, hence type: ignore - parquet_files.extend( - ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=config) - for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet") - ) - - # create the target revision if it does not exist yet - try: - refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE) - if all(ref.ref != target_revision for ref in refs.converts): - committer_hf_api.create_branch( - repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=source_revision - ) - except RepositoryNotFoundError as err: - raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err - - # delete: - # - the previous files, - target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False) - previous_files = {f.rfilename for f in target_dataset_info.siblings} - # except: - # - the files we will update, - files_to_add = {parquet_file.repo_file(): parquet_file.local_file for parquet_file in parquet_files} - # - .gitattributes if present. - files_to_delete = previous_files - set(files_to_add.keys()).union({".gitattributes"}) - delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete] - logging.debug(f"delete_operations={delete_operations}") - - # send the files to the target revision - add_operations: List[CommitOperation] = [ - CommitOperationAdd(path_in_repo=file, path_or_fileobj=local_file) - for (file, local_file) in files_to_add.items() - ] - logging.debug(f"add_operations={add_operations}") - - committer_hf_api.create_commit( - repo_id=dataset, - repo_type=DATASET_TYPE, - revision=target_revision, - operations=delete_operations + add_operations, - commit_message=commit_message, - parent_commit=target_dataset_info.sha, - ) - - # call the API again to get the list of parquet files - target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True) - repo_files = [repo_file for repo_file in target_dataset_info.siblings if repo_file.rfilename.endswith(".parquet")] - # we might want to check if the sha of the parquet files is the same as the one we just uploaded - # we could also check that the list of parquet files is exactly what we expect - # let's not over engineer this for now. After all, what is on the Hub is the source of truth - # and the /parquet response is more a helper to get the list of parquet files - return { - "parquet_files": [ - create_parquet_file_item( - repo_file=repo_file, - dataset=dataset, - hf_endpoint=hf_endpoint, - target_revision=target_revision, - url_template=url_template, - ) - for repo_file in repo_files - ], - "dataset_info": dataset_info, - } - - -class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner): - parquet_and_dataset_info_config: ParquetAndInfoConfig - - @staticmethod - def get_job_type() -> str: - return "/parquet-and-dataset-info" - - @staticmethod - def get_job_runner_version() -> int: - return PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION - - def __init__( - self, - job_info: JobInfo, - app_config: AppConfig, - processing_step: ProcessingStep, - hf_datasets_cache: Path, - ) -> None: - super().__init__( - job_info=job_info, - app_config=app_config, - processing_step=processing_step, - hf_datasets_cache=hf_datasets_cache, - ) - self.parquet_and_dataset_info_config = app_config.parquet_and_info - - def compute(self) -> CompleteJobResult: - if self.dataset is None: - raise ParameterMissingError("'dataset' parameter is required") - return CompleteJobResult( - compute_parquet_and_dataset_info_response( - dataset=self.dataset, - hf_endpoint=self.common_config.hf_endpoint, - hf_token=self.common_config.hf_token, - committer_hf_token=self.parquet_and_dataset_info_config.committer_hf_token, - source_revision=self.parquet_and_dataset_info_config.source_revision, - target_revision=self.parquet_and_dataset_info_config.target_revision, - commit_message=self.parquet_and_dataset_info_config.commit_message, - url_template=self.parquet_and_dataset_info_config.url_template, - supported_datasets=self.parquet_and_dataset_info_config.supported_datasets, - blocked_datasets=self.parquet_and_dataset_info_config.blocked_datasets, - max_dataset_size=self.parquet_and_dataset_info_config.max_dataset_size, - max_external_data_files=self.parquet_and_dataset_info_config.max_external_data_files, - ) - ) - - def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]: - """Get the set of new splits, from the content created by the compute.""" - return { - SplitFullName(dataset=parquet_file["dataset"], config=parquet_file["config"], split=parquet_file["split"]) - for parquet_file in content["parquet_files"] - } diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index fc3d2e1a..ee8f3452 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -266 +265,0 @@ class HubDatasetTest(TypedDict): - parquet_and_dataset_info_response: Any @@ -366,28 +364,0 @@ def create_dataset_info_response_for_audio() -> Any: -# TODO: remove when "/parquet-and-dataset-info" step is removed -def create_parquet_and_dataset_info_response(dataset: str, data_type: Literal["csv", "audio"]) -> Any: - dataset, config, split = get_default_config_split(dataset) - - filename = "csv-train.parquet" if data_type == "csv" else "parquet-train.parquet" - size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE - info = ( - create_dataset_info_response_for_csv(dataset, config) - if data_type == "csv" - else create_dataset_info_response_for_audio() - ) - return { - "parquet_files": [ - { - "dataset": dataset, - "config": config, - "split": split, - "url": CI_URL_TEMPLATE.format( - repo_id=f"datasets/{dataset}", revision="refs%2Fconvert%2Fparquet", filename=f"{config}/{filename}" - ), - "filename": filename, - "size": size, - } - ], - "dataset_info": {config: info}, - } - - @@ -417 +388 @@ def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "au - "dataset_info": info, # we don't have config key here, only dataset_info dictionary + "dataset_info": info, @@ -568 +538,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -576 +545,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -584 +552,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -592 +559,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -600,3 +566,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": create_parquet_and_dataset_info_response( - dataset=hub_public_csv, data_type="csv" - ), @@ -610,3 +573,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": create_parquet_and_dataset_info_response( - dataset=hub_private_csv, data_type="csv" - ), @@ -620,3 +580,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": create_parquet_and_dataset_info_response( - dataset=hub_gated_csv, data_type="csv" - ), @@ -630 +587,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -638,3 +594,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": create_parquet_and_dataset_info_response( - dataset=hub_gated_extra_fields_csv, data_type="csv" - ), @@ -652,3 +605,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": create_parquet_and_dataset_info_response( - dataset=hub_public_audio, data_type="audio" - ), @@ -664 +614,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -674 +623,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -682 +630,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -690 +637,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, @@ -700 +646,0 @@ def hub_datasets( - "parquet_and_dataset_info_response": None, diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py deleted file mode 100644 index 0f6e9344..00000000 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ /dev/null @@ -1,482 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import io -from http import HTTPStatus -from typing import Any, Callable, Iterator, List, Optional - -import datasets.builder -import datasets.info -import pandas as pd -import pytest -import requests -from datasets import Features, Image, Value -from libcommon.exceptions import CustomError -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Priority -from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import DoesNotExist, get_response - -from worker.config import AppConfig -from worker.job_runners.parquet_and_dataset_info import ( - DatasetInBlockListError, - DatasetTooBigFromDatasetsError, - DatasetTooBigFromHubError, - DatasetWithTooBigExternalFilesError, - DatasetWithTooManyExternalFilesError, - ParquetAndDatasetInfoJobRunner, - get_dataset_info_or_raise, - get_writer_batch_size, - parse_repo_filename, - raise_if_blocked, - raise_if_not_supported, - raise_if_too_big_from_datasets, - raise_if_too_big_from_external_data_files, - raise_if_too_big_from_hub, -) -from worker.resources import LibrariesResource - -from ..fixtures.hub import HubDatasets - - -# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200 [email protected](scope="module", autouse=True) -def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyPatch]: - mp = pytest.MonkeyPatch() - mp.setenv( - "PARQUET_AND_INFO_BLOCKED_DATASETS", - ",".join(value["name"] for value in hub_datasets.values() if "jsonl" in value["name"]), - ) - mp.setenv( - "PARQUET_AND_INFO_SUPPORTED_DATASETS", - ",".join(value["name"] for value in hub_datasets.values() if "big" not in value["name"]), - ) - yield mp - mp.undo() - - -GetJobRunner = Callable[[str, AppConfig, bool], ParquetAndDatasetInfoJobRunner] - - [email protected] -def get_job_runner( - libraries_resource: LibrariesResource, - cache_mongo_resource: CacheMongoResource, - queue_mongo_resource: QueueMongoResource, -) -> GetJobRunner: - def _get_job_runner( - dataset: str, - app_config: AppConfig, - force: bool = False, - ) -> ParquetAndDatasetInfoJobRunner: - return ParquetAndDatasetInfoJobRunner( - job_info={ - "type": ParquetAndDatasetInfoJobRunner.get_job_type(), - "dataset": dataset, - "config": None, - "split": None, - "job_id": "job_id", - "force": force, - "priority": Priority.NORMAL, - }, - app_config=app_config, - processing_step=ProcessingStep( - name=ParquetAndDatasetInfoJobRunner.get_job_type(), - input_type="dataset", - requires=[], - required_by_dataset_viewer=False, - ancestors=[], - children=[], - parents=[], - job_runner_version=ParquetAndDatasetInfoJobRunner.get_job_runner_version(), - ), - hf_datasets_cache=libraries_resource.hf_datasets_cache, - ) - - return _get_job_runner - - -def assert_content_is_equal(content: Any, expected: Any) -> None: - print(content) - assert set(content.keys()) == {"parquet_files", "dataset_info"}, content - assert content["parquet_files"] == expected["parquet_files"], content - assert len(content["dataset_info"]) == 1, content - content_value = list(content["dataset_info"].values())[0] - expected_value = list(expected["dataset_info"].values())[0] - assert set(content_value.keys()) == set(expected_value.keys()), content - for key in content_value.keys(): - if key != "download_checksums": - assert content_value[key] == expected_value[key], content - assert len(content_value["download_checksums"]) == 1, content - content_checksum = list(content_value["download_checksums"].values())[0] - expected_checksum = list(expected_value["download_checksums"].values())[0] - assert content_checksum == expected_checksum, content - - -def test_compute( - app_config: AppConfig, - get_job_runner: GetJobRunner, - hub_datasets: HubDatasets, -) -> None: - dataset = hub_datasets["public"]["name"] - job_runner = get_job_runner(dataset, app_config, False) - assert job_runner.process() - cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset) - assert cached_response["http_status"] == HTTPStatus.OK - assert cached_response["error_code"] is None - assert cached_response["job_runner_version"] == job_runner.get_job_runner_version() - assert cached_response["dataset_git_revision"] is not None - content = cached_response["content"] - assert len(content["parquet_files"]) == 1 - assert_content_is_equal(content, hub_datasets["public"]["parquet_and_dataset_info_response"]) - - -def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset = "doesnotexist" - job_runner = get_job_runner(dataset, app_config, False) - assert not job_runner.process() - with pytest.raises(DoesNotExist): - get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset) - - [email protected]( - "dataset,blocked,raises", - [ - ("public", ["public"], True), - ("public", ["public", "audio"], True), - ("public", ["audio"], False), - ("public", [], False), - ], -) -def test_raise_if_blocked(dataset: str, blocked: List[str], raises: bool) -> None: - if raises: - with pytest.raises(DatasetInBlockListError): - raise_if_blocked(dataset=dataset, blocked_datasets=blocked) - else: - raise_if_blocked(dataset=dataset, blocked_datasets=blocked) - - [email protected]( - "name,raises", - [("public", False), ("big", True)], -) -def test_raise_if_too_big_from_hub( - hub_datasets: HubDatasets, - name: str, - raises: bool, - app_config: AppConfig, -) -> None: - dataset = hub_datasets[name]["name"] - dataset_info = get_dataset_info_or_raise( - dataset=dataset, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - revision="main", - ) - if raises: - with pytest.raises(DatasetTooBigFromHubError): - raise_if_too_big_from_hub( - dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size - ) - else: - raise_if_too_big_from_hub( - dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size - ) - - [email protected]( - "name,raises", - [("public", False), ("big", True)], -) -def test_raise_if_too_big_from_datasets( - hub_datasets: HubDatasets, - name: str, - raises: bool, - app_config: AppConfig, -) -> None: - dataset = hub_datasets[name]["name"] - if raises: - with pytest.raises(DatasetTooBigFromDatasetsError): - raise_if_too_big_from_datasets( - dataset=dataset, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - revision="main", - max_dataset_size=app_config.parquet_and_info.max_dataset_size, - ) - else: - raise_if_too_big_from_datasets( - dataset=dataset, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - revision="main", - max_dataset_size=app_config.parquet_and_info.max_dataset_size, - ) - - [email protected]( - "max_dataset_size,max_external_data_files,raises", - [ - (None, None, False), - (10, None, True), - ], -) -def test_raise_if_too_big_external_files( - external_files_dataset_builder: "datasets.builder.DatasetBuilder", - raises: bool, - max_dataset_size: Optional[int], - max_external_data_files: Optional[int], - app_config: AppConfig, -) -> None: - max_dataset_size = max_dataset_size or app_config.parquet_and_info.max_dataset_size - max_external_data_files = max_external_data_files or app_config.parquet_and_info.max_external_data_files - if raises: - with pytest.raises(DatasetWithTooBigExternalFilesError): - raise_if_too_big_from_external_data_files( - builder=external_files_dataset_builder, - hf_token=app_config.common.hf_token, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - ) - else: - raise_if_too_big_from_external_data_files( - builder=external_files_dataset_builder, - hf_token=app_config.common.hf_token, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - ) - - [email protected]( - "max_dataset_size,max_external_data_files,raises", - [ - (None, None, False), - (None, 1, True), - ], -) -def test_raise_if_too_many_external_files( - external_files_dataset_builder: "datasets.builder.DatasetBuilder", - raises: bool, - max_dataset_size: Optional[int], - max_external_data_files: Optional[int], - app_config: AppConfig, -) -> None: - max_dataset_size = max_dataset_size or app_config.parquet_and_info.max_dataset_size - max_external_data_files = max_external_data_files or app_config.parquet_and_info.max_external_data_files - if raises: - with pytest.raises(DatasetWithTooManyExternalFilesError): - raise_if_too_big_from_external_data_files( - builder=external_files_dataset_builder, - hf_token=app_config.common.hf_token, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - ) - else: - raise_if_too_big_from_external_data_files( - builder=external_files_dataset_builder, - hf_token=app_config.common.hf_token, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - ) - - [email protected]( - "in_list,raises", - [ - (True, False), - (False, True), - ], -) -def test_raise_if_not_supported( - hub_public_big: str, - app_config: AppConfig, - in_list: bool, - raises: bool, -) -> None: - if raises: - with pytest.raises(DatasetTooBigFromDatasetsError): - raise_if_not_supported( - dataset=hub_public_big, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - committer_hf_token=app_config.parquet_and_info.committer_hf_token, - revision="main", - max_dataset_size=app_config.parquet_and_info.max_dataset_size, - supported_datasets=[hub_public_big] if in_list else ["another_dataset"], - blocked_datasets=[], - ) - else: - raise_if_not_supported( - dataset=hub_public_big, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - committer_hf_token=app_config.parquet_and_info.committer_hf_token, - revision="main", - max_dataset_size=app_config.parquet_and_info.max_dataset_size, - supported_datasets=[hub_public_big] if in_list else ["another_dataset"], - blocked_datasets=[], - ) - - -def test_not_supported_if_big( - app_config: AppConfig, - get_job_runner: GetJobRunner, - hub_public_big: str, -) -> None: - # Not in the list of supported datasets and bigger than the maximum size - dataset = hub_public_big - job_runner = get_job_runner(dataset, app_config, False) - assert not job_runner.process() - cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset) - assert cached_response["http_status"] == HTTPStatus.NOT_IMPLEMENTED - assert cached_response["error_code"] == "DatasetTooBigFromDatasetsError" - - -def test_supported_if_gated( - app_config: AppConfig, - get_job_runner: GetJobRunner, - hub_gated_csv: str, -) -> None: - # Access must be granted - dataset = hub_gated_csv - job_runner = get_job_runner(dataset, app_config, False) - assert job_runner.process() - cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset) - assert cached_response["http_status"] == HTTPStatus.OK - assert cached_response["error_code"] is None - - -def test_not_supported_if_gated_with_extra_fields( - app_config: AppConfig, - get_job_runner: GetJobRunner, - hub_gated_extra_fields_csv: str, -) -> None: - # Access request should fail because extra fields in gated datasets are not supported - dataset = hub_gated_extra_fields_csv - job_runner = get_job_runner(dataset, app_config, False) - assert not job_runner.process() - cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset) - assert cached_response["http_status"] == HTTPStatus.NOT_FOUND - assert cached_response["error_code"] == "GatedExtraFieldsError" - - -def test_blocked( - app_config: AppConfig, - get_job_runner: GetJobRunner, - hub_public_jsonl: str, -) -> None: - # In the list of blocked datasets - dataset = hub_public_jsonl - job_runner = get_job_runner(dataset, app_config, False) - assert not job_runner.process() - cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset) - assert cached_response["http_status"] == HTTPStatus.NOT_IMPLEMENTED - assert cached_response["error_code"] == "DatasetInBlockListError" - - [email protected]( - "name", - ["public", "audio", "gated"], -) -def test_compute_splits_response_simple_csv_ok( - hub_datasets: HubDatasets, - get_job_runner: GetJobRunner, - name: str, - app_config: AppConfig, - data_df: pd.DataFrame, -) -> None: - dataset = hub_datasets[name]["name"] - expected_parquet_and_dataset_info_response = hub_datasets[name]["parquet_and_dataset_info_response"] - job_runner = get_job_runner(dataset, app_config, False) - result = job_runner.compute().content - assert_content_is_equal(result, expected_parquet_and_dataset_info_response) - - # download the parquet file and check that it is valid - if name == "audio": - return - - if name == "public": - df = pd.read_parquet(result["parquet_files"][0]["url"], engine="auto") - else: - # in all these cases, the parquet files are not accessible without a token - with pytest.raises(Exception): - pd.read_parquet(result["parquet_files"][0]["url"], engine="auto") - r = requests.get( - result["parquet_files"][0]["url"], headers={"Authorization": f"Bearer {app_config.common.hf_token}"} - ) - assert r.status_code == HTTPStatus.OK, r.text - df = pd.read_parquet(io.BytesIO(r.content), engine="auto") - assert df.equals(data_df), df - - [email protected]( - "name,error_code,cause", - [ - ("empty", "EmptyDatasetError", "EmptyDatasetError"), - ("does_not_exist", "DatasetNotFoundError", "HTTPError"), - ("gated_extra_fields", "GatedExtraFieldsError", "HTTPError"), - ("private", "DatasetNotFoundError", None), - ], -) -def test_compute_splits_response_simple_csv_error( - hub_datasets: HubDatasets, - get_job_runner: GetJobRunner, - name: str, - error_code: str, - cause: str, - app_config: AppConfig, -) -> None: - dataset = hub_datasets[name]["name"] - job_runner = get_job_runner(dataset, app_config, False) - with pytest.raises(CustomError) as exc_info: - job_runner.compute() - assert exc_info.value.code == error_code - assert exc_info.value.cause_exception == cause - if exc_info.value.disclose_cause: - response = exc_info.value.as_response() - assert set(response.keys()) == {"error", "cause_exception", "cause_message", "cause_traceback"} - response_dict = dict(response) - # ^ to remove mypy warnings - assert response_dict["cause_exception"] == cause - assert isinstance(response_dict["cause_traceback"], list) - assert response_dict["cause_traceback"][0] == "Traceback (most recent call last):\n" - - [email protected]( - "filename,split,config,raises", - [ - ("config/builder-split.parquet", "split", "config", False), - ("config/builder-with-dashes-split.parquet", "split", "config", False), - ("config/builder-split-00000-of-00001.parquet", "split", "config", False), - ("config/builder-with-dashes-split-00000-of-00001.parquet", "split", "config", False), - ("config/builder-split.with.dots-00000-of-00001.parquet", "split.with.dots", "config", False), - ( - "config/builder-with-dashes-caveat-asplitwithdashesisnotsupported-00000-of-00001.parquet", - "asplitwithdashesisnotsupported", - "config", - False, - ), - ("builder-split-00000-of-00001.parquet", "split", "config", True), - ("plain_text/openwebtext-10k-train.parquet", "train", "plain_text", False), - ("plain_text/openwebtext-10k-train-00000-of-00001.parquet", "train", "plain_text", False), - ], -) -def test_parse_repo_filename(filename: str, split: str, config: str, raises: bool) -> None: - if raises: - with pytest.raises(Exception): - parse_repo_filename(filename) - else: - assert parse_repo_filename(filename) == (config, split) - - [email protected]( - "ds_info, with_image", - [ - (datasets.info.DatasetInfo(), False), - (datasets.info.DatasetInfo(features=Features({"text": Value("string")})), False), - (datasets.info.DatasetInfo(features=Features({"image": Image()})), True), - (datasets.info.DatasetInfo(features=Features({"nested": [{"image": Image()}]})), True), - ], -) -def test_get_writer_batch_size(ds_info: datasets.info.DatasetInfo, with_image: bool) -> None: - assert get_writer_batch_size(ds_info) == (100 if with_image else None) diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py index 7987caa7..9a61857c 100644 --- a/services/worker/tests/test_job_runner_factory.py +++ b/services/worker/tests/test_job_runner_factory.py @@ -26 +25,0 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph: - ("/parquet-and-dataset-info", "ParquetAndDatasetInfoJobRunner"),
bdda4ed42c3f884995917ae6e7575ef67e338f48
Andrea Francis Soria Jimenez
2023-04-21T19:50:13
Fix key name for split names (#1081)
diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index 4bd03ac7..a63891b8 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -55 +55 @@ def fetch_split_names(dataset: str, config: str) -> List[str]: - for split_name_item in best_response.response["content"]["split_names"]: + for split_name_item in best_response.response["content"]["splits"]: diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 19ade72c..8f203fad 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -93 +93 @@ def get_SPLIT_NAMES_CONTENT_OK(dataset: str, config: str, splits: List[str]) -> - return {"split_names": [{"dataset": dataset, "config": config, "split": split_name} for split_name in splits]} + return {"splits": [{"dataset": dataset, "config": config, "split": split_name} for split_name in splits]}
c39cf845badb77a820923d8a230acda5ea64e472
Sylvain Lesage
2023-04-21T15:25:54
fix: 🐛 remove the wrong concept of blocked_by_parent (#1078)
diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index 4174a8d2..4bd03ac7 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -246 +245,0 @@ class CacheStatus: - blocked_by_parent: Dict[str, ArtifactState] = field(default_factory=dict) @@ -256 +254,0 @@ class CacheStatus: - "blocked_by_parent": sorted(self.blocked_by_parent.keys()), @@ -336,6 +333,0 @@ class Plan: -@dataclass -class ParentState: - is_fan_in: bool - artifact_states: List[ArtifactState] - - @@ -440,24 +431,0 @@ class DatasetState: - parent_states = [ - ParentState( - artifact_states=self._get_artifact_states_for_step( - step=parent_step, config=artifact_state.config, split=artifact_state.split - ), - is_fan_in=(parent_step.input_type == "config" and step.input_type == "dataset") - or (parent_step.input_type == "split" and step.input_type == "dataset") - or (parent_step.input_type == "split" and step.input_type == "config"), - ) - for parent_step in step.parents - ] - - # blocked by a parent? - if any( - parent_artifact_state.id not in cache_status.up_to_date - for parent_state in parent_states - for parent_artifact_state in parent_state.artifact_states - if not parent_state.is_fan_in - # ^ the fan-in relations between steps do not block, because these steps consider the - # cases: ok, pending and error. So, if one of the previous artifacts is pending, we must not block - ): - cache_status.blocked_by_parent[artifact_state.id] = artifact_state - continue - @@ -467,2 +435,4 @@ class DatasetState: - for parent_state in parent_states - for parent_artifact_state in parent_state.artifact_states + for parent_step in step.parents + for parent_artifact_state in self._get_artifact_states_for_step( + step=parent_step, config=artifact_state.config, split=artifact_state.split + ) diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 6bb78997..19ade72c 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -12 +11,0 @@ from libcommon.constants import ( - PROCESSING_STEP_CONFIG_INFO_VERSION, @@ -18 +16,0 @@ from libcommon.constants import ( - PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, @@ -509,2 +506,0 @@ def test_plan() -> None: - # "dataset-is-valid" is also empty, but is marked as blocked because it depends on "dataset-split-names", - # which is not yet known. @@ -514 +510,5 @@ def test_plan() -> None: - "blocked_by_parent": [ + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/config-names,dataset", + "/parquet-and-dataset-info,dataset", @@ -523,6 +522,0 @@ def test_plan() -> None: - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/config-names,dataset", - "/parquet-and-dataset-info,dataset", - ], @@ -538,0 +533,7 @@ def test_plan() -> None: + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-is-valid,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -548,0 +550,7 @@ def test_plan_job_creation_and_termination() -> None: + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-is-valid,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -558 +566,5 @@ def test_plan_job_creation_and_termination() -> None: - "blocked_by_parent": [ + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/config-names,dataset", + "/parquet-and-dataset-info,dataset", @@ -567,6 +578,0 @@ def test_plan_job_creation_and_termination() -> None: - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/config-names,dataset", - "/parquet-and-dataset-info,dataset", - ], @@ -581,0 +588,7 @@ def test_plan_job_creation_and_termination() -> None: + "dataset-info,dataset", + "dataset-is-valid,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", @@ -608,2 +621 @@ def test_plan_job_creation_and_termination() -> None: - # Config-level artifacts appear in the "blocked_by_parent" status, because the config names are now known, but - # the parents are not ready yet. + # Config-level artifacts are empty and ready to be filled (even if some of their parents are still missing) @@ -612,11 +623,0 @@ def test_plan_job_creation_and_termination() -> None: - "blocked_by_parent": [ - "/split-names-from-dataset-info,dataset,config1", - "/split-names-from-dataset-info,dataset,config2", - "config-info,dataset,config1", - "config-info,dataset,config2", - "config-parquet,dataset,config1", - "config-parquet,dataset,config2", - "config-size,dataset,config1", - "config-size,dataset,config2", - "dataset-is-valid,dataset", - ], @@ -626,0 +628,2 @@ def test_plan_job_creation_and_termination() -> None: + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", @@ -628,0 +632,4 @@ def test_plan_job_creation_and_termination() -> None: + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", @@ -630,0 +638,2 @@ def test_plan_job_creation_and_termination() -> None: + "config-size,dataset,config1", + "config-size,dataset,config2", @@ -631,0 +641 @@ def test_plan_job_creation_and_termination() -> None: + "dataset-is-valid,dataset", @@ -643,60 +653,2 @@ def test_plan_job_creation_and_termination() -> None: - queue_status={"in_process": ["/parquet-and-dataset-info,dataset"]}, - # The children of "/config-names,dataset" are ready to be backfilled for each config (fan-out) - tasks=[ - "CreateJob[/split-names-from-streaming,dataset,config1]", - "CreateJob[/split-names-from-streaming,dataset,config2]", - "CreateJob[config-parquet-and-info,dataset,config1]", - "CreateJob[config-parquet-and-info,dataset,config2]", - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]", - ], - ) - - -def test_plan_only_one_config() -> None: - # Set the "/config-names,dataset" artifact in cache - upsert_response( - kind="/config-names", - dataset=DATASET_NAME, - config=None, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "config-parquet-and-info,dataset,config1" artifact in cache - upsert_response( - kind="config-parquet-and-info", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - - assert_dataset_state( - # The config names are known - config_names=TWO_CONFIG_NAMES, - # The split names are not yet known - split_names_in_first_config=[], - # The children of "config-parquet-and-info" are blocked for config2, but not for config1 - # The two cache entries we created are up-to-date - cache_status={ - "blocked_by_parent": [ - "/split-names-from-dataset-info,dataset,config1", - "/split-names-from-dataset-info,dataset,config2", - "config-info,dataset,config2", - "config-parquet,dataset,config2", - "config-size,dataset,config2", - "dataset-is-valid,dataset", - ], - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ + queue_status={ + "in_process": [ @@ -704,6 +655,0 @@ def test_plan_only_one_config() -> None: - "/split-names-from-streaming,dataset,config1", - "/split-names-from-streaming,dataset,config2", - "config-info,dataset,config1", - "config-parquet,dataset,config1", - "config-parquet-and-info,dataset,config2", - "config-size,dataset,config1", @@ -710,0 +657 @@ def test_plan_only_one_config() -> None: + "dataset-is-valid,dataset", @@ -716,4 +663 @@ def test_plan_only_one_config() -> None: - ], - "cache_is_error_to_retry": [], - "cache_is_job_runner_obsolete": [], - "up_to_date": ["/config-names,dataset", "config-parquet-and-info,dataset,config1"], + ] @@ -721,2 +664,0 @@ def test_plan_only_one_config() -> None: - queue_status={"in_process": []}, - # the children of "config-parquet-and-info" for config1 are ready to be backfilled @@ -724 +666,2 @@ def test_plan_only_one_config() -> None: - "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", @@ -727,0 +671 @@ def test_plan_only_one_config() -> None: + "CreateJob[config-info,dataset,config2]", @@ -728,0 +673,2 @@ def test_plan_only_one_config() -> None: + "CreateJob[config-parquet,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", @@ -731,6 +677 @@ def test_plan_only_one_config() -> None: - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]", + "CreateJob[config-size,dataset,config2]", @@ -765 +706,4 @@ def test_plan_retry_error() -> None: - "blocked_by_parent": [ + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", @@ -786,5 +729,0 @@ def test_plan_retry_error() -> None: - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/parquet-and-dataset-info,dataset", - ], @@ -799,0 +739,19 @@ def test_plan_retry_error() -> None: + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-info,dataset,config2]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[config-size,dataset,config1]", + "CreateJob[config-size,dataset,config2]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-is-valid,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -833 +791 @@ def test_plan_incoherent_state() -> None: - # The split names are known (even if it's from an outdated artifact) + # The split names are known @@ -835,2 +793 @@ def test_plan_incoherent_state() -> None: - # "/split-names-from-dataset-info,dataset,config1" is blocked by its parents, even if it's in cache, because - # two ancestors are missing. It's not listed in up-to-date, even if it has been cached after /config-names. + # The split level artifacts for config1 are ready to be backfilled @@ -838,17 +794,0 @@ def test_plan_incoherent_state() -> None: - "blocked_by_parent": [ - "/split-names-from-dataset-info,dataset,config1", - "/split-names-from-dataset-info,dataset,config2", - "config-info,dataset,config1", - "config-info,dataset,config2", - "config-parquet,dataset,config1", - "config-parquet,dataset,config2", - "config-size,dataset,config1", - "config-size,dataset,config2", - "dataset-is-valid,dataset", - "split-first-rows-from-parquet,dataset,config1,split1", - "split-first-rows-from-parquet,dataset,config1,split2", - "split-first-rows-from-streaming,dataset,config1,split1", - "split-first-rows-from-streaming,dataset,config1,split2", - "split-opt-in-out-urls-scan,dataset,config1,split1", - "split-opt-in-out-urls-scan,dataset,config1,split2", - ], @@ -858,0 +799 @@ def test_plan_incoherent_state() -> None: + "/split-names-from-dataset-info,dataset,config2", @@ -861,100 +802 @@ def test_plan_incoherent_state() -> None: - "config-parquet-and-info,dataset,config1", - "config-parquet-and-info,dataset,config2", - "dataset-info,dataset", - "dataset-parquet,dataset", - "dataset-size,dataset", - "dataset-split-names,dataset", - "dataset-split-names-from-dataset-info,dataset", - "dataset-split-names-from-streaming,dataset", - ], - "cache_is_error_to_retry": [], - "cache_is_job_runner_obsolete": [], - "up_to_date": ["/config-names,dataset"], - }, - queue_status={"in_process": []}, - # The children of "/split-names-from-dataset-info,dataset,config1" are not ready to be backfilled - # because it's blocked by its parents. - tasks=[ - "CreateJob[/parquet-and-dataset-info,dataset]", - "CreateJob[/split-names-from-streaming,dataset,config1]", - "CreateJob[/split-names-from-streaming,dataset,config2]", - "CreateJob[config-parquet-and-info,dataset,config1]", - "CreateJob[config-parquet-and-info,dataset,config2]", - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]", - ], - ) - - -def test_plan_get_splits() -> None: - # Set the "/config-names,dataset" artifact in cache - upsert_response( - kind="/config-names", - dataset=DATASET_NAME, - config=None, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "config-parquet-and-info,dataset,config1" artifact in cache - upsert_response( - kind="config-parquet-and-info", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "config-info,dataset,config1" artifact in cache - upsert_response( - kind="config-info", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_INFO_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "/split-names-from-dataset-info,dataset,config1" artifact in cache - upsert_response( - kind="/split-names-from-dataset-info", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "/split-names-from-streaming,dataset,config1" artifact in cache - upsert_response( - kind="/split-names-from-streaming", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - - assert_dataset_state( - # The config names are known - config_names=TWO_CONFIG_NAMES, - # The split names are known (even if it's from an outdated artifact) - split_names_in_first_config=SPLIT_NAMES_OK, - # The chains /config-names,dataset -> config-parquet-and-info,dataset,config1 -> config-info,dataset,config1 - # -> /split-names-from-dataset-info,dataset,config1 and - # /config-names,dataset -> /split-names-from-streaming,dataset,config1 are up to date - cache_status={ - "blocked_by_parent": [ - "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config1", @@ -962,13 +803,0 @@ def test_plan_get_splits() -> None: - "config-parquet,dataset,config2", - "config-size,dataset,config2", - "dataset-is-valid,dataset", - "split-first-rows-from-parquet,dataset,config1,split1", - "split-first-rows-from-parquet,dataset,config1,split2", - "split-opt-in-out-urls-scan,dataset,config1,split1", - "split-opt-in-out-urls-scan,dataset,config1,split2", - ], - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/parquet-and-dataset-info,dataset", - "/split-names-from-streaming,dataset,config2", @@ -975,0 +805,2 @@ def test_plan_get_splits() -> None: + "config-parquet,dataset,config2", + "config-parquet-and-info,dataset,config1", @@ -977,0 +809 @@ def test_plan_get_splits() -> None: + "config-size,dataset,config2", @@ -978,0 +811 @@ def test_plan_get_splits() -> None: + "dataset-is-valid,dataset", @@ -983,0 +817,2 @@ def test_plan_get_splits() -> None: + "split-first-rows-from-parquet,dataset,config1,split1", + "split-first-rows-from-parquet,dataset,config1,split2", @@ -985,0 +821,2 @@ def test_plan_get_splits() -> None: + "split-opt-in-out-urls-scan,dataset,config1,split1", + "split-opt-in-out-urls-scan,dataset,config1,split2", @@ -992,3 +828,0 @@ def test_plan_get_splits() -> None: - "/split-names-from-streaming,dataset,config1", - "config-info,dataset,config1", - "config-parquet-and-info,dataset,config1", @@ -998 +831,0 @@ def test_plan_get_splits() -> None: - # split-first-rows-from-streaming is now ready to be backfilled for split1 and split2 of dataset,config1 @@ -1000,0 +834,2 @@ def test_plan_get_splits() -> None: + "CreateJob[/split-names-from-dataset-info,dataset,config2]", + "CreateJob[/split-names-from-streaming,dataset,config1]", @@ -1001,0 +837,2 @@ def test_plan_get_splits() -> None: + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-info,dataset,config2]", @@ -1002,0 +840,2 @@ def test_plan_get_splits() -> None: + "CreateJob[config-parquet,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", @@ -1004,0 +844 @@ def test_plan_get_splits() -> None: + "CreateJob[config-size,dataset,config2]", @@ -1005,0 +846 @@ def test_plan_get_splits() -> None: + "CreateJob[dataset-is-valid,dataset]", @@ -1010,0 +852,2 @@ def test_plan_get_splits() -> None: + "CreateJob[split-first-rows-from-parquet,dataset,config1,split1]", + "CreateJob[split-first-rows-from-parquet,dataset,config1,split2]", @@ -1012,0 +856,2 @@ def test_plan_get_splits() -> None: + "CreateJob[split-opt-in-out-urls-scan,dataset,config1,split1]", + "CreateJob[split-opt-in-out-urls-scan,dataset,config1,split2]", @@ -1040,34 +884,0 @@ def test_plan_updated_at() -> None: - # Set the "config-info,dataset,config1" artifact in cache - upsert_response( - kind="config-info", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_INFO_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "/split-names-from-dataset-info,dataset,config1" artifact in cache - upsert_response( - kind="/split-names-from-dataset-info", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "/split-names-from-streaming,dataset,config1" artifact in cache - upsert_response( - kind="/split-names-from-streaming", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - @@ -1089,5 +900,3 @@ def test_plan_updated_at() -> None: - # The split names are known (even if it's from an outdated artifact) - split_names_in_first_config=SPLIT_NAMES_OK, - # config-parquet-and-info,dataset,config1 is marked as outdate by parent, - # as well as /split-names-from-streaming,dataset,config1 - # This means that all their children are marked as blocked by parent + # The split names are not yet known + split_names_in_first_config=[], + # config-parquet-and-info,dataset,config1 is marked as outdated by parent, @@ -1096 +905,4 @@ def test_plan_updated_at() -> None: - "blocked_by_parent": [ + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": ["config-parquet-and-info,dataset,config1"], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", @@ -1098,0 +911,2 @@ def test_plan_updated_at() -> None: + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", @@ -1102,0 +917 @@ def test_plan_updated_at() -> None: + "config-parquet-and-info,dataset,config2", @@ -1105,17 +919,0 @@ def test_plan_updated_at() -> None: - "dataset-is-valid,dataset", - "split-first-rows-from-parquet,dataset,config1,split1", - "split-first-rows-from-parquet,dataset,config1,split2", - "split-first-rows-from-streaming,dataset,config1,split1", - "split-first-rows-from-streaming,dataset,config1,split2", - "split-opt-in-out-urls-scan,dataset,config1,split1", - "split-opt-in-out-urls-scan,dataset,config1,split2", - ], - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [ - "/split-names-from-streaming,dataset,config1", - "config-parquet-and-info,dataset,config1", - ], - "cache_is_empty": [ - "/parquet-and-dataset-info,dataset", - "/split-names-from-streaming,dataset,config2", - "config-parquet-and-info,dataset,config2", @@ -1122,0 +921 @@ def test_plan_updated_at() -> None: + "dataset-is-valid,dataset", @@ -1134,2 +933 @@ def test_plan_updated_at() -> None: - # config-parquet-and-info and /split-names-from-streaming will be refreshed for both configs - # (even if config1 is already in the cache -> but it's outdated) + # config-parquet-and-info,dataset,config1 will be refreshed @@ -1137,0 +936,2 @@ def test_plan_updated_at() -> None: + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", @@ -1139,0 +940,4 @@ def test_plan_updated_at() -> None: + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-info,dataset,config2]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet,dataset,config2]", @@ -1141,0 +946,2 @@ def test_plan_updated_at() -> None: + "CreateJob[config-size,dataset,config1]", + "CreateJob[config-size,dataset,config2]", @@ -1142,0 +949 @@ def test_plan_updated_at() -> None: + "CreateJob[dataset-is-valid,dataset]", @@ -1161,23 +968 @@ def test_plan_job_runner_version() -> None: - job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "config-parquet-and-info,dataset,config1" artifact in cache - upsert_response( - kind="config-parquet-and-info", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - # Set the "config-parquet,dataset,config1" artifact in cache - upsert_response( - kind="config-parquet", - dataset=DATASET_NAME, - config=CONFIG_NAME_1, - split=None, - content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important - http_status=HTTPStatus.OK, - job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_VERSION - 1, # <- current version is 3, this one is 1 + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION - 1, # <- old version @@ -1186 +970,0 @@ def test_plan_job_runner_version() -> None: - @@ -1192 +976 @@ def test_plan_job_runner_version() -> None: - # config-parquet,dataset,config1 is in the category: "is_job_runner_obsolete" + # /config-names is in the category: "is_job_runner_obsolete" @@ -1194,8 +977,0 @@ def test_plan_job_runner_version() -> None: - "blocked_by_parent": [ - "/split-names-from-dataset-info,dataset,config1", - "/split-names-from-dataset-info,dataset,config2", - "config-info,dataset,config2", - "config-parquet,dataset,config2", - "config-size,dataset,config2", - "dataset-is-valid,dataset", - ], @@ -1205,0 +982,2 @@ def test_plan_job_runner_version() -> None: + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", @@ -1208,0 +987,4 @@ def test_plan_job_runner_version() -> None: + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-parquet-and-info,dataset,config1", @@ -1210,0 +993 @@ def test_plan_job_runner_version() -> None: + "config-size,dataset,config2", @@ -1211,0 +995 @@ def test_plan_job_runner_version() -> None: + "dataset-is-valid,dataset", @@ -1219,2 +1003,2 @@ def test_plan_job_runner_version() -> None: - "cache_is_job_runner_obsolete": ["config-parquet,dataset,config1"], - "up_to_date": ["/config-names,dataset", "config-parquet-and-info,dataset,config1"], + "cache_is_job_runner_obsolete": ["/config-names,dataset"], + "up_to_date": [], @@ -1223 +1007 @@ def test_plan_job_runner_version() -> None: - # config-parquet,dataset,config1 will be refreshed because its job runner has been upgraded + # "/config-names,dataset" will be refreshed because its job runner has been upgraded @@ -1224,0 +1009 @@ def test_plan_job_runner_version() -> None: + "CreateJob[/config-names,dataset]", @@ -1225,0 +1011,2 @@ def test_plan_job_runner_version() -> None: + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", @@ -1228,0 +1016 @@ def test_plan_job_runner_version() -> None: + "CreateJob[config-info,dataset,config2]", @@ -1229,0 +1018,2 @@ def test_plan_job_runner_version() -> None: + "CreateJob[config-parquet,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", @@ -1231,0 +1022 @@ def test_plan_job_runner_version() -> None: + "CreateJob[config-size,dataset,config2]", @@ -1232,0 +1024 @@ def test_plan_job_runner_version() -> None: + "CreateJob[dataset-is-valid,dataset]", @@ -1269 +1060,0 @@ def test_plan_git_revision( - # and its children will be blocked @@ -1277 +1068,4 @@ def test_plan_git_revision( - "blocked_by_parent": [ + "cache_has_different_git_revision": ["/config-names,dataset"], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", @@ -1298,5 +1091,0 @@ def test_plan_git_revision( - "cache_has_different_git_revision": ["/config-names,dataset"], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/parquet-and-dataset-info,dataset", - ], @@ -1310,0 +1100,19 @@ def test_plan_git_revision( + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-info,dataset,config2]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[config-size,dataset,config1]", + "CreateJob[config-size,dataset,config2]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-is-valid,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -1321,11 +1128,0 @@ def test_plan_git_revision( - "blocked_by_parent": [ - "/split-names-from-dataset-info,dataset,config1", - "/split-names-from-dataset-info,dataset,config2", - "config-info,dataset,config1", - "config-info,dataset,config2", - "config-parquet,dataset,config1", - "config-parquet,dataset,config2", - "config-size,dataset,config1", - "config-size,dataset,config2", - "dataset-is-valid,dataset", - ], @@ -1335,0 +1133,2 @@ def test_plan_git_revision( + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", @@ -1337,0 +1137,4 @@ def test_plan_git_revision( + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", @@ -1339,0 +1143,2 @@ def test_plan_git_revision( + "config-size,dataset,config1", + "config-size,dataset,config2", @@ -1340,0 +1146 @@ def test_plan_git_revision( + "dataset-is-valid,dataset", @@ -1353,0 +1160,2 @@ def test_plan_git_revision( + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", @@ -1355,0 +1164,4 @@ def test_plan_git_revision( + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-info,dataset,config2]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet,dataset,config2]", @@ -1357,0 +1170,2 @@ def test_plan_git_revision( + "CreateJob[config-size,dataset,config1]", + "CreateJob[config-size,dataset,config2]", @@ -1358,0 +1173 @@ def test_plan_git_revision( + "CreateJob[dataset-is-valid,dataset]", @@ -1421,8 +1235,0 @@ def test_plan_update_fan_in_parent() -> None: - "blocked_by_parent": [ - "/split-names-from-dataset-info,dataset,config1", - "/split-names-from-dataset-info,dataset,config2", - "config-info,dataset,config2", - "config-parquet,dataset,config2", - "config-size,dataset,config2", - "dataset-is-valid,dataset", - ], @@ -1434,0 +1242,2 @@ def test_plan_update_fan_in_parent() -> None: + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", @@ -1437,0 +1247,2 @@ def test_plan_update_fan_in_parent() -> None: + "config-info,dataset,config2", + "config-parquet,dataset,config2", @@ -1439,0 +1251 @@ def test_plan_update_fan_in_parent() -> None: + "config-size,dataset,config2", @@ -1440,0 +1253 @@ def test_plan_update_fan_in_parent() -> None: + "dataset-is-valid,dataset", @@ -1455 +1268 @@ def test_plan_update_fan_in_parent() -> None: - # config-parquet,dataset,config1 will be refreshed because its job runner has been upgraded + # dataset-parquet,dataset will be refreshed @@ -1457,0 +1271,2 @@ def test_plan_update_fan_in_parent() -> None: + "CreateJob[/split-names-from-dataset-info,dataset,config1]", + "CreateJob[/split-names-from-dataset-info,dataset,config2]", @@ -1460,0 +1276,2 @@ def test_plan_update_fan_in_parent() -> None: + "CreateJob[config-info,dataset,config2]", + "CreateJob[config-parquet,dataset,config2]", @@ -1462,0 +1280 @@ def test_plan_update_fan_in_parent() -> None: + "CreateJob[config-size,dataset,config2]", @@ -1463,0 +1282 @@ def test_plan_update_fan_in_parent() -> None: + "CreateJob[dataset-is-valid,dataset]",
c6e300d9f16f2efce24c2cc9d0a40e7070526ef2
Sylvain Lesage
2023-04-21T14:47:46
test: 💍 make the e2e tests on API clearer (#1076)
diff --git a/e2e/tests/test_11_auth.py b/e2e/tests/test_11_api.py similarity index 53% rename from e2e/tests/test_11_auth.py rename to e2e/tests/test_11_api.py index 7c018cc6..fbeae374 100644 --- a/e2e/tests/test_11_auth.py +++ b/e2e/tests/test_11_api.py @@ -3,0 +4,2 @@ +from typing import Literal + @@ -37,19 +39,47 @@ def test_auth_e2e( - endpoints = [ - f"/config-names?dataset={dataset}", - f"/splits?dataset={dataset}", - f"/first-rows?dataset={dataset}&config={config}&split={split}", - f"/parquet-and-dataset-info?dataset={dataset}&config={config}", - f"/parquet?dataset={dataset}", - f"/parquet?dataset={dataset}&config={config}", - f"/dataset-info?dataset={dataset}", - f"/dataset-info?dataset={dataset}&config={config}", - f"/size?dataset={dataset}", - f"/size?dataset={dataset}&config={config}", - ] - for endpoint in endpoints: - poll_until_ready_and_assert( - relative_url=endpoint, - expected_status_code=expected_status_code, - expected_error_code=expected_error_code, - headers=headers, - ) + poll_until_ready_and_assert( + relative_url=f"/config-names?dataset={dataset}", + expected_status_code=expected_status_code, + expected_error_code=expected_error_code, + headers=headers, + ) + + [email protected]( + "endpoint,input_type", + [ + ("/config-names", "dataset"), + ("/splits", "dataset"), + ("/splits", "config"), + ("/first-rows", "split"), + ("/parquet-and-dataset-info", "dataset"), + ("/parquet", "dataset"), + ("/parquet", "config"), + ("/dataset-info", "dataset"), + ("/dataset-info", "config"), + ("/size", "dataset"), + ("/size", "config"), + ("/is-valid", "dataset"), + ("/valid", "all"), + ], +) +def test_endpoint( + auth_headers: AuthHeaders, + hf_public_dataset_repo_csv_data: str, + endpoint: str, + input_type: Literal["all", "dataset", "config", "split"], +) -> None: + auth: AuthType = "none" + expected_status_code: int = 200 + expected_error_code = None + # TODO: add dataset with various splits, or various configs + dataset, config, split = get_default_config_split(hf_public_dataset_repo_csv_data) + headers = auth_headers[auth] + + # asking for the dataset will launch the jobs, without the need of a webhook + relative_url = endpoint + if input_type != "all": + relative_url += f"?dataset={dataset}" + if input_type != "dataset": + relative_url += f"&config={config}" + if input_type != "config": + relative_url += f"&split={split}" @@ -56,0 +87,18 @@ def test_auth_e2e( + poll_until_ready_and_assert( + relative_url=relative_url, + expected_status_code=expected_status_code, + expected_error_code=expected_error_code, + headers=headers, + ) + + +def test_rows_endpoint( + auth_headers: AuthHeaders, + hf_public_dataset_repo_csv_data: str, +) -> None: + auth: AuthType = "none" + expected_status_code: int = 200 + expected_error_code = None + # TODO: add dataset with various splits, or various configs + dataset, config, split = get_default_config_split(hf_public_dataset_repo_csv_data) + headers = auth_headers[auth] @@ -68,0 +117 @@ def test_auth_e2e( + assert "features" in content, rows_response @@ -69,0 +119 @@ def test_auth_e2e( + features = content["features"] @@ -70,0 +121 @@ def test_auth_e2e( + assert isinstance(features, list), features @@ -73,3 +123,0 @@ def test_auth_e2e( - assert "features" in content, rows_response - features = content["features"] - assert isinstance(features, list), features diff --git a/e2e/tests/utils.py b/e2e/tests/utils.py index cbcdc83b..cd6669cf 100644 --- a/e2e/tests/utils.py +++ b/e2e/tests/utils.py @@ -124 +124 @@ def poll_until_ready_and_assert( - should_retry = response.headers.get("X-Error-Code") == "ResponseNotReady" + should_retry = response.headers.get("X-Error-Code") in ["ResponseNotReady", "ResponseAlreadyComputedError"]
cffaca17e2650cb4596ef3ef8b12895bfad1470d
Quentin Lhoest
2023-04-21T14:09:51
Lower parquet row group size for image datasets (#833)
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index dfa58b19..840569ff 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -21 +21 @@ PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION = 2 -PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION = 1 +PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION = 2 @@ -29,0 +30,2 @@ PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION = 1 +PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS = 100 +PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS = 100 diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py index 7a02018b..a122c73a 100644 --- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py +++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py @@ -16,0 +17 @@ import datasets.config +import datasets.info @@ -42 +43,5 @@ from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFo -from libcommon.constants import PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION +from libcommon.constants import ( + PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS, + PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS, + PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION, +) @@ -636,0 +642,24 @@ def raise_if_too_big_from_external_data_files( +def get_writer_batch_size(ds_config_info: datasets.info.DatasetInfo) -> Optional[int]: + """ + Get the writer_batch_size that defines the maximum row group size in the parquet files. + The default in `datasets` is 1,000 but we lower it to 100 for image datasets. + This allows to optimize random access to parquet file, since accessing 1 row requires + to read its entire row group. + + Args: + ds_config_info (`datasets.info.DatasetInfo`): + Dataset info from `datasets`. + + Returns: + writer_batch_size (`Optional[int]`): + Writer batch size to pass to a dataset builder. + If `None`, then it will use the `datasets` default. + """ + if "Audio(" in str(ds_config_info.features): + return PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS + elif "Image(" in str(ds_config_info.features): + return PROCESSING_STEP_PARQUET_AND_DATASET_INFO_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS + else: + return None + + @@ -768,0 +798,5 @@ def compute_parquet_and_dataset_info_response( + writer_batch_size = get_writer_batch_size(builder.info) + if writer_batch_size is not None and ( + builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size + ): + builder._writer_batch_size = writer_batch_size diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index 8fe74c65..0f6e9344 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -8,0 +9 @@ import datasets.builder +import datasets.info @@ -11,0 +13 @@ import requests +from datasets import Features, Image, Value @@ -26,0 +29 @@ from worker.job_runners.parquet_and_dataset_info import ( + get_writer_batch_size, @@ -466,0 +470,13 @@ def test_parse_repo_filename(filename: str, split: str, config: str, raises: boo + + [email protected]( + "ds_info, with_image", + [ + (datasets.info.DatasetInfo(), False), + (datasets.info.DatasetInfo(features=Features({"text": Value("string")})), False), + (datasets.info.DatasetInfo(features=Features({"image": Image()})), True), + (datasets.info.DatasetInfo(features=Features({"nested": [{"image": Image()}]})), True), + ], +) +def test_get_writer_batch_size(ds_info: datasets.info.DatasetInfo, with_image: bool) -> None: + assert get_writer_batch_size(ds_info) == (100 if with_image else None)
422486083fceba21f5b56e64eb855744e3258209
Quentin Lhoest
2023-04-21T13:05:44
Opt-in / opt-out URLs scan with Spawning (#1044)
diff --git a/.github/workflows/_unit-tests-python.yml b/.github/workflows/_unit-tests-python.yml index 071c9c9e..78d2daef 100644 --- a/.github/workflows/_unit-tests-python.yml +++ b/.github/workflows/_unit-tests-python.yml @@ -49,0 +50 @@ jobs: + CI_SPAWNING_TOKEN: ${{ secrets.CI_SPAWNING_TOKEN }} diff --git a/.github/workflows/s-worker.yml b/.github/workflows/s-worker.yml index cd234c9c..ffb3be37 100644 --- a/.github/workflows/s-worker.yml +++ b/.github/workflows/s-worker.yml @@ -35,0 +36 @@ jobs: + secrets: inherit diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 013f9206..e413d595 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -66,0 +67,3 @@ secrets: + spawningToken: + fromSecret: true + secretName: "spawning-token" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index a8a28952..d610b212 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -64,0 +65,3 @@ secrets: + spawningToken: + fromSecret: true + secretName: "spawning-token" @@ -90,0 +94,6 @@ parquetAndInfo: +optInOutUrlsScan: + maxConcurrentRequestsNumber: 100 + maxRequestsPerSecond: 50 + rowsMaxNumber: 100_000 + urlsNumberPerBatch: 1000 + diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index 1a34a7c5..7275ea30 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -63,0 +64,24 @@ +# specific to the split-opt-in-out-urls-scan job runner +- name: OPT_IN_OUT_URLS_SCAN_COLUMNS_MAX_NUMBER + value: {{ .Values.optInOutUrlsScan.columnsMaxNumber | quote }} +- name: OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER + value: {{ .Values.optInOutUrlsScan.maxConcurrentRequestsNumber | quote }} +- name: OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND + value: {{ .Values.optInOutUrlsScan.maxRequestsPerSecond | quote }} +- name: OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER + value: {{ .Values.optInOutUrlsScan.rowsMaxNumber | quote }} +- name: OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN + {{- if .Values.secrets.spawningToken.fromSecret }} + valueFrom: + secretKeyRef: + name: {{ .Values.secrets.spawningToken.secretName | quote }} + key: SPAWNING_TOKEN + optional: false + {{- else }} + value: {{ .Values.secrets.spawningToken.value }} + {{- end }} +- name: OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH + value: {{ .Values.optInOutUrlsScan.urlsNumberPerBatch | quote }} +- name: OPT_IN_OUT_URLS_SCAN_SPAWNING_URL + value: {{ .Values.optInOutUrlsScan.spawningUrl | quote }} + diff --git a/chart/values.yaml b/chart/values.yaml index 7449c373..c37fca78 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -84,0 +85,3 @@ secrets: + spawningToken: + fromSecret: true + secretName: "spawning-token" @@ -167,0 +171,14 @@ parquetAndInfo: +optInOutUrlsScan: + columnsMaxNumber: 10 + # the max number of columns to scan + maxConcurrentRequestsNumber: 10 + # the max concurrent request number + maxRequestsPerSecond: 20 + # the max number of request allowed to process in parallel per second + rowsMaxNumber: 1_000 + # the max number of rows to scan + urlsNumberPerBatch: 1_000 + # the number of grouped urls to be send in every request to spawning + spawningUrl: "https://opts-api.spawningaiapi.com/api/v2/query/urls" + # the URL for spawning requests + diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index efacfa31..a95be214 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -28,0 +29 @@ from libcommon.constants import ( + PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION, @@ -271,0 +273,5 @@ class ProcessingGraphConfig: + "split-opt-in-out-urls-scan": { + "input_type": "split", + "requires": ["split-first-rows-from-streaming"], + "job_runner_version": PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION, + }, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index d8894972..dfa58b19 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -27,0 +28 @@ PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 2 +PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION = 1 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 28b46c2f..9c302ba9 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -147 +147 @@ def graph() -> ProcessingGraph: - ["dataset-is-valid"], + ["dataset-is-valid", "split-opt-in-out-urls-scan"], @@ -211,0 +212,13 @@ def graph() -> ProcessingGraph: + ( + "split-opt-in-out-urls-scan", + [], + ["split-first-rows-from-streaming"], + [ + "/config-names", + "/split-names-from-streaming", + "split-first-rows-from-streaming", + "/split-names-from-dataset-info", + "config-info", + "config-parquet-and-info", + ], + ), diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 328eef6a..6bb78997 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -283,0 +284,5 @@ def get_SPLIT_STATE_DICT(dataset: str, config: str, split: str) -> Any: + { + "id": f"split-opt-in-out-urls-scan,{dataset},{config},{split}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, @@ -846,0 +852,2 @@ def test_plan_incoherent_state() -> None: + "split-opt-in-out-urls-scan,dataset,config1,split1", + "split-opt-in-out-urls-scan,dataset,config1,split2", @@ -959,0 +967,2 @@ def test_plan_get_splits() -> None: + "split-opt-in-out-urls-scan,dataset,config1,split1", + "split-opt-in-out-urls-scan,dataset,config1,split2", @@ -1100,0 +1110,2 @@ def test_plan_updated_at() -> None: + "split-opt-in-out-urls-scan,dataset,config1,split1", + "split-opt-in-out-urls-scan,dataset,config1,split2", diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 5f5f7af6..0e7b2a76 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -142,0 +143,3 @@ class EndpointConfig: + "/opt-in-out-urls-scan": { + "split": ["split-opt-in-out-urls-scan"], + }, diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index a8e1a47b..c5c65517 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. +# This file is automatically @generated by Poetry and should not be changed by hand. @@ -123,0 +124,15 @@ speedups = ["Brotli", "aiodns", "cchardet"] +[[package]] +name = "aiolimiter" +version = "1.0.0" +description = "asyncio rate limiter, a leaky bucket implementation" +category = "main" +optional = false +python-versions = ">=3.6.1,<4.0.0" +files = [ + {file = "aiolimiter-1.0.0-py3-none-any.whl", hash = "sha256:f1c5ba2a2861cd4a126c1294f5282208383e67d5b128a4f32def0c702cae8039"}, + {file = "aiolimiter-1.0.0.tar.gz", hash = "sha256:9d40767e4476048145dfa9f61948445168d6e63cf42c95785a20b9aaff2e4564"}, +] + +[package.extras] +docs = ["aiohttp-theme (>=0.1.6,<0.2.0)", "sphinx (>=2.2.1,<5.0.0)", "sphinx-autodoc-typehints (>=1.10.3,<2.0.0)", "sphinxcontrib-spelling (>=4.3,<8.0)", "toml (>=0.10.0,<0.11.0)"] + @@ -991 +1006 @@ description = "Backport of PEP 654 (exception groups)" -category = "dev" +category = "main" @@ -1586 +1601 @@ description = "brain-dead simple config-ini parsing" -category = "dev" +category = "main" @@ -2750,0 +2766 @@ files = [ + {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -2942 +2958 @@ description = "plugin and hook calling mechanisms for python" -category = "dev" +category = "main" @@ -3623 +3639 @@ description = "pytest: simple powerful testing with Python" -category = "dev" +category = "main" @@ -3642,0 +3659,19 @@ testing = ["argcomplete", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2. +[[package]] +name = "pytest-asyncio" +version = "0.21.0" +description = "Pytest support for asyncio" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pytest-asyncio-0.21.0.tar.gz", hash = "sha256:2b38a496aef56f56b0e87557ec313e11e1ab9276fc3863f6a7be0f1d0e415e1b"}, + {file = "pytest_asyncio-0.21.0-py3-none-any.whl", hash = "sha256:f2b3366b7cd501a4056858bd39349d5af19742aed2d81660b7998b6341c7eb9c"}, +] + +[package.dependencies] +pytest = ">=7.0.0" + +[package.extras] +docs = ["sphinx (>=5.3)", "sphinx-rtd-theme (>=1.0)"] +testing = ["coverage (>=6.2)", "flaky (>=3.5.0)", "hypothesis (>=5.7.1)", "mypy (>=0.931)", "pytest-trio (>=0.7.0)"] + @@ -4643 +4678 @@ description = "A lil' TOML parser" -category = "dev" +category = "main" @@ -5395 +5430 @@ python-versions = "3.9.15" -content-hash = "02e169574090bd1891985027473ac6278439746c700ceb5a3136b0b99f6223f1" +content-hash = "6e0822d5ac7b76f433ac94f4840b17312a8684c031aee7a92ff0bf442d41430b" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index bd348c1e..1f869141 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -11,0 +12 @@ aiohttp = "^3.8.4" +aiolimiter = "^1.0.0" @@ -49,0 +51 @@ mirakuru = "^2.4.2" +pytest-asyncio = "^0.21.0" diff --git a/services/worker/src/worker/common_exceptions.py b/services/worker/src/worker/common_exceptions.py new file mode 100644 index 00000000..343f772b --- /dev/null +++ b/services/worker/src/worker/common_exceptions.py @@ -0,0 +1,18 @@ +from http import HTTPStatus +from typing import Optional + +from worker.job_runner import JobRunnerError + + +class StreamingRowsError(JobRunnerError): + """Raised when the rows could not be fetched in streaming mode.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "StreamingRowsError", cause, True) + + +class NormalRowsError(JobRunnerError): + """Raised when the rows could not be fetched in normal mode.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "NormalRowsError", cause, True) diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index a7d7dca1..20d3e7c6 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -125,0 +126,40 @@ class FirstRowsConfig: +OPT_IN_OUT_URLS_SCAN_COLUMNS_MAX_NUMBER = 10 +OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER = 100 +OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND = 50 +OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER = 100_000 +OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN = None +OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH = 1000 +OPT_IN_OUT_URLS_SCAN_SPAWNING_URL = "https://opts-api.spawningaiapi.com/api/v2/query/urls" + + +@dataclass(frozen=True) +class OptInOutUrlsScanConfig: + spawning_url: str = OPT_IN_OUT_URLS_SCAN_SPAWNING_URL + rows_max_number: int = OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER + columns_max_number: int = FIRST_ROWS_COLUMNS_MAX_NUMBER + urls_number_per_batch: int = OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH + spawning_token: Optional[str] = OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN + max_concurrent_requests_number: int = OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER + max_requests_per_second: int = OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND + + @classmethod + def from_env(cls) -> "OptInOutUrlsScanConfig": + env = Env(expand_vars=True) + with env.prefixed("OPT_IN_OUT_URLS_SCAN_"): + return cls( + rows_max_number=env.int(name="ROWS_MAX_NUMBER", default=OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER), + columns_max_number=env.int(name="COLUMNS_MAX_NUMBER", default=OPT_IN_OUT_URLS_SCAN_COLUMNS_MAX_NUMBER), + urls_number_per_batch=env.int( + name="URLS_NUMBER_PER_BATCH", default=OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH + ), + spawning_token=env.str(name="SPAWNING_TOKEN", default=OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN), + max_concurrent_requests_number=env.int( + name="MAX_CONCURRENT_REQUESTS_NUMBER", default=OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER + ), + max_requests_per_second=env.int( + name="MAX_REQUESTS_PER_SECOND", default=OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND + ), + spawning_url=env.str(name="SPAWNING_URL", default=OPT_IN_OUT_URLS_SCAN_SPAWNING_URL), + ) + + @@ -189,0 +230 @@ class AppConfig: + urls_scan: OptInOutUrlsScanConfig = field(default_factory=OptInOutUrlsScanConfig) @@ -204,0 +246 @@ class AppConfig: + urls_scan=OptInOutUrlsScanConfig.from_env(), diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index df32f5f0..33113bca 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -42,0 +43,3 @@ from worker.job_runners.split.first_rows_from_streaming import ( +from worker.job_runners.split.opt_in_out_urls_scan_from_streaming import ( + SplitOptInOutUrlsScanJobRunner, +) @@ -198,0 +202,9 @@ class JobRunnerFactory(BaseJobRunnerFactory): + + if job_type == SplitOptInOutUrlsScanJobRunner.get_job_type(): + return SplitOptInOutUrlsScanJobRunner( + job_info=job_info, + app_config=self.app_config, + processing_step=processing_step, + hf_datasets_cache=self.hf_datasets_cache, + ) + @@ -215,0 +228 @@ class JobRunnerFactory(BaseJobRunnerFactory): + SplitOptInOutUrlsScanJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index ae019a87..c10ba8d2 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -4,2 +3,0 @@ -import functools -import itertools @@ -7,2 +4,0 @@ import logging -import time -import warnings @@ -11,10 +7,3 @@ from pathlib import Path -from typing import Any, Callable, List, Literal, Mapping, Optional, TypeVar, Union, cast - -from datasets import ( - Dataset, - DownloadConfig, - Features, - IterableDataset, - get_dataset_config_info, - load_dataset, -) +from typing import Any, List, Literal, Mapping, Optional, Union + +from datasets import Features, IterableDataset, get_dataset_config_info, load_dataset @@ -40,0 +30 @@ from worker.utils import ( + Row, @@ -43,0 +34 @@ from worker.utils import ( + get_rows_or_raise, @@ -105,14 +95,0 @@ class FeaturesError(SplitFirstRowsFromStreamingJobRunnerError): -class StreamingRowsError(SplitFirstRowsFromStreamingJobRunnerError): - """Raised when the rows could not be fetched in streaming mode.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "StreamingRowsError", cause, True) - - -class NormalRowsError(SplitFirstRowsFromStreamingJobRunnerError): - """Raised when the rows could not be fetched in normal mode.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "NormalRowsError", cause, True) - - @@ -147,63 +123,0 @@ class PreviousStepFormatError(SplitFirstRowsFromStreamingJobRunnerError): -FuncT = TypeVar("FuncT", bound=Callable[..., Any]) - - -def retry(func: FuncT) -> FuncT: - """retries with an increasing sleep before every attempt""" - SLEEPS = [1, 7, 70, 7 * 60, 70 * 60] - MAX_ATTEMPTS = len(SLEEPS) - - @functools.wraps(func) - def decorator(*args: Any, **kwargs: Any) -> Any: - attempt = 0 - last_err = None - while attempt < MAX_ATTEMPTS: - try: - """always sleep before calling the function. It will prevent rate limiting in the first place""" - duration = SLEEPS[attempt] - logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.") - time.sleep(duration) - return func(*args, **kwargs) - except ConnectionError as err: - logging.info("Got a ConnectionError, possibly due to rate limiting. Let's retry.") - last_err = err - attempt += 1 - raise RuntimeError(f"Give up after {attempt} attempts with ConnectionError") from last_err - - return cast(FuncT, decorator) - - -Row = Mapping[str, Any] - - -@retry -def get_rows( - dataset: str, - config: str, - split: str, - streaming: bool, - rows_max_number: int, - use_auth_token: Union[bool, str, None] = False, -) -> List[Row]: - download_config = DownloadConfig(delete_extracted=True) - ds = load_dataset( - dataset, - name=config, - split=split, - streaming=streaming, - use_auth_token=use_auth_token, - download_config=download_config, - ) - if streaming: - if not isinstance(ds, IterableDataset): - raise TypeError("load_dataset should return an IterableDataset in streaming mode") - elif not isinstance(ds, Dataset): - raise TypeError("load_dataset should return a Dataset in normal mode") - rows_plus_one = list(itertools.islice(ds, rows_max_number + 1)) - # ^^ to be able to detect if a split has exactly ROWS_MAX_NUMBER rows - if len(rows_plus_one) <= rows_max_number: - logging.debug(f"all the rows in the split have been fetched ({len(rows_plus_one)})") - else: - logging.debug(f"the rows in the split have been truncated ({rows_max_number} rows)") - return rows_plus_one[:rows_max_number] - - @@ -295,4 +208,0 @@ def compute_first_rows_response( - - [`~job_runners.split.first_rows.StreamingRowsError`] - If the split rows could not be obtained using the datasets library in streaming mode. - - [`~job_runners.split.first_rows.NormalRowsError`] - If the split rows could not be obtained using the datasets library in normal mode. @@ -308,0 +219,4 @@ def compute_first_rows_response( + - [`~job_runners.common_exceptions.StreamingRowsError`] + If the split rows could not be obtained using the datasets library in streaming mode. + - [`~job_runners.common_exceptions.NormalRowsError`] + If the split rows could not be obtained using the datasets library in normal mode. @@ -388,38 +302,10 @@ def compute_first_rows_response( - try: - rows = get_rows( - dataset=dataset, - config=config, - split=split, - streaming=True, - rows_max_number=rows_max_number, - use_auth_token=use_auth_token, - ) - except Exception as err: - MAX_SIZE_FALLBACK = 100_000_000 - if max_size_fallback: - warnings.warn( - ( - f"The parameter 'max_size_fallback' is deprecated. The hard-coded value `{MAX_SIZE_FALLBACK}`" - " will be used instead." - ), - category=DeprecationWarning, - ) - if info.size_in_bytes is None or info.size_in_bytes > MAX_SIZE_FALLBACK: - raise StreamingRowsError( - "Cannot load the dataset split (in streaming mode) to extract the first rows.", - cause=err, - ) from err - try: - rows = get_rows( - dataset=dataset, - config=config, - split=split, - streaming=False, - rows_max_number=rows_max_number, - use_auth_token=use_auth_token, - ) - except Exception as err: - raise NormalRowsError( - "Cannot load the dataset split (in normal download mode) to extract the first rows.", - cause=err, - ) from err + rows = get_rows_or_raise( + dataset=dataset, + config=config, + split=split, + info=info, + max_size_fallback=max_size_fallback, + rows_max_number=rows_max_number, + use_auth_token=use_auth_token, + ) + diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py new file mode 100644 index 00000000..6cd4ad55 --- /dev/null +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py @@ -0,0 +1,364 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from asyncio import Semaphore, create_task, run, wait +from http import HTTPStatus +from pathlib import Path +from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict, Union + +from aiohttp import ClientSession +from aiolimiter import AsyncLimiter +from datasets import get_dataset_config_info +from libcommon.constants import PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import JobInfo +from libcommon.simple_cache import SplitFullName + +from worker.config import AppConfig, OptInOutUrlsScanConfig +from worker.job_runner import ( + CompleteJobResult, + JobRunnerError, + get_previous_step_or_raise, +) +from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.utils import SplitFirstRowsResponse, get_rows_or_raise + +SplitOptInOutUrlsScanJobRunnerErrorCode = Literal[ + "InfoError", + "TooManyColumnsError", + "PreviousStepStatusError", + "PreviousStepFormatError", + "MissingSpawningTokenError", + "ExternalServerError", +] + + +class SplitOptInOutUrlsScanJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: SplitOptInOutUrlsScanJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class InfoError(SplitOptInOutUrlsScanJobRunnerError): + """Raised when the info could not be fetched.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "InfoError", cause, True) + + +class TooManyColumnsError(SplitOptInOutUrlsScanJobRunnerError): + """Raised when the dataset exceeded the max number of columns.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "TooManyColumnsError", cause, True) + + +class PreviousStepStatusError(SplitOptInOutUrlsScanJobRunnerError): + """Raised when the previous step gave an error. The job should not have been created.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) + + +class PreviousStepFormatError(SplitOptInOutUrlsScanJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +class MissingSpawningTokenError(SplitOptInOutUrlsScanJobRunnerError): + """Raised when the spawning.ai token is not set.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "MissingSpawningTokenError", cause, False) + + +class ExternalServerError(SplitOptInOutUrlsScanJobRunnerError): + """Raised when the spawning.ai server is not responding.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ExternalServerError", cause, False) + + +class OptUrl(TypedDict): + url: str + row_idx: int + column_name: str + + +class OptInOutUrlsScanResponse(TypedDict): + urls_columns: List[str] + opt_in_urls: List[OptUrl] + opt_out_urls: List[OptUrl] + num_opt_in_urls: int + num_opt_out_urls: int + num_urls: int + num_scanned_rows: int + has_urls_columns: bool + + +async def check_spawning( + image_urls: List[str], session: ClientSession, semaphore: Semaphore, limiter: AsyncLimiter, spawning_url: str +) -> Any: + if not image_urls: + return {"urls": []} + elif len(image_urls) == 1: + image_urls = image_urls + [""] # the API requires >1 urls + async with semaphore: + async with limiter: + async with session.post(url=spawning_url, data="\n".join(image_urls)) as resp: + spawning_response = await resp.json() + return spawning_response + + +async def opt_in_out_task( + image_urls: List[str], session: ClientSession, semaphore: Semaphore, limiter: AsyncLimiter, spawning_url: str +) -> Tuple[List[Any], List[Any]]: + try: + spawning_response = await check_spawning(image_urls, session, semaphore, limiter, spawning_url) + except Exception: + raise ExternalServerError(message=f"Error when trying to connect to {spawning_url}") + if "urls" not in spawning_response: + raise ExternalServerError(message=f"Error when trying to connect to {spawning_url}: '{spawning_response}'") + opt_in_urls_indices = [i for i in range(len(image_urls)) if spawning_response["urls"][i]["optIn"]] + opt_out_urls_indices = [i for i in range(len(image_urls)) if spawning_response["urls"][i]["optOut"]] + return opt_in_urls_indices, opt_out_urls_indices + + +async def opt_in_out_scan_urls( + urls: List[str], + urls_number_per_batch: int, + spawning_token: str, + max_concurrent_requests_number: int, + max_requests_per_second: int, + spawning_url: str, +) -> Tuple[List[int], List[int]]: + offsets = [] + tasks = [] + semaphore = Semaphore(value=max_concurrent_requests_number) + limiter = AsyncLimiter(max_requests_per_second, time_period=1) + + headers = {"Authorization": f"API {spawning_token}"} + async with ClientSession(headers=headers) as session: + for offset in range(0, len(urls), urls_number_per_batch): + offsets.append(offset) + limit = offset + urls_number_per_batch + tasks.append( + create_task(opt_in_out_task(urls[offset:limit], session, semaphore, limiter, spawning_url)) + ) # noqa: E203 + await wait(tasks) + + opt_in_urls_indices = [] + opt_out_urls_indices = [] + for offset, task in zip(offsets, tasks): + batch_opt_in_urls_indices, batch_opt_out_urls_indices = task.result() + for batch_opt_in_urls_idx in batch_opt_in_urls_indices: + opt_in_urls_indices.append(offset + batch_opt_in_urls_idx) + for batch_opt_out_urls_idx in batch_opt_out_urls_indices: + opt_out_urls_indices.append(offset + batch_opt_out_urls_idx) + + return opt_in_urls_indices, opt_out_urls_indices + + +def compute_opt_in_out_urls_scan_response( + dataset: str, + config: str, + split: str, + hf_token: Optional[str], + rows_max_number: int, + columns_max_number: int, + urls_number_per_batch: int, + spawning_token: Optional[str], + max_concurrent_requests_number: int, + max_requests_per_second: int, + spawning_url: str, +) -> OptInOutUrlsScanResponse: + logging.info(f"get opt-in-out-urls-scan for dataset={dataset} config={config} split={split}") + + use_auth_token: Union[bool, str, None] = hf_token if hf_token is not None else False + if not spawning_token: + raise MissingSpawningTokenError("OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN is not set") + + # get the first rows from previous job + upstream_response = get_previous_step_or_raise( + kinds=["split-first-rows-from-streaming"], dataset=dataset, config=config, split=split + ) + try: + first_rows_response = upstream_response.response + upstream_response_content = SplitFirstRowsResponse( + dataset=dataset, + config=config, + split=split, + features=first_rows_response["content"]["features"], + rows=first_rows_response["content"]["rows"], + ) + + features = upstream_response_content["features"] + first_rows = upstream_response_content["rows"] + except KeyError as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + # get the info + try: + info = get_dataset_config_info( + path=dataset, + config_name=config, + use_auth_token=use_auth_token, + ) + except Exception as err: + raise InfoError( + f"The info cannot be fetched for the config '{config}' of the dataset.", + cause=err, + ) from err + + # look for URLs columns using the first rows + string_type_dict = {"dtype": "string", "_type": "Value"} + string_columns = [feature["name"] for feature in features if feature["type"] == string_type_dict] + urls_columns = [] + for string_column in string_columns: + urls_count = sum( + 1 + for row in first_rows + if isinstance(row["row"].get(string_column), str) + and (row["row"][string_column].startswith("https://") or row["row"][string_column].startswith("http://")) + ) + if urls_count and urls_count / len(first_rows) > 0.5: + urls_columns.append(string_column) + + if not urls_columns: + return OptInOutUrlsScanResponse( + urls_columns=[], + opt_in_urls=[], + opt_out_urls=[], + num_opt_in_urls=0, + num_opt_out_urls=0, + num_urls=0, + num_scanned_rows=0, + has_urls_columns=False, + ) + + if len(urls_columns) > columns_max_number: + raise TooManyColumnsError( + f"The number of columns ({len(urls_columns)}) exceeds the maximum supported number of columns to scan" + f" ({columns_max_number})." + ) + + # get the rows + rows = get_rows_or_raise( + dataset=dataset, + config=config, + split=split, + info=info, + rows_max_number=rows_max_number, + use_auth_token=use_auth_token, + column_names=urls_columns, + ) + + # get the urls + num_scanned_rows = len(rows) + urls = [row[urls_column] for row in rows for urls_column in urls_columns] + + # scan the urls + opt_in_urls_indices, opt_out_urls_indices = run( + opt_in_out_scan_urls( + urls, + urls_number_per_batch=urls_number_per_batch, + spawning_token=spawning_token, + max_concurrent_requests_number=max_concurrent_requests_number, + max_requests_per_second=max_requests_per_second, + spawning_url=spawning_url, + ) + ) + + opt_in_urls = [ + OptUrl( + url=urls[url_idx], + row_idx=url_idx // len(urls_columns), + column_name=urls_columns[url_idx % len(urls_columns)], + ) + for url_idx in opt_in_urls_indices + ] + opt_out_urls = [ + OptUrl( + url=urls[url_idx], + row_idx=url_idx // len(urls_columns), + column_name=urls_columns[url_idx % len(urls_columns)], + ) + for url_idx in opt_out_urls_indices + ] + + # return scan result + return OptInOutUrlsScanResponse( + urls_columns=urls_columns, + opt_in_urls=opt_in_urls, + opt_out_urls=opt_out_urls, + num_opt_in_urls=len(opt_in_urls), + num_opt_out_urls=len(opt_out_urls), + num_urls=len(urls), + num_scanned_rows=num_scanned_rows, + has_urls_columns=True, + ) + + +class SplitOptInOutUrlsScanJobRunner(DatasetsBasedJobRunner): + urls_scan_config: OptInOutUrlsScanConfig + + @staticmethod + def get_job_type() -> str: + return "split-opt-in-out-urls-scan" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION + + def __init__( + self, + job_info: JobInfo, + app_config: AppConfig, + processing_step: ProcessingStep, + hf_datasets_cache: Path, + ) -> None: + super().__init__( + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + hf_datasets_cache=hf_datasets_cache, + ) + self.urls_scan_config = app_config.urls_scan + + def compute(self) -> CompleteJobResult: + if self.config is None or self.split is None: + raise ValueError("config and split are required") + return CompleteJobResult( + compute_opt_in_out_urls_scan_response( + dataset=self.dataset, + config=self.config, + split=self.split, + hf_token=self.common_config.hf_token, + rows_max_number=self.urls_scan_config.rows_max_number, + columns_max_number=self.urls_scan_config.columns_max_number, + urls_number_per_batch=self.urls_scan_config.urls_number_per_batch, + spawning_token=self.urls_scan_config.spawning_token, + max_concurrent_requests_number=self.urls_scan_config.max_concurrent_requests_number, + max_requests_per_second=self.urls_scan_config.max_requests_per_second, + spawning_url=self.urls_scan_config.spawning_url, + ) + ) + + def get_new_splits(self, _: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by compute.""" + if self.config is None or self.split is None: + raise ValueError("config and split are required") + return {SplitFullName(dataset=self.dataset, config=self.config, split=self.split)} diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index a706ce3a..85966d05 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -4,3 +4,25 @@ -from typing import Any, List, Mapping, Optional, TypedDict - -from datasets import Features +import functools +import itertools +import logging +import time +import warnings +from typing import ( + Any, + Callable, + List, + Mapping, + Optional, + TypedDict, + TypeVar, + Union, + cast, +) + +from datasets import ( + Dataset, + DatasetInfo, + DownloadConfig, + Features, + IterableDataset, + load_dataset, +) @@ -8,0 +31,2 @@ from libcommon.utils import orjson_dumps +from worker.common_exceptions import NormalRowsError, StreamingRowsError + @@ -206,0 +231,114 @@ def create_truncated_row_items( + + +FuncT = TypeVar("FuncT", bound=Callable[..., Any]) + + +def retry(func: FuncT) -> FuncT: + """retries with an increasing sleep before every attempt""" + SLEEPS = [1, 7, 70, 7 * 60, 70 * 60] + MAX_ATTEMPTS = len(SLEEPS) + + @functools.wraps(func) + def decorator(*args: Any, **kwargs: Any) -> Any: + attempt = 0 + last_err = None + while attempt < MAX_ATTEMPTS: + try: + """always sleep before calling the function. It will prevent rate limiting in the first place""" + duration = SLEEPS[attempt] + logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.") + time.sleep(duration) + return func(*args, **kwargs) + except ConnectionError as err: + logging.info("Got a ConnectionError, possibly due to rate limiting. Let's retry.") + last_err = err + attempt += 1 + raise RuntimeError(f"Give up after {attempt} attempts with ConnectionError") from last_err + + return cast(FuncT, decorator) + + +@retry +def get_rows( + dataset: str, + config: str, + split: str, + streaming: bool, + rows_max_number: int, + use_auth_token: Union[bool, str, None] = False, + column_names: Optional[List[str]] = None, +) -> List[Row]: + download_config = DownloadConfig(delete_extracted=True) + ds = load_dataset( + dataset, + name=config, + split=split, + streaming=streaming, + use_auth_token=use_auth_token, + download_config=download_config, + ) + if streaming: + if not isinstance(ds, IterableDataset): + raise TypeError("load_dataset should return an IterableDataset in streaming mode") + elif not isinstance(ds, Dataset): + raise TypeError("load_dataset should return a Dataset in normal mode") + if column_names: + ds = ds.select_columns(column_names) + rows_plus_one = list(itertools.islice(ds, rows_max_number + 1)) + # ^^ to be able to detect if a split has exactly ROWS_MAX_NUMBER rows + if len(rows_plus_one) <= rows_max_number: + logging.debug(f"all the rows in the split have been fetched ({len(rows_plus_one)})") + else: + logging.debug(f"the rows in the split have been truncated ({rows_max_number} rows)") + return rows_plus_one[:rows_max_number] + + +def get_rows_or_raise( + dataset: str, + config: str, + split: str, + rows_max_number: int, + use_auth_token: Union[bool, str, None], + info: DatasetInfo, + max_size_fallback: Optional[int] = None, + column_names: Optional[List[str]] = [], +) -> List[Row]: + try: + return get_rows( + dataset=dataset, + config=config, + split=split, + streaming=True, + rows_max_number=rows_max_number, + use_auth_token=use_auth_token, + column_names=column_names, + ) + except Exception as err: + MAX_SIZE_FALLBACK = 100_000_000 + if max_size_fallback: + warnings.warn( + ( + f"The parameter 'max_size_fallback' is deprecated. The hard-coded value `{MAX_SIZE_FALLBACK}`" + " will be used instead." + ), + category=DeprecationWarning, + ) + if info.size_in_bytes is None or info.size_in_bytes > MAX_SIZE_FALLBACK: + raise StreamingRowsError( + "Cannot load the dataset split (in streaming mode) to extract the first rows.", + cause=err, + ) from err + try: + return get_rows( + dataset=dataset, + config=config, + split=split, + streaming=False, + rows_max_number=rows_max_number, + use_auth_token=use_auth_token, + ) + except Exception as err: + raise NormalRowsError( + "Cannot load the dataset split (in normal download mode) to extract the first rows.", + cause=err, + ) from err diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 9b06592f..0148daf3 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -69,0 +70 @@ def set_env_vars( + mp.setenv("OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN", "dummy_spawning_token") diff --git a/services/worker/tests/constants.py b/services/worker/tests/constants.py index 66dbb47a..ebacd2d6 100644 --- a/services/worker/tests/constants.py +++ b/services/worker/tests/constants.py @@ -2,0 +3 @@ +import os @@ -9,0 +11 @@ CI_USER_TOKEN = "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" +CI_SPAWNING_TOKEN = os.getenv("CI_SPAWNING_TOKEN", "unset") diff --git a/services/worker/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py index ea07ce7f..88f47df7 100644 --- a/services/worker/tests/fixtures/datasets.py +++ b/services/worker/tests/fixtures/datasets.py @@ -132,0 +133,13 @@ def datasets() -> Mapping[str, Dataset]: + "spawning_opt_in_out": Dataset.from_pandas( + pd.DataFrame( + { + "col": [ + "http://testurl.test/test_image-optOut.jpg", + "http://testurl.test/test_image2.jpg", + "other", + "http://testurl.test/test_image3-optIn.jpg", + ] + }, + dtype=pd.StringDtype(storage="python"), + ) + ), diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 3fa69d89..fc3d2e1a 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -253,0 +254,7 @@ def hub_public_legacy_configs(dataset_script_with_two_configs_path: str) -> Iter [email protected](scope="session") +def hub_public_spawning_opt_in_out(datasets: Mapping[str, Dataset]) -> Iterator[str]: + repo_id = create_hub_dataset_repo(prefix="spawning_opt_in_out", dataset=datasets["spawning_opt_in_out"]) + yield repo_id + delete_hub_dataset_repo(repo_id=repo_id) + + @@ -532,0 +540,7 @@ TEXT_rows = [ +SPAWNING_OPT_IN_OUT_cols = { + "col": [{"_type": "Value", "dtype": "string"}], +} + +SPAWNING_OPT_IN_OUT_rows = ["http://testurl.test/test_image.jpg", "http://testurl.test/test_image2.jpg", "other"] + + @@ -545,0 +560 @@ def hub_datasets( + hub_public_spawning_opt_in_out: str, @@ -677,0 +693,10 @@ def hub_datasets( + "spawning_opt_in_out": { + "name": hub_public_spawning_opt_in_out, + "config_names_response": create_config_names_response(hub_public_spawning_opt_in_out), + "splits_response": create_splits_response(hub_public_spawning_opt_in_out), + "first_rows_response": create_first_rows_response( + hub_public_spawning_opt_in_out, SPAWNING_OPT_IN_OUT_cols, SPAWNING_OPT_IN_OUT_rows + ), + "parquet_and_dataset_info_response": None, + "parquet_and_info_response": None, + }, diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py new file mode 100644 index 00000000..706f456f --- /dev/null +++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py @@ -0,0 +1,314 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from asyncio import Semaphore +from dataclasses import replace +from http import HTTPStatus +from typing import Any, Callable, List, Mapping +from unittest.mock import patch + +import pytest +from aiohttp import ClientSession +from aiolimiter import AsyncLimiter +from libcommon.constants import PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION +from libcommon.exceptions import CustomError +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import get_response, upsert_response + +from worker.config import AppConfig +from worker.job_runners.split.opt_in_out_urls_scan_from_streaming import ( + ExternalServerError, + SplitOptInOutUrlsScanJobRunner, + check_spawning, +) +from worker.resources import LibrariesResource + +from ...constants import CI_SPAWNING_TOKEN +from ...fixtures.hub import HubDatasets, get_default_config_split + +GetJobRunner = Callable[[str, str, str, AppConfig, bool], SplitOptInOutUrlsScanJobRunner] + + +async def mock_check_spawning( + image_urls: List[str], session: ClientSession, semaphore: Semaphore, limiter: AsyncLimiter, url: str +) -> Any: + return {"urls": [{"url": url, "optIn": "optIn" in url, "optOut": "optOut" in url} for url in image_urls]} + + [email protected] +def get_job_runner( + libraries_resource: LibrariesResource, + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + split: str, + app_config: AppConfig, + force: bool = False, + ) -> SplitOptInOutUrlsScanJobRunner: + return SplitOptInOutUrlsScanJobRunner( + job_info={ + "type": SplitOptInOutUrlsScanJobRunner.get_job_type(), + "dataset": dataset, + "config": config, + "split": split, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + app_config=app_config, + processing_step=ProcessingStep( + name=SplitOptInOutUrlsScanJobRunner.get_job_type(), + input_type="split", + requires=[], + required_by_dataset_viewer=True, + ancestors=[], + children=[], + parents=[], + job_runner_version=SplitOptInOutUrlsScanJobRunner.get_job_runner_version(), + ), + hf_datasets_cache=libraries_resource.hf_datasets_cache, + ) + + return _get_job_runner + + +FIRST_ROWS_WITHOUT_OPT_IN_OUT_URLS = { + "features": [ + { + "feature_idx": 0, + "name": "col1", + "type": { + "dtype": "int64", + "_type": "Value", + }, + }, + { + "feature_idx": 1, + "name": "col2", + "type": { + "dtype": "int64", + "_type": "Value", + }, + }, + { + "feature_idx": 2, + "name": "col3", + "type": { + "dtype": "float64", + "_type": "Value", + }, + }, + ], + "rows": [], +} + + +FIRST_ROWS_WITH_OPT_IN_OUT_URLS = { + "features": [ + { + "feature_idx": 0, + "name": "col", + "type": { + "dtype": "string", + "_type": "Value", + }, + } + ], + "rows": [ + {"row_idx": 0, "row": {"col": "http://testurl.test/test_image.jpg"}, "truncated_cells": []}, + {"row_idx": 1, "row": {"col": "http://testurl.test/test_image2.jpg"}, "truncated_cells": []}, + {"row_idx": 2, "row": {"col": "other"}, "truncated_cells": []}, + ], +} + + [email protected]( + "name,upstream_content,expected_content", + [ + ( + "public", + FIRST_ROWS_WITHOUT_OPT_IN_OUT_URLS, + { + "has_urls_columns": False, + "num_scanned_rows": 0, + "opt_in_urls": [], + "opt_out_urls": [], + "urls_columns": [], + "num_opt_out_urls": 0, + "num_opt_in_urls": 0, + "num_urls": 0, + }, + ), + ( + "spawning_opt_in_out", + FIRST_ROWS_WITH_OPT_IN_OUT_URLS, + { + "has_urls_columns": True, + "num_scanned_rows": 4, + "opt_in_urls": [ + {"url": "http://testurl.test/test_image3-optIn.jpg", "row_idx": 3, "column_name": "col"} + ], + "opt_out_urls": [ + {"url": "http://testurl.test/test_image-optOut.jpg", "row_idx": 0, "column_name": "col"} + ], + "urls_columns": ["col"], + "num_opt_out_urls": 1, + "num_opt_in_urls": 1, + "num_urls": 4, + }, + ), + ], +) +def test_compute( + hub_datasets: HubDatasets, + app_config: AppConfig, + get_job_runner: GetJobRunner, + name: str, + upstream_content: Mapping[str, Any], + expected_content: Mapping[str, Any], +) -> None: + dataset, config, split = get_default_config_split(hub_datasets[name]["name"]) + job_runner = get_job_runner(dataset, config, split, app_config, False) + upsert_response( + kind="split-first-rows-from-streaming", + dataset=dataset, + config=config, + split=split, + content=upstream_content, + dataset_git_revision="dataset_git_revision", + job_runner_version=PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION, + progress=1.0, + http_status=HTTPStatus.OK, + ) + with patch("worker.job_runners.split.opt_in_out_urls_scan_from_streaming.check_spawning", mock_check_spawning): + assert job_runner.process() + cached_response = get_response( + kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config, split=split + ) + assert cached_response + assert cached_response["content"] == expected_content + assert cached_response["http_status"] == HTTPStatus.OK + assert cached_response["error_code"] is None + + [email protected]( + "dataset,columns_max_number,upstream_content,upstream_status,error_code,status_code", + [ + ("doesnotexist", 10, {}, HTTPStatus.OK, "CachedResponseNotFound", HTTPStatus.NOT_FOUND), + ("wrong_format", 10, {}, HTTPStatus.OK, "PreviousStepFormatError", HTTPStatus.INTERNAL_SERVER_ERROR), + ( + "upstream_failed", + 10, + {}, + HTTPStatus.INTERNAL_SERVER_ERROR, + "PreviousStepError", + HTTPStatus.INTERNAL_SERVER_ERROR, + ), + ( + "info_error", + 10, + FIRST_ROWS_WITHOUT_OPT_IN_OUT_URLS, + HTTPStatus.OK, + "InfoError", + HTTPStatus.INTERNAL_SERVER_ERROR, + ), + ( + "too_many_columns", + 0, + FIRST_ROWS_WITH_OPT_IN_OUT_URLS, + HTTPStatus.OK, + "TooManyColumnsError", + HTTPStatus.INTERNAL_SERVER_ERROR, + ), + ], +) +def test_compute_failed( + app_config: AppConfig, + hub_datasets: HubDatasets, + get_job_runner: GetJobRunner, + dataset: str, + columns_max_number: int, + upstream_content: Mapping[str, Any], + upstream_status: HTTPStatus, + error_code: str, + status_code: HTTPStatus, +) -> None: + if dataset == "too_many_columns": + dataset = hub_datasets["spawning_opt_in_out"]["name"] + dataset, config, split = get_default_config_split(dataset) + job_runner = get_job_runner( + dataset, + config, + split, + replace(app_config, urls_scan=replace(app_config.urls_scan, columns_max_number=columns_max_number)), + False, + ) + if dataset != "doesnotexist": + upsert_response( + kind="split-first-rows-from-streaming", + dataset=dataset, + config=config, + split=split, + content=upstream_content, + dataset_git_revision="dataset_git_revision", + job_runner_version=PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION, + progress=1.0, + http_status=upstream_status, + ) + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.status_code == status_code + assert exc_info.value.code == error_code + + +def test_compute_error_from_spawning( + app_config: AppConfig, + get_job_runner: GetJobRunner, + hub_public_spawning_opt_in_out: str, +) -> None: + dataset, config, split = get_default_config_split(hub_public_spawning_opt_in_out) + job_runner = get_job_runner( + dataset, + config, + split, + replace(app_config, urls_scan=replace(app_config.urls_scan, spawning_url="wrong_url")), + False, + ) + upsert_response( + kind="split-first-rows-from-streaming", + dataset=dataset, + config=config, + split=split, + content=FIRST_ROWS_WITH_OPT_IN_OUT_URLS, + dataset_git_revision="dataset_git_revision", + job_runner_version=PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION, + progress=1.0, + http_status=HTTPStatus.OK, + ) + with pytest.raises(ExternalServerError): + job_runner.compute() + + [email protected] +async def test_real_check_spawning_response(app_config: AppConfig) -> None: + semaphore = Semaphore(value=10) + limiter = AsyncLimiter(10, time_period=1) + + headers = {"Authorization": f"API {CI_SPAWNING_TOKEN}"} + async with ClientSession(headers=headers) as session: + image_url = "http://testurl.test/test_image.jpg" + image_urls = [image_url] + spawning_url = app_config.urls_scan.spawning_url + spawning_response = await check_spawning(image_urls, session, semaphore, limiter, spawning_url) + assert spawning_response and isinstance(spawning_response, dict) + assert spawning_response["urls"] and isinstance(spawning_response["urls"], list) + assert len(spawning_response["urls"]) == 2 # the API requires >1 urls + first_url = spawning_response["urls"][0] + assert first_url and isinstance(first_url, dict) + assert first_url["url"] and isinstance(first_url["url"], str) + assert first_url["url"] == image_url diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml index 485c15eb..4b10f16c 100644 --- a/tools/docker-compose-datasets-server.yml +++ b/tools/docker-compose-datasets-server.yml @@ -110,0 +111,7 @@ services: + OPT_IN_OUT_URLS_SCAN_COLUMNS_MAX_NUMBER: ${OPT_IN_OUT_URLS_SCAN_COLUMNS_MAX_NUMBER-10} + OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER: ${OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER-100} + OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND: ${OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND-50} + OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER: ${OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER-100_000} + OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN: ${OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN-} + OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH: ${OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH-1000} + OPT_IN_OUT_URLS_SCAN_SPAWNING_URL: ${OPT_IN_OUT_URLS_SCAN_SPAWNING_URL-https://opts-api.spawningaiapi.com/api/v2/query/urls} diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index eacbf287..81806954 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -114,0 +115,7 @@ services: + OPT_IN_OUT_URLS_SCAN_COLUMNS_MAX_NUMBER: ${OPT_IN_OUT_URLS_SCAN_COLUMNS_MAX_NUMBER-10} + OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER: ${OPT_IN_OUT_URLS_SCAN_MAX_CONCURRENT_REQUESTS_NUMBER-100} + OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND: ${OPT_IN_OUT_URLS_SCAN_MAX_REQUESTS_PER_SECOND-50} + OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER: ${OPT_IN_OUT_URLS_SCAN_ROWS_MAX_NUMBER-100_000} + OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN: ${OPT_IN_OUT_URLS_SCAN_SPAWNING_TOKEN-} + OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH: ${OPT_IN_OUT_URLS_SCAN_URLS_NUMBER_PER_BATCH-1000} + OPT_IN_OUT_URLS_SCAN_SPAWNING_URL: ${OPT_IN_OUT_URLS_SCAN_SPAWNING_URL-https://opts-api.spawningaiapi.com/api/v2/query/urls}
b5ac81280310aa43f568cf4e79d532b1f5b6963b
Sylvain Lesage
2023-04-21T12:25:29
refactor: 💡 create the configs at startup (#1075)
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index 716b8476..a7d7dca1 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -182,0 +183 @@ class AppConfig: + first_rows: FirstRowsConfig = field(default_factory=FirstRowsConfig) @@ -184,0 +186 @@ class AppConfig: + parquet_and_info: ParquetAndInfoConfig = field(default_factory=ParquetAndInfoConfig) @@ -195,0 +198 @@ class AppConfig: + first_rows=FirstRowsConfig.from_env(), @@ -197,0 +201 @@ class AppConfig: + parquet_and_info=ParquetAndInfoConfig.from_env(), diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 0a3ea836..df32f5f0 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -12 +12 @@ from libcommon.storage import StrPath -from worker.config import AppConfig, FirstRowsConfig, ParquetAndInfoConfig +from worker.config import AppConfig @@ -93 +92,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - first_rows_config = FirstRowsConfig.from_env() @@ -99 +97,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - first_rows_config=first_rows_config, @@ -108 +105,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - parquet_and_dataset_info_config=ParquetAndInfoConfig.from_env(), @@ -116 +112,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - parquet_and_info_config=ParquetAndInfoConfig.from_env(), @@ -189 +184,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - first_rows_config = FirstRowsConfig.from_env() @@ -195 +189,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - first_rows_config=first_rows_config, diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 68ca603f..28e879ac 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -923 +922,0 @@ class ConfigParquetAndInfoJobRunner(DatasetsBasedJobRunner): - parquet_and_info_config: ParquetAndInfoConfig, @@ -931 +930 @@ class ConfigParquetAndInfoJobRunner(DatasetsBasedJobRunner): - self.parquet_and_info_config = parquet_and_info_config + self.parquet_and_info_config = app_config.parquet_and_info diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py index 6702deff..7a02018b 100644 --- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py +++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py @@ -862 +861,0 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner): - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -870 +869 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner): - self.parquet_and_dataset_info_config = parquet_and_dataset_info_config + self.parquet_and_dataset_info_config = app_config.parquet_and_info diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index 1c87b574..7ecf9185 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -300 +299,0 @@ class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): - first_rows_config: FirstRowsConfig, @@ -310 +309 @@ class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): - self.first_rows_config = first_rows_config + self.first_rows_config = app_config.first_rows diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 7f370f84..ae019a87 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -475 +474,0 @@ class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): - first_rows_config: FirstRowsConfig, @@ -485 +484 @@ class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): - self.first_rows_config = first_rows_config + self.first_rows_config = app_config.first_rows diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index b56e4e0f..9b06592f 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -14 +14 @@ from pytest import MonkeyPatch, fixture -from worker.config import AppConfig, FirstRowsConfig +from worker.config import AppConfig @@ -106,5 +105,0 @@ def libraries_resource(app_config: AppConfig) -> Iterator[LibrariesResource]: -@fixture -def first_rows_config(set_env_vars: MonkeyPatch) -> FirstRowsConfig: - return FirstRowsConfig.from_env() - - diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 35cb6f29..d13c2c91 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -20 +20 @@ from libcommon.simple_cache import DoesNotExist, get_response, upsert_response -from worker.config import AppConfig, ParquetAndInfoConfig +from worker.config import AppConfig @@ -58,8 +58 @@ def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyP [email protected] -def parquet_and_info_config( - set_env_vars: pytest.MonkeyPatch, set_supported_datasets: pytest.MonkeyPatch -) -> ParquetAndInfoConfig: - return ParquetAndInfoConfig.from_env() - - -GetJobRunner = Callable[[str, str, AppConfig, ParquetAndInfoConfig, bool], ConfigParquetAndInfoJobRunner] +GetJobRunner = Callable[[str, str, AppConfig, bool], ConfigParquetAndInfoJobRunner] @@ -78 +70,0 @@ def get_job_runner( - parquet_and_info_config: ParquetAndInfoConfig, @@ -103 +94,0 @@ def get_job_runner( - parquet_and_info_config=parquet_and_info_config, @@ -129 +119,0 @@ def test_compute( - parquet_and_info_config: ParquetAndInfoConfig, @@ -140 +130 @@ def test_compute( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -155 +144,0 @@ def test_compute_legacy_configs( - parquet_and_info_config: ParquetAndInfoConfig, @@ -173 +162 @@ def test_compute_legacy_configs( - job_runner = get_job_runner(dataset_name, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset_name, config, app_config, False) @@ -177 +166 @@ def test_compute_legacy_configs( - repo_id=hub_public_legacy_configs, revision=parquet_and_info_config.target_revision, files_metadata=False + repo_id=hub_public_legacy_configs, revision=app_config.parquet_and_info.target_revision, files_metadata=False @@ -202 +191 @@ def test_compute_legacy_configs( - job_runner = get_job_runner(dataset_name, "first", app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset_name, "first", app_config, False) @@ -205 +194 @@ def test_compute_legacy_configs( - repo_id=hub_public_legacy_configs, revision=parquet_and_info_config.target_revision, files_metadata=False + repo_id=hub_public_legacy_configs, revision=app_config.parquet_and_info.target_revision, files_metadata=False @@ -219,3 +208 @@ def test_compute_legacy_configs( -def test_doesnotexist( - app_config: AppConfig, get_job_runner: GetJobRunner, parquet_and_info_config: ParquetAndInfoConfig -) -> None: +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: @@ -223 +210 @@ def test_doesnotexist( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -255 +241,0 @@ def test_raise_if_too_big_from_hub( - parquet_and_info_config: ParquetAndInfoConfig, @@ -267 +253 @@ def test_raise_if_too_big_from_hub( - dataset_info=dataset_info, max_dataset_size=parquet_and_info_config.max_dataset_size + dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size @@ -270 +256,3 @@ def test_raise_if_too_big_from_hub( - raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=parquet_and_info_config.max_dataset_size) + raise_if_too_big_from_hub( + dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size + ) @@ -282 +269,0 @@ def test_raise_if_too_big_from_datasets( - parquet_and_info_config: ParquetAndInfoConfig, @@ -294 +281 @@ def test_raise_if_too_big_from_datasets( - max_dataset_size=parquet_and_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -303 +290 @@ def test_raise_if_too_big_from_datasets( - max_dataset_size=parquet_and_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -320 +306,0 @@ def test_raise_if_too_big_external_files( - parquet_and_info_config: ParquetAndInfoConfig, @@ -322,2 +308,2 @@ def test_raise_if_too_big_external_files( - max_dataset_size = max_dataset_size or parquet_and_info_config.max_dataset_size - max_external_data_files = max_external_data_files or parquet_and_info_config.max_external_data_files + max_dataset_size = max_dataset_size or app_config.parquet_and_info.max_dataset_size + max_external_data_files = max_external_data_files or app_config.parquet_and_info.max_external_data_files @@ -354 +339,0 @@ def test_raise_if_too_many_external_files( - parquet_and_info_config: ParquetAndInfoConfig, @@ -356,2 +341,2 @@ def test_raise_if_too_many_external_files( - max_dataset_size = max_dataset_size or parquet_and_info_config.max_dataset_size - max_external_data_files = max_external_data_files or parquet_and_info_config.max_external_data_files + max_dataset_size = max_dataset_size or app_config.parquet_and_info.max_dataset_size + max_external_data_files = max_external_data_files or app_config.parquet_and_info.max_external_data_files @@ -385 +369,0 @@ def test_raise_if_not_supported( - parquet_and_info_config: ParquetAndInfoConfig, @@ -398 +382 @@ def test_raise_if_not_supported( - committer_hf_token=parquet_and_info_config.committer_hf_token, + committer_hf_token=app_config.parquet_and_info.committer_hf_token, @@ -400 +384 @@ def test_raise_if_not_supported( - max_dataset_size=parquet_and_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -410 +394 @@ def test_raise_if_not_supported( - committer_hf_token=parquet_and_info_config.committer_hf_token, + committer_hf_token=app_config.parquet_and_info.committer_hf_token, @@ -412 +396 @@ def test_raise_if_not_supported( - max_dataset_size=parquet_and_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -421 +404,0 @@ def test_not_supported_if_big( - parquet_and_info_config: ParquetAndInfoConfig, @@ -434 +417 @@ def test_not_supported_if_big( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -444 +426,0 @@ def test_supported_if_gated( - parquet_and_info_config: ParquetAndInfoConfig, @@ -456 +438 @@ def test_supported_if_gated( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -466 +447,0 @@ def test_not_supported_if_gated_with_extra_fields( - parquet_and_info_config: ParquetAndInfoConfig, @@ -478 +459 @@ def test_not_supported_if_gated_with_extra_fields( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -488 +468,0 @@ def test_blocked( - parquet_and_info_config: ParquetAndInfoConfig, @@ -500 +480 @@ def test_blocked( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -516 +495,0 @@ def test_compute_splits_response_simple_csv_ok( - parquet_and_info_config: ParquetAndInfoConfig, @@ -528 +507 @@ def test_compute_splits_response_simple_csv_ok( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -566 +544,0 @@ def test_compute_splits_response_simple_csv_error( - parquet_and_info_config: ParquetAndInfoConfig, @@ -571 +549 @@ def test_compute_splits_response_simple_csv_error( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) @@ -602 +579,0 @@ def test_previous_step_error( - parquet_and_info_config: ParquetAndInfoConfig, @@ -606 +583 @@ def test_previous_step_error( - job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + job_runner = get_job_runner(dataset, config, app_config, False) diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index 1b848140..5fd059fc 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -20 +20 @@ from pyarrow.fs import LocalFileSystem -from worker.config import AppConfig, FirstRowsConfig +from worker.config import AppConfig @@ -29 +29 @@ from ...fixtures.hub import get_default_config_split -GetJobRunner = Callable[[str, str, str, AppConfig, FirstRowsConfig, bool], SplitFirstRowsFromParquetJobRunner] +GetJobRunner = Callable[[str, str, str, AppConfig, bool], SplitFirstRowsFromParquetJobRunner] @@ -44 +43,0 @@ def get_job_runner( - first_rows_config: FirstRowsConfig, @@ -69 +67,0 @@ def get_job_runner( - first_rows_config=first_rows_config, @@ -76 +74 @@ def get_job_runner( -def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig) -> None: +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: @@ -79 +77 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first - job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False) + job_runner = get_job_runner(dataset, config, split, app_config, False) @@ -96 +93,0 @@ def test_compute( - first_rows_config: FirstRowsConfig, @@ -131 +127,0 @@ def test_compute( - replace(app_config, common=replace(app_config.common, hf_token=None)), @@ -133,6 +129,10 @@ def test_compute( - first_rows_config, - max_number=1_000_000, - min_number=10, - max_bytes=rows_max_bytes, - min_cell_bytes=10, - columns_max_number=columns_max_number, + app_config, + common=replace(app_config.common, hf_token=None), + first_rows=replace( + app_config.first_rows, + max_number=1_000_000, + min_number=10, + max_bytes=rows_max_bytes, + min_cell_bytes=10, + columns_max_number=columns_max_number, + ), @@ -186 +185,0 @@ def test_response_already_computed( - first_rows_config: FirstRowsConfig, @@ -213 +211,0 @@ def test_response_already_computed( - first_rows_config, diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index 93496c33..46b69729 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -19 +19 @@ from libcommon.storage import StrPath -from worker.config import AppConfig, FirstRowsConfig +from worker.config import AppConfig @@ -28 +28 @@ from ...fixtures.hub import HubDatasets, get_default_config_split -GetJobRunner = Callable[[str, str, str, AppConfig, FirstRowsConfig, bool], SplitFirstRowsFromStreamingJobRunner] +GetJobRunner = Callable[[str, str, str, AppConfig, bool], SplitFirstRowsFromStreamingJobRunner] @@ -43 +42,0 @@ def get_job_runner( - first_rows_config: FirstRowsConfig, @@ -68 +66,0 @@ def get_job_runner( - first_rows_config=first_rows_config, @@ -75,3 +73 @@ def get_job_runner( -def test_should_skip_job( - app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig, hub_public_csv: str -) -> None: +def test_should_skip_job(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public_csv: str) -> None: @@ -79 +75 @@ def test_should_skip_job( - job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False) + job_runner = get_job_runner(dataset, config, split, app_config, False) @@ -91 +87 @@ def test_should_skip_job( - job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, True) + job_runner = get_job_runner(dataset, config, split, app_config, True) @@ -95,3 +91 @@ def test_should_skip_job( -def test_compute( - app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig, hub_public_csv: str -) -> None: +def test_compute(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public_csv: str) -> None: @@ -99 +93 @@ def test_compute( - job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False) + job_runner = get_job_runner(dataset, config, split, app_config, False) @@ -124 +118 @@ def test_compute( -def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig) -> None: +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: @@ -127 +121 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first - job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False) + job_runner = get_job_runner(dataset, config, split, app_config, False) @@ -159 +152,0 @@ def test_number_rows( - first_rows_config: FirstRowsConfig, @@ -175 +167,0 @@ def test_number_rows( - first_rows_config, @@ -239 +230,0 @@ def test_truncation( - first_rows_config: FirstRowsConfig, @@ -250 +240,0 @@ def test_truncation( - replace(app_config, common=replace(app_config.common, hf_token=None)), @@ -252,6 +242,10 @@ def test_truncation( - first_rows_config, - max_number=1_000_000, - min_number=10, - max_bytes=rows_max_bytes, - min_cell_bytes=10, - columns_max_number=columns_max_number, + app_config, + common=replace(app_config.common, hf_token=None), + first_rows=replace( + app_config.first_rows, + max_number=1_000_000, + min_number=10, + max_bytes=rows_max_bytes, + min_cell_bytes=10, + columns_max_number=columns_max_number, + ), @@ -291 +284,0 @@ def test_response_already_computed( - first_rows_config: FirstRowsConfig, @@ -318 +310,0 @@ def test_response_already_computed( - first_rows_config, diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index 931ea63b..8fe74c65 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -18 +18 @@ from libcommon.simple_cache import DoesNotExist, get_response -from worker.config import AppConfig, ParquetAndInfoConfig +from worker.config import AppConfig @@ -55,8 +55 @@ def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyP [email protected] -def parquet_and_dataset_info_config( - set_env_vars: pytest.MonkeyPatch, set_supported_datasets: pytest.MonkeyPatch -) -> ParquetAndInfoConfig: - return ParquetAndInfoConfig.from_env() - - -GetJobRunner = Callable[[str, AppConfig, ParquetAndInfoConfig, bool], ParquetAndDatasetInfoJobRunner] +GetJobRunner = Callable[[str, AppConfig, bool], ParquetAndDatasetInfoJobRunner] @@ -74 +66,0 @@ def get_job_runner( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -99 +90,0 @@ def get_job_runner( - parquet_and_dataset_info_config=parquet_and_dataset_info_config, @@ -125 +115,0 @@ def test_compute( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -129 +119 @@ def test_compute( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False) @@ -141,3 +131 @@ def test_compute( -def test_doesnotexist( - app_config: AppConfig, get_job_runner: GetJobRunner, parquet_and_dataset_info_config: ParquetAndInfoConfig -) -> None: +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: @@ -145 +133 @@ def test_doesnotexist( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False) @@ -177 +164,0 @@ def test_raise_if_too_big_from_hub( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -189 +176 @@ def test_raise_if_too_big_from_hub( - dataset_info=dataset_info, max_dataset_size=parquet_and_dataset_info_config.max_dataset_size + dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size @@ -193 +180 @@ def test_raise_if_too_big_from_hub( - dataset_info=dataset_info, max_dataset_size=parquet_and_dataset_info_config.max_dataset_size + dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size @@ -206 +192,0 @@ def test_raise_if_too_big_from_datasets( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -216 +202 @@ def test_raise_if_too_big_from_datasets( - max_dataset_size=parquet_and_dataset_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -224 +210 @@ def test_raise_if_too_big_from_datasets( - max_dataset_size=parquet_and_dataset_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -241 +226,0 @@ def test_raise_if_too_big_external_files( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -243,2 +228,2 @@ def test_raise_if_too_big_external_files( - max_dataset_size = max_dataset_size or parquet_and_dataset_info_config.max_dataset_size - max_external_data_files = max_external_data_files or parquet_and_dataset_info_config.max_external_data_files + max_dataset_size = max_dataset_size or app_config.parquet_and_info.max_dataset_size + max_external_data_files = max_external_data_files or app_config.parquet_and_info.max_external_data_files @@ -275 +259,0 @@ def test_raise_if_too_many_external_files( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -277,2 +261,2 @@ def test_raise_if_too_many_external_files( - max_dataset_size = max_dataset_size or parquet_and_dataset_info_config.max_dataset_size - max_external_data_files = max_external_data_files or parquet_and_dataset_info_config.max_external_data_files + max_dataset_size = max_dataset_size or app_config.parquet_and_info.max_dataset_size + max_external_data_files = max_external_data_files or app_config.parquet_and_info.max_external_data_files @@ -306 +289,0 @@ def test_raise_if_not_supported( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -316 +299 @@ def test_raise_if_not_supported( - committer_hf_token=parquet_and_dataset_info_config.committer_hf_token, + committer_hf_token=app_config.parquet_and_info.committer_hf_token, @@ -318 +301 @@ def test_raise_if_not_supported( - max_dataset_size=parquet_and_dataset_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -327 +310 @@ def test_raise_if_not_supported( - committer_hf_token=parquet_and_dataset_info_config.committer_hf_token, + committer_hf_token=app_config.parquet_and_info.committer_hf_token, @@ -329 +312 @@ def test_raise_if_not_supported( - max_dataset_size=parquet_and_dataset_info_config.max_dataset_size, + max_dataset_size=app_config.parquet_and_info.max_dataset_size, @@ -338 +320,0 @@ def test_not_supported_if_big( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -343 +325 @@ def test_not_supported_if_big( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False) @@ -353 +334,0 @@ def test_supported_if_gated( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -358 +339 @@ def test_supported_if_gated( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False) @@ -368 +348,0 @@ def test_not_supported_if_gated_with_extra_fields( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -373 +353 @@ def test_not_supported_if_gated_with_extra_fields( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False) @@ -383 +362,0 @@ def test_blocked( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -388 +367 @@ def test_blocked( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False) @@ -404 +382,0 @@ def test_compute_splits_response_simple_csv_ok( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -409 +387 @@ def test_compute_splits_response_simple_csv_ok( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False) @@ -447 +424,0 @@ def test_compute_splits_response_simple_csv_error( - parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -450 +427 @@ def test_compute_splits_response_simple_csv_error( - job_runner = get_job_runner(dataset, app_config, parquet_and_dataset_info_config, False) + job_runner = get_job_runner(dataset, app_config, False)
44dabb7f58dc68e578f86328b47ee343114cd757
Sylvain Lesage
2023-04-20T19:18:41
refactor: 💡 change name of parameter to be more precise (#1070)
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index e7546d63..0f4669e4 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -641 +641 @@ class JobRunner(ABC): - def raise_if_parallel_response_exists(self, parallel_job_type: str, parallel_job_version: int) -> None: + def raise_if_parallel_response_exists(self, parallel_cache_kind: str, parallel_job_version: int) -> None: @@ -644 +644 @@ class JobRunner(ABC): - kind=parallel_job_type, dataset=self.dataset, config=self.config, split=self.split + kind=parallel_cache_kind, dataset=self.dataset, config=self.config, split=self.split @@ -655 +655,2 @@ class JobRunner(ABC): - f"Response has already been computed by {parallel_job_type}. Compute will be skipped." + f"Response has already been computed and stored in cache kind: {parallel_cache_kind}. Compute will" + " be skipped." @@ -658 +659 @@ class JobRunner(ABC): - logging.debug(f"no cache found for {parallel_job_type}.") + logging.debug(f"no cache found for {parallel_cache_kind}.") diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index d8a232d9..5a042943 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -117 +117 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): - parallel_job_type="/split-names-from-streaming", + parallel_cache_kind="/split-names-from-streaming", diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index 318e75ad..7b5e1bf6 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -137 +137 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): - parallel_job_type="/split-names-from-dataset-info", + parallel_cache_kind="/split-names-from-dataset-info", diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index 24f63fe9..1c87b574 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -318 +318 @@ class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): - parallel_job_type="split-first-rows-from-streaming", + parallel_cache_kind="split-first-rows-from-streaming", diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 2f026f70..7f370f84 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -497 +497 @@ class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): - parallel_job_type="split-first-rows-from-parquet", + parallel_cache_kind="split-first-rows-from-parquet", diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py index 3398be07..2ad532cb 100644 --- a/services/worker/tests/test_job_runner.py +++ b/services/worker/tests/test_job_runner.py @@ -374 +374 @@ def test_raise_if_parallel_response_exists( - job_runner.raise_if_parallel_response_exists(parallel_job_type="dummy-parallel", parallel_job_version=1) + job_runner.raise_if_parallel_response_exists(parallel_cache_kind="dummy-parallel", parallel_job_version=1)
eb7cd6d3b8feb1826a54b7c603af865e815545d1
Sylvain Lesage
2023-04-20T15:26:04
Propagate previous step error (#1066)
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 7771ec0e..d8894972 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -12,9 +12,9 @@ PROCESSING_STEP_CONFIG_NAMES_VERSION = 1 -PROCESSING_STEP_CONFIG_PARQUET_VERSION = 3 -PROCESSING_STEP_CONFIG_SIZE_VERSION = 1 -PROCESSING_STEP_CONFIG_INFO_VERSION = 1 -PROCESSING_STEP_DATASET_INFO_VERSION = 1 -PROCESSING_STEP_DATASET_IS_VALID_VERSION = 1 -PROCESSING_STEP_DATASET_PARQUET_VERSION = 1 -PROCESSING_STEP_DATASET_SIZE_VERSION = 1 -PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION = 1 -PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION = 1 +PROCESSING_STEP_CONFIG_PARQUET_VERSION = 4 +PROCESSING_STEP_CONFIG_SIZE_VERSION = 2 +PROCESSING_STEP_CONFIG_INFO_VERSION = 2 +PROCESSING_STEP_DATASET_INFO_VERSION = 2 +PROCESSING_STEP_DATASET_IS_VALID_VERSION = 2 +PROCESSING_STEP_DATASET_PARQUET_VERSION = 2 +PROCESSING_STEP_DATASET_SIZE_VERSION = 2 +PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION = 2 +PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION = 2 @@ -22,2 +22,2 @@ PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION = 1 -PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION = 1 -PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 1 +PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION = 2 +PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 2 @@ -27 +27 @@ PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION = 3 -PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 1 +PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 2 diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index 4b6fbb41..24141b27 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -212,0 +213,4 @@ class CacheEntry(CacheEntryWithoutContent): +class CacheEntryWithDetails(CacheEntry): + details: Mapping[str, str] + + @@ -229,0 +234,22 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O +# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them +def get_response_with_details( + kind: str, dataset: str, config: Optional[str] = None, split: Optional[str] = None +) -> CacheEntryWithDetails: + response = ( + CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + .only( + "content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "details" + ) + .get() + ) + return { + "content": response.content, + "http_status": response.http_status, + "error_code": response.error_code, + "job_runner_version": response.job_runner_version, + "dataset_git_revision": response.dataset_git_revision, + "progress": response.progress, + "details": response.details, + } + + @@ -232 +258 @@ def get_response_or_missing_error( -) -> CacheEntry: +) -> CacheEntryWithDetails: @@ -234 +260 @@ def get_response_or_missing_error( - response = get_response(kind=kind, dataset=dataset, config=config, split=split) + response = get_response_with_details(kind=kind, dataset=dataset, config=config, split=split) @@ -236 +262 @@ def get_response_or_missing_error( - response = CacheEntry( + response = CacheEntryWithDetails( @@ -246,0 +273 @@ def get_response_or_missing_error( + details={}, @@ -254 +281 @@ class BestResponse: - response: CacheEntry + response: CacheEntryWithDetails diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index c6d14f91..e7546d63 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -12 +12,5 @@ from libcommon.dataset import DatasetNotFoundError, get_dataset_git_revision -from libcommon.exceptions import CustomError +from libcommon.exceptions import ( + CustomError, + ErrorResponseWithCause, + ErrorResponseWithoutCause, +) @@ -17 +21 @@ from libcommon.simple_cache import ( - CacheEntry, + CacheEntryWithDetails, @@ -32 +35,0 @@ GeneralJobRunnerErrorCode = Literal[ - "ConfigNotFoundError", @@ -95,13 +97,0 @@ class GeneralJobRunnerError(JobRunnerError): -class ConfigNotFoundError(GeneralJobRunnerError): - """Raised when the config does not exist.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__( - message=message, - status_code=HTTPStatus.NOT_FOUND, - code="ConfigNotFoundError", - cause=cause, - disclose_cause=False, - ) - - @@ -213 +203 @@ class ResponseAlreadyComputedError(GeneralJobRunnerError): -class PreviousStepError(Exception): +class PreviousStepError(JobRunnerError): @@ -217,0 +208,3 @@ class PreviousStepError(Exception): + error_with_cause: ErrorResponseWithCause + error_without_cause: ErrorResponseWithoutCause + @@ -219 +212,8 @@ class PreviousStepError(Exception): - self, response: CacheEntry, kind: str, dataset: str, config: Optional[str] = None, split: Optional[str] = None + self, + message: str, + status_code: HTTPStatus, + code: str, + cause: Optional[BaseException], + disclose_cause: bool, + error_with_cause: ErrorResponseWithCause, + error_without_cause: ErrorResponseWithoutCause, @@ -221,2 +221,15 @@ class PreviousStepError(Exception): - super().__init__("Error in previous step") - if response.get("http_status") == HTTPStatus.OK or response["content"].get("error") is None: + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + self.error_with_cause = error_with_cause + self.error_without_cause = error_without_cause + + @staticmethod + def from_response( + response: CacheEntryWithDetails, + kind: str, + dataset: str, + config: Optional[str] = None, + split: Optional[str] = None, + ) -> "PreviousStepError": + if response.get("http_status") == HTTPStatus.OK: @@ -225,6 +238,14 @@ class PreviousStepError(Exception): - self.content = response["content"] - self.http_status = response["http_status"] - self.error_code = response["error_code"] - self.details = { - "message": "The previous step failed", - "previous_step": {"kind": kind, "dataset": dataset, "config": config, "split": split}, + message = response["content"]["error"] if "error" in response["content"] else "Unknown error" + status_code = response["http_status"] + error_code = response["error_code"] or "PreviousStepError" + cause = None # No way to create the same exception + disclose_cause = response["details"] == response["content"] + error_without_cause: ErrorResponseWithoutCause = {"error": message} + error_with_cause: ErrorResponseWithCause = { + "error": message, + # Add lines in the traceback to give some info about the previous step error (a bit hacky) + "cause_traceback": [ + "The previous step failed, the error is copied to this step:", + f" {kind=} {dataset=} {config=} {split=}", + "---", + ], @@ -231,0 +253,25 @@ class PreviousStepError(Exception): + if "cause_exception" in response["details"] and isinstance(response["details"]["cause_exception"], str): + error_with_cause["cause_exception"] = response["details"]["cause_exception"] + if "cause_message" in response["details"] and isinstance(response["details"]["cause_message"], str): + error_with_cause["cause_message"] = response["details"]["cause_message"] + if ( + "cause_traceback" in response["details"] + and isinstance(response["details"]["cause_traceback"], list) + and all(isinstance(line, str) for line in response["details"]["cause_traceback"]) + ): + error_with_cause["cause_traceback"].extend(*response["details"]["cause_traceback"]) + return PreviousStepError( + message=message, + status_code=status_code, + code=error_code, + cause=cause, + disclose_cause=disclose_cause, + error_without_cause=error_without_cause, + error_with_cause=error_with_cause, + ) + + def as_response_with_cause(self) -> ErrorResponseWithCause: + return self.error_with_cause + + def as_response_without_cause(self) -> ErrorResponseWithoutCause: + return self.error_without_cause @@ -240 +286 @@ def get_previous_step_or_raise( - raise PreviousStepError( + raise PreviousStepError.from_response( @@ -445,10 +491,3 @@ class JobRunner(ABC): - except ( - DatasetNotFoundError, - ConfigNotFoundError, - SplitNotFoundError, - ): - # To avoid filling the cache, we don't save these errors. Otherwise, DoS is possible. - self.debug( - f"the dataset={self.dataset}, config {self.config} or split {self.split} could not be found, don't" - " update the cache" - ) + except DatasetNotFoundError: + # To avoid filling the cache, we don't save this error. Otherwise, DoS is possible. + self.debug(f"the dataset={self.dataset} could not be found, don't update the cache") @@ -457,11 +496 @@ class JobRunner(ABC): - if isinstance(err, PreviousStepError): - content = err.content - http_status = err.http_status - error_code = err.error_code - details: Mapping[str, Any] = err.details - else: - e = err if isinstance(err, CustomError) else UnexpectedError(str(err), err) - content = dict(e.as_response()) - http_status = e.status_code - error_code = e.code - details = dict(e.as_response_with_cause()) + e = err if isinstance(err, CustomError) else UnexpectedError(str(err), err) @@ -473,4 +502,4 @@ class JobRunner(ABC): - content=content, - http_status=http_status, - error_code=error_code, - details=details, + content=dict(e.as_response()), + http_status=e.status_code, + error_code=e.code, + details=dict(e.as_response_with_cause()), diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py index a0944827..a419b0fe 100644 --- a/services/worker/src/worker/job_runners/config/info.py +++ b/services/worker/src/worker/job_runners/config/info.py @@ -6,2 +6 @@ from libcommon.constants import PROCESSING_STEP_CONFIG_INFO_VERSION -from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import SplitFullName @@ -13,0 +13 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -16,5 +16 @@ from worker.job_runner import ( -ConfigInfoJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", - "MissingInfoForConfigError", -] +ConfigInfoJobRunnerErrorCode = Literal["PreviousStepFormatError"] @@ -39,7 +34,0 @@ class ConfigInfoJobRunnerError(JobRunnerError): -class PreviousStepStatusError(ConfigInfoJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -53,7 +41,0 @@ class PreviousStepFormatError(ConfigInfoJobRunnerError): -class MissingInfoForConfigError(ConfigInfoJobRunnerError): - """Raised when the dataset info from the parquet export is missing the requested dataset configuration.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "MissingInfoForConfigError", cause, False) - - @@ -77,2 +59,2 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - - [`~job_runners.config_info.PreviousStepStatusError`] - `If the previous step gave an error. + - [`~job_runner.PreviousStepError`] + If the previous step gave an error. @@ -81,6 +62,0 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - - [`~job_runners.config.info.MissingInfoForConfigError`] - If the dataset info from the parquet export doesn't have the requested dataset configuration - - [`~libcommon.dataset.DatasetNotFoundError`] - If the dataset does not exist, or if the - token does not give the sufficient access to the dataset, or if the dataset is private - (private datasets are not supported by the datasets server) @@ -91,9 +67,2 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - try: - response = get_response(kind=previous_step, dataset=dataset, config=config) - except DoesNotExist as e: - raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e - - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError(f"Previous step '{previous_step}' raised an error: {response['http_status']}..") - - content = response["content"] + dataset_info_best_response = get_previous_step_or_raise(kinds=[previous_step], dataset=dataset, config=config) + content = dataset_info_best_response.response["content"] diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py index aca97cdc..e9aafe26 100644 --- a/services/worker/src/worker/job_runners/config/parquet.py +++ b/services/worker/src/worker/job_runners/config/parquet.py @@ -9,2 +9 @@ from libcommon.constants import PROCESSING_STEP_CONFIG_PARQUET_VERSION -from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import SplitFullName @@ -16,0 +16 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -20,4 +20 @@ from worker.job_runners.config.parquet_and_info import ParquetFileItem -ConfigParquetJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] +ConfigParquetJobRunnerErrorCode = Literal["PreviousStepFormatError"] @@ -46,7 +42,0 @@ class ConfigParquetJobRunnerError(JobRunnerError): -class PreviousStepStatusError(ConfigParquetJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -73,2 +63,2 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - - [`~job_runners.parquet.PreviousStepStatusError`] - If the previous step gave an error. + - [`~job_runner.PreviousStepError`] + If the previous step gave an error. @@ -77,3 +66,0 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - - [`~libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the - token does not give the sufficient access to the dataset, or if the dataset is private - (private datasets are not supported by the datasets server) @@ -85,8 +72,4 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - try: - response = get_response(kind=previous_step, dataset=dataset, config=config) - except DoesNotExist as e: - raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError(f"Previous step gave an error: {response['http_status']}.") - - content = response["content"] + config_parquet_and_info_best_response = get_previous_step_or_raise( + kinds=[previous_step], dataset=dataset, config=config + ) + content = config_parquet_and_info_best_response.response["content"] diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index bae66220..68ca603f 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -41 +41 @@ from libcommon.queue import JobInfo -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import SplitFullName @@ -44 +44,6 @@ from worker.config import AppConfig, ParquetAndInfoConfig -from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError +from worker.job_runner import ( + CompleteJobResult, + JobRunnerError, + ParameterMissingError, + get_previous_step_or_raise, +) @@ -61 +65,0 @@ ConfigParquetAndInfoJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", @@ -117,7 +120,0 @@ class DatasetTooBigFromDatasetsError(ConfigParquetAndInfoJobRunnerError): -class PreviousStepStatusError(ConfigParquetAndInfoJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -763,2 +760,2 @@ def compute_config_parquet_and_info_response( - - [`~job_runners.config.parquet_and_info.PreviousStepStatusError`] - If the previous step gave an error. + - [`~job_runner.PreviousStepError`] + If the previous step gave an error. @@ -786,6 +783 @@ def compute_config_parquet_and_info_response( - try: - response = get_response(kind=previous_step, dataset=dataset) - except DoesNotExist as e: - raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError(f"Previous step {previous_step} gave an error: {response['http_status']}..") + config_names_best_response = get_previous_step_or_raise(kinds=[previous_step], dataset=dataset) @@ -793 +785 @@ def compute_config_parquet_and_info_response( - config_names_content = response["content"] + config_names_content = config_names_best_response.response["content"] diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py index 662dbfec..6487ad17 100644 --- a/services/worker/src/worker/job_runners/config/size.py +++ b/services/worker/src/worker/job_runners/config/size.py @@ -88 +88 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - If the previous step gave an error. + If the previous step gave an error. diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index 09a6a5d7..d8a232d9 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -12,2 +12 @@ from libcommon.constants import ( -from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import SplitFullName @@ -15 +14,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError +from worker.job_runner import ( + CompleteJobResult, + JobRunnerError, + ParameterMissingError, + get_previous_step_or_raise, +) @@ -20 +23,0 @@ SplitNamesFromDatasetInfoJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", @@ -42,7 +44,0 @@ class SplitNamesFromDatasetInfoJobRunnerError(JobRunnerError): -class PreviousStepStatusError(SplitNamesFromDatasetInfoJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -81,2 +77,2 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - - [`~job_runners.config.split_names_from_dataset_info.PreviousStepStatusError`] - If the the previous step gave an error. + - [`~job_runner.PreviousStepError`] + If the previous step gave an error. @@ -85,4 +80,0 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - - [`~libcommon.dataset.DatasetNotFoundError`] - If previous step content was not found for the dataset - - [`~job_runners.config.split_names_from_dataset_info.ResponseAlreadyComputedError`] - If response has been already computed by /split-names-from-streaming job runner. @@ -92,8 +84 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - try: - response = get_response(kind="config-info", dataset=dataset, config=config) - except DoesNotExist as e: - raise DatasetNotFoundError("No response found in previous step for this dataset: 'config-info'.", e) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step gave an error: {response['http_status']}. This job should not have been created." - ) + config_info_best_response = get_previous_step_or_raise(kinds=["config-info"], dataset=dataset, config=config) @@ -102 +87 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - splits_content = response["content"]["dataset_info"]["splits"] + splits_content = config_info_best_response.response["content"]["dataset_info"]["splits"] @@ -126,0 +112,4 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): + """ + Raises [`~job_runners.config.split_names_from_dataset_info.ResponseAlreadyComputedError`] + If response has been already computed by /split-names-from-streaming job runner. + """ diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index d40c275a..318e75ad 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -98,2 +97,0 @@ def compute_split_names_from_streaming_response( - - [`~job_runners.config.split_names_from_streaming.ResponseAlreadyComputedError`] - If response has been already computed by /split-names-from-dataset-info job runner. @@ -133,0 +132,4 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): + """ + Raises [`~job_runners.config.split_names_from_streaming.ResponseAlreadyComputedError`] + If response has been already computed by /split-names-from-dataset-info job runner. + """ diff --git a/services/worker/src/worker/job_runners/dataset/is_valid.py b/services/worker/src/worker/job_runners/dataset/is_valid.py index 049e3844..53e6392c 100644 --- a/services/worker/src/worker/job_runners/dataset/is_valid.py +++ b/services/worker/src/worker/job_runners/dataset/is_valid.py @@ -5 +5 @@ import logging -from typing import Any, Literal, Mapping, Tuple, TypedDict +from typing import Any, Mapping, Tuple, TypedDict @@ -12,5 +11,0 @@ from worker.job_runner import JobResult, JobRunner -IsValidJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] - diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index d9a292a0..24f63fe9 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -21 +21 @@ from libcommon.queue import JobInfo -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import SplitFullName @@ -28 +28,5 @@ from worker.config import AppConfig, FirstRowsConfig -from worker.job_runner import CompleteJobResult, ConfigNotFoundError, JobRunnerError +from worker.job_runner import ( + CompleteJobResult, + JobRunnerError, + get_previous_step_or_raise, +) @@ -43 +46,0 @@ SplitFirstRowsFromParquetJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", @@ -87,7 +89,0 @@ class TooBigContentError(SplitFirstRowsFromParquetJobRunnerError): -class PreviousStepStatusError(SplitFirstRowsFromParquetJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -171,7 +166,0 @@ def compute_first_rows_response( - try: - upstream_response = get_response(kind="config-parquet", dataset=dataset, config=config) - if upstream_response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step gave an error: {upstream_response['http_status']}. This job should not have been" - " created." - ) @@ -179 +168,3 @@ def compute_first_rows_response( - parquet_files_content = upstream_response["content"]["parquet_files"] + config_parquet_best_response = get_previous_step_or_raise(kinds=["config-parquet"], dataset=dataset, config=config) + try: + parquet_files_content = config_parquet_best_response.response["content"]["parquet_files"] @@ -187,2 +177,0 @@ def compute_first_rows_response( - except DoesNotExist: - raise ConfigNotFoundError(f"The config '{config}' does not exist for the dataset.'") diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index d4194e36..2f026f70 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -27 +27 @@ from libcommon.queue import JobInfo -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import SplitFullName @@ -34 +33,0 @@ from worker.job_runner import ( - ConfigNotFoundError, @@ -37,0 +37 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -57 +56,0 @@ SplitFirstRowsFromStreamingJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", @@ -141,7 +139,0 @@ class TooBigContentError(SplitFirstRowsFromStreamingJobRunnerError): -class PreviousStepStatusError(SplitFirstRowsFromStreamingJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -297,2 +288,0 @@ def compute_first_rows_response( - - [`~job_runner.ConfigNotFoundError`] - If the config does not exist in the dataset. @@ -314,0 +305,4 @@ def compute_first_rows_response( + - [`~job_runner.PreviousStepError`] + If the previous step gave an error. + - [`~job_runners.split.first_rows.PreviousStepFormatError`] + If the content of the previous step has not the expected format @@ -319,0 +314,3 @@ def compute_first_rows_response( + split_names_best_response = get_previous_step_or_raise( + kinds=["/split-names-from-streaming", "/split-names-from-dataset-info"], dataset=dataset, config=config + ) @@ -321,15 +318,3 @@ def compute_first_rows_response( - upstream_response = get_response(kind="/split-names-from-streaming", dataset=dataset, config=config) - splits_content = upstream_response["content"]["splits"] - except Exception: - try: - upstream_response = get_response(kind="/split-names-from-dataset-info", dataset=dataset, config=config) - splits_content = upstream_response["content"]["splits"] - except DoesNotExist as e: - raise ConfigNotFoundError(f"The config '{config}' does not exist for the dataset.'", e) from e - except Exception as e: - raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e - - if upstream_response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step gave an error: {upstream_response['http_status']}. This job should not have been created." - ) + splits_content = split_names_best_response.response["content"]["splits"] + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py index ca1b51b1..897f4a5c 100644 --- a/services/worker/tests/job_runners/config/test_info.py +++ b/services/worker/tests/job_runners/config/test_info.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -15,5 +14,2 @@ from worker.config import AppConfig -from worker.job_runners.config.info import ( - ConfigInfoJobRunner, - PreviousStepFormatError, - PreviousStepStatusError, -) +from worker.job_runner import PreviousStepError +from worker.job_runners.config.info import ConfigInfoJobRunner, PreviousStepFormatError @@ -196 +192 @@ def get_job_runner( - PreviousStepStatusError.__name__, + PreviousStepError.__name__, @@ -241 +237 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index 08542e00..30cd9942 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -14,0 +14 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -19 +18,0 @@ from worker.job_runners.config.parquet import ( - PreviousStepStatusError, @@ -110 +109 @@ def get_job_runner( - PreviousStepStatusError.__name__, + PreviousStepError.__name__, @@ -155 +154 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 9f2c6e70..35cb6f29 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -556 +556 @@ def test_compute_splits_response_simple_csv_ok( - ("public", "DatasetNotFoundError", "DoesNotExist"), # no cache for /config-names -> DatasetNotFoundError + ("public", "CachedResponseNotFound", None), # no cache for /config-names -> CachedResponseNotFound @@ -589 +589 @@ def test_compute_splits_response_simple_csv_error( - (HTTPStatus.NOT_FOUND, {"error": "error"}, "PreviousStepStatusError"), + (HTTPStatus.NOT_FOUND, {"error": "error"}, "PreviousStepError"), diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index 03b81dfb..e8ab3e7e 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -10 +9,0 @@ from libcommon.constants import PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSI -from libcommon.dataset import DatasetNotFoundError @@ -17,0 +17 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -20 +19,0 @@ from worker.job_runners.config.split_names_from_dataset_info import ( - PreviousStepStatusError, @@ -95 +94 @@ def get_job_runner( - PreviousStepStatusError.__name__, + PreviousStepError.__name__, @@ -152 +151 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - assert exc_info.value.code == DatasetNotFoundError.__name__ + assert exc_info.value.code == "CachedResponseNotFound" @@ -159,2 +158,2 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "DatasetNotFoundError", HTTPStatus.NOT_FOUND), - (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "DatasetNotFoundError", HTTPStatus.NOT_FOUND), + (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "CachedResponseNotFound", HTTPStatus.NOT_FOUND), + (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "CachedResponseNotFound", HTTPStatus.NOT_FOUND), diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index f9ec3285..1b848140 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -180,2 +180,2 @@ def test_compute( - (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), - (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), + (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "CachedResponseNotFound", HTTPStatus.NOT_FOUND), + (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "CachedResponseNotFound", HTTPStatus.NOT_FOUND), diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index 2f69fc9c..93496c33 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -143 +143 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first - ("does_not_exist_config", False, "ConfigNotFoundError", "DoesNotExist"), + ("does_not_exist_config", False, "CachedResponseNotFound", None), @@ -285,2 +285,2 @@ def test_truncation( - (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), - (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), + (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "CachedResponseNotFound", HTTPStatus.NOT_FOUND), + (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "CachedResponseNotFound", HTTPStatus.NOT_FOUND),
f3e48d9f53d2b3a7fde7441710abfbf3e6400585
Sylvain Lesage
2023-04-20T11:30:41
fix: 🐛 add missing "config" parameter (#1065)
diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index ad9dcd7e..09a6a5d7 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -93 +93 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - response = get_response(kind="config-info", dataset=dataset) + response = get_response(kind="config-info", dataset=dataset, config=config) diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index 9555ea5c..03b81dfb 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -130,2 +130,5 @@ def test_compute( - upsert_response(kind="config-info", dataset=dataset, content=upstream_content, http_status=upstream_status) - job_runner = get_job_runner(dataset, "config_name", app_config, False) + config = "config_name" + upsert_response( + kind="config-info", dataset=dataset, config=config, content=upstream_content, http_status=upstream_status + ) + job_runner = get_job_runner(dataset, config, app_config, False)
7242bf47d02840b60642cd01562eaf9cc77e0191
Andrea Francis Soria Jimenez
2023-04-19T16:44:52
Adding ttls time for metrics cron job (#1064)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index f6c7c1dc..05776ee9 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.10.0 +version: 1.11.0 diff --git a/chart/templates/cron-jobs/metrics-collector/job.yaml b/chart/templates/cron-jobs/metrics-collector/job.yaml index 102a090e..a924c8ee 100644 --- a/chart/templates/cron-jobs/metrics-collector/job.yaml +++ b/chart/templates/cron-jobs/metrics-collector/job.yaml @@ -14,0 +15 @@ spec: + ttlSecondsAfterFinished: 300
b995126eda2a13b9d2d559030604689da3fc40bb
Andrea Francis Soria Jimenez
2023-04-19T16:07:09
Add missing cache config (#1063)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 750c0cbc..f6c7c1dc 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.9.0 +version: 1.10.0 diff --git a/chart/templates/cron-jobs/metrics-collector/_container.tpl b/chart/templates/cron-jobs/metrics-collector/_container.tpl index b9391105..eb13d893 100644 --- a/chart/templates/cron-jobs/metrics-collector/_container.tpl +++ b/chart/templates/cron-jobs/metrics-collector/_container.tpl @@ -12,0 +13 @@ + {{ include "envCache" . | nindent 2 }} diff --git a/chart/templates/jobs/cache-maintenance/_container.tpl b/chart/templates/jobs/cache-maintenance/_container.tpl index 2d45e7f8..5394c7e1 100644 --- a/chart/templates/jobs/cache-maintenance/_container.tpl +++ b/chart/templates/jobs/cache-maintenance/_container.tpl @@ -9,0 +10 @@ + {{ include "envCache" . | nindent 2 }}
c23599b52717d958b3d911b7d22fd7ac9f1107a3
Rémy
2023-04-19T15:45:16
fix: cron job definition (#1062)
diff --git a/chart/templates/cron-jobs/metrics-collector/_container.tpl b/chart/templates/cron-jobs/metrics-collector/_container.tpl index e159d00b..b9391105 100644 --- a/chart/templates/cron-jobs/metrics-collector/_container.tpl +++ b/chart/templates/cron-jobs/metrics-collector/_container.tpl @@ -7,0 +8,3 @@ + securityContext: + allowPrivilegeEscalation: false + resources: {{ toYaml .Values.metricsCollector.resources | nindent 4 }} @@ -9,4 +12,4 @@ - {{ include "envLog" . | nindent 2 }} - {{ include "envQueue" . | nindent 2 }} - {{ include "envCommon" . | nindent 2 }} - {{ include "envMetrics" . | nindent 2 }} + {{ include "envLog" . | nindent 2 }} + {{ include "envQueue" . | nindent 2 }} + {{ include "envCommon" . | nindent 2 }} + {{ include "envMetrics" . | nindent 2 }} @@ -15,3 +17,0 @@ - securityContext: - allowPrivilegeEscalation: false - resources: {{ toYaml .Values.metricsCollector.resources | nindent 4 }} diff --git a/chart/templates/cron-jobs/metrics-collector/job.yaml b/chart/templates/cron-jobs/metrics-collector/job.yaml index e57c4e09..102a090e 100644 --- a/chart/templates/cron-jobs/metrics-collector/job.yaml +++ b/chart/templates/cron-jobs/metrics-collector/job.yaml @@ -12,4 +12,2 @@ spec: - schedule: {{ toYaml .Values.metricsCollector.schedule }} - template: - metadata: - labels: {{ include "labels.metricsCollector" . | nindent 8 }} + schedule: {{ .Values.metricsCollector.schedule | quote }} + jobTemplate: @@ -17,6 +15,8 @@ spec: - restartPolicy: OnFailure - {{- include "image.imagePullSecrets" . | nindent 6 }} - containers: {{ include "containerMetricsCollector" . | nindent 8 }} - nodeSelector: {{ toYaml .Values.metricsCollector.nodeSelector | nindent 8 }} - tolerations: {{ toYaml .Values.metricsCollector.tolerations | nindent 8 }} - securityContext: {{ include "securityContext" . | nindent 8 }} + template: + spec: + restartPolicy: OnFailure + {{- include "image.imagePullSecrets" . | nindent 6 }} + nodeSelector: {{ toYaml .Values.metricsCollector.nodeSelector | nindent 12 }} + tolerations: {{ toYaml .Values.metricsCollector.tolerations | nindent 12 }} + containers: {{ include "containerMetricsCollector" . | nindent 12 }} + securityContext: {{ include "securityContext" . | nindent 12 }}
a1f3400f66f65c9d3750b700385a91a582e80308
Sylvain Lesage
2023-04-19T13:44:17
feat: 🎸 add backfill tasks, and button to launch backfill (#1061)
diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py index b7f69f66..c3868c7c 100644 --- a/front/admin_ui/app.py +++ b/front/admin_ui/app.py @@ -80,0 +81,4 @@ with gr.Blocks() as demo: + backfill_message = gr.Markdown("", visible=False) + backfill_plan_table = gr.DataFrame(visible=False) + backfill_execute_button = gr.Button("Execute backfill plan", visible=False) + backfill_execute_error = gr.Markdown("", visible=False) @@ -177,0 +182,40 @@ with gr.Blocks() as demo: + + def get_backfill_plan(token, dataset): + headers = {"Authorization": f"Bearer {token}"} + response = requests.get(f"{DSS_ENDPOINT}/admin/dataset-state?dataset={dataset}", headers=headers, timeout=60) + if response.status_code != 200: + return { + backfill_message: gr.update(value=f"❌ Failed to get backfill plan for {dataset} (error {response.status_code})", visible=True), + backfill_plan_table: gr.update(value=None,visible=False), + backfill_execute_button: gr.update( visible=False), + backfill_execute_error: gr.update( visible=False) + } + dataset_state = response.json() + tasks_df = pd.DataFrame(dataset_state["plan"]) + has_tasks = len(tasks_df) > 0 + return { + backfill_message: gr.update( + value="""### Backfill plan + +The cache is outdated or in an incoherent state. Here is the plan to backfill the cache.""" + ,visible=has_tasks), + backfill_plan_table: gr.update(value=tasks_df,visible=has_tasks), + backfill_execute_button: gr.update(visible=has_tasks), + backfill_execute_error: gr.update(visible=False), + } + + def get_dataset_status_and_backfill_plan(token, dataset): + return {**get_dataset_status(token, dataset), **get_backfill_plan(token, dataset)} + + + def execute_backfill_plan(token, dataset): + headers = {"Authorization": f"Bearer {token}"} + response = requests.post(f"{DSS_ENDPOINT}/admin/dataset-backfill?dataset={dataset}", headers=headers, timeout=60) + state = get_dataset_status_and_backfill_plan(token, dataset) + message = ( + "✅ Backfill plan executed" + if response.status_code == 200 + else f"❌ Failed to execute backfill plan (error {response.status_code})<pre>{response.text}</pre>" + ) + state[backfill_execute_error] = gr.update(value=message, visible=True) + return state @@ -226 +270,2 @@ with gr.Blocks() as demo: - dataset_status_button.click(get_dataset_status, inputs=[token_box, dataset_name], outputs=[cached_responses_table, jobs_table]) + dataset_status_button.click(get_dataset_status_and_backfill_plan, inputs=[token_box, dataset_name], outputs=[cached_responses_table, jobs_table, backfill_message, backfill_plan_table, backfill_execute_button, backfill_execute_error]) + backfill_execute_button.click(execute_backfill_plan, inputs=[token_box, dataset_name], outputs=[cached_responses_table, jobs_table, backfill_message, backfill_plan_table, backfill_execute_button, backfill_execute_error])
43b2992779055243563189755818b7ab0191e290
Sylvain Lesage
2023-04-19T13:43:51
feat: 🎸 handle the fan-in steps correctly (#1060)
diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index fc2e5a6d..4174a8d2 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -166,33 +165,0 @@ class ArtifactState: - def get_parent_artifact_states(self, parent_step: ProcessingStep) -> List["ArtifactState"]: - if parent_step.input_type == "dataset": - return [ArtifactState(step=parent_step, dataset=self.dataset, config=None, split=None)] - elif parent_step.input_type == "config": - return ( - [ - ArtifactState( - step=parent_step, - dataset=self.dataset, - config=self.config, - split=None, - ) - ] - if self.step.input_type in ["config", "split"] - else [] - # ^ fan-in: config->dataset. For now, we don't return the list of parent artifact states in that case - ) - else: - return ( - [ - ArtifactState( - step=parent_step, - dataset=self.dataset, - config=self.config, - split=self.split, - ) - ] - if self.step.input_type == "split" - else [] - # ^ fan-in: split->config, or split->dataset. For now, we don't return the list of parent artifact - # states in that case - ) - @@ -207,3 +173,0 @@ class ArtifactState: - def get_all_parents_artifact_states(self) -> List[List["ArtifactState"]]: - return [self.get_parent_artifact_states(parent_step) for parent_step in self.step.parents] - @@ -371,0 +336,6 @@ class Plan: +@dataclass +class ParentState: + is_fan_in: bool + artifact_states: List[ArtifactState] + + @@ -405 +375,15 @@ class DatasetState: - def _get_artifact_states_for_step(self, step: ProcessingStep) -> List[ArtifactState]: + def _get_artifact_states_for_step( + self, step: ProcessingStep, config: Optional[str] = None, split: Optional[str] = None + ) -> List[ArtifactState]: + """Get the artifact states for a step. + + Args: + step: the processing step + config: if not None, and step input type is config or split, only return the artifact states for this + config + split: if not None, and step input type is split, only return the artifact states for this split (config + must be specified) + + Returns: + the artifact states for the step + """ @@ -409 +393,10 @@ class DatasetState: - artifact_states = [config_state.artifact_state_by_step[step.name] for config_state in self.config_states] + if config is None: + artifact_states = [ + config_state.artifact_state_by_step[step.name] for config_state in self.config_states + ] + else: + artifact_states = [ + config_state.artifact_state_by_step[step.name] + for config_state in self.config_states + if config_state.config == config + ] @@ -411,5 +404,21 @@ class DatasetState: - artifact_states = [ - split_state.artifact_state_by_step[step.name] - for config_state in self.config_states - for split_state in config_state.split_states - ] + if config is None: + artifact_states = [ + split_state.artifact_state_by_step[step.name] + for config_state in self.config_states + for split_state in config_state.split_states + ] + elif split is None: + artifact_states = [ + split_state.artifact_state_by_step[step.name] + for config_state in self.config_states + if config_state.config == config + for split_state in config_state.split_states + ] + else: + artifact_states = [ + split_state.artifact_state_by_step[step.name] + for config_state in self.config_states + if config_state.config == config + for split_state in config_state.split_states + if split_state.split == split + ] @@ -426,0 +436,2 @@ class DatasetState: + # Every step can have one or multiple artifacts, for example config-level steps have one artifact per + # config @@ -429,4 +440,10 @@ class DatasetState: - # (fan-in steps: config -> dataset, split -> config, split -> dataset) - # what should we do? always recompute? test the progress? - all_not_none_parents_artifact_states = [ - x for x in artifact_state.get_all_parents_artifact_states() if x is not None + parent_states = [ + ParentState( + artifact_states=self._get_artifact_states_for_step( + step=parent_step, config=artifact_state.config, split=artifact_state.split + ), + is_fan_in=(parent_step.input_type == "config" and step.input_type == "dataset") + or (parent_step.input_type == "split" and step.input_type == "dataset") + or (parent_step.input_type == "split" and step.input_type == "config"), + ) + for parent_step in step.parents @@ -438,2 +455,5 @@ class DatasetState: - for all_parents_artifact_state in all_not_none_parents_artifact_states - for parent_artifact_state in all_parents_artifact_state + for parent_state in parent_states + for parent_artifact_state in parent_state.artifact_states + if not parent_state.is_fan_in + # ^ the fan-in relations between steps do not block, because these steps consider the + # cases: ok, pending and error. So, if one of the previous artifacts is pending, we must not block @@ -447,2 +467,2 @@ class DatasetState: - for all_parents_artifact_state in all_not_none_parents_artifact_states - for parent_artifact_state in all_parents_artifact_state + for parent_state in parent_states + for parent_artifact_state in parent_state.artifact_states diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 0a8e9f5e..328eef6a 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -15,0 +16 @@ from libcommon.constants import ( + PROCESSING_STEP_DATASET_PARQUET_VERSION, @@ -1353,0 +1355,105 @@ def test_plan_git_revision( + + +def test_plan_update_fan_in_parent() -> None: + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "dataset-parquet,dataset" artifact in cache + upsert_response( + kind="dataset-parquet", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=PROCESSING_STEP_DATASET_PARQUET_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-parquet-and-info,dataset,config1" artifact in cache + upsert_response( + kind="config-parquet-and-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-parquet,dataset,config1" artifact in cache + upsert_response( + kind="config-parquet", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + assert_dataset_state( + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are not known + split_names_in_first_config=[], + # dataset-parquet,dataset is in the category: "cache_is_outdated_by_parent" + # because one of the "config-parquet" artifacts is more recent + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config2", + "config-parquet,dataset,config2", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [ + "dataset-parquet,dataset", + ], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "config-size,dataset,config1", + "dataset-info,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": [ + "/config-names,dataset", + "config-parquet,dataset,config1", + "config-parquet-and-info,dataset,config1", + ], + }, + queue_status={"in_process": []}, + # config-parquet,dataset,config1 will be refreshed because its job runner has been upgraded + tasks=[ + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[config-size,dataset,config1]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + )
1194e0dfeb0b96d6068308b3d103ddf49318a422
Sylvain Lesage
2023-04-19T13:29:30
fix: 🐛 remove wrong filtering of the step parents (#1059)
diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index cd32e0f8..58d40e7a 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -126,6 +126 @@ class ProcessingGraph: - required_steps = [self.get_step(name) for name in graph.predecessors(step.name)] - step.parents = list(required_steps) - for parent_candidate in required_steps: - for other_parent_candidate in required_steps: - if other_parent_candidate in parent_candidate.ancestors and other_parent_candidate in step.parents: - step.parents.remove(other_parent_candidate) + step.parents = [self.get_step(name) for name in graph.predecessors(step.name)] diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 7157afa0..28b46c2f 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -53 +53 @@ def test_graph() -> None: - assert_step(a, children=[c, f], parents=[], ancestors=[]) + assert_step(a, children=[c, d, f], parents=[], ancestors=[]) @@ -56 +56 @@ def test_graph() -> None: - assert_step(d, children=[], parents=[c], ancestors=[a, c]) + assert_step(d, children=[], parents=[a, c], ancestors=[a, c]) @@ -70 +70,15 @@ def graph() -> ProcessingGraph: - ("/config-names", ["/split-names-from-streaming", "config-parquet-and-info"], [], []), + ( + "/config-names", + [ + "/split-names-from-streaming", + "config-parquet-and-info", + "dataset-split-names-from-dataset-info", + "dataset-split-names-from-streaming", + "dataset-split-names", + "dataset-parquet", + "dataset-info", + "dataset-size", + ], + [], + [], + ), @@ -73 +87,5 @@ def graph() -> ProcessingGraph: - ["config-parquet", "config-info", "config-size"], + [ + "config-parquet", + "config-info", + "config-size", + ], @@ -79 +97,5 @@ def graph() -> ProcessingGraph: - ["dataset-split-names-from-dataset-info", "split-first-rows-from-streaming", "dataset-split-names"], + [ + "dataset-split-names-from-dataset-info", + "split-first-rows-from-streaming", + "dataset-split-names", + ], @@ -92 +114 @@ def graph() -> ProcessingGraph: - ["/split-names-from-dataset-info"], + ["/config-names", "/split-names-from-dataset-info"], @@ -98 +120 @@ def graph() -> ProcessingGraph: - ["/split-names-from-streaming"], + ["/config-names", "/split-names-from-streaming"], @@ -104,0 +127 @@ def graph() -> ProcessingGraph: + "/config-names", @@ -144 +167,6 @@ def graph() -> ProcessingGraph: - ("dataset-parquet", [], ["config-parquet"], ["/config-names", "config-parquet-and-info", "config-parquet"]), + ( + "dataset-parquet", + [], + ["/config-names", "config-parquet"], + ["/config-names", "config-parquet-and-info", "config-parquet"], + ), @@ -151 +179,6 @@ def graph() -> ProcessingGraph: - ("dataset-info", [], ["config-info"], ["/config-names", "config-parquet-and-info", "config-info"]), + ( + "dataset-info", + [], + ["/config-names", "config-info"], + ["/config-names", "config-parquet-and-info", "config-info"], + ), @@ -153 +186,6 @@ def graph() -> ProcessingGraph: - ("dataset-size", [], ["config-size"], ["/config-names", "config-parquet-and-info", "config-size"]), + ( + "dataset-size", + [], + ["/config-names", "config-size"], + ["/config-names", "config-parquet-and-info", "config-size"], + ), diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index f0d56984..0a8e9f5e 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -470,19 +469,0 @@ CONFIG_INFO_OK = {"config": CONFIG_NAME_1, "content": "not important"} -OLD_JOB_RUNNER_VERSION = 1 -NEW_JOB_RUNNER_VERSION = 2 - - -def finish_job(job_type: str, content: Any, job_runner_version: Optional[int] = NEW_JOB_RUNNER_VERSION) -> None: - job_info = Queue().start_job(only_job_types=[job_type]) - upsert_response( - kind=job_info["type"], - dataset=job_info["dataset"], - config=job_info["config"], - split=job_info["split"], - content=content, - http_status=HTTPStatus.OK, - job_runner_version=job_runner_version, - dataset_git_revision=CURRENT_GIT_REVISION, - ) - Queue().finish_job(job_id=job_info["job_id"], finished_status=Status.SUCCESS) - - @@ -527,6 +508 @@ def test_plan() -> None: - "blocked_by_parent": ["dataset-is-valid,dataset"], - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/config-names,dataset", - "/parquet-and-dataset-info,dataset", + "blocked_by_parent": [ @@ -533,0 +510 @@ def test_plan() -> None: + "dataset-is-valid,dataset", @@ -539,0 +517,6 @@ def test_plan() -> None: + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/config-names,dataset", + "/parquet-and-dataset-info,dataset", + ], @@ -550,6 +532,0 @@ def test_plan() -> None: - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -566,6 +542,0 @@ def test_plan_job_creation_and_termination() -> None: - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -581,6 +552 @@ def test_plan_job_creation_and_termination() -> None: - "blocked_by_parent": ["dataset-is-valid,dataset"], - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/config-names,dataset", - "/parquet-and-dataset-info,dataset", + "blocked_by_parent": [ @@ -587,0 +554 @@ def test_plan_job_creation_and_termination() -> None: + "dataset-is-valid,dataset", @@ -593,0 +561,6 @@ def test_plan_job_creation_and_termination() -> None: + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/config-names,dataset", + "/parquet-and-dataset-info,dataset", + ], @@ -603,6 +575,0 @@ def test_plan_job_creation_and_termination() -> None: - "dataset-info,dataset", - "dataset-parquet,dataset", - "dataset-size,dataset", - "dataset-split-names,dataset", - "dataset-split-names-from-dataset-info,dataset", - "dataset-split-names-from-streaming,dataset", @@ -616 +583,12 @@ def test_plan_job_creation_and_termination() -> None: - finish_job("/config-names", TWO_CONFIG_NAMES_CONTENT_OK, job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION) + job_info = Queue().start_job(only_job_types=["/config-names"]) + upsert_response( + kind=job_info["type"], + dataset=job_info["dataset"], + config=job_info["config"], + split=job_info["split"], + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + Queue().finish_job(job_id=job_info["job_id"], finished_status=Status.SUCCESS) @@ -659,11 +637 @@ def test_plan_job_creation_and_termination() -> None: - queue_status={ - "in_process": [ - "/parquet-and-dataset-info,dataset", - "dataset-info,dataset", - "dataset-parquet,dataset", - "dataset-size,dataset", - "dataset-split-names,dataset", - "dataset-split-names-from-dataset-info,dataset", - "dataset-split-names-from-streaming,dataset", - ] - }, + queue_status={"in_process": ["/parquet-and-dataset-info,dataset"]}, @@ -675,0 +644,6 @@ def test_plan_job_creation_and_termination() -> None: + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -798,6 +771,0 @@ def test_plan_retry_error() -> None: - "dataset-is-valid,dataset", - ], - "cache_has_different_git_revision": [], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/parquet-and-dataset-info,dataset", @@ -804,0 +773 @@ def test_plan_retry_error() -> None: + "dataset-is-valid,dataset", @@ -810,0 +780,5 @@ def test_plan_retry_error() -> None: + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + ], @@ -820,6 +793,0 @@ def test_plan_retry_error() -> None: - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]", @@ -1310,6 +1277,0 @@ def test_plan_git_revision( - "dataset-is-valid,dataset", - ], - "cache_has_different_git_revision": ["/config-names,dataset"], - "cache_is_outdated_by_parent": [], - "cache_is_empty": [ - "/parquet-and-dataset-info,dataset", @@ -1316,0 +1279 @@ def test_plan_git_revision( + "dataset-is-valid,dataset", @@ -1322,0 +1286,5 @@ def test_plan_git_revision( + "cache_has_different_git_revision": ["/config-names,dataset"], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + ], @@ -1331,6 +1298,0 @@ def test_plan_git_revision( - "CreateJob[dataset-info,dataset]", - "CreateJob[dataset-parquet,dataset]", - "CreateJob[dataset-size,dataset]", - "CreateJob[dataset-split-names,dataset]", - "CreateJob[dataset-split-names-from-dataset-info,dataset]", - "CreateJob[dataset-split-names-from-streaming,dataset]",
553eec2580b38159c275b70d6779e38ae18ae9bb
Sylvain Lesage
2023-04-19T07:28:29
Fix dataset level aggregators (#1056)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index ad8ba6ea..efacfa31 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -220 +220 @@ class ProcessingGraphConfig: - "requires": "config-parquet", + "requires": ["config-parquet", "/config-names"], @@ -230 +230 @@ class ProcessingGraphConfig: - "requires": "config-info", + "requires": ["config-info", "/config-names"], @@ -245 +245 @@ class ProcessingGraphConfig: - "requires": "config-size", + "requires": ["config-size", "/config-names"], @@ -250 +250 @@ class ProcessingGraphConfig: - "requires": "/split-names-from-streaming", + "requires": ["/split-names-from-streaming", "/config-names"], @@ -255 +255 @@ class ProcessingGraphConfig: - "requires": "/split-names-from-dataset-info", + "requires": ["/split-names-from-dataset-info", "/config-names"], @@ -260 +260 @@ class ProcessingGraphConfig: - "requires": ["/split-names-from-dataset-info", "/split-names-from-streaming"], + "requires": ["/split-names-from-dataset-info", "/split-names-from-streaming", "/config-names"], diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index c6b2de41..7771ec0e 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -24,3 +24,3 @@ PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 1 -PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION = 2 -PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION = 2 -PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION = 2 +PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION = 3 +PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION = 3 +PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION = 3 diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index 20a7db07..cd32e0f8 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -130 +130 @@ class ProcessingGraph: - if other_parent_candidate in parent_candidate.ancestors: + if other_parent_candidate in parent_candidate.ancestors and other_parent_candidate in step.parents: diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index b2d7b1bf..4b6fbb41 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -4 +3,0 @@ -import contextlib @@ -230,0 +230,21 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O +def get_response_or_missing_error( + kind: str, dataset: str, config: Optional[str] = None, split: Optional[str] = None +) -> CacheEntry: + try: + response = get_response(kind=kind, dataset=dataset, config=config, split=split) + except DoesNotExist: + response = CacheEntry( + content={ + "error": ( + f"Cached response not found for kind {kind}, dataset {dataset}, config {config}, split {split}" + ) + }, + http_status=HTTPStatus.NOT_FOUND, + error_code="CachedResponseNotFound", + dataset_git_revision=None, + job_runner_version=None, + progress=None, + ) + return response + + @@ -234,9 +254 @@ class BestResponse: - response: Any - - -def _get_progress(best_response: BestResponse) -> float: - return ( - best_response.response["progress"] - if ("progress" in best_response.response and isinstance(best_response.response["progress"], float)) - else 0.0 - ) + response: CacheEntry @@ -252,4 +264,2 @@ def get_best_response( - - the first success response with progress=1.0 is returned - - if no success response with progress=1.0 is found, the success response with the highest progress is returned - - if no success response is found, the first error response is returned - - if no response is found, a `~libcommon.simple_cache.DoesNotExist` error is raised + - the first success response with the highest progress, + - else: the first error response (including cache miss) @@ -259 +269 @@ def get_best_response( - A list of cache kinds to look responses for. + A non-empty list of cache kinds to look responses for. @@ -267,6 +277,2 @@ def get_best_response( - BestResponse: The best response (object with fields: kind and response). The response can be an error. - <Tip> - Raises the following errors: - - [`~libcommon.simple_cache.DoesNotExist`] - If no response was found - </Tip> + BestResponse: The best response (object with fields: kind and response). The response can be an error, + including a cache miss (error code: `CachedResponseNotFound`) @@ -274,17 +280,23 @@ def get_best_response( - in_progress_success_responses = [] - error_responses = [] - for kind in kinds: - with contextlib.suppress(DoesNotExist): - response = get_response(kind=kind, dataset=dataset, config=config, split=split) - best_response = BestResponse(kind=kind, response=response) - if best_response.response["http_status"] == HTTPStatus.OK: - if best_response.response["progress"] == 1.0: - return best_response - in_progress_success_responses.append(best_response) - else: - error_responses.append(best_response) - if in_progress_success_responses: - return sorted(in_progress_success_responses, key=_get_progress, reverse=True)[0] - if error_responses: - return error_responses[0] - raise DoesNotExist("No cached response found.") + if not kinds: + raise ValueError("kinds must be a non-empty list") + best_response_candidates = [ + BestResponse( + kind=kind, response=get_response_or_missing_error(kind=kind, dataset=dataset, config=config, split=split) + ) + for kind in kinds + ] + max_index = 0 + max_value = float("-inf") + for index, candidate in enumerate(best_response_candidates): + if candidate.response["http_status"] >= HTTPStatus.BAD_REQUEST.value: + # only the first error response is considered + continue + value = ( + 0.0 + if candidate.response["progress"] is None or candidate.response["progress"] < 0.0 + else candidate.response["progress"] + ) + if value > max_value: + max_value = value + max_index = index + return best_response_candidates[max_index] diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 0b4355f8..7157afa0 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -26,0 +27 @@ def assert_step( + parents: List[ProcessingStep], @@ -30,0 +32 @@ def assert_step( + assert_lists_are_equal(step.parents, parents) @@ -51,6 +53,6 @@ def test_graph() -> None: - assert_step(a, children=[c, f], ancestors=[]) - assert_step(b, children=[f], ancestors=[]) - assert_step(c, children=[d, e], ancestors=[a]) - assert_step(d, children=[], ancestors=[a, c]) - assert_step(e, children=[], ancestors=[a, c]) - assert_step(f, children=[], ancestors=[a, b]) + assert_step(a, children=[c, f], parents=[], ancestors=[]) + assert_step(b, children=[f], parents=[], ancestors=[]) + assert_step(c, children=[d, e], parents=[a], ancestors=[a]) + assert_step(d, children=[], parents=[c], ancestors=[a, c]) + assert_step(e, children=[], parents=[c], ancestors=[a, c]) + assert_step(f, children=[], parents=[a, b], ancestors=[a, b]) @@ -66 +68 @@ def graph() -> ProcessingGraph: - "step_name,children,ancestors", + "step_name,children,parents,ancestors", @@ -68,2 +70,7 @@ def graph() -> ProcessingGraph: - ("/config-names", ["/split-names-from-streaming", "config-parquet-and-info"], []), - ("config-parquet-and-info", ["config-parquet", "config-info", "config-size"], ["/config-names"]), + ("/config-names", ["/split-names-from-streaming", "config-parquet-and-info"], [], []), + ( + "config-parquet-and-info", + ["config-parquet", "config-info", "config-size"], + ["/config-names"], + ["/config-names"], + ), @@ -72,0 +80 @@ def graph() -> ProcessingGraph: + ["config-info"], @@ -78,0 +87 @@ def graph() -> ProcessingGraph: + ["/config-names"], @@ -82,0 +92 @@ def graph() -> ProcessingGraph: + ["/split-names-from-dataset-info"], @@ -85 +95,6 @@ def graph() -> ProcessingGraph: - ("dataset-split-names-from-streaming", [], ["/config-names", "/split-names-from-streaming"]), + ( + "dataset-split-names-from-streaming", + [], + ["/split-names-from-streaming"], + ["/config-names", "/split-names-from-streaming"], + ), @@ -88,0 +104,4 @@ def graph() -> ProcessingGraph: + [ + "/split-names-from-dataset-info", + "/split-names-from-streaming", + ], @@ -99,0 +119 @@ def graph() -> ProcessingGraph: + ["config-parquet"], @@ -104,0 +125,4 @@ def graph() -> ProcessingGraph: + [ + "/split-names-from-streaming", + "/split-names-from-dataset-info", + ], @@ -113 +137 @@ def graph() -> ProcessingGraph: - ("/parquet-and-dataset-info", [], []), + ("/parquet-and-dataset-info", [], [], []), @@ -116,0 +141 @@ def graph() -> ProcessingGraph: + ["config-parquet-and-info"], @@ -119 +144 @@ def graph() -> ProcessingGraph: - ("dataset-parquet", [], ["/config-names", "config-parquet-and-info", "config-parquet"]), + ("dataset-parquet", [], ["config-parquet"], ["/config-names", "config-parquet-and-info", "config-parquet"]), @@ -122,0 +148 @@ def graph() -> ProcessingGraph: + ["config-parquet-and-info"], @@ -125,3 +151,3 @@ def graph() -> ProcessingGraph: - ("dataset-info", [], ["/config-names", "config-parquet-and-info", "config-info"]), - ("config-size", ["dataset-size"], ["/config-names", "config-parquet-and-info"]), - ("dataset-size", [], ["/config-names", "config-parquet-and-info", "config-size"]), + ("dataset-info", [], ["config-info"], ["/config-names", "config-parquet-and-info", "config-info"]), + ("config-size", ["dataset-size"], ["config-parquet-and-info"], ["/config-names", "config-parquet-and-info"]), + ("dataset-size", [], ["config-size"], ["/config-names", "config-parquet-and-info", "config-size"]), @@ -130,0 +157,5 @@ def graph() -> ProcessingGraph: + [ + "dataset-split-names", + "split-first-rows-from-parquet", + "split-first-rows-from-streaming", + ], @@ -146 +177 @@ def test_default_graph_steps( - graph: ProcessingGraph, step_name: str, children: List[str], ancestors: List[str] + graph: ProcessingGraph, step_name: str, children: List[str], parents: List[str], ancestors: List[str] @@ -150,0 +182 @@ def test_default_graph_steps( + parents=[graph.get_step(parent) for parent in parents], diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index c8391532..1c6b699c 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -658 +658 @@ class EntrySpec(TypedDict): - progress: float + progress: Optional[float] @@ -679,4 +679,4 @@ class EntrySpec(TypedDict): - # - if no response is found, a `~libcommon.simple_cache.DoesNotExist` error is raised - ([], ["kind1"], "dataset", None, None), - (["ok_config1"], ["kind1"], "dataset", None, None), - (["ok1"], ["kind1"], "dataset", "config", None), + # - if no response is found, an error response is returned + ([], ["kind1"], "dataset", None, "cache_miss"), + (["ok_config1"], ["kind1"], "dataset", None, "cache_miss"), + (["ok1"], ["kind1"], "dataset", "config", "cache_miss"), @@ -686 +686 @@ def test_get_best_response( - selected_entries: List[str], kinds: List[str], dataset: str, config: Optional[str], best_entry: Optional[str] + selected_entries: List[str], kinds: List[str], dataset: str, config: Optional[str], best_entry: str @@ -738,0 +739,7 @@ def test_get_best_response( + "cache_miss": { + "kind": "kind1", + "dataset": "dataset", + "config": None, + "http_status": HTTPStatus.NOT_FOUND, + "progress": None, + }, @@ -748 +755 @@ def test_get_best_response( - content={"entry": entry}, + content={"error": "some_error"} if (entries[entry]["http_status"] >= HTTPStatus.BAD_REQUEST.value) else {}, @@ -752,4 +758,0 @@ def test_get_best_response( - if best_entry is None: - with pytest.raises(DoesNotExist): - get_best_response(kinds, dataset, config) - return @@ -760 +763,3 @@ def test_get_best_response( - assert best_response.response["content"]["entry"] == best_entry + assert ("error" in best_response.response["content"]) is ( + entries[best_entry]["http_status"] >= HTTPStatus.BAD_REQUEST.value + ) diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py index 5d87af38..f0d56984 100644 --- a/libs/libcommon/tests/test_state.py +++ b/libs/libcommon/tests/test_state.py @@ -10,0 +11,8 @@ from libcommon.config import ProcessingGraphConfig +from libcommon.constants import ( + PROCESSING_STEP_CONFIG_INFO_VERSION, + PROCESSING_STEP_CONFIG_NAMES_VERSION, + PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, + PROCESSING_STEP_CONFIG_PARQUET_VERSION, + PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, +) @@ -608 +616 @@ def test_plan_job_creation_and_termination() -> None: - finish_job("/config-names", TWO_CONFIG_NAMES_CONTENT_OK) + finish_job("/config-names", TWO_CONFIG_NAMES_CONTENT_OK, job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION) @@ -681 +689 @@ def test_plan_only_one_config() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, @@ -692 +700 @@ def test_plan_only_one_config() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, @@ -763 +771 @@ def test_plan_retry_error() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, @@ -831 +839 @@ def test_plan_incoherent_state() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, @@ -844 +852 @@ def test_plan_incoherent_state() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, @@ -918 +926 @@ def test_plan_get_splits() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, @@ -929 +937 @@ def test_plan_get_splits() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, @@ -940 +948 @@ def test_plan_get_splits() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_INFO_VERSION, @@ -951 +959 @@ def test_plan_get_splits() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, @@ -962 +970 @@ def test_plan_get_splits() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, @@ -1040 +1048 @@ def test_plan_updated_at() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, @@ -1051 +1059 @@ def test_plan_updated_at() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, @@ -1062 +1070 @@ def test_plan_updated_at() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_INFO_VERSION, @@ -1073 +1081 @@ def test_plan_updated_at() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, @@ -1084 +1092 @@ def test_plan_updated_at() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, @@ -1096 +1104 @@ def test_plan_updated_at() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, @@ -1173 +1181 @@ def test_plan_job_runner_version() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, @@ -1184 +1192 @@ def test_plan_job_runner_version() -> None: - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, @@ -1195 +1203 @@ def test_plan_job_runner_version() -> None: - job_runner_version=OLD_JOB_RUNNER_VERSION, # <- current version is 3, this one is 1 + job_runner_version=PROCESSING_STEP_CONFIG_PARQUET_VERSION - 1, # <- current version is 3, this one is 1 @@ -1275 +1283 @@ def test_plan_git_revision( - job_runner_version=NEW_JOB_RUNNER_VERSION, + job_runner_version=PROCESSING_STEP_CONFIG_NAMES_VERSION, diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index 7c963ca3..c6d14f91 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -8 +8 @@ from http import HTTPStatus -from typing import Any, Literal, Mapping, Optional +from typing import Any, List, Literal, Mapping, Optional @@ -15,0 +16,2 @@ from libcommon.simple_cache import ( + BestResponse, + CacheEntry, @@ -18,0 +21 @@ from libcommon.simple_cache import ( + get_best_response, @@ -198 +201 @@ class ResponseAlreadyComputedError(GeneralJobRunnerError): - """Raised when reponse has been already computed by another job runner.""" + """Raised when response has been already computed by another job runner.""" @@ -209,0 +213,37 @@ class ResponseAlreadyComputedError(GeneralJobRunnerError): +class PreviousStepError(Exception): + """Raised when the previous step failed. It contains the contents of the error response, + and the details contain extra information about the previous step. + """ + + def __init__( + self, response: CacheEntry, kind: str, dataset: str, config: Optional[str] = None, split: Optional[str] = None + ): + super().__init__("Error in previous step") + if response.get("http_status") == HTTPStatus.OK or response["content"].get("error") is None: + raise ValueError("Cannot create a PreviousStepError, the response should contain an error") + + self.content = response["content"] + self.http_status = response["http_status"] + self.error_code = response["error_code"] + self.details = { + "message": "The previous step failed", + "previous_step": {"kind": kind, "dataset": dataset, "config": config, "split": split}, + } + + +def get_previous_step_or_raise( + kinds: List[str], dataset: str, config: Optional[str] = None, split: Optional[str] = None +) -> BestResponse: + """Get the previous step from the cache, or raise an exception if it failed.""" + best_response = get_best_response(kinds=kinds, dataset=dataset, config=config, split=split) + if best_response.response["http_status"] != HTTPStatus.OK: + raise PreviousStepError( + response=best_response.response, + kind=best_response.kind, + dataset=dataset, + config=config, + split=split, + ) + return best_response + + @@ -417 +457,11 @@ class JobRunner(ABC): - e = err if isinstance(err, CustomError) else UnexpectedError(str(err), err) + if isinstance(err, PreviousStepError): + content = err.content + http_status = err.http_status + error_code = err.error_code + details: Mapping[str, Any] = err.details + else: + e = err if isinstance(err, CustomError) else UnexpectedError(str(err), err) + content = dict(e.as_response()) + http_status = e.status_code + error_code = e.code + details = dict(e.as_response_with_cause()) @@ -423,4 +473,4 @@ class JobRunner(ABC): - content=dict(e.as_response()), - http_status=e.status_code, - error_code=e.code, - details=dict(e.as_response_with_cause()), + content=content, + http_status=http_status, + error_code=error_code, + details=details, diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py index b9f4650d..662dbfec 100644 --- a/services/worker/src/worker/job_runners/config/size.py +++ b/services/worker/src/worker/job_runners/config/size.py @@ -9,2 +9 @@ from libcommon.constants import PROCESSING_STEP_CONFIG_SIZE_VERSION -from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import SplitFullName @@ -16,0 +16 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -19,4 +19 @@ from worker.job_runner import ( -ConfigSizeJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] +ConfigSizeJobRunnerErrorCode = Literal["PreviousStepFormatError"] @@ -70,7 +66,0 @@ class ConfigSizeJobRunnerError(JobRunnerError): -class PreviousStepStatusError(ConfigSizeJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -97 +87 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - - [`~job_runners.config.size.PreviousStepStatusError`] + - [`~job_runner.PreviousStepError`] @@ -101,3 +90,0 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - - [`~libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the - token does not give the sufficient access to the dataset, or if the dataset is private - (private datasets are not supported by the datasets server) @@ -108,9 +95,4 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - previous_step = "config-parquet-and-info" - try: - response = get_response(kind=previous_step, dataset=dataset, config=config) - except DoesNotExist as e: - raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError(f"Previous step {previous_step} gave an error: {response['http_status']}..") - - content = response["content"] + dataset_info_best_response = get_previous_step_or_raise( + kinds=["config-parquet-and-info"], dataset=dataset, config=config + ) + content = dataset_info_best_response.response["content"] @@ -119 +100,0 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index 81283100..ad9dcd7e 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -88 +88 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - If reponse has been already computed by /split-names-from-streaming job runner. + If response has been already computed by /split-names-from-streaming job runner. diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index ca798ff9..d40c275a 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -99 +99 @@ def compute_split_names_from_streaming_response( - If reponse has been already computed by /split-names-from-dataset-info job runner. + If response has been already computed by /split-names-from-dataset-info job runner. diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py index ab01427a..b9e1b366 100644 --- a/services/worker/src/worker/job_runners/dataset/info.py +++ b/services/worker/src/worker/job_runners/dataset/info.py @@ -9 +8,0 @@ from libcommon.constants import PROCESSING_STEP_DATASET_INFO_VERSION -from libcommon.dataset import DatasetNotFoundError @@ -16,0 +16 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -20,4 +20 @@ from worker.utils import PreviousJob -DatasetInfoJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] +DatasetInfoJobRunnerErrorCode = Literal["PreviousStepFormatError"] @@ -48,7 +44,0 @@ class DatasetInfoJobRunnerError(JobRunnerError): -class PreviousStepStatusError(DatasetInfoJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -76 +66 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - - [`~job_runners.dataset.info.PreviousStepStatusError`] + - [`~job_runner.PreviousStepError`] @@ -80,4 +69,0 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - - [`~libcommon.dataset.DatasetNotFoundError`] - If the dataset does not exist, or if the - token does not give the sufficient access to the dataset, or if the dataset is private - (private datasets are not supported by the datasets server) @@ -88,9 +74,2 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - try: - response = get_response(kind="/config-names", dataset=dataset) - except DoesNotExist as e: - raise DatasetNotFoundError("No response for '/config-names' found for this dataset: .", e) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step raised an error: {response['http_status']}. This job should not have been created." - ) - content = response["content"] + config_names_best_response = get_previous_step_or_raise(kinds=["/config-names"], dataset=dataset) + content = config_names_best_response.response["content"] @@ -98 +77 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + raise PreviousStepFormatError("Previous step did not return the expected content: 'config_names'.") @@ -121 +100 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - logging.debug(f"Previous step gave an error: {response['http_status']}") + logging.debug(f"Previous step gave an error: {config_response['http_status']}") diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index 4163d564..e8420667 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -9 +8,0 @@ from libcommon.constants import PROCESSING_STEP_DATASET_PARQUET_VERSION -from libcommon.dataset import DatasetNotFoundError @@ -16,0 +16 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -22,4 +22 @@ from worker.utils import PreviousJob -SizesJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] +SizesJobRunnerErrorCode = Literal["PreviousStepFormatError"] @@ -50,7 +46,0 @@ class DatasetParquetJobRunnerError(JobRunnerError): -class PreviousStepStatusError(DatasetParquetJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -75 +65 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - - [`~job_runners.dataset.parquet.PreviousStepStatusError`] + - [`~job_runner.PreviousStepError`] @@ -83,9 +73,2 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - try: - response = get_response(kind="/config-names", dataset=dataset) - except DoesNotExist as e: - raise DatasetNotFoundError("No response for '/config-names' found for this dataset: .", e) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step raised an error: {response['http_status']}. This job should not have been created." - ) - content = response["content"] + config_names_best_response = get_previous_step_or_raise(kinds=["/config-names"], dataset=dataset) + content = config_names_best_response.response["content"] @@ -93 +76 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + raise PreviousStepFormatError("Previous step did not return the expected content: 'config_names'.") diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py index 958807bb..250d3cfc 100644 --- a/services/worker/src/worker/job_runners/dataset/size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -9 +8,0 @@ from libcommon.constants import PROCESSING_STEP_DATASET_SIZE_VERSION -from libcommon.dataset import DatasetNotFoundError @@ -16,0 +16 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -21,4 +21 @@ from worker.utils import PreviousJob -SizesJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] +SizesJobRunnerErrorCode = Literal["PreviousStepFormatError"] @@ -63,7 +59,0 @@ class DatasetSizeJobRunnerError(JobRunnerError): -class PreviousStepStatusError(DatasetSizeJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -88 +78 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - - [`~job_runners.dataset.size.PreviousStepStatusError`] + - [`~job_runner.PreviousStepError`] @@ -96,9 +86,2 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - try: - response = get_response(kind="/config-names", dataset=dataset) - except DoesNotExist as e: - raise DatasetNotFoundError("No response for '/config-names' found for this dataset: .", e) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step raised an error: {response['http_status']}. This job should not have been created." - ) - content = response["content"] + config_names_best_response = get_previous_step_or_raise(kinds=["/config-names"], dataset=dataset) + content = config_names_best_response.response["content"] @@ -106 +89 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + raise PreviousStepFormatError("Previous step did not return the expected content: 'config_names'.") diff --git a/services/worker/src/worker/job_runners/dataset/split_names.py b/services/worker/src/worker/job_runners/dataset/split_names.py index c2087b02..bf0dfcfb 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names.py +++ b/services/worker/src/worker/job_runners/dataset/split_names.py @@ -9,2 +9 @@ from libcommon.constants import PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION -from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_best_response +from libcommon.simple_cache import SplitFullName, get_best_response @@ -12 +11,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_best_respons -from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + get_previous_step_or_raise, +) @@ -20,4 +24 @@ from worker.utils import ( -DatasetSplitNamesErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] +DatasetSplitNamesErrorCode = Literal["PreviousStepFormatError"] @@ -42,7 +42,0 @@ class DatasetSplitNamesJobRunnerError(JobRunnerError): -class PreviousStepStatusError(DatasetSplitNamesJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -68 +62 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - - [`~job_runners.dataset_split_names.PreviousStepStatusError`] + - [`~job_runner.PreviousStepError`] @@ -72,2 +65,0 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - - [`~libcommon.dataset.DatasetNotFoundError`] - If previous step content was not found for the dataset @@ -79,17 +71,12 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - try: - best_response = get_best_response(["/config-names", "dataset-info"], dataset) - if best_response.kind == "/config-names": - config_names = [ - config_name_item["config"] for config_name_item in best_response.response["content"]["config_names"] - ] - elif best_response.kind == "dataset-info": - config_names = list(best_response.response["content"]["dataset-info"].keys()) - else: - raise PreviousStepFormatError( - "Previous step '/config-names' or 'dataset-info' did not return the expected content." - ) - except DoesNotExist as e: - raise DatasetNotFoundError( - "No response found in previous step for this dataset: '/config-names' or 'dataset-info'.", e - ) from e - except KeyError as e: + + config_names_best_response = get_previous_step_or_raise(kinds=["/config-names", "dataset-info"], dataset=dataset) + content = config_names_best_response.response["content"] + if config_names_best_response.kind == "/config-names": + if "config_names" not in content: + raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + config_names = [config_name_item["config"] for config_name_item in content["config_names"]] + elif config_names_best_response.kind == "dataset-info": + if "dataset_info" not in content: + raise PreviousStepFormatError("'dataset-info' did not return the expected content: 'dataset_info'.") + config_names = list(content["dataset-info"].keys()) + else: @@ -97,2 +84,2 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - "Previous steps '/config-names' or 'dataset-info' did not return the expected content.", e - ) from e + "Previous step '/config-names' or 'dataset-info' did not return the expected content." + ) @@ -112,3 +99,2 @@ def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitName - try: - best_response = get_best_response(split_names_cache_kinds, dataset=dataset, config=config) - except DoesNotExist: + best_response = get_best_response(split_names_cache_kinds, dataset=dataset, config=config) + if best_response.response["error_code"] == "CachedResponseNotFound": diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py index a93d2798..809cbb11 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py @@ -11 +10,0 @@ from libcommon.constants import ( -from libcommon.dataset import DatasetNotFoundError @@ -18,0 +18 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -27,4 +27 @@ from worker.utils import ( -DatasetSplitNamesFromDatasetInfoErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] +DatasetSplitNamesFromDatasetInfoErrorCode = Literal["PreviousStepFormatError"] @@ -49,7 +45,0 @@ class DatasetSplitNamesFromDatasetInfoJobRunnerError(JobRunnerError): -class PreviousStepStatusError(DatasetSplitNamesFromDatasetInfoJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -75 +65 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - - [`~job_runners.dataset.split_names_from_dataset_info.PreviousStepStatusError`] + - [`~job_runner.PreviousStepError`] @@ -79,2 +68,0 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - - [`~libcommon.dataset.DatasetNotFoundError`] - If previous step content was not found for the dataset @@ -85,12 +73,5 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - try: - response = get_response(kind="dataset-info", dataset=dataset) - dataset_info_content = response["content"]["dataset_info"] - except DoesNotExist as e: - raise DatasetNotFoundError("No response found in previous step for this dataset: 'dataset-info'.", e) from e - except KeyError as e: - raise PreviousStepFormatError("Previous step 'dataset-info' did not return the expected content.") from e - - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step gave an error: {response['http_status']}. This job should not have been created." - ) + best_dataset_info_response = get_previous_step_or_raise(kinds=["dataset-info"], dataset=dataset) + content = best_dataset_info_response.response["content"] + if "dataset_info" not in content: + raise PreviousStepFormatError("Previous step did not return the expected content: 'dataset_info'.") + dataset_info_content = content["dataset_info"] diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py index f486683a..8db0d3ff 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py @@ -11 +10,0 @@ from libcommon.constants import ( -from libcommon.dataset import DatasetNotFoundError @@ -18,0 +18 @@ from worker.job_runner import ( + get_previous_step_or_raise, @@ -22,3 +22 @@ from worker.utils import ConfigItem, SplitItem -DatasetSplitNamesFromStreamingJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", "PreviousStepFormatError", "ResponseNotReady" -] +DatasetSplitNamesFromStreamingJobRunnerErrorCode = Literal["PreviousStepFormatError", "ResponseNotReady"] @@ -43,7 +40,0 @@ class DatasetSplitNamesFromStreamingJobRunnerError(JobRunnerError): -class PreviousStepStatusError(DatasetSplitNamesFromStreamingJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - @@ -89 +80 @@ def compute_dataset_split_names_from_streaming_response( - - [`~job_runners.dataset.split_names_from_streaming.PreviousStepStatusError`] + - [`~job_runner.PreviousStepError`] @@ -93,2 +83,0 @@ def compute_dataset_split_names_from_streaming_response( - - [`~libcommon.dataset.DatasetNotFoundError`] - If previous step content was not found for the dataset @@ -98,12 +87,7 @@ def compute_dataset_split_names_from_streaming_response( - try: - config_names = get_response(kind="/config-names", dataset=dataset) - config_content = config_names["content"]["config_names"] - except DoesNotExist as e: - raise DatasetNotFoundError("No response found in previous step '/config-names' for this dataset.", e) from e - except KeyError as e: - raise PreviousStepFormatError("Previous step '/config-names' did not return the expected content.") from e - - if config_names["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step gave an error: {config_names['http_status']}. This job should not have been created." - ) + + config_names_best_response = get_previous_step_or_raise(kinds=["/config-names"], dataset=dataset) + content = config_names_best_response.response["content"] + if "config_names" not in content: + raise PreviousStepFormatError("Previous step did not return the expected content: 'config_names'.") + config_content = content["config_names"] + diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py index 4de49942..23208685 100644 --- a/services/worker/tests/job_runners/config/test_size.py +++ b/services/worker/tests/job_runners/config/test_size.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -15,5 +14,2 @@ from worker.config import AppConfig -from worker.job_runners.config.size import ( - ConfigSizeJobRunner, - PreviousStepFormatError, - PreviousStepStatusError, -) +from worker.job_runner import PreviousStepError +from worker.job_runners.config.size import ConfigSizeJobRunner, PreviousStepFormatError @@ -167 +163 @@ def get_job_runner( - PreviousStepStatusError.__name__, + PreviousStepError.__name__, @@ -212 +208 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py index be2f1f86..58f5002b 100644 --- a/services/worker/tests/job_runners/dataset/test_info.py +++ b/services/worker/tests/job_runners/dataset/test_info.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -14,0 +14 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -18 +17,0 @@ from worker.job_runners.dataset.info import ( - PreviousStepStatusError, @@ -196 +195 @@ def get_job_runner( - PreviousStepStatusError.__name__, + PreviousStepError.__name__, @@ -242 +241 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index fa7a782c..885379e1 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -14,0 +14 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -21 +20,0 @@ from worker.job_runners.dataset.parquet import ( - PreviousStepStatusError, @@ -154 +153 @@ def get_job_runner( - PreviousStepStatusError.__name__, + PreviousStepError.__name__, @@ -198 +197 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/dataset/test_size.py b/services/worker/tests/job_runners/dataset/test_size.py index fbe4c79e..ab3a3074 100644 --- a/services/worker/tests/job_runners/dataset/test_size.py +++ b/services/worker/tests/job_runners/dataset/test_size.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -14,0 +14 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -18 +17,0 @@ from worker.job_runners.dataset.size import ( - PreviousStepStatusError, @@ -252 +251 @@ def get_job_runner( - PreviousStepStatusError.__name__, + PreviousStepError.__name__, @@ -296 +295 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/dataset/test_split_names.py b/services/worker/tests/job_runners/dataset/test_split_names.py index 913c4f4f..8f1046e6 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names.py +++ b/services/worker/tests/job_runners/dataset/test_split_names.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -14,0 +14 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -272 +272 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py index 72bfd8e9..7444910d 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -14,0 +14 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -218 +218 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError): diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py index 8dccf5f3..ec18d17c 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py @@ -8 +7,0 @@ import pytest -from libcommon.dataset import DatasetNotFoundError @@ -14,0 +14 @@ from worker.config import AppConfig +from worker.job_runner import PreviousStepError @@ -240 +240 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - with pytest.raises(DatasetNotFoundError): + with pytest.raises(PreviousStepError):
5dbe4a07257c13b55fc46caf59dfe8ab4f722fc4
Andrea Francis Soria Jimenez
2023-04-18T18:18:26
Change restart policy for metrics collector cron job (#1058)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 577023ad..750c0cbc 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.8.0 +version: 1.9.0 diff --git a/chart/templates/cron-jobs/metrics-collector/job.yaml b/chart/templates/cron-jobs/metrics-collector/job.yaml index 892af386..e57c4e09 100644 --- a/chart/templates/cron-jobs/metrics-collector/job.yaml +++ b/chart/templates/cron-jobs/metrics-collector/job.yaml @@ -17 +17 @@ spec: - restartPolicy: Never + restartPolicy: OnFailure @@ -23 +22,0 @@ spec: - backoffLimit: 0
d7f702e08976cfafceb01b00eccaebc42bebcc6a
Andrea Francis Soria Jimenez
2023-04-18T14:54:04
Fix container for metrics collector (#1057)
diff --git a/chart/templates/cron-jobs/metrics-collector/job.yaml b/chart/templates/cron-jobs/metrics-collector/job.yaml index 130e754e..892af386 100644 --- a/chart/templates/cron-jobs/metrics-collector/job.yaml +++ b/chart/templates/cron-jobs/metrics-collector/job.yaml @@ -19 +19 @@ spec: - containers: {{ include "containerCacheMaintenance" . | nindent 8 }} + containers: {{ include "containerMetricsCollector" . | nindent 8 }}
4c0878f54ac88f38ff11b65fb7d6cfc72f104c97
Sylvain Lesage
2023-04-18T08:23:32
feat: 🎸 get the dataset state, and backfill the missing parts (#1045)
diff --git a/jobs/cache_maintenance/tests/test_backfill_cache.py b/jobs/cache_maintenance/tests/test_backfill_cache.py index c89d8870..8e0d6626 100644 --- a/jobs/cache_maintenance/tests/test_backfill_cache.py +++ b/jobs/cache_maintenance/tests/test_backfill_cache.py @@ -24,0 +25 @@ def test_backfill_cache() -> None: + parents=[], diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py index fffc11d4..1e29e4bd 100644 --- a/jobs/cache_maintenance/tests/test_collect_metrics.py +++ b/jobs/cache_maintenance/tests/test_collect_metrics.py @@ -35,0 +36 @@ def test_collect_metrics() -> None: + parents=[], diff --git a/jobs/cache_maintenance/tests/test_upgrade_cache.py b/jobs/cache_maintenance/tests/test_upgrade_cache.py index 5fb6a24c..34c8caf1 100644 --- a/jobs/cache_maintenance/tests/test_upgrade_cache.py +++ b/jobs/cache_maintenance/tests/test_upgrade_cache.py @@ -33,0 +34 @@ def test_upgrade_cache( + parents=[], diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index 65f33c78..20a7db07 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -46,0 +47 @@ class ProcessingStep: + parents: List[ProcessingStep] @@ -93,0 +95 @@ class ProcessingGraph: + topologically_ordered_steps: List[ProcessingStep] @@ -103,0 +106 @@ class ProcessingGraph: + parents=[], @@ -124 +127 @@ class ProcessingGraph: - parents = list(required_steps) + step.parents = list(required_steps) @@ -128,2 +131,2 @@ class ProcessingGraph: - parents.remove(other_parent_candidate) - for parent in parents: + step.parents.remove(other_parent_candidate) + for parent in step.parents: @@ -132,0 +136 @@ class ProcessingGraph: + self.topologically_ordered_steps = [self.get_step(name) for name in nx.topological_sort(graph)] diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index ae51c3e9..0bb98dce 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -282,6 +282,8 @@ class Queue: - existing = Job.objects(type=job_type, dataset=dataset, config=config, split=split, status=Status.WAITING) - if existing(force=True).count() > 0: - force = True - if existing(priority=Priority.NORMAL).count() > 0: - priority = Priority.NORMAL - existing.update(finished_at=get_datetime(), status=Status.CANCELLED) + canceled_jobs = self.cancel_jobs( + job_type=job_type, dataset=dataset, config=config, split=split, statuses_to_cancel=[Status.WAITING] + ) + for job in canceled_jobs: + if job["force"]: + force = True + if job["priority"] == Priority.NORMAL: + priority = Priority.NORMAL @@ -291,0 +294,33 @@ class Queue: + def cancel_jobs( + self, + job_type: str, + dataset: str, + config: Optional[str] = None, + split: Optional[str] = None, + statuses_to_cancel: Optional[List[Status]] = None, + ) -> List[JobDict]: + """Cancel jobs from the queue. + + Returns the list of canceled jobs (as JobDict, before they are canceled, to be able to know their previous + status) + + Args: + job_type (`str`): The type of the job + dataset (`str`): The dataset on which to apply the job. + config (`str`, optional): The config on which to apply the job. + split (`str`, optional): The config on which to apply the job. + statuses_to_cancel (`list[Status]`, optional): The list of statuses to cancel. Defaults to + [Status.WAITING, Status.STARTED]. + + Returns: + `list[JobDict]`: The list of canceled jobs + """ + if statuses_to_cancel is None: + statuses_to_cancel = [Status.WAITING, Status.STARTED] + existing = Job.objects( + type=job_type, dataset=dataset, config=config, split=split, status__in=statuses_to_cancel + ) + job_dicts = [job.to_dict() for job in existing] + existing.update(finished_at=get_datetime(), status=Status.CANCELLED) + return job_dicts + diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index c08be874..b2d7b1bf 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -186,0 +187,23 @@ def get_response_without_content( +class CacheEntryMetadata(CacheEntryWithoutContent): + updated_at: datetime + + +# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them +def get_response_metadata( + kind: str, dataset: str, config: Optional[str] = None, split: Optional[str] = None +) -> CacheEntryMetadata: + response = ( + CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "updated_at") + .get() + ) + return { + "http_status": response.http_status, + "error_code": response.error_code, + "dataset_git_revision": response.dataset_git_revision, + "job_runner_version": response.job_runner_version, + "progress": response.progress, + "updated_at": response.updated_at, + } + + diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py new file mode 100644 index 00000000..fc2e5a6d --- /dev/null +++ b/libs/libcommon/src/libcommon/state.py @@ -0,0 +1,525 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from __future__ import annotations + +import contextlib +from abc import ABC, abstractmethod +from dataclasses import dataclass, field +from typing import Any, Dict, List, Optional + +from libcommon.processing_graph import ProcessingGraph, ProcessingStep +from libcommon.queue import Priority, Queue, Status +from libcommon.simple_cache import ( + CacheEntryMetadata, + DoesNotExist, + get_best_response, + get_response, + get_response_metadata, +) + +# TODO: use the term Artifact elsewhere in the code (a Step should produce one or several Artifacts, depending on the +# input level: one, one per dataset, one per config, or one per split) +# A job and a cache entry is related to an Artifact, not to a Step +# TODO: assets, cached_assets, parquet files +# TODO: obsolete/dangling cache entries and jobs + +HARD_CODED_CONFIG_NAMES_CACHE_KIND = "/config-names" +HARD_CODED_SPLIT_NAMES_FROM_STREAMING_CACHE_KIND = "/split-names-from-streaming" +HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND = "/split-names-from-dataset-info" + + +def fetch_config_names(dataset: str) -> List[str]: + """Fetch the list of config names from the database.""" + config_names = [] + + response = get_response(HARD_CODED_CONFIG_NAMES_CACHE_KIND, dataset=dataset, config=None, split=None) + for config_name_item in response["content"]["config_names"]: + config_name = config_name_item["config"] + if not isinstance(config_name, str): + raise ValueError(f"Invalid config name: {config_name}, type should be str, got: {type(config_name)}") + config_names.append(config_name) + return config_names + + +def fetch_split_names(dataset: str, config: str) -> List[str]: + """Fetch the list of config names from the database.""" + split_names = [] + + best_response = get_best_response( + [HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND, HARD_CODED_SPLIT_NAMES_FROM_STREAMING_CACHE_KIND], + dataset=dataset, + config=config, + split=None, + ) + for split_name_item in best_response.response["content"]["split_names"]: + split_name = split_name_item["split"] + if not isinstance(split_name, str): + raise ValueError(f"Invalid split name: {split_name}, type should be str, got: {type(split_name)}") + split_names.append(split_name) + return split_names + + +@dataclass +class JobState: + """The state of a job for a given input.""" + + dataset: str + config: Optional[str] + split: Optional[str] + job_type: str + is_in_process: bool = field(init=False) + + def __post_init__(self) -> None: + self.is_in_process = Queue().is_job_in_process( + job_type=self.job_type, dataset=self.dataset, config=self.config, split=self.split + ) + + def as_dict(self) -> Dict[str, Any]: + return { + "is_in_process": self.is_in_process, + } + + +ERROR_CODES_TO_RETRY: List[str] = [] + + +@dataclass +class CacheState: + """The state of a cache entry for a given input.""" + + dataset: str + config: Optional[str] + split: Optional[str] + cache_kind: str + cache_entry_metadata: Optional[CacheEntryMetadata] = field(init=False) + exists: bool = field(init=False) + is_success: bool = field(init=False) + + def __post_init__(self) -> None: + self.cache_entry_metadata = None + with contextlib.suppress(DoesNotExist): + self.cache_entry_metadata = get_response_metadata( + kind=self.cache_kind, dataset=self.dataset, config=self.config, split=self.split + ) + """Whether the cache entry exists.""" + self.exists = self.cache_entry_metadata is not None + self.is_success = self.cache_entry_metadata is not None and self.cache_entry_metadata["http_status"] < 400 + + def as_dict(self) -> Dict[str, Any]: + return { + "exists": self.exists, + "is_success": self.is_success, + } + + def is_empty(self) -> bool: + return self.cache_entry_metadata is None + + def is_error_to_retry(self) -> bool: + return self.cache_entry_metadata is not None and ( + self.cache_entry_metadata["http_status"] >= 400 + and self.cache_entry_metadata["error_code"] in ERROR_CODES_TO_RETRY + ) + + def is_older_than(self, other: "CacheState") -> bool: + if self.cache_entry_metadata is None or other.cache_entry_metadata is None: + return False + return self.cache_entry_metadata["updated_at"] < other.cache_entry_metadata["updated_at"] + + def is_git_revision_different_from(self, git_revision: Optional[str]) -> bool: + return self.cache_entry_metadata is None or self.cache_entry_metadata["dataset_git_revision"] != git_revision + + +@dataclass +class ArtifactState: + """The state of an artifact.""" + + step: ProcessingStep + dataset: str + config: Optional[str] + split: Optional[str] + + job_state: JobState = field(init=False) + cache_state: CacheState = field(init=False) + + def __post_init__(self) -> None: + if self.step.input_type == "dataset": + if self.config is not None or self.split is not None: + raise ValueError("Step input type is dataset, but config or split is not None") + elif self.step.input_type == "config": + if self.config is None or self.split is not None: + raise ValueError("Step input type is config, but config is None or split is not None") + elif self.step.input_type == "split": + if self.config is None or self.split is None: + raise ValueError("Step input type is split, but config or split is None") + else: + raise ValueError(f"Invalid step input type: {self.step.input_type}") + self.id = ",".join([p for p in (self.step.name, self.dataset, self.config, self.split) if p]) + + self.job_state = JobState( + job_type=self.step.job_type, dataset=self.dataset, config=self.config, split=self.split + ) + self.cache_state = CacheState( + cache_kind=self.step.cache_kind, dataset=self.dataset, config=self.config, split=self.split + ) + + def get_parent_artifact_states(self, parent_step: ProcessingStep) -> List["ArtifactState"]: + if parent_step.input_type == "dataset": + return [ArtifactState(step=parent_step, dataset=self.dataset, config=None, split=None)] + elif parent_step.input_type == "config": + return ( + [ + ArtifactState( + step=parent_step, + dataset=self.dataset, + config=self.config, + split=None, + ) + ] + if self.step.input_type in ["config", "split"] + else [] + # ^ fan-in: config->dataset. For now, we don't return the list of parent artifact states in that case + ) + else: + return ( + [ + ArtifactState( + step=parent_step, + dataset=self.dataset, + config=self.config, + split=self.split, + ) + ] + if self.step.input_type == "split" + else [] + # ^ fan-in: split->config, or split->dataset. For now, we don't return the list of parent artifact + # states in that case + ) + + def is_job_runner_obsolete(self) -> bool: + if self.cache_state.cache_entry_metadata is None: + return False + job_runner_version = self.cache_state.cache_entry_metadata["job_runner_version"] + if job_runner_version is None: + return True + return job_runner_version < self.step.job_runner_version + + def get_all_parents_artifact_states(self) -> List[List["ArtifactState"]]: + return [self.get_parent_artifact_states(parent_step) for parent_step in self.step.parents] + + def as_dict(self) -> Dict[str, Any]: + return { + "id": self.id, + "job_state": self.job_state.as_dict(), + "cache_state": self.cache_state.as_dict(), + } + + +@dataclass +class SplitState: + """The state of a split.""" + + dataset: str + config: str + split: str + processing_graph: ProcessingGraph + + artifact_state_by_step: Dict[str, ArtifactState] = field(init=False) + + def __post_init__(self) -> None: + self.artifact_state_by_step = { + step.name: ArtifactState(step=step, dataset=self.dataset, config=self.config, split=self.split) + for step in self.processing_graph.steps.values() + if step.input_type == "split" + } + + def as_dict(self) -> Dict[str, Any]: + return { + "split": self.split, + "artifact_states": [artifact_state.as_dict() for artifact_state in self.artifact_state_by_step.values()], + } + + +@dataclass +class ConfigState: + """The state of a config.""" + + dataset: str + config: str + processing_graph: ProcessingGraph + + split_names: List[str] = field(init=False) + split_states: List[SplitState] = field(init=False) + artifact_state_by_step: Dict[str, ArtifactState] = field(init=False) + + def __post_init__(self) -> None: + self.artifact_state_by_step = { + step.name: ArtifactState(step=step, dataset=self.dataset, config=self.config, split=None) + for step in self.processing_graph.steps.values() + if step.input_type == "config" + } + + try: + self.split_names = fetch_split_names(self.dataset, self.config) + except Exception: + self.split_names = [] + + self.split_states = [ + SplitState(self.dataset, self.config, split_name, processing_graph=self.processing_graph) + for split_name in self.split_names + ] + + def as_dict(self) -> Dict[str, Any]: + return { + "config": self.config, + "split_states": [split_state.as_dict() for split_state in self.split_states], + "artifact_states": [artifact_state.as_dict() for artifact_state in self.artifact_state_by_step.values()], + } + + +@dataclass +class CacheStatus: + blocked_by_parent: Dict[str, ArtifactState] = field(default_factory=dict) + cache_has_different_git_revision: Dict[str, ArtifactState] = field(default_factory=dict) + cache_is_outdated_by_parent: Dict[str, ArtifactState] = field(default_factory=dict) + cache_is_empty: Dict[str, ArtifactState] = field(default_factory=dict) + cache_is_error_to_retry: Dict[str, ArtifactState] = field(default_factory=dict) + cache_is_job_runner_obsolete: Dict[str, ArtifactState] = field(default_factory=dict) + up_to_date: Dict[str, ArtifactState] = field(default_factory=dict) + + def as_response(self) -> Dict[str, List[str]]: + return { + "blocked_by_parent": sorted(self.blocked_by_parent.keys()), + "cache_has_different_git_revision": sorted(self.cache_has_different_git_revision.keys()), + "cache_is_outdated_by_parent": sorted(self.cache_is_outdated_by_parent.keys()), + "cache_is_empty": sorted(self.cache_is_empty.keys()), + "cache_is_error_to_retry": sorted(self.cache_is_error_to_retry.keys()), + "cache_is_job_runner_obsolete": sorted(self.cache_is_job_runner_obsolete.keys()), + "up_to_date": sorted(self.up_to_date.keys()), + } + + +@dataclass +class QueueStatus: + in_process: Dict[str, ArtifactState] = field(default_factory=dict) + + def as_response(self) -> Dict[str, List[str]]: + return {"in_process": sorted(self.in_process.keys())} + + +@dataclass +class Task(ABC): + artifact_state: ArtifactState + + id: str = field(init=False) + + @abstractmethod + def run(self) -> None: + pass + + +@dataclass +class CreateJobTask(Task): + force: bool + priority: Priority + + def __post_init__(self) -> None: + self.id = f"CreateJob[{self.artifact_state.id}]" + + def run(self) -> None: + Queue().upsert_job( + job_type=self.artifact_state.step.job_type, + dataset=self.artifact_state.dataset, + config=self.artifact_state.config, + split=self.artifact_state.split, + force=self.force, + priority=self.priority, + ) + + +@dataclass +class DeleteJobTask(Task): + def __post_init__(self) -> None: + self.id = f"DeleteJob[{self.artifact_state.id}]" + + def run(self) -> None: + # TODO: the started jobs are also canceled: we need to ensure the job runners will + # not try to update the cache when they finish + Queue().cancel_jobs( + job_type=self.artifact_state.step.job_type, + dataset=self.artifact_state.dataset, + config=self.artifact_state.config, + split=self.artifact_state.split, + statuses_to_cancel=[Status.WAITING, Status.STARTED], + ) + + +@dataclass +class Plan: + tasks: List[Task] = field(default_factory=list) + + def add(self, task: Task) -> None: + self.tasks.append(task) + + def run(self) -> None: + for task in self.tasks: + task.run() + + def as_response(self) -> List[str]: + return sorted(task.id for task in self.tasks) + + +@dataclass +class DatasetState: + """The state of a dataset.""" + + dataset: str + processing_graph: ProcessingGraph + revision: Optional[str] + + config_names: List[str] = field(init=False) + config_states: List[ConfigState] = field(init=False) + artifact_state_by_step: Dict[str, ArtifactState] = field(init=False) + cache_status: CacheStatus = field(init=False) + queue_status: QueueStatus = field(init=False) + plan: Plan = field(init=False) + + def __post_init__(self) -> None: + self.artifact_state_by_step = { + step.name: ArtifactState(step=step, dataset=self.dataset, config=None, split=None) + for step in self.processing_graph.steps.values() + if step.input_type == "dataset" + } + try: + self.config_names = fetch_config_names(self.dataset) + except Exception: + self.config_names = [] + self.config_states = [ + ConfigState(dataset=self.dataset, config=config_name, processing_graph=self.processing_graph) + for config_name in self.config_names + ] + self.cache_status = self._get_cache_status() + self.queue_status = self._get_queue_status() + self.plan = self._create_plan() + + def _get_artifact_states_for_step(self, step: ProcessingStep) -> List[ArtifactState]: + if step.input_type == "dataset": + artifact_states = [self.artifact_state_by_step[step.name]] + elif step.input_type == "config": + artifact_states = [config_state.artifact_state_by_step[step.name] for config_state in self.config_states] + elif step.input_type == "split": + artifact_states = [ + split_state.artifact_state_by_step[step.name] + for config_state in self.config_states + for split_state in config_state.split_states + ] + else: + raise ValueError(f"Invalid input type: {step.input_type}") + artifact_states_ids = {artifact_state.id for artifact_state in artifact_states} + if len(artifact_states_ids) != len(artifact_states): + raise ValueError(f"Duplicate artifact states for step {step.name}") + return artifact_states + + def _get_cache_status(self) -> CacheStatus: + cache_status = CacheStatus() + + for step in self.processing_graph.topologically_ordered_steps: + artifact_states = self._get_artifact_states_for_step(step) + for artifact_state in artifact_states: + # (fan-in steps: config -> dataset, split -> config, split -> dataset) + # what should we do? always recompute? test the progress? + all_not_none_parents_artifact_states = [ + x for x in artifact_state.get_all_parents_artifact_states() if x is not None + ] + + # blocked by a parent? + if any( + parent_artifact_state.id not in cache_status.up_to_date + for all_parents_artifact_state in all_not_none_parents_artifact_states + for parent_artifact_state in all_parents_artifact_state + ): + cache_status.blocked_by_parent[artifact_state.id] = artifact_state + continue + + # any of the parents is more recent? + if any( + artifact_state.cache_state.is_older_than(parent_artifact_state.cache_state) + for all_parents_artifact_state in all_not_none_parents_artifact_states + for parent_artifact_state in all_parents_artifact_state + ): + cache_status.cache_is_outdated_by_parent[artifact_state.id] = artifact_state + continue + + # is empty? + if artifact_state.cache_state.is_empty(): + cache_status.cache_is_empty[artifact_state.id] = artifact_state + continue + + # is an error that can be retried? + if artifact_state.cache_state.is_error_to_retry(): + cache_status.cache_is_error_to_retry[artifact_state.id] = artifact_state + continue + + # was created with an obsolete version of the job runner? + if artifact_state.is_job_runner_obsolete(): + cache_status.cache_is_job_runner_obsolete[artifact_state.id] = artifact_state + continue + + # has a different git revision from the dataset current revision? + if artifact_state.cache_state.is_git_revision_different_from(self.revision): + cache_status.cache_has_different_git_revision[artifact_state.id] = artifact_state + continue + + # ok + cache_status.up_to_date[artifact_state.id] = artifact_state + + return cache_status + + def _get_queue_status(self) -> QueueStatus: + queue_status = QueueStatus() + + for step in self.processing_graph.topologically_ordered_steps: + artifact_states = self._get_artifact_states_for_step(step) + for artifact_state in artifact_states: + if artifact_state.job_state.is_in_process: + queue_status.in_process[artifact_state.id] = artifact_state + + return queue_status + + def _create_plan(self) -> Plan: + plan = Plan() + remaining_in_process_artifact_state_ids = list(self.queue_status.in_process.keys()) + artifact_states = ( + list(self.cache_status.cache_is_empty.values()) + + list(self.cache_status.cache_is_error_to_retry.values()) + + list(self.cache_status.cache_is_outdated_by_parent.values()) + + list(self.cache_status.cache_is_job_runner_obsolete.values()) + + list(self.cache_status.cache_has_different_git_revision.values()) + ) + for artifact_state in artifact_states: + if artifact_state.id in remaining_in_process_artifact_state_ids: + # the job already exists + remaining_in_process_artifact_state_ids.remove(artifact_state.id) + continue + plan.add(CreateJobTask(artifact_state=artifact_state, force=True, priority=Priority.LOW)) + for artifact_state_id in remaining_in_process_artifact_state_ids: + plan.add(DeleteJobTask(artifact_state=self.queue_status.in_process[artifact_state_id])) + return plan + + def backfill(self) -> None: + self.plan.run() + + def as_dict(self) -> Dict[str, Any]: + return { + "dataset": self.dataset, + "config_states": [config_state.as_dict() for config_state in self.config_states], + "artifact_states": [artifact_state.as_dict() for artifact_state in self.artifact_state_by_step.values()], + } + + def as_response(self) -> Dict[str, Any]: + return { + "dataset": self.dataset, + "cache_status": self.cache_status.as_response(), + "queue_status": self.queue_status.as_response(), + "plan": self.plan.as_response(), + } diff --git a/libs/libcommon/tests/conftest.py b/libs/libcommon/tests/conftest.py index 0c30c347..0147594e 100644 --- a/libs/libcommon/tests/conftest.py +++ b/libs/libcommon/tests/conftest.py @@ -3,0 +4 @@ from pathlib import Path +from typing import Iterator @@ -7,0 +9,3 @@ from pytest import fixture +from libcommon.queue import _clean_queue_database +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import _clean_cache_database @@ -55,0 +60,24 @@ def cached_assets_directory(tmp_path: Path) -> StrPath: + + +@fixture +def queue_mongo_resource(queue_mongo_host: str) -> Iterator[QueueMongoResource]: + database = "datasets_server_queue_test" + host = queue_mongo_host + if "test" not in database: + raise ValueError("Test must be launched on a test mongo database") + with QueueMongoResource(database=database, host=host, server_selection_timeout_ms=3_000) as queue_mongo_resource: + if not queue_mongo_resource.is_available(): + raise RuntimeError("Mongo resource is not available") + yield queue_mongo_resource + _clean_queue_database() + + +@fixture +def cache_mongo_resource(cache_mongo_host: str) -> Iterator[CacheMongoResource]: + database = "datasets_server_cache_test" + host = cache_mongo_host + if "test" not in database: + raise ValueError("Test must be launched on a test mongo database") + with CacheMongoResource(database=database, host=host) as cache_mongo_resource: + yield cache_mongo_resource + _clean_cache_database() diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 3db33a5a..67a0f6a2 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -6 +6 @@ from datetime import datetime, timedelta -from typing import Iterator, Optional +from typing import List, Optional @@ -12,7 +12 @@ import pytz -from libcommon.queue import ( - EmptyQueueError, - Priority, - Queue, - Status, - _clean_queue_database, -) +from libcommon.queue import EmptyQueueError, Priority, Queue, Status @@ -28,10 +22,2 @@ def get_old_datetime() -> datetime: -def queue_mongo_resource(queue_mongo_host: str) -> Iterator[QueueMongoResource]: - database = "datasets_server_queue_test" - host = queue_mongo_host - if "test" not in database: - raise ValueError("Test must be launched on a test mongo database") - with QueueMongoResource(database=database, host=host, server_selection_timeout_ms=3_000) as queue_mongo_resource: - if not queue_mongo_resource.is_available(): - raise RuntimeError("Mongo resource is not available") - yield queue_mongo_resource - _clean_queue_database() +def queue_mongo_resource_autouse(queue_mongo_resource: QueueMongoResource) -> QueueMongoResource: + return queue_mongo_resource @@ -124,0 +111,31 @@ def test_upsert_job() -> None: [email protected]( + "statuses_to_cancel, expected_remaining_number", + [ + (None, 0), + ([Status.WAITING], 1), + ([Status.WAITING, Status.STARTED], 0), + ([Status.STARTED], 1), + ([Status.SUCCESS], 2), + ], +) +def test_cancel_jobs(statuses_to_cancel: Optional[List[Status]], expected_remaining_number: int) -> None: + test_type = "test_type" + test_dataset = "test_dataset" + queue = Queue() + queue._add_job(job_type=test_type, dataset=test_dataset, force=True) + queue._add_job(job_type=test_type, dataset=test_dataset) + queue.start_job() + + canceled_job_dicts = queue.cancel_jobs( + job_type=test_type, dataset=test_dataset, statuses_to_cancel=statuses_to_cancel + ) + assert len(canceled_job_dicts) == 2 - expected_remaining_number + + if expected_remaining_number == 0: + assert not queue.is_job_in_process(job_type=test_type, dataset=test_dataset) + with pytest.raises(EmptyQueueError): + queue.start_job() + else: + assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset) + + diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index 04fa6b68..c8391532 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -7 +7 @@ from time import process_time -from typing import Dict, Iterator, List, Optional, TypedDict +from typing import Dict, List, Optional, TypedDict @@ -21 +20,0 @@ from libcommon.simple_cache import ( - _clean_cache_database, @@ -40,8 +39,2 @@ from libcommon.simple_cache import ( -def cache_mongo_resource(cache_mongo_host: str) -> Iterator[CacheMongoResource]: - database = "datasets_server_cache_test" - host = cache_mongo_host - if "test" not in database: - raise ValueError("Test must be launched on a test mongo database") - with CacheMongoResource(database=database, host=host) as cache_mongo_resource: - yield cache_mongo_resource - _clean_cache_database() +def cache_mongo_resource_autouse(cache_mongo_resource: CacheMongoResource) -> CacheMongoResource: + return cache_mongo_resource diff --git a/libs/libcommon/tests/test_state.py b/libs/libcommon/tests/test_state.py new file mode 100644 index 00000000..5d87af38 --- /dev/null +++ b/libs/libcommon/tests/test_state.py @@ -0,0 +1,1383 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Dict, List, Mapping, Optional, TypedDict +from unittest.mock import patch + +import pytest + +from libcommon.config import ProcessingGraphConfig +from libcommon.processing_graph import ProcessingGraph +from libcommon.queue import Queue, Status +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response +from libcommon.state import ( + HARD_CODED_CONFIG_NAMES_CACHE_KIND, + HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND, + HARD_CODED_SPLIT_NAMES_FROM_STREAMING_CACHE_KIND, + ArtifactState, + CacheState, + ConfigState, + DatasetState, + JobState, + SplitState, + fetch_config_names, + fetch_split_names, +) + + [email protected](autouse=True) +def queue_mongo_resource_autouse(queue_mongo_resource: QueueMongoResource) -> QueueMongoResource: + return queue_mongo_resource + + [email protected](autouse=True) +def cache_mongo_resource_autouse(cache_mongo_resource: CacheMongoResource) -> CacheMongoResource: + return cache_mongo_resource + + +DATASET_NAME = "dataset" +CONFIG_NAMES_OK = ["config1", "config2"] +CONFIG_NAMES_CONTENT_OK = {"config_names": [{"config": config_name} for config_name in CONFIG_NAMES_OK]} +CONTENT_ERROR = {"error": "error"} + + [email protected]( + "content,http_status,expected_config_names", + [ + (CONFIG_NAMES_CONTENT_OK, HTTPStatus.OK, CONFIG_NAMES_OK), + (CONTENT_ERROR, HTTPStatus.INTERNAL_SERVER_ERROR, None), + (None, HTTPStatus.OK, None), + ], +) +def test_fetch_config_names( + content: Optional[Mapping[str, Any]], http_status: HTTPStatus, expected_config_names: Optional[List[str]] +) -> None: + raises = expected_config_names is None + if content: + upsert_response( + kind=HARD_CODED_CONFIG_NAMES_CACHE_KIND, + dataset=DATASET_NAME, + config=None, + split=None, + content=content, + http_status=http_status, + ) + + if raises: + with pytest.raises(Exception): + fetch_config_names(dataset=DATASET_NAME) + else: + config_names = fetch_config_names(dataset=DATASET_NAME) + assert config_names == expected_config_names + + +class ResponseSpec(TypedDict): + content: Mapping[str, Any] + http_status: HTTPStatus + + +CONFIG_NAME_1 = "config1" +SPLIT_NAMES_OK = ["split1", "split2"] + + +def get_SPLIT_NAMES_CONTENT_OK(dataset: str, config: str, splits: List[str]) -> Any: + return {"split_names": [{"dataset": dataset, "config": config, "split": split_name} for split_name in splits]} + + +SPLIT_NAMES_RESPONSE_OK = ResponseSpec( + content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), + http_status=HTTPStatus.OK, +) +SPLIT_NAMES_RESPONSE_ERROR = ResponseSpec(content={"error": "error"}, http_status=HTTPStatus.INTERNAL_SERVER_ERROR) + + [email protected]( + "response_spec_by_kind,expected_split_names", + [ + ({HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND: SPLIT_NAMES_RESPONSE_OK}, SPLIT_NAMES_OK), + ({HARD_CODED_SPLIT_NAMES_FROM_STREAMING_CACHE_KIND: SPLIT_NAMES_RESPONSE_OK}, SPLIT_NAMES_OK), + ( + { + HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND: SPLIT_NAMES_RESPONSE_ERROR, + HARD_CODED_SPLIT_NAMES_FROM_STREAMING_CACHE_KIND: SPLIT_NAMES_RESPONSE_OK, + }, + SPLIT_NAMES_OK, + ), + ({HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND: SPLIT_NAMES_RESPONSE_ERROR}, None), + ({}, None), + ], +) +def test_fetch_split_names( + response_spec_by_kind: Mapping[str, Mapping[str, Any]], + expected_split_names: Optional[List[str]], +) -> None: + raises = expected_split_names is None + for kind, response_spec in response_spec_by_kind.items(): + upsert_response( + kind=kind, + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=response_spec["content"], + http_status=response_spec["http_status"], + ) + + if raises: + with pytest.raises(Exception): + fetch_split_names(dataset=DATASET_NAME, config=CONFIG_NAME_1) + else: + split_names = fetch_split_names(dataset=DATASET_NAME, config=CONFIG_NAME_1) + assert split_names == expected_split_names + + +SPLIT_NAME = "split" +JOB_TYPE = "job_type" + + [email protected]( + "dataset,config,split,job_type", + [ + (DATASET_NAME, None, None, JOB_TYPE), + (DATASET_NAME, CONFIG_NAME_1, None, JOB_TYPE), + (DATASET_NAME, CONFIG_NAME_1, SPLIT_NAME, JOB_TYPE), + ], +) +def test_job_state_is_in_process(dataset: str, config: Optional[str], split: Optional[str], job_type: str) -> None: + queue = Queue() + queue.upsert_job(job_type=job_type, dataset=dataset, config=config, split=split) + assert JobState(dataset=dataset, config=config, split=split, job_type=job_type).is_in_process + job_info = queue.start_job() + assert JobState(dataset=dataset, config=config, split=split, job_type=job_type).is_in_process + queue.finish_job(job_id=job_info["job_id"], finished_status=Status.SUCCESS) + assert not JobState(dataset=dataset, config=config, split=split, job_type=job_type).is_in_process + + [email protected]( + "dataset,config,split,job_type", + [ + (DATASET_NAME, None, None, JOB_TYPE), + (DATASET_NAME, CONFIG_NAME_1, None, JOB_TYPE), + (DATASET_NAME, CONFIG_NAME_1, SPLIT_NAME, JOB_TYPE), + ], +) +def test_job_state_as_dict(dataset: str, config: Optional[str], split: Optional[str], job_type: str) -> None: + queue = Queue() + queue.upsert_job(job_type=job_type, dataset=dataset, config=config, split=split) + assert JobState(dataset=dataset, config=config, split=split, job_type=job_type).as_dict() == { + "is_in_process": True, + } + + +CACHE_KIND = "cache_kind" + + [email protected]( + "dataset,config,split,cache_kind", + [ + (DATASET_NAME, None, None, CACHE_KIND), + (DATASET_NAME, CONFIG_NAME_1, None, CACHE_KIND), + (DATASET_NAME, CONFIG_NAME_1, SPLIT_NAME, CACHE_KIND), + ], +) +def test_cache_state_exists(dataset: str, config: Optional[str], split: Optional[str], cache_kind: str) -> None: + assert not CacheState(dataset=dataset, config=config, split=split, cache_kind=cache_kind).exists + upsert_response( + kind=cache_kind, dataset=dataset, config=config, split=split, content={}, http_status=HTTPStatus.OK + ) + assert CacheState(dataset=dataset, config=config, split=split, cache_kind=cache_kind).exists + + [email protected]( + "dataset,config,split,cache_kind", + [ + (DATASET_NAME, None, None, CACHE_KIND), + (DATASET_NAME, CONFIG_NAME_1, None, CACHE_KIND), + (DATASET_NAME, CONFIG_NAME_1, SPLIT_NAME, CACHE_KIND), + ], +) +def test_cache_state_is_success(dataset: str, config: Optional[str], split: Optional[str], cache_kind: str) -> None: + upsert_response( + kind=cache_kind, dataset=dataset, config=config, split=split, content={}, http_status=HTTPStatus.OK + ) + assert CacheState(dataset=dataset, config=config, split=split, cache_kind=cache_kind).is_success + upsert_response( + kind=cache_kind, + dataset=dataset, + config=config, + split=split, + content={}, + http_status=HTTPStatus.INTERNAL_SERVER_ERROR, + ) + assert not CacheState(dataset=dataset, config=config, split=split, cache_kind=cache_kind).is_success + + [email protected]( + "dataset,config,split,cache_kind", + [ + (DATASET_NAME, None, None, CACHE_KIND), + (DATASET_NAME, CONFIG_NAME_1, None, CACHE_KIND), + (DATASET_NAME, CONFIG_NAME_1, SPLIT_NAME, CACHE_KIND), + ], +) +def test_cache_state_as_dict(dataset: str, config: Optional[str], split: Optional[str], cache_kind: str) -> None: + assert CacheState(dataset=dataset, config=config, split=split, cache_kind=cache_kind).as_dict() == { + "exists": False, + "is_success": False, + } + upsert_response( + kind=cache_kind, + dataset=dataset, + config=config, + split=split, + content={"some": "content"}, + http_status=HTTPStatus.OK, + ) + assert CacheState(dataset=dataset, config=config, split=split, cache_kind=cache_kind).as_dict() == { + "exists": True, + "is_success": True, + } + + +PROCESSING_GRAPH = ProcessingGraph(processing_graph_specification=ProcessingGraphConfig().specification) + + +def test_artifact_state() -> None: + dataset = DATASET_NAME + config = None + split = None + step = PROCESSING_GRAPH.get_step(name="/config-names") + artifact_state = ArtifactState(dataset=dataset, config=config, split=split, step=step) + assert artifact_state.as_dict() == { + "id": f"/config-names,{dataset}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + } + assert not artifact_state.cache_state.exists + assert not artifact_state.job_state.is_in_process + + +def get_SPLIT_STATE_DICT(dataset: str, config: str, split: str) -> Any: + return { + "split": split, + "artifact_states": [ + { + "id": f"split-first-rows-from-streaming,{dataset},{config},{split}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"split-first-rows-from-parquet,{dataset},{config},{split}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + ], + } + + +SPLIT1_NAME = "split1" + + +def test_split_state_as_dict() -> None: + dataset = DATASET_NAME + config = CONFIG_NAME_1 + split = SPLIT1_NAME + processing_graph = PROCESSING_GRAPH + assert SplitState( + dataset=dataset, config=config, split=split, processing_graph=processing_graph + ).as_dict() == get_SPLIT_STATE_DICT(dataset=dataset, config=config, split=split) + + +SPLIT2_NAME = "split2" + + +def get_CONFIG_STATE_DICT(dataset: str, config: str, split_states: List[Any], cache_exists: bool) -> Any: + return { + "config": config, + "split_states": split_states, + "artifact_states": [ + { + "id": f"/split-names-from-streaming,{dataset},{config}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"config-parquet-and-info,{dataset},{config}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"config-parquet,{dataset},{config}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"config-info,{dataset},{config}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"/split-names-from-dataset-info,{dataset},{config}", + "job_state": {"is_in_process": False}, + "cache_state": { + "exists": cache_exists, + "is_success": cache_exists, + }, # ^ if this entry is in the cache + }, + { + "id": f"config-size,{dataset},{config}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + ], + } + + +def test_config_state_as_dict() -> None: + dataset = DATASET_NAME + config = CONFIG_NAME_1 + upsert_response( + kind=HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND, + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=SPLIT_NAMES_RESPONSE_OK["content"], + http_status=SPLIT_NAMES_RESPONSE_OK["http_status"], + ) + processing_graph = PROCESSING_GRAPH + assert ConfigState( + dataset=dataset, config=config, processing_graph=processing_graph + ).as_dict() == get_CONFIG_STATE_DICT( + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split_states=[ + get_SPLIT_STATE_DICT(dataset=dataset, config=config, split=SPLIT1_NAME), + get_SPLIT_STATE_DICT(dataset=dataset, config=config, split=SPLIT2_NAME), + ], + cache_exists=True, + ) + + +CONFIG_NAME_2 = "config2" +TWO_CONFIG_NAMES = [CONFIG_NAME_1, CONFIG_NAME_2] +TWO_CONFIG_NAMES_CONTENT_OK = {"config_names": [{"config": config} for config in TWO_CONFIG_NAMES]} +CURRENT_GIT_REVISION = "current_git_revision" + + +def test_dataset_state_as_dict() -> None: + dataset = DATASET_NAME + upsert_response( + kind=HARD_CODED_CONFIG_NAMES_CACHE_KIND, + dataset=dataset, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + ) + upsert_response( + kind=HARD_CODED_SPLIT_NAMES_FROM_DATASET_INFO_CACHE_KIND, + dataset=dataset, + config=CONFIG_NAME_1, + split=None, + content=SPLIT_NAMES_RESPONSE_OK["content"], + http_status=SPLIT_NAMES_RESPONSE_OK["http_status"], + ) + processing_graph = PROCESSING_GRAPH + assert DatasetState( + dataset=dataset, processing_graph=processing_graph, revision=CURRENT_GIT_REVISION + ).as_dict() == { + "dataset": "dataset", + "config_states": [ + get_CONFIG_STATE_DICT( + dataset=dataset, + config=CONFIG_NAME_1, + split_states=[ + get_SPLIT_STATE_DICT(dataset=dataset, config=CONFIG_NAME_1, split=SPLIT1_NAME), + get_SPLIT_STATE_DICT(dataset=dataset, config=CONFIG_NAME_1, split=SPLIT2_NAME), + ], + cache_exists=True, + ), + get_CONFIG_STATE_DICT( + dataset=dataset, + config=CONFIG_NAME_2, + split_states=[], + cache_exists=False, + ), + ], + "artifact_states": [ + { + "id": f"/config-names,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": True, "is_success": True}, # <- this entry is in the cache + }, + { + "id": f"/parquet-and-dataset-info,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"dataset-parquet,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"dataset-info,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"dataset-size,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"dataset-split-names-from-streaming,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"dataset-split-names-from-dataset-info,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"dataset-split-names,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + { + "id": f"dataset-is-valid,{DATASET_NAME}", + "job_state": {"is_in_process": False}, + "cache_state": {"exists": False, "is_success": False}, + }, + ], + } + + +CONFIG_PARQUET_AND_INFO_OK = {"config": CONFIG_NAME_1, "content": "not important"} +CONFIG_INFO_OK = {"config": CONFIG_NAME_1, "content": "not important"} + + +OLD_JOB_RUNNER_VERSION = 1 +NEW_JOB_RUNNER_VERSION = 2 + + +def finish_job(job_type: str, content: Any, job_runner_version: Optional[int] = NEW_JOB_RUNNER_VERSION) -> None: + job_info = Queue().start_job(only_job_types=[job_type]) + upsert_response( + kind=job_info["type"], + dataset=job_info["dataset"], + config=job_info["config"], + split=job_info["split"], + content=content, + http_status=HTTPStatus.OK, + job_runner_version=job_runner_version, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + Queue().finish_job(job_id=job_info["job_id"], finished_status=Status.SUCCESS) + + +def get_dataset_state(git_revision: Optional[str] = CURRENT_GIT_REVISION) -> DatasetState: + return DatasetState(dataset=DATASET_NAME, processing_graph=PROCESSING_GRAPH, revision=git_revision) + + +def assert_dataset_state( + config_names: List[str], + split_names_in_first_config: List[str], + cache_status: Dict[str, List[str]], + queue_status: Dict[str, List[str]], + tasks: List[str], + git_revision: Optional[str] = CURRENT_GIT_REVISION, +) -> DatasetState: + dataset_state = get_dataset_state(git_revision=git_revision) + assert dataset_state.config_names == config_names + assert len(dataset_state.config_states) == len(config_names) + if len(config_names): + assert dataset_state.config_states[0].split_names == split_names_in_first_config + else: + # this case is just to check the test, not the code + assert not split_names_in_first_config + assert dataset_state.cache_status.as_response() == cache_status + assert dataset_state.queue_status.as_response() == queue_status + assert dataset_state.plan.as_response() == tasks + return dataset_state + + +def test_plan() -> None: + assert_dataset_state( + # The config names are not yet known + config_names=[], + # The split names are not yet known + split_names_in_first_config=[], + # All the dataset-level cache entries are empty + # "dataset-is-valid" is also empty, but is marked as blocked because it depends on "dataset-split-names", + # which is not yet known. + # No config-level and split-level cache entries is listed, because the config names and splits + # names are not yet known. + cache_status={ + "blocked_by_parent": ["dataset-is-valid,dataset"], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/config-names,dataset", + "/parquet-and-dataset-info,dataset", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": [], + }, + # The queue is empty, so no step is in process. + queue_status={"in_process": []}, + # The root dataset-level steps, as well as the "fan-in" steps, are ready to be backfilled. + tasks=[ + "CreateJob[/config-names,dataset]", + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) + + +def test_plan_job_creation_and_termination() -> None: + # we launch all the backfill tasks + dataset_state = get_dataset_state() + assert dataset_state.plan.as_response() == [ + "CreateJob[/config-names,dataset]", + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ] + dataset_state.backfill() + assert_dataset_state( + # The config names are not yet known + config_names=[], + # The split names are not yet known + split_names_in_first_config=[], + # the cache has not changed + cache_status={ + "blocked_by_parent": ["dataset-is-valid,dataset"], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/config-names,dataset", + "/parquet-and-dataset-info,dataset", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": [], + }, + # the jobs have been created and are in process + queue_status={ + "in_process": [ + "/config-names,dataset", + "/parquet-and-dataset-info,dataset", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ] + }, + # thus: no new task + tasks=[], + ) + + # we simulate the job for "/config-names,dataset" has finished + finish_job("/config-names", TWO_CONFIG_NAMES_CONTENT_OK) + + assert_dataset_state( + # The config names are now known + config_names=TWO_CONFIG_NAMES, + # The split names are not yet known + split_names_in_first_config=[], + # The "/config-names" step is up-to-date + # Config-level artifacts appear in the "blocked_by_parent" status, because the config names are now known, but + # the parents are not ready yet. + # The split-level artifacts are still missing, because the splits names are not yet known, for any config. + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-size,dataset,config1", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-parquet-and-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": ["/config-names,dataset"], + }, + # the job "/config-names,dataset" is no more in process + queue_status={ + "in_process": [ + "/parquet-and-dataset-info,dataset", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ] + }, + # The children of "/config-names,dataset" are ready to be backfilled for each config (fan-out) + tasks=[ + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + ], + ) + + +def test_plan_only_one_config() -> None: + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-parquet-and-info,dataset,config1" artifact in cache + upsert_response( + kind="config-parquet-and-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + + assert_dataset_state( + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are not yet known + split_names_in_first_config=[], + # The children of "config-parquet-and-info" are blocked for config2, but not for config1 + # The two cache entries we created are up-to-date + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config2", + "config-parquet,dataset,config2", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-info,dataset,config1", + "config-parquet,dataset,config1", + "config-parquet-and-info,dataset,config2", + "config-size,dataset,config1", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": ["/config-names,dataset", "config-parquet-and-info,dataset,config1"], + }, + queue_status={"in_process": []}, + # the children of "config-parquet-and-info" for config1 are ready to be backfilled + tasks=[ + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[config-size,dataset,config1]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) + + +def test_plan_retry_error() -> None: + ERROR_CODE_TO_RETRY = "ERROR_CODE_TO_RETRY" + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.INTERNAL_SERVER_ERROR, + job_runner_version=NEW_JOB_RUNNER_VERSION, + error_code=ERROR_CODE_TO_RETRY, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + + with patch("libcommon.state.ERROR_CODES_TO_RETRY", [ERROR_CODE_TO_RETRY]): + assert_dataset_state( + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are not yet known + split_names_in_first_config=[], + # "/config-names,dataset" is in the cache, but it's not categorized in up to date, + # but in "cache_is_error_to_retry" due to the error code + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-parquet-and-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "config-size,dataset,config1", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": ["/config-names,dataset"], + "cache_is_job_runner_obsolete": [], + "up_to_date": [], + }, + queue_status={"in_process": []}, + # The "/config-names,dataset" artifact will be retried + tasks=[ + "CreateJob[/config-names,dataset]", + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) + + +def test_plan_incoherent_state() -> None: + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "/split-names-from-dataset-info,dataset,config1" artifact in cache + # -> It's not a coherent state for the cache: the ancestors artifacts are missing: + # "config-parquet-and-info,dataset" and "config-info,dataset,config1" + upsert_response( + kind="/split-names-from-dataset-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + + assert_dataset_state( + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are known (even if it's from an outdated artifact) + split_names_in_first_config=SPLIT_NAMES_OK, + # "/split-names-from-dataset-info,dataset,config1" is blocked by its parents, even if it's in cache, because + # two ancestors are missing. It's not listed in up-to-date, even if it has been cached after /config-names. + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-size,dataset,config1", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + "split-first-rows-from-parquet,dataset,config1,split1", + "split-first-rows-from-parquet,dataset,config1,split2", + "split-first-rows-from-streaming,dataset,config1,split1", + "split-first-rows-from-streaming,dataset,config1,split2", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-parquet-and-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": ["/config-names,dataset"], + }, + queue_status={"in_process": []}, + # The children of "/split-names-from-dataset-info,dataset,config1" are not ready to be backfilled + # because it's blocked by its parents. + tasks=[ + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) + + +def test_plan_get_splits() -> None: + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-parquet-and-info,dataset,config1" artifact in cache + upsert_response( + kind="config-parquet-and-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-info,dataset,config1" artifact in cache + upsert_response( + kind="config-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "/split-names-from-dataset-info,dataset,config1" artifact in cache + upsert_response( + kind="/split-names-from-dataset-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "/split-names-from-streaming,dataset,config1" artifact in cache + upsert_response( + kind="/split-names-from-streaming", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + + assert_dataset_state( + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are known (even if it's from an outdated artifact) + split_names_in_first_config=SPLIT_NAMES_OK, + # The chains /config-names,dataset -> config-parquet-and-info,dataset,config1 -> config-info,dataset,config1 + # -> /split-names-from-dataset-info,dataset,config1 and + # /config-names,dataset -> /split-names-from-streaming,dataset,config1 are up to date + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config2", + "config-parquet,dataset,config2", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + "split-first-rows-from-parquet,dataset,config1,split1", + "split-first-rows-from-parquet,dataset,config1,split2", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet-and-info,dataset,config2", + "config-size,dataset,config1", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + "split-first-rows-from-streaming,dataset,config1,split1", + "split-first-rows-from-streaming,dataset,config1,split2", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": [ + "/config-names,dataset", + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-streaming,dataset,config1", + "config-info,dataset,config1", + "config-parquet-and-info,dataset,config1", + ], + }, + queue_status={"in_process": []}, + # split-first-rows-from-streaming is now ready to be backfilled for split1 and split2 of dataset,config1 + tasks=[ + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[config-size,dataset,config1]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + "CreateJob[split-first-rows-from-streaming,dataset,config1,split1]", + "CreateJob[split-first-rows-from-streaming,dataset,config1,split2]", + ], + ) + + +def test_plan_updated_at() -> None: + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-parquet-and-info,dataset,config1" artifact in cache + upsert_response( + kind="config-parquet-and-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-info,dataset,config1" artifact in cache + upsert_response( + kind="config-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "/split-names-from-dataset-info,dataset,config1" artifact in cache + upsert_response( + kind="/split-names-from-dataset-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "/split-names-from-streaming,dataset,config1" artifact in cache + upsert_response( + kind="/split-names-from-streaming", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=get_SPLIT_NAMES_CONTENT_OK(dataset=DATASET_NAME, config=CONFIG_NAME_1, splits=SPLIT_NAMES_OK), + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + + # Now: refresh again the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + + assert_dataset_state( + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are known (even if it's from an outdated artifact) + split_names_in_first_config=SPLIT_NAMES_OK, + # config-parquet-and-info,dataset,config1 is marked as outdate by parent, + # as well as /split-names-from-streaming,dataset,config1 + # This means that all their children are marked as blocked by parent + # Only "/config-names,dataset" is marked as up to date + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-size,dataset,config1", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + "split-first-rows-from-parquet,dataset,config1,split1", + "split-first-rows-from-parquet,dataset,config1,split2", + "split-first-rows-from-streaming,dataset,config1,split1", + "split-first-rows-from-streaming,dataset,config1,split2", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [ + "/split-names-from-streaming,dataset,config1", + "config-parquet-and-info,dataset,config1", + ], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config2", + "config-parquet-and-info,dataset,config2", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": ["/config-names,dataset"], + }, + queue_status={"in_process": []}, + # config-parquet-and-info and /split-names-from-streaming will be refreshed for both configs + # (even if config1 is already in the cache -> but it's outdated) + tasks=[ + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) + + +def test_plan_job_runner_version() -> None: + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-parquet-and-info,dataset,config1" artifact in cache + upsert_response( + kind="config-parquet-and-info", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=CURRENT_GIT_REVISION, + ) + # Set the "config-parquet,dataset,config1" artifact in cache + upsert_response( + kind="config-parquet", + dataset=DATASET_NAME, + config=CONFIG_NAME_1, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, # <- not important + http_status=HTTPStatus.OK, + job_runner_version=OLD_JOB_RUNNER_VERSION, # <- current version is 3, this one is 1 + dataset_git_revision=CURRENT_GIT_REVISION, + ) + + assert_dataset_state( + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are not known + split_names_in_first_config=[], + # config-parquet,dataset,config1 is in the category: "is_job_runner_obsolete" + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config2", + "config-parquet,dataset,config2", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "config-size,dataset,config1", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": ["config-parquet,dataset,config1"], + "up_to_date": ["/config-names,dataset", "config-parquet-and-info,dataset,config1"], + }, + queue_status={"in_process": []}, + # config-parquet,dataset,config1 will be refreshed because its job runner has been upgraded + tasks=[ + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-info,dataset,config1]", + "CreateJob[config-parquet,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[config-size,dataset,config1]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) + + [email protected]( + "dataset_git_revision,cached_dataset_get_revision,expect_refresh", + [ + (None, None, False), + ("a", "a", False), + (None, "b", True), + ("a", None, True), + ("a", "b", True), + ], +) +def test_plan_git_revision( + dataset_git_revision: Optional[str], cached_dataset_get_revision: Optional[str], expect_refresh: bool +) -> None: + # Set the "/config-names,dataset" artifact in cache + upsert_response( + kind="/config-names", + dataset=DATASET_NAME, + config=None, + split=None, + content=TWO_CONFIG_NAMES_CONTENT_OK, + http_status=HTTPStatus.OK, + job_runner_version=NEW_JOB_RUNNER_VERSION, + dataset_git_revision=cached_dataset_get_revision, + ) + + if expect_refresh: + # if the git revision is different from the current dataset git revision, the artifact will be refreshed + # and its children will be blocked + assert_dataset_state( + git_revision=dataset_git_revision, + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are not known + split_names_in_first_config=[], + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-parquet-and-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "config-size,dataset,config1", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + ], + "cache_has_different_git_revision": ["/config-names,dataset"], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": [], + }, + queue_status={"in_process": []}, + tasks=[ + "CreateJob[/config-names,dataset]", + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) + else: + assert_dataset_state( + git_revision=dataset_git_revision, + # The config names are known + config_names=TWO_CONFIG_NAMES, + # The split names are not known + split_names_in_first_config=[], + cache_status={ + "blocked_by_parent": [ + "/split-names-from-dataset-info,dataset,config1", + "/split-names-from-dataset-info,dataset,config2", + "config-info,dataset,config1", + "config-info,dataset,config2", + "config-parquet,dataset,config1", + "config-parquet,dataset,config2", + "config-size,dataset,config1", + "config-size,dataset,config2", + "dataset-is-valid,dataset", + ], + "cache_has_different_git_revision": [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [ + "/parquet-and-dataset-info,dataset", + "/split-names-from-streaming,dataset,config1", + "/split-names-from-streaming,dataset,config2", + "config-parquet-and-info,dataset,config1", + "config-parquet-and-info,dataset,config2", + "dataset-info,dataset", + "dataset-parquet,dataset", + "dataset-size,dataset", + "dataset-split-names,dataset", + "dataset-split-names-from-dataset-info,dataset", + "dataset-split-names-from-streaming,dataset", + ], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": ["/config-names,dataset"], + }, + queue_status={"in_process": []}, + tasks=[ + "CreateJob[/parquet-and-dataset-info,dataset]", + "CreateJob[/split-names-from-streaming,dataset,config1]", + "CreateJob[/split-names-from-streaming,dataset,config2]", + "CreateJob[config-parquet-and-info,dataset,config1]", + "CreateJob[config-parquet-and-info,dataset,config2]", + "CreateJob[dataset-info,dataset]", + "CreateJob[dataset-parquet,dataset]", + "CreateJob[dataset-size,dataset]", + "CreateJob[dataset-split-names,dataset]", + "CreateJob[dataset-split-names-from-dataset-info,dataset]", + "CreateJob[dataset-split-names-from-streaming,dataset]", + ], + ) diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py index 49ba9efd..f911055b 100644 --- a/services/admin/src/admin/app.py +++ b/services/admin/src/admin/app.py @@ -27,0 +28,2 @@ from admin.routes.cancel_jobs import create_cancel_jobs_endpoint +from admin.routes.dataset_backfill import create_dataset_backfill_endpoint +from admin.routes.dataset_state import create_dataset_state_endpoint @@ -83,0 +86,22 @@ def create_app() -> Starlette: + Route( + "/dataset-backfill", + endpoint=create_dataset_backfill_endpoint( + processing_graph=processing_graph, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + external_auth_url=app_config.admin.external_auth_url, + organization=app_config.admin.hf_organization, + ), + methods=["POST"], + ), + Route( + "/dataset-state", + endpoint=create_dataset_state_endpoint( + processing_graph=processing_graph, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + max_age=app_config.admin.max_age, + external_auth_url=app_config.admin.external_auth_url, + organization=app_config.admin.hf_organization, + ), + ), diff --git a/services/admin/src/admin/routes/dataset_backfill.py b/services/admin/src/admin/routes/dataset_backfill.py new file mode 100644 index 00000000..35c5fdff --- /dev/null +++ b/services/admin/src/admin/routes/dataset_backfill.py @@ -0,0 +1,60 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from typing import Optional + +from libcommon.dataset import DatasetError, get_dataset_git_revision +from libcommon.processing_graph import ProcessingGraph +from libcommon.state import DatasetState +from starlette.requests import Request +from starlette.responses import Response + +from admin.authentication import auth_check +from admin.utils import ( + AdminCustomError, + Endpoint, + MissingRequiredParameterError, + UnexpectedError, + are_valid_parameters, + get_json_admin_error_response, + get_json_ok_response, +) + + +def create_dataset_backfill_endpoint( + processing_graph: ProcessingGraph, + hf_endpoint: str, + external_auth_url: Optional[str] = None, + organization: Optional[str] = None, + hf_token: Optional[str] = None, + hf_timeout_seconds: Optional[float] = None, +) -> Endpoint: + async def dataset_backfill_endpoint(request: Request) -> Response: + try: + dataset = request.query_params.get("dataset") + if not are_valid_parameters([dataset]) or not dataset: + raise MissingRequiredParameterError("Parameter 'dataset' is required") + logging.info(f"/dataset-backfill, dataset={dataset}") + + # if auth_check fails, it will raise an exception that will be caught below + auth_check(external_auth_url=external_auth_url, request=request, organization=organization) + + dataset_git_revision = get_dataset_git_revision( + dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds + ) + dataset_state = DatasetState( + dataset=dataset, processing_graph=processing_graph, revision=dataset_git_revision + ) + dataset_state.backfill() + tasks_list = ", ".join(dataset_state.plan.as_response()) + return get_json_ok_response( + {"status": "ok", "message": f"Backfilling dataset. Tasks: {tasks_list}"}, + max_age=0, + ) + except (DatasetError, AdminCustomError) as e: + return get_json_admin_error_response(e, max_age=0) + except Exception as e: + return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=0) + + return dataset_backfill_endpoint diff --git a/services/admin/src/admin/routes/dataset_state.py b/services/admin/src/admin/routes/dataset_state.py new file mode 100644 index 00000000..384f0ce1 --- /dev/null +++ b/services/admin/src/admin/routes/dataset_state.py @@ -0,0 +1,56 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from typing import Optional + +from libcommon.dataset import get_dataset_git_revision +from libcommon.processing_graph import ProcessingGraph +from libcommon.state import DatasetState +from starlette.requests import Request +from starlette.responses import Response + +from admin.authentication import auth_check +from admin.utils import ( + AdminCustomError, + Endpoint, + MissingRequiredParameterError, + UnexpectedError, + are_valid_parameters, + get_json_admin_error_response, + get_json_ok_response, +) + + +def create_dataset_state_endpoint( + processing_graph: ProcessingGraph, + max_age: int, + hf_endpoint: str, + external_auth_url: Optional[str] = None, + organization: Optional[str] = None, + hf_token: Optional[str] = None, + hf_timeout_seconds: Optional[float] = None, +) -> Endpoint: + async def dataset_state_endpoint(request: Request) -> Response: + try: + dataset = request.query_params.get("dataset") + if not are_valid_parameters([dataset]) or not dataset: + raise MissingRequiredParameterError("Parameter 'dataset' is required") + logging.info(f"/dataset-state, dataset={dataset}") + + # if auth_check fails, it will raise an exception that will be caught below + auth_check(external_auth_url=external_auth_url, request=request, organization=organization) + + dataset_git_revision = get_dataset_git_revision( + dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds + ) + dataset_state = DatasetState( + dataset=dataset, processing_graph=processing_graph, revision=dataset_git_revision + ) + return get_json_ok_response(dataset_state.as_response(), max_age=max_age) + except AdminCustomError as e: + return get_json_admin_error_response(e, max_age=max_age) + except Exception as e: + return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age) + + return dataset_state_endpoint diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py index f5396b97..83965821 100644 --- a/services/api/tests/routes/test_valid.py +++ b/services/api/tests/routes/test_valid.py @@ -17,0 +18 @@ dataset_step = ProcessingStep( + parents=[], @@ -26,0 +28 @@ config_step = ProcessingStep( + parents=[], @@ -35,0 +38 @@ split_step = ProcessingStep( + parents=[], diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 2c8bac4d..b56e4e0f 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -124,0 +125 @@ def test_processing_step() -> ProcessingStep: + parents=[], diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py index ed0ddf7c..ca1b51b1 100644 --- a/services/worker/tests/job_runners/config/test_info.py +++ b/services/worker/tests/job_runners/config/test_info.py @@ -167,0 +168 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index 7282747d..08542e00 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -65,0 +66 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 5e76eb95..9f2c6e70 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -98,0 +99 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py index fa19e076..4de49942 100644 --- a/services/worker/tests/job_runners/config/test_size.py +++ b/services/worker/tests/job_runners/config/test_size.py @@ -60,0 +61 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index 4ab3c6a2..9555ea5c 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -57,0 +58 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index 150b1036..283963c9 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -57,0 +58 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py index 880ba02a..be2f1f86 100644 --- a/services/worker/tests/job_runners/dataset/test_info.py +++ b/services/worker/tests/job_runners/dataset/test_info.py @@ -138,0 +139 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/dataset/test_is_valid.py b/services/worker/tests/job_runners/dataset/test_is_valid.py index b1abc270..4f825c52 100644 --- a/services/worker/tests/job_runners/dataset/test_is_valid.py +++ b/services/worker/tests/job_runners/dataset/test_is_valid.py @@ -95,0 +96 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index 8ded0a66..fa7a782c 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -64,0 +65 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/dataset/test_size.py b/services/worker/tests/job_runners/dataset/test_size.py index c8f9760f..fbe4c79e 100644 --- a/services/worker/tests/job_runners/dataset/test_size.py +++ b/services/worker/tests/job_runners/dataset/test_size.py @@ -61,0 +62 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/dataset/test_split_names.py b/services/worker/tests/job_runners/dataset/test_split_names.py index e473ab4e..913c4f4f 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names.py +++ b/services/worker/tests/job_runners/dataset/test_split_names.py @@ -51,0 +52 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py index 8e56cfb5..72bfd8e9 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py @@ -51,0 +52 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py index 3f1919f9..8dccf5f3 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py @@ -51,0 +52 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index 0bad2976..f9ec3285 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -64,0 +65 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index 6719cb68..2f69fc9c 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -63,0 +64 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py index 40554d94..a073b151 100644 --- a/services/worker/tests/job_runners/test__datasets_based_worker.py +++ b/services/worker/tests/job_runners/test__datasets_based_worker.py @@ -76,0 +77 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py index bcf755bc..7eea5ecb 100644 --- a/services/worker/tests/job_runners/test_config_names.py +++ b/services/worker/tests/job_runners/test_config_names.py @@ -52,0 +53 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index 9bb30190..931ea63b 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -94,0 +95 @@ def get_job_runner( + parents=[], diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py index 3c50f2d9..3398be07 100644 --- a/services/worker/tests/test_job_runner.py +++ b/services/worker/tests/test_job_runner.py @@ -231,0 +232 @@ def test_check_type( + parents=[],
ab6f4ba27ac553d589de371049db484a7a6a3825
Sylvain Lesage
2023-04-17T15:54:33
Avoid disclosing private datasets (#1053)
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py index 54b39585..706c0f10 100644 --- a/libs/libcommon/src/libcommon/dataset.py +++ b/libs/libcommon/src/libcommon/dataset.py @@ -17,0 +18 @@ DatasetErrorCode = Literal[ + "DatasetRevisionNotFoundError", @@ -78,0 +80,13 @@ class DatasetNotFoundError(DatasetError): +class DatasetRevisionNotFoundError(DatasetError): + """Raised when the dataset revision (git branch) does not exist.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__( + message=message, + status_code=HTTPStatus.NOT_FOUND, + code="DatasetRevisionNotFoundError", + cause=cause, + disclose_cause=True, + ) + + @@ -174 +188 @@ def ask_access( - if r.status_code in [401, 404]: + if r.status_code in {401, 404}: @@ -212,2 +226,3 @@ def get_dataset_info_for_supported_datasets( - (private datasets are not supported by the datasets server), or if the default branch cannot - be found in the dataset. + (private datasets are not supported by the datasets server). + - [`~libcommon.dataset.DatasetRevisionNotFoundError`]: if the git revision (branch, commit) does not + exist in the repository. @@ -234 +249 @@ def get_dataset_info_for_supported_datasets( - raise DatasetNotFoundError( + raise DatasetRevisionNotFoundError( @@ -246 +261 @@ def get_dataset_info_for_supported_datasets( - raise DatasetNotFoundError("The dataset is private and private datasets are not yet supported.") + raise DatasetNotFoundError(DOES_NOT_EXIST_OR_PRIVATE_DATASET_ERROR_MESSAGE)
03e15c72e539db79cfcde2b0fc1c96e43b6fa9ca
Sylvain Lesage
2023-04-17T15:54:22
chore: 🤖 add DEV_NETWORK_MODE and DEV_MONGO_HOST (#1052)
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index 28fbd274..7e11cf4d 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -13 +13 @@ services: - CACHE_MONGO_URL: ${CACHE_MONGO_URL-mongodb://host.docker.internal:${MONGO_PORT-27017}} # use mongo container by default + CACHE_MONGO_URL: ${CACHE_MONGO_URL-mongodb://${DEV_MONGO_HOST-host.docker.internal}:${MONGO_PORT-27017}} # use mongo container by default @@ -17 +17 @@ services: - QUEUE_MONGO_URL: ${QUEUE_MONGO_URL-mongodb://host.docker.internal:${MONGO_PORT-27017}} # use mongo container by default + QUEUE_MONGO_URL: ${QUEUE_MONGO_URL-mongodb://${DEV_MONGO_HOST-host.docker.internal}:${MONGO_PORT-27017}} # use mongo container by default @@ -20 +20 @@ services: - METRICS_MONGO_URL: ${METRICS_MONGO_URL-mongodb://host.docker.internal:${MONGO_PORT-27017}} # use mongo container by default + METRICS_MONGO_URL: ${METRICS_MONGO_URL-mongodb://${DEV_MONGO_HOST-host.docker.internal}:${MONGO_PORT-27017}} # use mongo container by default diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index ffcd1dec..eacbf287 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -47,0 +48 @@ services: + network_mode: ${DEV_NETWORK_MODE} @@ -78,0 +80 @@ services: + network_mode: ${DEV_NETWORK_MODE} @@ -112,0 +115 @@ services: + network_mode: ${DEV_NETWORK_MODE}
4f3e52d5b1f3f0d8551716aef964516f7fd501f2
Andrea Francis Soria Jimenez
2023-04-17T11:51:30
Separate computation metrics for jobs and cache in a cron job (#1046)
diff --git a/.github/workflows/_unit-tests-python.yml b/.github/workflows/_unit-tests-python.yml index dd05f1da..071c9c9e 100644 --- a/.github/workflows/_unit-tests-python.yml +++ b/.github/workflows/_unit-tests-python.yml @@ -47,0 +48 @@ jobs: + METRICS_MONGO_URL: mongodb://localhost:${{ env.mongo-port }} diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 2f782c84..577023ad 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.7.0 +version: 1.8.0 diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 96ef959c..013f9206 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -107,0 +108,13 @@ cacheMaintenance: +# --- cron jobs --- +metricsCollector: + action: "collect-metrics" + schedule: "*/5 * * * *" + # every five minutes + nodeSelector: {} + resources: + requests: + cpu: 0 + limits: + cpu: 0 + tolerations: [] + diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 6229aae0..a8a28952 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -115,0 +116,14 @@ cacheMaintenance: +# --- cron jobs --- +metricsCollector: + action: "collect-metrics" + schedule: "*/2 * * * *" + # every two minutes + nodeSelector: {} + resources: + requests: + cpu: 1 + limits: + cpu: 1 + memory: "512Mi" + tolerations: [] + diff --git a/chart/templates/_envMetrics.tpl b/chart/templates/_envMetrics.tpl new file mode 100644 index 00000000..3966ee78 --- /dev/null +++ b/chart/templates/_envMetrics.tpl @@ -0,0 +1,9 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "envMetrics" -}} +- name: METRICS_MONGO_DATABASE + value: {{ .Values.metrics.mongoDatabase | quote }} +- name: METRICS_MONGO_URL + {{ include "datasetServer.mongo.url" . | nindent 2 }} +{{- end -}} diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index b6b294fb..bfc62fbb 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -80,0 +81,5 @@ app.kubernetes.io/component: "{{ include "name" . }}-cache-maintenance" +{{- define "labels.metricsCollector" -}} +{{ include "hf.labels.commons" . }} +app.kubernetes.io/component: "{{ include "name" . }}-metrics-collector" +{{- end -}} + diff --git a/chart/templates/cron-jobs/metrics-collector/_container.tpl b/chart/templates/cron-jobs/metrics-collector/_container.tpl new file mode 100644 index 00000000..e159d00b --- /dev/null +++ b/chart/templates/cron-jobs/metrics-collector/_container.tpl @@ -0,0 +1,18 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "containerMetricsCollector" -}} +- name: "{{ include "name" . }}-metrics-collector" + image: {{ include "jobs.cacheMaintenance.image" . }} + imagePullPolicy: {{ .Values.images.pullPolicy }} + env: + {{ include "envLog" . | nindent 2 }} + {{ include "envQueue" . | nindent 2 }} + {{ include "envCommon" . | nindent 2 }} + {{ include "envMetrics" . | nindent 2 }} + - name: CACHE_MAINTENANCE_ACTION + value: {{ .Values.metricsCollector.action | quote }} + securityContext: + allowPrivilegeEscalation: false + resources: {{ toYaml .Values.metricsCollector.resources | nindent 4 }} +{{- end -}} diff --git a/chart/templates/cron-jobs/metrics-collector/job.yaml b/chart/templates/cron-jobs/metrics-collector/job.yaml new file mode 100644 index 00000000..130e754e --- /dev/null +++ b/chart/templates/cron-jobs/metrics-collector/job.yaml @@ -0,0 +1,24 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- if .Values.images.jobs.cacheMaintenance }} +apiVersion: batch/v1 +kind: CronJob +metadata: + labels: {{ include "labels.metricsCollector" . | nindent 4 }} + name: "{{ include "name" . }}-job-metrics-collector" + namespace: {{ .Release.Namespace }} +spec: + schedule: {{ toYaml .Values.metricsCollector.schedule }} + template: + metadata: + labels: {{ include "labels.metricsCollector" . | nindent 8 }} + spec: + restartPolicy: Never + {{- include "image.imagePullSecrets" . | nindent 6 }} + containers: {{ include "containerCacheMaintenance" . | nindent 8 }} + nodeSelector: {{ toYaml .Values.metricsCollector.nodeSelector | nindent 8 }} + tolerations: {{ toYaml .Values.metricsCollector.tolerations | nindent 8 }} + securityContext: {{ include "securityContext" . | nindent 8 }} + backoffLimit: 0 +{{- end}} diff --git a/chart/templates/jobs/cache-maintenance/_container.tpl b/chart/templates/jobs/cache-maintenance/_container.tpl index 648d0b06..2d45e7f8 100644 --- a/chart/templates/jobs/cache-maintenance/_container.tpl +++ b/chart/templates/jobs/cache-maintenance/_container.tpl @@ -11,0 +12 @@ + {{ include "envMetrics" . | nindent 2 }} diff --git a/chart/templates/services/admin/_container.tpl b/chart/templates/services/admin/_container.tpl index 04925e82..da6a4bc5 100644 --- a/chart/templates/services/admin/_container.tpl +++ b/chart/templates/services/admin/_container.tpl @@ -14,0 +15 @@ + {{ include "envMetrics" . | nindent 2 }} diff --git a/chart/values.yaml b/chart/values.yaml index 0a12c555..7449c373 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -115,0 +116,4 @@ queue: +metrics: + # Name of the mongo db database used to store metrics + mongoDatabase: "datasets_server_metrics" + @@ -204 +208,14 @@ cacheMaintenance: - # ^ allowed values are {skip,backfill,upgrade} + # ^ allowed values are {skip,backfill,upgrade,collect-metrics} + nodeSelector: {} + resources: + requests: + cpu: 0 + limits: + cpu: 0 + tolerations: [] + +# --- cron jobs --- +metricsCollector: + action: "collect-metrics" + schedule: "*/5 * * * *" + # every five minutes diff --git a/jobs/cache_maintenance/Makefile b/jobs/cache_maintenance/Makefile index a56a0338..1b1c4fb0 100644 --- a/jobs/cache_maintenance/Makefile +++ b/jobs/cache_maintenance/Makefile @@ -5,0 +6 @@ export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +export METRICS_MONGO_URL := mongodb://localhost:${MONGO_PORT} diff --git a/jobs/cache_maintenance/src/cache_maintenance/config.py b/jobs/cache_maintenance/src/cache_maintenance/config.py index 69ded5c9..60858020 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/config.py +++ b/jobs/cache_maintenance/src/cache_maintenance/config.py @@ -11,0 +12 @@ from libcommon.config import ( + MetricsConfig, @@ -23,0 +25 @@ class JobConfig: + metrics: MetricsConfig = field(default_factory=MetricsConfig) @@ -35,0 +38 @@ class JobConfig: + metrics=MetricsConfig.from_env(), diff --git a/jobs/cache_maintenance/src/cache_maintenance/main.py b/jobs/cache_maintenance/src/cache_maintenance/main.py index 0479017d..41513a40 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/main.py +++ b/jobs/cache_maintenance/src/cache_maintenance/main.py @@ -10 +10,5 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.resources import ( + CacheMongoResource, + MetricsMongoResource, + QueueMongoResource, +) @@ -13,0 +18 @@ from cache_maintenance.config import JobConfig +from cache_maintenance.metrics import collect_metrics @@ -20 +25 @@ def run_job() -> None: - supported_actions = ["backfill", "upgrade"] + supported_actions = ["backfill", "upgrade", "collect-metrics"] @@ -37,0 +43,3 @@ def run_job() -> None: + MetricsMongoResource( + database=job_config.metrics.mongo_database, host=job_config.metrics.mongo_url + ) as metrics_resource, @@ -40,3 +48 @@ def run_job() -> None: - logging.warning( - "The connection to the cache database could not be established. The cache refresh job is skipped." - ) + logging.warning("The connection to the cache database could not be established. The action is skipped.") @@ -45,3 +51,4 @@ def run_job() -> None: - logging.warning( - "The connection to the queue database could not be established. The cache refresh job is skipped." - ) + logging.warning("The connection to the queue database could not be established. The action is skipped.") + return + if not metrics_resource.is_available(): + logging.warning("The connection to the metrics database could not be established. The action is skipped.") @@ -63,0 +71,3 @@ def run_job() -> None: + if action == "collect-metrics": + collect_metrics(processing_steps) + diff --git a/jobs/cache_maintenance/src/cache_maintenance/metrics.py b/jobs/cache_maintenance/src/cache_maintenance/metrics.py new file mode 100644 index 00000000..e5f8a097 --- /dev/null +++ b/jobs/cache_maintenance/src/cache_maintenance/metrics.py @@ -0,0 +1,24 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from libcommon.metrics import CacheTotalMetric, JobTotalMetric +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Queue +from libcommon.simple_cache import get_responses_count_by_kind_status_and_error_code + + +def collect_metrics(processing_steps: list[ProcessingStep]) -> None: + logging.info("collecting jobs metrics") + queue = Queue() + for processing_step in processing_steps: + for status, total in queue.get_jobs_count_by_status(job_type=processing_step.job_type).items(): + JobTotalMetric.objects(queue=processing_step.job_type, status=status).upsert_one(total=total) + + logging.info("collecting cache metrics") + for metric in get_responses_count_by_kind_status_and_error_code(): + CacheTotalMetric.objects( + kind=metric["kind"], http_status=metric["http_status"], error_code=metric["error_code"] + ).upsert_one(total=metric["count"]) + logging.info("metrics have been collected") diff --git a/jobs/cache_maintenance/tests/conftest.py b/jobs/cache_maintenance/tests/conftest.py index ad491836..46639105 100644 --- a/jobs/cache_maintenance/tests/conftest.py +++ b/jobs/cache_maintenance/tests/conftest.py @@ -5,0 +6 @@ from typing import Iterator +from libcommon.metrics import _clean_metrics_database @@ -7 +8,5 @@ from libcommon.queue import _clean_queue_database -from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.resources import ( + CacheMongoResource, + MetricsMongoResource, + QueueMongoResource, +) @@ -19,0 +25 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]: + monkeypatch_session.setenv("METRICS_MONGO_DATABASE", "datasets_server_metrics_test") @@ -27 +33,5 @@ def job_config(monkeypatch_session: MonkeyPatch) -> JobConfig: - if "test" not in job_config.cache.mongo_database or "test" not in job_config.queue.mongo_database: + if ( + "test" not in job_config.cache.mongo_database + or "test" not in job_config.queue.mongo_database + or "test" not in job_config.metrics.mongo_database + ): @@ -43,0 +54,9 @@ def queue_mongo_resource(job_config: JobConfig) -> Iterator[QueueMongoResource]: + + +@fixture(autouse=True) +def metrics_mongo_resource(job_config: JobConfig) -> Iterator[MetricsMongoResource]: + with MetricsMongoResource( + database=job_config.metrics.mongo_database, host=job_config.metrics.mongo_url + ) as resource: + yield resource + _clean_metrics_database() diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py new file mode 100644 index 00000000..fffc11d4 --- /dev/null +++ b/jobs/cache_maintenance/tests/test_collect_metrics.py @@ -0,0 +1,55 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus + +from libcommon.metrics import CacheTotalMetric, JobTotalMetric +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Queue +from libcommon.simple_cache import upsert_response + +from cache_maintenance.metrics import collect_metrics + + +def test_collect_metrics() -> None: + dataset = "test_dataset" + config = None + split = None + content = {"some": "content"} + + test_type = "test_type" + queue = Queue() + queue.upsert_job(job_type=test_type, dataset="dataset", config="config", split="split") + + upsert_response( + kind=test_type, dataset=dataset, config=config, split=split, content=content, http_status=HTTPStatus.OK + ) + + processing_steps = [ + ProcessingStep( + name=test_type, + input_type="dataset", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + job_runner_version=1, + ) + ] + + collect_metrics(processing_steps=processing_steps) + + cache_metrics = CacheTotalMetric.objects() + assert cache_metrics + assert len(cache_metrics) == 1 + + job_metrics = JobTotalMetric.objects() + assert job_metrics + assert len(job_metrics) == 6 # One by each job state, see libcommon.queue.get_jobs_count_by_status + waiting_job = next((job for job in job_metrics if job.status == "waiting"), None) + assert waiting_job + assert waiting_job.total == 1 + + remaining_status = [job for job in job_metrics if job.status != "waiting"] + assert remaining_status + assert all(job.total == 0 for job in remaining_status) diff --git a/libs/libcommon/Makefile b/libs/libcommon/Makefile index 932267a9..59cb1772 100644 --- a/libs/libcommon/Makefile +++ b/libs/libcommon/Makefile @@ -4,0 +5 @@ export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +export METRICS_MONGO_URL := mongodb://localhost:${MONGO_PORT} diff --git a/libs/libcommon/README.md b/libs/libcommon/README.md index 04b6fba1..f8b39f10 100644 --- a/libs/libcommon/README.md +++ b/libs/libcommon/README.md @@ -38,0 +39,7 @@ Set environment variables to configure the job queues to precompute API response + +## Metrics configuration + +Set environment variables to configure the storage of calculated metrics in a MongoDB database: + +- `METRICS_MONGO_DATABASE`: name of the database used for storing the metrics. Defaults to `datasets_server_metrics`. +- `METRICS_MONGO_URL`: URL used to connect to the MongoDB server. Defaults to `mongodb://localhost:27017`. diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index ddd7ed31..ad8ba6ea 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -163,0 +164,19 @@ class QueueConfig: +METRICS_MONGO_DATABASE = "datasets_server_metrics" +METRICS_MONGO_URL = "mongodb://localhost:27017" + + +@dataclass(frozen=True) +class MetricsConfig: + mongo_database: str = METRICS_MONGO_DATABASE + mongo_url: str = METRICS_MONGO_URL + + @classmethod + def from_env(cls) -> "MetricsConfig": + env = Env(expand_vars=True) + with env.prefixed("METRICS_"): + return cls( + mongo_database=env.str(name="MONGO_DATABASE", default=METRICS_MONGO_DATABASE), + mongo_url=env.str(name="MONGO_URL", default=METRICS_MONGO_URL), + ) + + diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 83327cc6..c6b2de41 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -7,0 +8 @@ QUEUE_MONGOENGINE_ALIAS = "queue" +METRICS_MONGOENGINE_ALIAS = "metrics" diff --git a/libs/libcommon/src/libcommon/metrics.py b/libs/libcommon/src/libcommon/metrics.py new file mode 100644 index 00000000..6d60a107 --- /dev/null +++ b/libs/libcommon/src/libcommon/metrics.py @@ -0,0 +1,88 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import types +from typing import Generic, Type, TypeVar + +from bson import ObjectId +from mongoengine import Document +from mongoengine.fields import DateTimeField, IntField, ObjectIdField, StringField +from mongoengine.queryset.queryset import QuerySet + +from libcommon.constants import METRICS_MONGOENGINE_ALIAS +from libcommon.utils import get_datetime + +# START monkey patching ### hack ### +# see https://github.com/sbdchd/mongo-types#install +U = TypeVar("U", bound=Document) + + +def no_op(self, x): # type: ignore + return self + + +QuerySet.__class_getitem__ = types.MethodType(no_op, QuerySet) + + +class QuerySetManager(Generic[U]): + def __get__(self, instance: object, cls: Type[U]) -> QuerySet[U]: + return QuerySet(cls, cls._get_collection()) + + +# END monkey patching ### hack ### + + +class JobTotalMetric(Document): + """Jobs total metric in mongoDB database, used to compute prometheus metrics. + + Args: + queue (`str`): queue name + status (`str`): job status see libcommon.queue.Status + total (`int`): total of jobs + created_at (`datetime`): when the metric has been created. + """ + + id = ObjectIdField(db_field="_id", primary_key=True, default=ObjectId) + queue = StringField(required=True) + status = StringField(required=True) + total = IntField(required=True, default=0) + created_at = DateTimeField(default=get_datetime) + + meta = { + "collection": "jobTotalMetric", + "db_alias": METRICS_MONGOENGINE_ALIAS, + "indexes": [("queue", "status")], + } + objects = QuerySetManager["JobTotalMetric"]() + + +class CacheTotalMetric(Document): + """Cache total metric in the mongoDB database, used to compute prometheus metrics. + + Args: + kind (`str`): kind name + http_status (`int`): cache http_status + error_code (`str`): error code name + total (`int`): total of jobs + created_at (`datetime`): when the metric has been created. + """ + + id = ObjectIdField(db_field="_id", primary_key=True, default=ObjectId) + kind = StringField(required=True) + http_status = IntField(required=True) + error_code = StringField() + total = IntField(required=True, default=0) + created_at = DateTimeField(default=get_datetime) + + meta = { + "collection": "cacheTotalMetric", + "db_alias": METRICS_MONGOENGINE_ALIAS, + "indexes": [("kind", "http_status", "error_code")], + } + objects = QuerySetManager["CacheTotalMetric"]() + + +# only for the tests +def _clean_metrics_database() -> None: + CacheTotalMetric.drop_collection() # type: ignore + JobTotalMetric.drop_collection() # type: ignore diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 492a80d5..ae51c3e9 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -9 +9 @@ from collections import Counter -from datetime import datetime, timedelta, timezone +from datetime import datetime, timedelta @@ -19,0 +20 @@ from libcommon.constants import QUEUE_MONGOENGINE_ALIAS, QUEUE_TTL_SECONDS +from libcommon.utils import get_datetime @@ -99,4 +99,0 @@ class EmptyQueueError(Exception): -def get_datetime() -> datetime: - return datetime.now(timezone.utc) - - diff --git a/libs/libcommon/src/libcommon/resources.py b/libs/libcommon/src/libcommon/resources.py index bb227a5a..e89121ce 100644 --- a/libs/libcommon/src/libcommon/resources.py +++ b/libs/libcommon/src/libcommon/resources.py @@ -12 +12,5 @@ from pymongo.errors import ServerSelectionTimeoutError -from libcommon.constants import CACHE_MONGOENGINE_ALIAS, QUEUE_MONGOENGINE_ALIAS +from libcommon.constants import ( + CACHE_MONGOENGINE_ALIAS, + METRICS_MONGOENGINE_ALIAS, + QUEUE_MONGOENGINE_ALIAS, +) @@ -127,0 +132,13 @@ class QueueMongoResource(MongoResource): + + +@dataclass +class MetricsMongoResource(MongoResource): + """ + A resource that represents a connection to the metrics mongo database. + + Args: + database (:obj:`str`): The name of the mongo database. + host (:obj:`str`): The host of the mongo database. It must start with ``mongodb://`` or ``mongodb+srv://``. + """ + + mongoengine_alias: str = field(default=METRICS_MONGOENGINE_ALIAS, init=False) diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index bbf451c4..c08be874 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -7 +7 @@ from dataclasses import dataclass -from datetime import datetime, timezone +from datetime import datetime @@ -36,0 +37 @@ from libcommon.constants import CACHE_MONGOENGINE_ALIAS +from libcommon.utils import get_datetime @@ -58,4 +58,0 @@ class QuerySetManager(Generic[U]): -def get_datetime() -> datetime: - return datetime.now(timezone.utc) - - diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py index a03b6138..39db0ed7 100644 --- a/libs/libcommon/src/libcommon/utils.py +++ b/libs/libcommon/src/libcommon/utils.py @@ -4,0 +5 @@ import base64 +from datetime import datetime, timezone @@ -21,0 +23,4 @@ def orjson_dumps(content: Any) -> bytes: + + +def get_datetime() -> datetime: + return datetime.now(timezone.utc) diff --git a/libs/libcommon/tests/conftest.py b/libs/libcommon/tests/conftest.py index 8e81561b..0c30c347 100644 --- a/libs/libcommon/tests/conftest.py +++ b/libs/libcommon/tests/conftest.py @@ -40,0 +41,11 @@ def queue_mongo_host(env: Env) -> str: +@fixture(scope="session") +def metrics_mongo_host(env: Env) -> str: + try: + url = env.str(name="METRICS_MONGO_URL") + if type(url) is not str: + raise ValueError("METRICS_MONGO_URL is not set") + return url + except Exception as e: + raise ValueError("METRICS_MONGO_URL is not set") from e + + diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 3bb6a1fc..3db33a5a 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -18 +17,0 @@ from libcommon.queue import ( - get_datetime, @@ -20,0 +20 @@ from libcommon.resources import QueueMongoResource +from libcommon.utils import get_datetime diff --git a/libs/libcommon/tests/test_resources.py b/libs/libcommon/tests/test_resources.py index 48450db8..cb6ce2da 100644 --- a/libs/libcommon/tests/test_resources.py +++ b/libs/libcommon/tests/test_resources.py @@ -12,0 +13 @@ from libcommon.resources import ( + MetricsMongoResource, @@ -117,0 +119,14 @@ def test_queue_database(queue_mongo_host: str) -> None: + + +def test_metrics_database(metrics_mongo_host: str) -> None: + resource = MetricsMongoResource(database="test_metrics_database", host=metrics_mongo_host) + + class User(Document): + name = StringField() + meta = {"db_alias": resource.mongoengine_alias} + + assert len(User.objects()) == 0 # type: ignore + # clean + User.drop_collection() # type: ignore + assert resource.is_available() + resource.release() diff --git a/services/admin/Makefile b/services/admin/Makefile index 1dc6ae35..cd552232 100644 --- a/services/admin/Makefile +++ b/services/admin/Makefile @@ -5,0 +6 @@ export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +export METRICS_MONGO_URL := mongodb://localhost:${MONGO_PORT} diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py index 0e29841f..49ba9efd 100644 --- a/services/admin/src/admin/app.py +++ b/services/admin/src/admin/app.py @@ -7 +7,6 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource +from libcommon.resources import ( + CacheMongoResource, + MetricsMongoResource, + QueueMongoResource, + Resource, +) @@ -44 +49,5 @@ def create_app() -> Starlette: - resources: list[Resource] = [cache_resource, queue_resource] + metrics_resource = MetricsMongoResource( + database=app_config.metrics.mongo_database, host=app_config.metrics.mongo_url + ) + + resources: list[Resource] = [cache_resource, queue_resource, metrics_resource] @@ -48,0 +58,2 @@ def create_app() -> Starlette: + if not metrics_resource.is_available(): + raise RuntimeError("The connection to the metrics database could not be established. Exiting.") diff --git a/services/admin/src/admin/config.py b/services/admin/src/admin/config.py index 4434dcb0..7aa5a232 100644 --- a/services/admin/src/admin/config.py +++ b/services/admin/src/admin/config.py @@ -12,0 +13 @@ from libcommon.config import ( + MetricsConfig, @@ -84,0 +86 @@ class AppConfig: + metrics: MetricsConfig = field(default_factory=MetricsConfig) @@ -96,0 +99 @@ class AppConfig: + metrics=MetricsConfig.from_env(), diff --git a/services/admin/src/admin/prometheus.py b/services/admin/src/admin/prometheus.py index a30b9647..690b5854 100644 --- a/services/admin/src/admin/prometheus.py +++ b/services/admin/src/admin/prometheus.py @@ -7,0 +8 @@ from typing import Any, List +from libcommon.metrics import CacheTotalMetric, JobTotalMetric @@ -9,2 +10 @@ from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Queue -from libcommon.simple_cache import get_responses_count_by_kind_status_and_error_code +from libcommon.queue import Status @@ -65,4 +65,11 @@ class Prometheus: - queue = Queue() - for processing_step in self.processing_steps: - for status, total in queue.get_jobs_count_by_status(job_type=processing_step.job_type).items(): - QUEUE_JOBS_TOTAL.labels(queue=processing_step.job_type, status=status).set(total) + queue_jobs_total = JobTotalMetric.objects() + if not queue_jobs_total: + # TODO: Move this logic to a metrics manager + # In case job collected metrics do not exist, fill with 0 + for processing_step in self.processing_steps: + for status in Status: + QUEUE_JOBS_TOTAL.labels(queue=processing_step.job_type, status=status.value).set(0) + else: + for job_metric in queue_jobs_total: + QUEUE_JOBS_TOTAL.labels(queue=job_metric.queue, status=job_metric.status).set(job_metric.total) + @@ -70,4 +77,14 @@ class Prometheus: - for metric in get_responses_count_by_kind_status_and_error_code(): - RESPONSES_IN_CACHE_TOTAL.labels( - kind=metric["kind"], http_status=metric["http_status"], error_code=metric["error_code"] - ).set(metric["count"]) + responses_in_cache_total = CacheTotalMetric.objects() + if not responses_in_cache_total: + # TODO: Move this logic to a metrics manager + # In case cache collected metrics do not exist, fill with 0 + for processing_step in self.processing_steps: + RESPONSES_IN_CACHE_TOTAL.labels( + kind=processing_step.cache_kind, http_status="200", error_code="None" + ).set(0) + else: + for cache_metric in responses_in_cache_total: + RESPONSES_IN_CACHE_TOTAL.labels( + kind=cache_metric.kind, http_status=cache_metric.http_status, error_code=cache_metric.error_code + ).set(cache_metric.total) + diff --git a/services/admin/tests/conftest.py b/services/admin/tests/conftest.py index 1e21b047..3359a15d 100644 --- a/services/admin/tests/conftest.py +++ b/services/admin/tests/conftest.py @@ -5,0 +6 @@ from typing import Iterator, List +from libcommon.metrics import _clean_metrics_database @@ -8 +9,5 @@ from libcommon.queue import _clean_queue_database -from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.resources import ( + CacheMongoResource, + MetricsMongoResource, + QueueMongoResource, +) @@ -24,0 +30 @@ def monkeypatch_session(hf_endpoint: str, hf_token: str) -> Iterator[MonkeyPatch + monkeypatch_session.setenv("METRICS_MONGO_DATABASE", "datasets_server_metrics_test") @@ -61,0 +68,9 @@ def queue_mongo_resource(app_config: AppConfig) -> Iterator[QueueMongoResource]: + + +@fixture(autouse=True) +def metrics_mongo_resource(app_config: AppConfig) -> Iterator[MetricsMongoResource]: + with MetricsMongoResource( + database=app_config.metrics.mongo_database, host=app_config.metrics.mongo_url + ) as resource: + yield resource + _clean_metrics_database() diff --git a/services/admin/tests/test_prometheus.py b/services/admin/tests/test_prometheus.py index f417fb73..3850d2a0 100644 --- a/services/admin/tests/test_prometheus.py +++ b/services/admin/tests/test_prometheus.py @@ -4,0 +5 @@ import os +from http import HTTPStatus @@ -6,0 +8 @@ from typing import List +from libcommon.metrics import CacheTotalMetric, JobTotalMetric @@ -16,0 +19,19 @@ def test_prometheus( + cache_metric = { + "kind": "dummy", + "http_status": HTTPStatus.OK, + "error_code": None, + "total": 1, + } + + collection = CacheTotalMetric._get_collection() + collection.insert_one(cache_metric) + + job_metric = { + "queue": "dummy", + "status": "waiting", + "total": 1, + } + + collection = JobTotalMetric._get_collection() + collection.insert_one(job_metric) + @@ -36,5 +57,3 @@ def test_prometheus( - for processing_step in processing_steps: - assert ( - "queue_jobs_total{" + additional_field + 'queue="' + processing_step.job_type + '",status="started"}' - in metrics - ) + + assert 'responses_in_cache_total{error_code="None",http_status="200",kind="dummy"}' in metrics + assert 'queue_jobs_total{queue="dummy",status="waiting"}' in metrics diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py index aea8a152..bb8a6af9 100644 --- a/services/worker/src/worker/executor.py +++ b/services/worker/src/worker/executor.py @@ -12 +12,2 @@ from filelock import FileLock -from libcommon.queue import Queue, get_datetime +from libcommon.queue import Queue +from libcommon.utils import get_datetime diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py index e961fdde..6b7dade9 100644 --- a/services/worker/src/worker/loop.py +++ b/services/worker/src/worker/loop.py @@ -13 +13,2 @@ from filelock import FileLock -from libcommon.queue import EmptyQueueError, JobInfo, Queue, get_datetime +from libcommon.queue import EmptyQueueError, JobInfo, Queue +from libcommon.utils import get_datetime diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py index ee8db769..5f32f3b6 100644 --- a/services/worker/tests/test_executor.py +++ b/services/worker/tests/test_executor.py @@ -15,9 +15 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.queue import ( - DoesNotExist, - Job, - JobInfo, - Priority, - Queue, - Status, - get_datetime, -) +from libcommon.queue import DoesNotExist, Job, JobInfo, Priority, Queue, Status @@ -26,0 +19 @@ from libcommon.storage import StrPath +from libcommon.utils import get_datetime diff --git a/tools/docker-compose-base.yml b/tools/docker-compose-base.yml index a8a383e4..1cf2b09a 100644 --- a/tools/docker-compose-base.yml +++ b/tools/docker-compose-base.yml @@ -18,0 +19,3 @@ services: + # metrics + METRICS_MONGO_URL: ${METRICS_MONGO_URL-mongodb://mongodb} # use mongo container by default + METRICS_MONGO_DATABASE: ${METRICS_MONGO_DATABASE-datasets_server_metrics} diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index cbb6a092..28fbd274 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -18,0 +19,3 @@ services: + # metrics + METRICS_MONGO_URL: ${METRICS_MONGO_URL-mongodb://host.docker.internal:${MONGO_PORT-27017}} # use mongo container by default + METRICS_MONGO_DATABASE: ${METRICS_MONGO_DATABASE-datasets_server_metrics}
880959a69dbf5f0f546c8c359f67c9d7bec47b08
Sylvain Lesage
2023-04-12T16:47:17
fix: 🐛 add missing environment variable (#1048)
diff --git a/chart/templates/reverse-proxy/_container.tpl b/chart/templates/reverse-proxy/_container.tpl index 01018420..5f1a0952 100644 --- a/chart/templates/reverse-proxy/_container.tpl +++ b/chart/templates/reverse-proxy/_container.tpl @@ -10,0 +11,2 @@ + - name: CACHED_ASSETS_DIRECTORY + value: {{ .Values.cachedAssets.storageDirectory | quote }}
3c68950d658a2b97b8cd33a79a8dbbe715df6139
Quentin Lhoest
2023-04-12T16:32:37
Random access: image and audio files support (#1026)
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml index 9c1af905..ecd866c3 100644 --- a/.github/workflows/_quality-python.yml +++ b/.github/workflows/_quality-python.yml @@ -11,3 +10,0 @@ on: - is-datasets-worker: - required: false - type: boolean @@ -37,2 +34 @@ jobs: - - name: Install packages for workers that use datasets - if: ${{ inputs.is-datasets-worker }} + - name: Install packages @@ -56 +52 @@ jobs: - run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^pymongo==/,+109 d' | sed '/^requests==2.28.2 ;/,+2 d' | sed '/^kenlm @/d' | sed '/^fsspec==/,+2 d' | sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" + run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" diff --git a/.github/workflows/_unit-tests-python.yml b/.github/workflows/_unit-tests-python.yml index 232c189b..dd05f1da 100644 --- a/.github/workflows/_unit-tests-python.yml +++ b/.github/workflows/_unit-tests-python.yml @@ -11,3 +10,0 @@ on: - is-datasets-worker: - required: false - type: boolean @@ -38,2 +35 @@ jobs: - - name: Install packages for workers that use datasets - if: ${{ inputs.is-datasets-worker }} + - name: Install packages diff --git a/.github/workflows/s-worker.yml b/.github/workflows/s-worker.yml index 20f562e9..cd234c9c 100644 --- a/.github/workflows/s-worker.yml +++ b/.github/workflows/s-worker.yml @@ -32 +31,0 @@ jobs: - is-datasets-worker: true @@ -37 +35,0 @@ jobs: - is-datasets-worker: true diff --git a/chart/nginx-templates/default.conf.template b/chart/nginx-templates/default.conf.template index 3c8d9d89..ac2edf9c 100644 --- a/chart/nginx-templates/default.conf.template +++ b/chart/nginx-templates/default.conf.template @@ -35,0 +36,4 @@ server { + location /cached-assets/ { + alias ${CACHED_ASSETS_DIRECTORY}/; + } + diff --git a/chart/templates/_envCachedAssets.tpl b/chart/templates/_envCachedAssets.tpl new file mode 100644 index 00000000..52445b5b --- /dev/null +++ b/chart/templates/_envCachedAssets.tpl @@ -0,0 +1,17 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "envCachedAssets" -}} +- name: CACHED_ASSETS_BASE_URL + value: "{{ include "cachedAssets.baseUrl" . }}" +- name: CACHED_ASSETS_STORAGE_DIRECTORY + value: {{ .Values.cachedAssets.storageDirectory | quote }} +- name: CACHED_ASSETS_CLEAN_CACHE_PROBA + value: {{ .Values.cachedAssets.cleanCacheProba | quote }} +- name: CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER + value: {{ .Values.cachedAssets.keepFirstRowsNumber | quote }} +- name: CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER + value: {{ .Values.cachedAssets.keepMostRecentRowsNumber | quote }} +- name: CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER + value: {{ .Values.cachedAssets.maxCleanedRowsNumber | quote }} +{{- end -}} diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index 98cf42eb..b6b294fb 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -136,0 +137,16 @@ The assets/ subpath in the NFS +{{/* +The cached-assets base URL +*/}} +{{- define "cachedAssets.baseUrl" -}} +{{- printf "%s%s/cached-assets" (include "datasetsServer.ingress.scheme" .) (include "datasetsServer.ingress.hostname" .) }} +{{- end }} + +{{/* +The cached-assets/ subpath in the NFS +- in a subdirectory named as the chart (datasets-server/), and below it, +- in a subdirectory named as the Release, so that Releases will not share the same dir +*/}} +{{- define "cachedAssets.subpath" -}} +{{- printf "%s/%s/%s/" .Chart.Name .Release.Name "cached-assets" }} +{{- end }} + diff --git a/chart/templates/_initContainerCachedAssets.tpl b/chart/templates/_initContainerCachedAssets.tpl new file mode 100644 index 00000000..d4a603b2 --- /dev/null +++ b/chart/templates/_initContainerCachedAssets.tpl @@ -0,0 +1,21 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "initContainerCachedAssets" -}} +- name: prepare-cached-assets + image: ubuntu:focal + imagePullPolicy: {{ .Values.images.pullPolicy }} + command: ["/bin/sh", "-c"] + args: + - chown {{ .Values.uid }}:{{ .Values.gid }} /mounted-path; + volumeMounts: + - mountPath: /mounted-path + mountPropagation: None + name: data + subPath: "{{ include "cachedAssets.subpath" . }}" + readOnly: false + securityContext: + runAsNonRoot: false + runAsUser: 0 + runAsGroup: 0 +{{- end -}} diff --git a/chart/templates/_volumeMountCachedAssets.tpl b/chart/templates/_volumeMountCachedAssets.tpl new file mode 100644 index 00000000..7ee8dcf1 --- /dev/null +++ b/chart/templates/_volumeMountCachedAssets.tpl @@ -0,0 +1,18 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "volumeMountCachedAssetsRO" -}} +- mountPath: {{ .Values.cachedAssets.storageDirectory | quote }} + mountPropagation: None + name: data + subPath: "{{ include "cachedAssets.subpath" . }}" + readOnly: true +{{- end -}} + +{{- define "volumeMountCachedAssetsRW" -}} +- mountPath: {{ .Values.cachedAssets.storageDirectory | quote }} + mountPropagation: None + name: data + subPath: "{{ include "cachedAssets.subpath" . }}" + readOnly: false +{{- end -}} diff --git a/chart/templates/reverse-proxy/_container.tpl b/chart/templates/reverse-proxy/_container.tpl index 19bda663..01018420 100644 --- a/chart/templates/reverse-proxy/_container.tpl +++ b/chart/templates/reverse-proxy/_container.tpl @@ -20,0 +21 @@ + {{ include "volumeMountCachedAssetsRO" . | nindent 2 }} diff --git a/chart/templates/reverse-proxy/deployment.yaml b/chart/templates/reverse-proxy/deployment.yaml index 3d0398be..b20f26e0 100644 --- a/chart/templates/reverse-proxy/deployment.yaml +++ b/chart/templates/reverse-proxy/deployment.yaml @@ -29 +29,3 @@ spec: - initContainers: {{ include "initContainerAssets" . | nindent 8 }} + initContainers: + {{ include "initContainerAssets" . | nindent 8 }} + {{ include "initContainerCachedAssets" . | nindent 8 }} diff --git a/chart/templates/services/admin/_container.tpl b/chart/templates/services/admin/_container.tpl index dfd0d6df..04925e82 100644 --- a/chart/templates/services/admin/_container.tpl +++ b/chart/templates/services/admin/_container.tpl @@ -9,0 +10 @@ + {{ include "envCachedAssets" . | nindent 2 }} @@ -35 +36,3 @@ - volumeMounts: {{ include "volumeMountAssetsRO" . | nindent 2 }} + volumeMounts: + {{ include "volumeMountAssetsRO" . | nindent 2 }} + {{ include "volumeMountCachedAssetsRO" . | nindent 2 }} diff --git a/chart/templates/services/admin/deployment.yaml b/chart/templates/services/admin/deployment.yaml index c701e356..79ddd1f0 100644 --- a/chart/templates/services/admin/deployment.yaml +++ b/chart/templates/services/admin/deployment.yaml @@ -26 +26,3 @@ spec: - initContainers: {{ include "initContainerAssets" . | nindent 8 }} + initContainers: + {{ include "initContainerAssets" . | nindent 8 }} + {{ include "initContainerCachedAssets" . | nindent 8 }} diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl index c28f92fa..317695ef 100644 --- a/chart/templates/services/api/_container.tpl +++ b/chart/templates/services/api/_container.tpl @@ -8,0 +9 @@ + {{ include "envCachedAssets" . | nindent 2 }} @@ -45,0 +47,2 @@ + volumeMounts: + {{ include "volumeMountCachedAssetsRW" . | nindent 2 }} diff --git a/chart/templates/services/api/deployment.yaml b/chart/templates/services/api/deployment.yaml index e77b45e3..3977649a 100644 --- a/chart/templates/services/api/deployment.yaml +++ b/chart/templates/services/api/deployment.yaml @@ -25,0 +26,2 @@ spec: + initContainers: + {{ include "initContainerCachedAssets" . | nindent 8 }} diff --git a/chart/values.yaml b/chart/values.yaml index a88727f4..0a12c555 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -169,0 +170,14 @@ assets: +cachedAssets: + # base URL for the cached assets files. It should be set accordingly to the datasets-server domain, eg https://datasets-server.huggingface.co/cached-assets + # baseUrl: "not used for now" + # Directory on the cached shared storage (audio files and images) + storageDirectory: "/cached-assets" + # Probability of cleaning the cached assets directory at each request. + cleanCacheProba: 0.05 + # When cleaning the cached assets directory: keep the rows with an index below a certain number. + keepFirstRowsNumber: 100 + # When cleaning the cached assets directory: keep the most recently accessed rows. + keepMostRecentRowsNumber: 200 + # When cleaning the cached assets directory: maximum number of rows to discard. + maxCleanedRowsNumber: 10000 + diff --git a/e2e/poetry.lock b/e2e/poetry.lock index 8cebc09a..d84fc362 100644 --- a/e2e/poetry.lock +++ b/e2e/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -305 +305 @@ name = "cyclonedx-python-lib" -version = "3.1.5" +version = "2.7.1" @@ -311,2 +311,2 @@ files = [ - {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, - {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, + {file = "cyclonedx-python-lib-2.7.1.tar.gz", hash = "sha256:493bf2f30e26c48f305f745ed8580ce10d05a8d68d62a598fe95f05a0d9007dc"}, + {file = "cyclonedx_python_lib-2.7.1-py3-none-any.whl", hash = "sha256:fabc4c8baf722faeea01c3bbca83730e3489dfb37d85c6036baa67a9a7519d40"}, @@ -890 +890 @@ name = "pip-audit" -version = "2.4.14" +version = "2.5.4" @@ -896,2 +896,2 @@ files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, + {file = "pip_audit-2.5.4-py3-none-any.whl", hash = "sha256:bee3748030c895488b4dd8a6196fa44f484da000cfd52d8fc64dfa3a2e121624"}, + {file = "pip_audit-2.5.4.tar.gz", hash = "sha256:a4cb03f9e2896d626f5b153973d3ac0d32fdb18594d78d393b153c83bb8089b6"}, @@ -902 +902 @@ CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +cyclonedx-python-lib = ">=2.0,<2.5.0 || >2.5.0,<3.0" @@ -907 +906,0 @@ pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" @@ -914 +913 @@ doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.260)", "types-html5lib", "types-requests", "types-toml"] @@ -1176,18 +1174,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "resolvelib" -version = "0.9.0" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"}, - {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"}, -] - -[package.extras] -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] - @@ -1412 +1393 @@ python-versions = "3.9.15" -content-hash = "e903d839571339372972d5d533fc7386aa70b612240f5ee6f3d0f1673be1d0ba" +content-hash = "417444e7564752194ddc94c1ac78a99891f4ef9acb1b14bd477867151f4bbb60" diff --git a/e2e/pyproject.toml b/e2e/pyproject.toml index 189e1efa..70736558 100644 --- a/e2e/pyproject.toml +++ b/e2e/pyproject.toml @@ -19 +19 @@ mypy = "^1.0.0" -pip-audit = "^2.4.14" +pip-audit = "^2.5.4" diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock index ee4e2a3a..1527c84b 100644 --- a/jobs/cache_maintenance/poetry.lock +++ b/jobs/cache_maintenance/poetry.lock @@ -1 +1,125 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. + +[[package]] +name = "aiohttp" +version = "3.8.4" +description = "Async http client/server framework (asyncio)" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"}, + {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"}, + {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"}, + {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"}, + {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"}, + {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"}, + {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"}, + {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"}, + {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"}, + {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"}, + {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"}, + {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"}, +] + +[package.dependencies] +aiosignal = ">=1.1.2" +async-timeout = ">=4.0.0a3,<5.0" +attrs = ">=17.3.0" +charset-normalizer = ">=2.0,<4.0" +frozenlist = ">=1.1.1" +multidict = ">=4.5,<7.0" +yarl = ">=1.0,<2.0" + +[package.extras] +speedups = ["Brotli", "aiodns", "cchardet"] + +[[package]] +name = "aiosignal" +version = "1.3.1" +description = "aiosignal: a list of registered asynchronous callbacks" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, + {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, +] + +[package.dependencies] +frozenlist = ">=1.1.0" @@ -14,0 +139,12 @@ files = [ +[[package]] +name = "async-timeout" +version = "4.0.2" +description = "Timeout context manager for asyncio programs" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"}, + {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"}, +] + @@ -19 +155 @@ description = "Classes Without Boilerplate" -category = "dev" +category = "main" @@ -33,0 +170,11 @@ tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy +[[package]] +name = "audioread" +version = "3.0.0" +description = "multi-library, cross-platform audio decoding" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"}, +] + @@ -126,0 +274,77 @@ files = [ +[[package]] +name = "cffi" +version = "1.15.1" +description = "Foreign Function Interface for Python calling C code." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"}, + {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"}, + {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"}, + {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"}, + {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"}, + {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"}, + {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"}, + {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"}, + {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"}, + {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"}, + {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"}, + {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"}, + {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"}, + {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"}, + {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"}, + {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"}, + {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"}, + {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"}, + {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"}, +] + +[package.dependencies] +pycparser = "*" + @@ -320,0 +545,74 @@ toml = ">=0.10.0,<0.11.0" +[[package]] +name = "datasets" +version = "2.11.0" +description = "HuggingFace community-driven open-source library of datasets" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "datasets-2.11.0-py3-none-any.whl", hash = "sha256:d946cdb8c4885d3016a2ab3129c9403dd3358fe9107e8ab5e549ceab672774af"}, + {file = "datasets-2.11.0.tar.gz", hash = "sha256:1ca53b9cd6ece7a3fdb81176dadd5b9e646420e52e68e85307b27db3a36ca18c"}, +] + +[package.dependencies] +aiohttp = "*" +dill = ">=0.3.0,<0.3.7" +fsspec = {version = ">=2021.11.1", extras = ["http"]} +huggingface-hub = ">=0.11.0,<1.0.0" +librosa = {version = "*", optional = true, markers = "extra == \"audio\""} +multiprocess = "*" +numpy = ">=1.17" +packaging = "*" +pandas = "*" +Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} +pyarrow = ">=8.0.0" +pyyaml = ">=5.1" +requests = ">=2.19.0" +responses = "<0.19" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} +tqdm = ">=4.62.1" +xxhash = "*" + +[package.extras] +apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] +audio = ["librosa", "soundfile (>=0.12.1)"] +benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +docs = ["s3fs"] +jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"] +s3 = ["s3fs"] +tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"] +tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +torch = ["torch"] +vision = ["Pillow (>=6.2.1)"] + +[[package]] +name = "decorator" +version = "5.1.1" +description = "Decorators for Humans" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"}, + {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"}, +] + +[[package]] +name = "dill" +version = "0.3.6" +description = "serialize all of python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"}, + {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"}, +] + +[package.extras] +graph = ["objgraph (>=1.7.2)"] + @@ -406,0 +705,122 @@ pyflakes = ">=2.3.0,<2.4.0" +[[package]] +name = "frozenlist" +version = "1.3.3" +description = "A list-like structure which implements collections.abc.MutableSequence" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"}, + {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"}, + {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"}, + {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"}, + {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"}, + {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"}, + {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"}, + {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"}, + {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"}, + {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"}, + {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"}, +] + +[[package]] +name = "fsspec" +version = "2023.3.0" +description = "File-system specification" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "fsspec-2023.3.0-py3-none-any.whl", hash = "sha256:bf57215e19dbfa4fe7edae53040cc1deef825e3b1605cca9a8d2c2fadd2328a0"}, + {file = "fsspec-2023.3.0.tar.gz", hash = "sha256:24e635549a590d74c6c18274ddd3ffab4753341753e923408b1904eaabafe04d"}, +] + +[package.dependencies] +aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""} +requests = {version = "*", optional = true, markers = "extra == \"http\""} + +[package.extras] +abfs = ["adlfs"] +adl = ["adlfs"] +arrow = ["pyarrow (>=1)"] +dask = ["dask", "distributed"] +dropbox = ["dropbox", "dropboxdrivefs", "requests"] +fuse = ["fusepy"] +gcs = ["gcsfs"] +git = ["pygit2"] +github = ["requests"] +gs = ["gcsfs"] +gui = ["panel"] +hdfs = ["pyarrow (>=1)"] +http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"] +libarchive = ["libarchive-c"] +oci = ["ocifs"] +s3 = ["s3fs"] +sftp = ["paramiko"] +smb = ["smbprotocol"] +ssh = ["paramiko"] +tqdm = ["tqdm"] + @@ -531,0 +952,28 @@ requirements-deprecated-finder = ["pip-api", "pipreqs"] +[[package]] +name = "joblib" +version = "1.2.0" +description = "Lightweight pipelining with Python functions" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "joblib-1.2.0-py3-none-any.whl", hash = "sha256:091138ed78f800342968c523bdde947e7a305b8594b910a0fea2ab83c3c6d385"}, + {file = "joblib-1.2.0.tar.gz", hash = "sha256:e1cee4a79e4af22881164f218d4311f60074197fb707e082e803b61f6d137018"}, +] + +[[package]] +name = "lazy-loader" +version = "0.2" +description = "lazy_loader" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "lazy_loader-0.2-py3-none-any.whl", hash = "sha256:c35875f815c340f823ce3271ed645045397213f961b40ad0c0d395c3f5218eeb"}, + {file = "lazy_loader-0.2.tar.gz", hash = "sha256:0edc7a5175c400acb108f283749951fefdadedeb00adcec6e88b974a9254f18a"}, +] + +[package.extras] +lint = ["pre-commit (>=3.1)"] +test = ["pytest (>=7.2)", "pytest-cov (>=4.0)"] + @@ -543,0 +992 @@ appdirs = "^1.4.4" +datasets = {version = "^2.11.0", extras = ["audio", "vision"]} @@ -548,0 +998 @@ networkx = "^3.0" +numba = "0.56.4" @@ -550,0 +1001 @@ psutil = "^5.9.4" +pydub = "^0.25.1" @@ -553,0 +1005 @@ requests = "^2.28.2" +soundfile = ">=0.12.1" @@ -558,0 +1011,70 @@ url = "../../libs/libcommon" +[[package]] +name = "librosa" +version = "0.10.0.post2" +description = "Python module for audio and music processing" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"}, + {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"}, +] + +[package.dependencies] +audioread = ">=2.1.9" +decorator = ">=4.3.0" +joblib = ">=0.14" +lazy-loader = ">=0.1" +msgpack = ">=1.0" +numba = ">=0.51.0" +numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2" +pooch = ">=1.0,<1.7" +scikit-learn = ">=0.20.0" +scipy = ">=1.2.0" +soundfile = ">=0.12.1" +soxr = ">=0.3.2" +typing-extensions = ">=4.1.1" + +[package.extras] +display = ["matplotlib (>=3.3.0)"] +docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"] +tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"] + +[[package]] +name = "llvmlite" +version = "0.39.1" +description = "lightweight wrapper around basic LLVM functionality" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"}, + {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"}, + {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"}, + {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"}, + {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"}, + {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"}, + {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"}, + {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"}, + {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"}, + {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"}, +] + @@ -672 +1194 @@ description = "MessagePack serializer" -category = "dev" +category = "main" @@ -740,0 +1263,111 @@ files = [ +[[package]] +name = "multidict" +version = "6.0.4" +description = "multidict implementation" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"}, + {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"}, + {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"}, + {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"}, + {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"}, + {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"}, + {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"}, + {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"}, + {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"}, + {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"}, + {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"}, + {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"}, + {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"}, +] + +[[package]] +name = "multiprocess" +version = "0.70.14" +description = "better multiprocessing and multithreading in python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"}, + {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"}, + {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"}, + {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"}, + {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"}, + {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"}, +] + +[package.dependencies] +dill = ">=0.3.6" + @@ -818,0 +1452,81 @@ test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] +[[package]] +name = "numba" +version = "0.56.4" +description = "compiling Python code using LLVM" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"}, + {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"}, + {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"}, + {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"}, + {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"}, + {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"}, + {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"}, + {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"}, + {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"}, + {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"}, + {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"}, + {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"}, + {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"}, + {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"}, + {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"}, + {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"}, +] + +[package.dependencies] +llvmlite = ">=0.39.0dev0,<0.40" +numpy = ">=1.18,<1.24" +setuptools = "*" + +[[package]] +name = "numpy" +version = "1.23.5" +description = "NumPy is the fundamental package for array computing with Python." +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"}, + {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"}, + {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"}, + {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"}, + {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"}, + {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"}, + {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"}, + {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"}, + {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"}, + {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"}, +] + @@ -906,0 +1621,64 @@ files = [ +[[package]] +name = "pandas" +version = "2.0.0" +description = "Powerful data structures for data analysis, time series, and statistics" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "pandas-2.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:bbb2c5e94d6aa4e632646a3bacd05c2a871c3aa3e85c9bec9be99cb1267279f2"}, + {file = "pandas-2.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b5337c87c4e963f97becb1217965b6b75c6fe5f54c4cf09b9a5ac52fc0bd03d3"}, + {file = "pandas-2.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6ded51f7e3dd9b4f8b87f2ceb7bd1a8df2491f7ee72f7074c6927a512607199e"}, + {file = "pandas-2.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:52c858de9e9fc422d25e67e1592a6e6135d7bcf9a19fcaf4d0831a0be496bf21"}, + {file = "pandas-2.0.0-cp310-cp310-win32.whl", hash = "sha256:2d1d138848dd71b37e3cbe7cd952ff84e2ab04d8988972166e18567dcc811245"}, + {file = "pandas-2.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:d08e41d96bc4de6f500afe80936c68fce6099d5a434e2af7c7fd8e7c72a3265d"}, + {file = "pandas-2.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:24472cfc7ced511ac90608728b88312be56edc8f19b9ed885a7d2e47ffaf69c0"}, + {file = "pandas-2.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4ffb14f50c74ee541610668137830bb93e9dfa319b1bef2cedf2814cd5ac9c70"}, + {file = "pandas-2.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c24c7d12d033a372a9daf9ff2c80f8b0af6f98d14664dbb0a4f6a029094928a7"}, + {file = "pandas-2.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8318de0f886e4dcb8f9f36e45a3d6a6c3d1cfdc508354da85e739090f0222991"}, + {file = "pandas-2.0.0-cp311-cp311-win32.whl", hash = "sha256:57c34b79c13249505e850d0377b722961b99140f81dafbe6f19ef10239f6284a"}, + {file = "pandas-2.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:8f987ec26e96a8490909bc5d98c514147236e49830cba7df8690f6087c12bbae"}, + {file = "pandas-2.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b3ba8f5dd470d8bfbc4259829589f4a32881151c49e36384d9eb982b35a12020"}, + {file = "pandas-2.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:fcd471c9d9f60926ab2f15c6c29164112f458acb42280365fbefa542d0c2fc74"}, + {file = "pandas-2.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9253edfd015520ce77a9343eb7097429479c039cd3ebe81d7810ea11b4b24695"}, + {file = "pandas-2.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:977326039bd1ded620001a1889e2ed4798460a6bc5a24fbaebb5f07a41c32a55"}, + {file = "pandas-2.0.0-cp38-cp38-win32.whl", hash = "sha256:78425ca12314b23356c28b16765639db10ebb7d8983f705d6759ff7fe41357fa"}, + {file = "pandas-2.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:d93b7fcfd9f3328072b250d6d001dcfeec5d3bb66c1b9c8941e109a46c0c01a8"}, + {file = "pandas-2.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:425705cee8be54db2504e8dd2a730684790b15e5904b750c367611ede49098ab"}, + {file = "pandas-2.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a4f789b7c012a608c08cda4ff0872fd979cb18907a37982abe884e6f529b8793"}, + {file = "pandas-2.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3bb9d840bf15656805f6a3d87eea9dcb7efdf1314a82adcf7f00b820427c5570"}, + {file = "pandas-2.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0778ab54c8f399d83d98ffb674d11ec716449956bc6f6821891ab835848687f2"}, + {file = "pandas-2.0.0-cp39-cp39-win32.whl", hash = "sha256:70db5c278bbec0306d32bf78751ff56b9594c05a5098386f6c8a563659124f91"}, + {file = "pandas-2.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:4f3320bb55f34af4193020158ef8118ee0fb9aec7cc47d2084dbfdd868a0a24f"}, + {file = "pandas-2.0.0.tar.gz", hash = "sha256:cda9789e61b44463c1c4fe17ef755de77bcd13b09ba31c940d20f193d63a5dc8"}, +] + +[package.dependencies] +numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""} +python-dateutil = ">=2.8.2" +pytz = ">=2020.1" +tzdata = ">=2022.1" + +[package.extras] +all = ["PyQt5 (>=5.15.1)", "SQLAlchemy (>=1.4.16)", "beautifulsoup4 (>=4.9.3)", "bottleneck (>=1.3.2)", "brotlipy (>=0.7.0)", "fastparquet (>=0.6.3)", "fsspec (>=2021.07.0)", "gcsfs (>=2021.07.0)", "html5lib (>=1.1)", "hypothesis (>=6.34.2)", "jinja2 (>=3.0.0)", "lxml (>=4.6.3)", "matplotlib (>=3.6.1)", "numba (>=0.53.1)", "numexpr (>=2.7.3)", "odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pandas-gbq (>=0.15.0)", "psycopg2 (>=2.8.6)", "pyarrow (>=7.0.0)", "pymysql (>=1.0.2)", "pyreadstat (>=1.1.2)", "pytest (>=7.0.0)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)", "python-snappy (>=0.6.0)", "pyxlsb (>=1.0.8)", "qtpy (>=2.2.0)", "s3fs (>=2021.08.0)", "scipy (>=1.7.1)", "tables (>=3.6.1)", "tabulate (>=0.8.9)", "xarray (>=0.21.0)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)", "zstandard (>=0.15.2)"] +aws = ["s3fs (>=2021.08.0)"] +clipboard = ["PyQt5 (>=5.15.1)", "qtpy (>=2.2.0)"] +compression = ["brotlipy (>=0.7.0)", "python-snappy (>=0.6.0)", "zstandard (>=0.15.2)"] +computation = ["scipy (>=1.7.1)", "xarray (>=0.21.0)"] +excel = ["odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pyxlsb (>=1.0.8)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)"] +feather = ["pyarrow (>=7.0.0)"] +fss = ["fsspec (>=2021.07.0)"] +gcp = ["gcsfs (>=2021.07.0)", "pandas-gbq (>=0.15.0)"] +hdf5 = ["tables (>=3.6.1)"] +html = ["beautifulsoup4 (>=4.9.3)", "html5lib (>=1.1)", "lxml (>=4.6.3)"] +mysql = ["SQLAlchemy (>=1.4.16)", "pymysql (>=1.0.2)"] +output-formatting = ["jinja2 (>=3.0.0)", "tabulate (>=0.8.9)"] +parquet = ["pyarrow (>=7.0.0)"] +performance = ["bottleneck (>=1.3.2)", "numba (>=0.53.1)", "numexpr (>=2.7.1)"] +plot = ["matplotlib (>=3.6.1)"] +postgresql = ["SQLAlchemy (>=1.4.16)", "psycopg2 (>=2.8.6)"] +spss = ["pyreadstat (>=1.1.2)"] +sql-other = ["SQLAlchemy (>=1.4.16)"] +test = ["hypothesis (>=6.34.2)", "pytest (>=7.0.0)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)"] +xml = ["lxml (>=4.6.3)"] + @@ -930,0 +1709,80 @@ files = [ +[[package]] +name = "pillow" +version = "9.5.0" +description = "Python Imaging Library (Fork)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pillow-9.5.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:ace6ca218308447b9077c14ea4ef381ba0b67ee78d64046b3f19cf4e1139ad16"}, + {file = "Pillow-9.5.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d3d403753c9d5adc04d4694d35cf0391f0f3d57c8e0030aac09d7678fa8030aa"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5ba1b81ee69573fe7124881762bb4cd2e4b6ed9dd28c9c60a632902fe8db8b38"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe7e1c262d3392afcf5071df9afa574544f28eac825284596ac6db56e6d11062"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f36397bf3f7d7c6a3abdea815ecf6fd14e7fcd4418ab24bae01008d8d8ca15e"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:252a03f1bdddce077eff2354c3861bf437c892fb1832f75ce813ee94347aa9b5"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:85ec677246533e27770b0de5cf0f9d6e4ec0c212a1f89dfc941b64b21226009d"}, + {file = "Pillow-9.5.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:b416f03d37d27290cb93597335a2f85ed446731200705b22bb927405320de903"}, + {file = "Pillow-9.5.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:1781a624c229cb35a2ac31cc4a77e28cafc8900733a864870c49bfeedacd106a"}, + {file = "Pillow-9.5.0-cp310-cp310-win32.whl", hash = "sha256:8507eda3cd0608a1f94f58c64817e83ec12fa93a9436938b191b80d9e4c0fc44"}, + {file = "Pillow-9.5.0-cp310-cp310-win_amd64.whl", hash = "sha256:d3c6b54e304c60c4181da1c9dadf83e4a54fd266a99c70ba646a9baa626819eb"}, + {file = "Pillow-9.5.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:7ec6f6ce99dab90b52da21cf0dc519e21095e332ff3b399a357c187b1a5eee32"}, + {file = "Pillow-9.5.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:560737e70cb9c6255d6dcba3de6578a9e2ec4b573659943a5e7e4af13f298f5c"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:96e88745a55b88a7c64fa49bceff363a1a27d9a64e04019c2281049444a571e3"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9c206c29b46cfd343ea7cdfe1232443072bbb270d6a46f59c259460db76779a"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cfcc2c53c06f2ccb8976fb5c71d448bdd0a07d26d8e07e321c103416444c7ad1"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:a0f9bb6c80e6efcde93ffc51256d5cfb2155ff8f78292f074f60f9e70b942d99"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:8d935f924bbab8f0a9a28404422da8af4904e36d5c33fc6f677e4c4485515625"}, + {file = "Pillow-9.5.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:fed1e1cf6a42577953abbe8e6cf2fe2f566daebde7c34724ec8803c4c0cda579"}, + {file = "Pillow-9.5.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:c1170d6b195555644f0616fd6ed929dfcf6333b8675fcca044ae5ab110ded296"}, + {file = "Pillow-9.5.0-cp311-cp311-win32.whl", hash = "sha256:54f7102ad31a3de5666827526e248c3530b3a33539dbda27c6843d19d72644ec"}, + {file = "Pillow-9.5.0-cp311-cp311-win_amd64.whl", hash = "sha256:cfa4561277f677ecf651e2b22dc43e8f5368b74a25a8f7d1d4a3a243e573f2d4"}, + {file = "Pillow-9.5.0-cp311-cp311-win_arm64.whl", hash = "sha256:965e4a05ef364e7b973dd17fc765f42233415974d773e82144c9bbaaaea5d089"}, + {file = "Pillow-9.5.0-cp312-cp312-win32.whl", hash = "sha256:22baf0c3cf0c7f26e82d6e1adf118027afb325e703922c8dfc1d5d0156bb2eeb"}, + {file = "Pillow-9.5.0-cp312-cp312-win_amd64.whl", hash = "sha256:432b975c009cf649420615388561c0ce7cc31ce9b2e374db659ee4f7d57a1f8b"}, + {file = "Pillow-9.5.0-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:5d4ebf8e1db4441a55c509c4baa7a0587a0210f7cd25fcfe74dbbce7a4bd1906"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:375f6e5ee9620a271acb6820b3d1e94ffa8e741c0601db4c0c4d3cb0a9c224bf"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:99eb6cafb6ba90e436684e08dad8be1637efb71c4f2180ee6b8f940739406e78"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2dfaaf10b6172697b9bceb9a3bd7b951819d1ca339a5ef294d1f1ac6d7f63270"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_28_aarch64.whl", hash = "sha256:763782b2e03e45e2c77d7779875f4432e25121ef002a41829d8868700d119392"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:35f6e77122a0c0762268216315bf239cf52b88865bba522999dc38f1c52b9b47"}, + {file = "Pillow-9.5.0-cp37-cp37m-win32.whl", hash = "sha256:aca1c196f407ec7cf04dcbb15d19a43c507a81f7ffc45b690899d6a76ac9fda7"}, + {file = "Pillow-9.5.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322724c0032af6692456cd6ed554bb85f8149214d97398bb80613b04e33769f6"}, + {file = "Pillow-9.5.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:a0aa9417994d91301056f3d0038af1199eb7adc86e646a36b9e050b06f526597"}, + {file = "Pillow-9.5.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:f8286396b351785801a976b1e85ea88e937712ee2c3ac653710a4a57a8da5d9c"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c830a02caeb789633863b466b9de10c015bded434deb3ec87c768e53752ad22a"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbd359831c1657d69bb81f0db962905ee05e5e9451913b18b831febfe0519082"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f8fc330c3370a81bbf3f88557097d1ea26cd8b019d6433aa59f71195f5ddebbf"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:7002d0797a3e4193c7cdee3198d7c14f92c0836d6b4a3f3046a64bd1ce8df2bf"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:229e2c79c00e85989a34b5981a2b67aa079fd08c903f0aaead522a1d68d79e51"}, + {file = "Pillow-9.5.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:9adf58f5d64e474bed00d69bcd86ec4bcaa4123bfa70a65ce72e424bfb88ed96"}, + {file = "Pillow-9.5.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:662da1f3f89a302cc22faa9f14a262c2e3951f9dbc9617609a47521c69dd9f8f"}, + {file = "Pillow-9.5.0-cp38-cp38-win32.whl", hash = "sha256:6608ff3bf781eee0cd14d0901a2b9cc3d3834516532e3bd673a0a204dc8615fc"}, + {file = "Pillow-9.5.0-cp38-cp38-win_amd64.whl", hash = "sha256:e49eb4e95ff6fd7c0c402508894b1ef0e01b99a44320ba7d8ecbabefddcc5569"}, + {file = "Pillow-9.5.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:482877592e927fd263028c105b36272398e3e1be3269efda09f6ba21fd83ec66"}, + {file = "Pillow-9.5.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3ded42b9ad70e5f1754fb7c2e2d6465a9c842e41d178f262e08b8c85ed8a1d8e"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c446d2245ba29820d405315083d55299a796695d747efceb5717a8b450324115"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8aca1152d93dcc27dc55395604dcfc55bed5f25ef4c98716a928bacba90d33a3"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:608488bdcbdb4ba7837461442b90ea6f3079397ddc968c31265c1e056964f1ef"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:60037a8db8750e474af7ffc9faa9b5859e6c6d0a50e55c45576bf28be7419705"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:07999f5834bdc404c442146942a2ecadd1cb6292f5229f4ed3b31e0a108746b1"}, + {file = "Pillow-9.5.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:a127ae76092974abfbfa38ca2d12cbeddcdeac0fb71f9627cc1135bedaf9d51a"}, + {file = "Pillow-9.5.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:489f8389261e5ed43ac8ff7b453162af39c3e8abd730af8363587ba64bb2e865"}, + {file = "Pillow-9.5.0-cp39-cp39-win32.whl", hash = "sha256:9b1af95c3a967bf1da94f253e56b6286b50af23392a886720f563c547e48e964"}, + {file = "Pillow-9.5.0-cp39-cp39-win_amd64.whl", hash = "sha256:77165c4a5e7d5a284f10a6efaa39a0ae8ba839da344f20b111d62cc932fa4e5d"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:833b86a98e0ede388fa29363159c9b1a294b0905b5128baf01db683672f230f5"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:aaf305d6d40bd9632198c766fb64f0c1a83ca5b667f16c1e79e1661ab5060140"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0852ddb76d85f127c135b6dd1f0bb88dbb9ee990d2cd9aa9e28526c93e794fba"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:91ec6fe47b5eb5a9968c79ad9ed78c342b1f97a091677ba0e012701add857829"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cb841572862f629b99725ebaec3287fc6d275be9b14443ea746c1dd325053cbd"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:c380b27d041209b849ed246b111b7c166ba36d7933ec6e41175fd15ab9eb1572"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7c9af5a3b406a50e313467e3565fc99929717f780164fe6fbb7704edba0cebbe"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5671583eab84af046a397d6d0ba25343c00cd50bce03787948e0fff01d4fd9b1"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:84a6f19ce086c1bf894644b43cd129702f781ba5751ca8572f08aa40ef0ab7b7"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:1e7723bd90ef94eda669a3c2c19d549874dd5badaeefabefd26053304abe5799"}, + {file = "Pillow-9.5.0.tar.gz", hash = "sha256:bf548479d336726d7a0eceb6e767e179fbde37833ae42794602631a070d630f1"}, +] + +[package.extras] +docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-removed-in", "sphinxext-opengraph"] +tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"] + @@ -960 +1818 @@ name = "pip-audit" -version = "2.4.14" +version = "2.5.4" @@ -966,2 +1824,2 @@ files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, + {file = "pip_audit-2.5.4-py3-none-any.whl", hash = "sha256:bee3748030c895488b4dd8a6196fa44f484da000cfd52d8fc64dfa3a2e121624"}, + {file = "pip_audit-2.5.4.tar.gz", hash = "sha256:a4cb03f9e2896d626f5b153973d3ac0d32fdb18594d78d393b153c83bb8089b6"}, @@ -972 +1830 @@ CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +cyclonedx-python-lib = ">=2.0,<2.5.0 || >2.5.0,<3.0" @@ -977 +1834,0 @@ pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" @@ -984 +1841 @@ doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.260)", "types-html5lib", "types-requests", "types-toml"] @@ -1038,0 +1896,22 @@ testing = ["pytest", "pytest-benchmark"] +[[package]] +name = "pooch" +version = "1.6.0" +description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\"" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"}, + {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"}, +] + +[package.dependencies] +appdirs = ">=1.3.0" +packaging = ">=20.0" +requests = ">=2.19.0" + +[package.extras] +progress = ["tqdm (>=4.41.0,<5.0.0)"] +sftp = ["paramiko (>=2.7.0)"] +xxhash = ["xxhash (>=1.4.3)"] + @@ -1065,0 +1945,38 @@ test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] +[[package]] +name = "pyarrow" +version = "11.0.0" +description = "Python library for Apache Arrow" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"}, + {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"}, + {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"}, + {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"}, + {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"}, + {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"}, + {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"}, + {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"}, +] + +[package.dependencies] +numpy = ">=1.16.6" + @@ -1077,0 +1995,24 @@ files = [ +[[package]] +name = "pycparser" +version = "2.21" +description = "C parser in Python" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, + {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, +] + +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + @@ -1295,0 +2237,15 @@ testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtuale +[[package]] +name = "python-dateutil" +version = "2.8.2" +description = "Extensions to the standard Python datetime module" +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" +files = [ + {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"}, + {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"}, +] + +[package.dependencies] +six = ">=1.5" + @@ -1396,4 +2352,4 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -name = "resolvelib" -version = "1.0.1" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" +name = "responses" +version = "0.18.0" +description = "A utility library for mocking out the `requests` Python library." +category = "main" @@ -1401 +2357 @@ optional = false -python-versions = "*" +python-versions = ">=3.7" @@ -1403,2 +2359,2 @@ files = [ - {file = "resolvelib-1.0.1-py2.py3-none-any.whl", hash = "sha256:d2da45d1a8dfee81bdd591647783e340ef3bcb104b54c383f70d422ef5cc7dbf"}, - {file = "resolvelib-1.0.1.tar.gz", hash = "sha256:04ce76cbd63fded2078ce224785da6ecd42b9564b1390793f64ddecbe997b309"}, + {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, + {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, @@ -1406,0 +2363,4 @@ files = [ +[package.dependencies] +requests = ">=2.0,<3.0" +urllib3 = ">=1.25.10" + @@ -1408,4 +2368 @@ files = [ -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] +tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] @@ -1431,0 +2389,82 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] +[[package]] +name = "scikit-learn" +version = "1.2.2" +description = "A set of python modules for machine learning and data mining" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "scikit-learn-1.2.2.tar.gz", hash = "sha256:8429aea30ec24e7a8c7ed8a3fa6213adf3814a6efbea09e16e0a0c71e1a1a3d7"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:99cc01184e347de485bf253d19fcb3b1a3fb0ee4cea5ee3c43ec0cc429b6d29f"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:e6e574db9914afcb4e11ade84fab084536a895ca60aadea3041e85b8ac963edb"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6fe83b676f407f00afa388dd1fdd49e5c6612e551ed84f3b1b182858f09e987d"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e2642baa0ad1e8f8188917423dd73994bf25429f8893ddbe115be3ca3183584"}, + {file = "scikit_learn-1.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:ad66c3848c0a1ec13464b2a95d0a484fd5b02ce74268eaa7e0c697b904f31d6c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:dfeaf8be72117eb61a164ea6fc8afb6dfe08c6f90365bde2dc16456e4bc8e45f"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:fe0aa1a7029ed3e1dcbf4a5bc675aa3b1bc468d9012ecf6c6f081251ca47f590"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:065e9673e24e0dc5113e2dd2b4ca30c9d8aa2fa90f4c0597241c93b63130d233"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf036ea7ef66115e0d49655f16febfa547886deba20149555a41d28f56fd6d3c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:8b0670d4224a3c2d596fd572fb4fa673b2a0ccfb07152688ebd2ea0b8c61025c"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9c710ff9f9936ba8a3b74a455ccf0dcf59b230caa1e9ba0223773c490cab1e51"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:2dd3ffd3950e3d6c0c0ef9033a9b9b32d910c61bd06cb8206303fb4514b88a49"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:44b47a305190c28dd8dd73fc9445f802b6ea716669cfc22ab1eb97b335d238b1"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:953236889928d104c2ef14027539f5f2609a47ebf716b8cbe4437e85dce42744"}, + {file = "scikit_learn-1.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:7f69313884e8eb311460cc2f28676d5e400bd929841a2c8eb8742ae78ebf7c20"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8156db41e1c39c69aa2d8599ab7577af53e9e5e7a57b0504e116cc73c39138dd"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:fe175ee1dab589d2e1033657c5b6bec92a8a3b69103e3dd361b58014729975c3"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d5312d9674bed14f73773d2acf15a3272639b981e60b72c9b190a0cffed5bad"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ea061bf0283bf9a9f36ea3c5d3231ba2176221bbd430abd2603b1c3b2ed85c89"}, + {file = "scikit_learn-1.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:6477eed40dbce190f9f9e9d0d37e020815825b300121307942ec2110302b66a3"}, +] + +[package.dependencies] +joblib = ">=1.1.1" +numpy = ">=1.17.3" +scipy = ">=1.3.2" +threadpoolctl = ">=2.0.0" + +[package.extras] +benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"] +docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=4.0.1)", "sphinx-gallery (>=0.7.0)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"] +examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"] +tests = ["black (>=22.3.0)", "flake8 (>=3.8.2)", "matplotlib (>=3.1.3)", "mypy (>=0.961)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=5.3.1)", "pytest-cov (>=2.9.0)", "scikit-image (>=0.16.2)"] + +[[package]] +name = "scipy" +version = "1.10.1" +description = "Fundamental algorithms for scientific computing in Python" +category = "main" +optional = false +python-versions = "<3.12,>=3.8" +files = [ + {file = "scipy-1.10.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e7354fd7527a4b0377ce55f286805b34e8c54b91be865bac273f527e1b839019"}, + {file = "scipy-1.10.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:4b3f429188c66603a1a5c549fb414e4d3bdc2a24792e061ffbd607d3d75fd84e"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1553b5dcddd64ba9a0d95355e63fe6c3fc303a8fd77c7bc91e77d61363f7433f"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c0ff64b06b10e35215abce517252b375e580a6125fd5fdf6421b98efbefb2d2"}, + {file = "scipy-1.10.1-cp310-cp310-win_amd64.whl", hash = "sha256:fae8a7b898c42dffe3f7361c40d5952b6bf32d10c4569098d276b4c547905ee1"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:0f1564ea217e82c1bbe75ddf7285ba0709ecd503f048cb1236ae9995f64217bd"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d925fa1c81b772882aa55bcc10bf88324dadb66ff85d548c71515f6689c6dac5"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaea0a6be54462ec027de54fca511540980d1e9eea68b2d5c1dbfe084797be35"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15a35c4242ec5f292c3dd364a7c71a61be87a3d4ddcc693372813c0b73c9af1d"}, + {file = "scipy-1.10.1-cp311-cp311-win_amd64.whl", hash = "sha256:43b8e0bcb877faf0abfb613d51026cd5cc78918e9530e375727bf0625c82788f"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5678f88c68ea866ed9ebe3a989091088553ba12c6090244fdae3e467b1139c35"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:39becb03541f9e58243f4197584286e339029e8908c46f7221abeea4b749fa88"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bce5869c8d68cf383ce240e44c1d9ae7c06078a9396df68ce88a1230f93a30c1"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:07c3457ce0b3ad5124f98a86533106b643dd811dd61b548e78cf4c8786652f6f"}, + {file = "scipy-1.10.1-cp38-cp38-win_amd64.whl", hash = "sha256:049a8bbf0ad95277ffba9b3b7d23e5369cc39e66406d60422c8cfef40ccc8415"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cd9f1027ff30d90618914a64ca9b1a77a431159df0e2a195d8a9e8a04c78abf9"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:79c8e5a6c6ffaf3a2262ef1be1e108a035cf4f05c14df56057b64acc5bebffb6"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:51af417a000d2dbe1ec6c372dfe688e041a7084da4fdd350aeb139bd3fb55353"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b4735d6c28aad3cdcf52117e0e91d6b39acd4272f3f5cd9907c24ee931ad601"}, + {file = "scipy-1.10.1-cp39-cp39-win_amd64.whl", hash = "sha256:7ff7f37b1bf4417baca958d254e8e2875d0cc23aaadbe65b3d5b3077b0eb23ea"}, + {file = "scipy-1.10.1.tar.gz", hash = "sha256:2cf9dfb80a7b4589ba4c40ce7588986d6d5cebc5457cad2c2880f6bc2d42f3a5"}, +] + +[package.dependencies] +numpy = ">=1.19.5,<1.27.0" + +[package.extras] +dev = ["click", "doit (>=0.36.0)", "flake8", "mypy", "pycodestyle", "pydevtool", "rich-click", "typing_extensions"] +doc = ["matplotlib (>2)", "numpydoc", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"] +test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"] + @@ -1436 +2475 @@ description = "Easily download, build, install, upgrade, and uninstall Python pa -category = "dev" +category = "main" @@ -1453 +2492 @@ description = "Python 2 and 3 compatibility utilities" -category = "dev" +category = "main" @@ -1484,0 +2524,68 @@ files = [ +[[package]] +name = "soundfile" +version = "0.12.1" +description = "An audio library based on libsndfile, CFFI and NumPy" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"}, + {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"}, + {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"}, + {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"}, +] + +[package.dependencies] +cffi = ">=1.0" + +[package.extras] +numpy = ["numpy"] + +[[package]] +name = "soxr" +version = "0.3.4" +description = "High quality, one-dimensional sample-rate conversion library" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "soxr-0.3.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b7b84126643c063d5daa203f7f9137e21734dabbd7e68c097607b2ef457e2f2e"}, + {file = "soxr-0.3.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:380d2d43871a68e8b1ef1702a0abe6f9e48ddb3933c7a303c45d67e121503e7c"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4a1b4019c9972f57612482c4f85523d6e832e3d10935e2f070a9dcd334a4dcb"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e613cee023b7c3f162b9da3f6b169cd7f58de345275be1fde9f19adc9cf144df"}, + {file = "soxr-0.3.4-cp310-cp310-win32.whl", hash = "sha256:182c02a7ba45a159a0dbb0a297335df2381ead03a65377b19663ea0ff720ecb7"}, + {file = "soxr-0.3.4-cp310-cp310-win_amd64.whl", hash = "sha256:1e95c96ce94524fae453b4331c9910d33f97506f99bae06d76a9c0649710619e"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2678d2719e7496803983584d661eb5fddc7017154a8dda4a774407c56ff07973"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:11bd1396052049e6d389225a0e96a9df15f706da501c619b35d3c72ac6bc7257"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7e23de4dfe54ac30e583bbc9cc3feda1cd776fedce13206bc4b3115b75ecab82"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3e7396498a5f5b7d8f23b656f65c24517a6ff5bdc3ee0623ccd491036a43ea08"}, + {file = "soxr-0.3.4-cp311-cp311-win32.whl", hash = "sha256:e57e9703c2bff834cabc06800d3c11a259544891d2c24a78949f3cf2f5492cc5"}, + {file = "soxr-0.3.4-cp311-cp311-win_amd64.whl", hash = "sha256:7c8350acd7150f74261a0569b47ccb1bb4aa39b2d575860bc97cfa69aab8aead"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:941f7355116fe77fe6a82938fa7799a0e466a494ebc093f676969ce32b2815b1"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:00fdbf24f64d8c3fb800425c383048cb24c32defac80901cde4a57fb6ce5d431"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7bb6d4dc807d04c536674429e2b05ae08a1efac9815c4595e41ffd6b57c2c662"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ff15853895b54f1b627799c6600be1ce5f7286724e7a93e4b7b9d79e5d4166f5"}, + {file = "soxr-0.3.4-cp38-cp38-win32.whl", hash = "sha256:d858becbc1fcc7b38c3436d3276290fae09403cdcbdf1d5986a18dab7023a6c3"}, + {file = "soxr-0.3.4-cp38-cp38-win_amd64.whl", hash = "sha256:068ab4df549df5783cc1eb4eb6c94f53823b164dc27134fc621fc9f5097f38cd"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:20130329985f9767c8417bbd125fe138790a71802b000481c386a800e2ad2bca"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:78090e97abfb326b7cf14ef37d08a17252b07d438388dcbbd82a6836a9d551b1"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84e590e75b7e5dca12bf68bfb090276f34a88fbcd793781c62d47f5d7dbe525e"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3479d265574b960e12bca0878baba0862c43278915e0319d84679bb4d4fcd33"}, + {file = "soxr-0.3.4-cp39-cp39-win32.whl", hash = "sha256:83de825d6a713c7b2e76d9ec3f229a58a9ed290237e7adc05d80e8b39be995a6"}, + {file = "soxr-0.3.4-cp39-cp39-win_amd64.whl", hash = "sha256:2082f88cae89de854c3e0d62f55d0cb31eb11764f5c2a28299121fb642a22472"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:fe8b5f92c802f1e7793c40344f5368dc6163718c9ffa82e79ee6ad779d318ac5"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d0063d5f9a4e1a367084f4705301e9da131cf4d2d32aa3fe0072a1245e18088f"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:a680bab57adae462cdc86abcc7330beb5daa3ba5101165583eedcda88b7ba551"}, + {file = "soxr-0.3.4.tar.gz", hash = "sha256:fe68daf00e8f020977b187699903d219f9e39b9fb3d915f3f923eed8ba431449"}, +] + +[package.dependencies] +numpy = "*" + +[package.extras] +docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"] +test = ["pytest"] + @@ -1499,0 +2607,12 @@ pbr = ">=2.0.0,<2.1.0 || >2.1.0" +[[package]] +name = "threadpoolctl" +version = "3.1.0" +description = "threadpoolctl" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, + {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"}, +] + @@ -1556,0 +2676,12 @@ files = [ +[[package]] +name = "tzdata" +version = "2023.3" +description = "Provider of IANA time zone data" +category = "main" +optional = false +python-versions = ">=2" +files = [ + {file = "tzdata-2023.3-py2.py3-none-any.whl", hash = "sha256:7e65763eef3120314099b6939b5546db7adce1e7d6f2e179e3df563c70511eda"}, + {file = "tzdata-2023.3.tar.gz", hash = "sha256:11ef1e08e54acb0d4f95bdb1be05da659673de4acbd21bf9c69e94cc5e907a3a"}, +] + @@ -1585,0 +2717,196 @@ files = [ +[[package]] +name = "xxhash" +version = "3.2.0" +description = "Python binding for xxHash" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"}, + {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"}, + {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"}, + {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"}, + {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"}, + {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"}, + {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"}, + {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"}, + {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"}, + {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"}, + {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"}, + {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"}, + {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"}, + {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"}, + {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"}, + {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"}, + {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"}, +] + +[[package]] +name = "yarl" +version = "1.8.2" +description = "Yet another URL library" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:bb81f753c815f6b8e2ddd2eef3c855cf7da193b82396ac013c661aaa6cc6b0a5"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:47d49ac96156f0928f002e2424299b2c91d9db73e08c4cd6742923a086f1c863"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3fc056e35fa6fba63248d93ff6e672c096f95f7836938241ebc8260e062832fe"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58a3c13d1c3005dbbac5c9f0d3210b60220a65a999b1833aa46bd6677c69b08e"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:10b08293cda921157f1e7c2790999d903b3fd28cd5c208cf8826b3b508026996"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:de986979bbd87272fe557e0a8fcb66fd40ae2ddfe28a8b1ce4eae22681728fef"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c4fcfa71e2c6a3cb568cf81aadc12768b9995323186a10827beccf5fa23d4f8"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae4d7ff1049f36accde9e1ef7301912a751e5bae0a9d142459646114c70ecba6"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:bf071f797aec5b96abfc735ab97da9fd8f8768b43ce2abd85356a3127909d146"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:74dece2bfc60f0f70907c34b857ee98f2c6dd0f75185db133770cd67300d505f"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:df60a94d332158b444301c7f569659c926168e4d4aad2cfbf4bce0e8fb8be826"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:63243b21c6e28ec2375f932a10ce7eda65139b5b854c0f6b82ed945ba526bff3"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:cfa2bbca929aa742b5084fd4663dd4b87c191c844326fcb21c3afd2d11497f80"}, + {file = "yarl-1.8.2-cp310-cp310-win32.whl", hash = "sha256:b05df9ea7496df11b710081bd90ecc3a3db6adb4fee36f6a411e7bc91a18aa42"}, + {file = "yarl-1.8.2-cp310-cp310-win_amd64.whl", hash = "sha256:24ad1d10c9db1953291f56b5fe76203977f1ed05f82d09ec97acb623a7976574"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:2a1fca9588f360036242f379bfea2b8b44cae2721859b1c56d033adfd5893634"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:f37db05c6051eff17bc832914fe46869f8849de5b92dc4a3466cd63095d23dfd"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:77e913b846a6b9c5f767b14dc1e759e5aff05502fe73079f6f4176359d832581"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0978f29222e649c351b173da2b9b4665ad1feb8d1daa9d971eb90df08702668a"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:388a45dc77198b2460eac0aca1efd6a7c09e976ee768b0d5109173e521a19daf"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2305517e332a862ef75be8fad3606ea10108662bc6fe08509d5ca99503ac2aee"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:42430ff511571940d51e75cf42f1e4dbdded477e71c1b7a17f4da76c1da8ea76"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3150078118f62371375e1e69b13b48288e44f6691c1069340081c3fd12c94d5b"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:c15163b6125db87c8f53c98baa5e785782078fbd2dbeaa04c6141935eb6dab7a"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:4d04acba75c72e6eb90745447d69f84e6c9056390f7a9724605ca9c56b4afcc6"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:e7fd20d6576c10306dea2d6a5765f46f0ac5d6f53436217913e952d19237efc4"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:75c16b2a900b3536dfc7014905a128a2bea8fb01f9ee26d2d7d8db0a08e7cb2c"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:6d88056a04860a98341a0cf53e950e3ac9f4e51d1b6f61a53b0609df342cc8b2"}, + {file = "yarl-1.8.2-cp311-cp311-win32.whl", hash = "sha256:fb742dcdd5eec9f26b61224c23baea46c9055cf16f62475e11b9b15dfd5c117b"}, + {file = "yarl-1.8.2-cp311-cp311-win_amd64.whl", hash = "sha256:8c46d3d89902c393a1d1e243ac847e0442d0196bbd81aecc94fcebbc2fd5857c"}, + {file = "yarl-1.8.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:ceff9722e0df2e0a9e8a79c610842004fa54e5b309fe6d218e47cd52f791d7ef"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3f6b4aca43b602ba0f1459de647af954769919c4714706be36af670a5f44c9c1"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1684a9bd9077e922300ecd48003ddae7a7474e0412bea38d4631443a91d61077"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ebb78745273e51b9832ef90c0898501006670d6e059f2cdb0e999494eb1450c2"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3adeef150d528ded2a8e734ebf9ae2e658f4c49bf413f5f157a470e17a4a2e89"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:57a7c87927a468e5a1dc60c17caf9597161d66457a34273ab1760219953f7f4c"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:efff27bd8cbe1f9bd127e7894942ccc20c857aa8b5a0327874f30201e5ce83d0"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:a783cd344113cb88c5ff7ca32f1f16532a6f2142185147822187913eb989f739"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:705227dccbe96ab02c7cb2c43e1228e2826e7ead880bb19ec94ef279e9555b5b"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:34c09b43bd538bf6c4b891ecce94b6fa4f1f10663a8d4ca589a079a5018f6ed7"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:a48f4f7fea9a51098b02209d90297ac324241bf37ff6be6d2b0149ab2bd51b37"}, + {file = "yarl-1.8.2-cp37-cp37m-win32.whl", hash = "sha256:0414fd91ce0b763d4eadb4456795b307a71524dbacd015c657bb2a39db2eab89"}, + {file = "yarl-1.8.2-cp37-cp37m-win_amd64.whl", hash = "sha256:d881d152ae0007809c2c02e22aa534e702f12071e6b285e90945aa3c376463c5"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5df5e3d04101c1e5c3b1d69710b0574171cc02fddc4b23d1b2813e75f35a30b1"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7a66c506ec67eb3159eea5096acd05f5e788ceec7b96087d30c7d2865a243918"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2b4fa2606adf392051d990c3b3877d768771adc3faf2e117b9de7eb977741229"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1e21fb44e1eff06dd6ef971d4bdc611807d6bd3691223d9c01a18cec3677939e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:93202666046d9edadfe9f2e7bf5e0782ea0d497b6d63da322e541665d65a044e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fc77086ce244453e074e445104f0ecb27530d6fd3a46698e33f6c38951d5a0f1"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:64dd68a92cab699a233641f5929a40f02a4ede8c009068ca8aa1fe87b8c20ae3"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1b372aad2b5f81db66ee7ec085cbad72c4da660d994e8e590c997e9b01e44901"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e6f3515aafe0209dd17fb9bdd3b4e892963370b3de781f53e1746a521fb39fc0"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:dfef7350ee369197106805e193d420b75467b6cceac646ea5ed3049fcc950a05"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:728be34f70a190566d20aa13dc1f01dc44b6aa74580e10a3fb159691bc76909d"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ff205b58dc2929191f68162633d5e10e8044398d7a45265f90a0f1d51f85f72c"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:baf211dcad448a87a0d9047dc8282d7de59473ade7d7fdf22150b1d23859f946"}, + {file = "yarl-1.8.2-cp38-cp38-win32.whl", hash = "sha256:272b4f1599f1b621bf2aabe4e5b54f39a933971f4e7c9aa311d6d7dc06965165"}, + {file = "yarl-1.8.2-cp38-cp38-win_amd64.whl", hash = "sha256:326dd1d3caf910cd26a26ccbfb84c03b608ba32499b5d6eeb09252c920bcbe4f"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:f8ca8ad414c85bbc50f49c0a106f951613dfa5f948ab69c10ce9b128d368baf8"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:418857f837347e8aaef682679f41e36c24250097f9e2f315d39bae3a99a34cbf"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:ae0eec05ab49e91a78700761777f284c2df119376e391db42c38ab46fd662b77"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:009a028127e0a1755c38b03244c0bea9d5565630db9c4cf9572496e947137a87"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3edac5d74bb3209c418805bda77f973117836e1de7c000e9755e572c1f7850d0"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:da65c3f263729e47351261351b8679c6429151ef9649bba08ef2528ff2c423b2"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ef8fb25e52663a1c85d608f6dd72e19bd390e2ecaf29c17fb08f730226e3a08"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bcd7bb1e5c45274af9a1dd7494d3c52b2be5e6bd8d7e49c612705fd45420b12d"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:44ceac0450e648de86da8e42674f9b7077d763ea80c8ceb9d1c3e41f0f0a9951"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:97209cc91189b48e7cfe777237c04af8e7cc51eb369004e061809bcdf4e55220"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:48dd18adcf98ea9cd721a25313aef49d70d413a999d7d89df44f469edfb38a06"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:e59399dda559688461762800d7fb34d9e8a6a7444fd76ec33220a926c8be1516"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d617c241c8c3ad5c4e78a08429fa49e4b04bedfc507b34b4d8dceb83b4af3588"}, + {file = "yarl-1.8.2-cp39-cp39-win32.whl", hash = "sha256:cb6d48d80a41f68de41212f3dfd1a9d9898d7841c8f7ce6696cf2fd9cb57ef83"}, + {file = "yarl-1.8.2-cp39-cp39-win_amd64.whl", hash = "sha256:6604711362f2dbf7160df21c416f81fac0de6dbcf0b5445a2ef25478ecc4c778"}, + {file = "yarl-1.8.2.tar.gz", hash = "sha256:49d43402c6e3013ad0978602bf6bf5328535c48d192304b91b97a3c6790b1562"}, +] + +[package.dependencies] +idna = ">=2.0" +multidict = ">=4.0" + @@ -1589 +2916 @@ python-versions = "3.9.15" -content-hash = "67ae1eb17c3921f944b323ddcaa58919c1deff1e7cc25e63cf3a82ec23340c62" +content-hash = "f916a260ab2e57ba4b18f783a88b5fbb1631a59795bed892690777d2c7ee0cd6" diff --git a/jobs/cache_maintenance/pyproject.toml b/jobs/cache_maintenance/pyproject.toml index d7481fff..c15f2dd3 100644 --- a/jobs/cache_maintenance/pyproject.toml +++ b/jobs/cache_maintenance/pyproject.toml @@ -19 +19 @@ mypy = "^1.0.0" -pip-audit = "^2.4.14" +pip-audit = "^2.5.4" diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock index d99b420d..9430bb98 100644 --- a/jobs/mongodb_migration/poetry.lock +++ b/jobs/mongodb_migration/poetry.lock @@ -1 +1,125 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. + +[[package]] +name = "aiohttp" +version = "3.8.4" +description = "Async http client/server framework (asyncio)" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"}, + {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"}, + {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"}, + {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"}, + {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"}, + {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"}, + {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"}, + {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"}, + {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"}, + {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"}, + {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"}, + {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"}, +] + +[package.dependencies] +aiosignal = ">=1.1.2" +async-timeout = ">=4.0.0a3,<5.0" +attrs = ">=17.3.0" +charset-normalizer = ">=2.0,<4.0" +frozenlist = ">=1.1.1" +multidict = ">=4.5,<7.0" +yarl = ">=1.0,<2.0" + +[package.extras] +speedups = ["Brotli", "aiodns", "cchardet"] + +[[package]] +name = "aiosignal" +version = "1.3.1" +description = "aiosignal: a list of registered asynchronous callbacks" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, + {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, +] + +[package.dependencies] +frozenlist = ">=1.1.0" @@ -14,0 +139,12 @@ files = [ +[[package]] +name = "async-timeout" +version = "4.0.2" +description = "Timeout context manager for asyncio programs" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"}, + {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"}, +] + @@ -19 +155 @@ description = "Classes Without Boilerplate" -category = "dev" +category = "main" @@ -33,0 +170,11 @@ tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy +[[package]] +name = "audioread" +version = "3.0.0" +description = "multi-library, cross-platform audio decoding" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"}, +] + @@ -125,0 +273,77 @@ files = [ +[[package]] +name = "cffi" +version = "1.15.1" +description = "Foreign Function Interface for Python calling C code." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"}, + {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"}, + {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"}, + {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"}, + {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"}, + {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"}, + {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"}, + {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"}, + {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"}, + {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"}, + {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"}, + {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"}, + {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"}, + {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"}, + {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"}, + {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"}, + {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"}, + {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"}, + {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"}, +] + +[package.dependencies] +pycparser = "*" + @@ -317 +541 @@ name = "cyclonedx-python-lib" -version = "3.1.5" +version = "2.7.1" @@ -323,2 +547,2 @@ files = [ - {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, - {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, + {file = "cyclonedx-python-lib-2.7.1.tar.gz", hash = "sha256:493bf2f30e26c48f305f745ed8580ce10d05a8d68d62a598fe95f05a0d9007dc"}, + {file = "cyclonedx_python_lib-2.7.1-py3-none-any.whl", hash = "sha256:fabc4c8baf722faeea01c3bbca83730e3489dfb37d85c6036baa67a9a7519d40"}, @@ -332,0 +557,74 @@ toml = ">=0.10.0,<0.11.0" +[[package]] +name = "datasets" +version = "2.11.0" +description = "HuggingFace community-driven open-source library of datasets" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "datasets-2.11.0-py3-none-any.whl", hash = "sha256:d946cdb8c4885d3016a2ab3129c9403dd3358fe9107e8ab5e549ceab672774af"}, + {file = "datasets-2.11.0.tar.gz", hash = "sha256:1ca53b9cd6ece7a3fdb81176dadd5b9e646420e52e68e85307b27db3a36ca18c"}, +] + +[package.dependencies] +aiohttp = "*" +dill = ">=0.3.0,<0.3.7" +fsspec = {version = ">=2021.11.1", extras = ["http"]} +huggingface-hub = ">=0.11.0,<1.0.0" +librosa = {version = "*", optional = true, markers = "extra == \"audio\""} +multiprocess = "*" +numpy = ">=1.17" +packaging = "*" +pandas = "*" +Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} +pyarrow = ">=8.0.0" +pyyaml = ">=5.1" +requests = ">=2.19.0" +responses = "<0.19" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} +tqdm = ">=4.62.1" +xxhash = "*" + +[package.extras] +apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] +audio = ["librosa", "soundfile (>=0.12.1)"] +benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +docs = ["s3fs"] +jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"] +s3 = ["s3fs"] +tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"] +tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +torch = ["torch"] +vision = ["Pillow (>=6.2.1)"] + +[[package]] +name = "decorator" +version = "5.1.1" +description = "Decorators for Humans" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"}, + {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"}, +] + +[[package]] +name = "dill" +version = "0.3.6" +description = "serialize all of python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"}, + {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"}, +] + +[package.extras] +graph = ["objgraph (>=1.7.2)"] + @@ -418,0 +717,122 @@ pyflakes = ">=2.3.0,<2.4.0" +[[package]] +name = "frozenlist" +version = "1.3.3" +description = "A list-like structure which implements collections.abc.MutableSequence" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"}, + {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"}, + {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"}, + {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"}, + {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"}, + {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"}, + {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"}, + {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"}, + {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"}, + {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"}, + {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"}, +] + +[[package]] +name = "fsspec" +version = "2023.3.0" +description = "File-system specification" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "fsspec-2023.3.0-py3-none-any.whl", hash = "sha256:bf57215e19dbfa4fe7edae53040cc1deef825e3b1605cca9a8d2c2fadd2328a0"}, + {file = "fsspec-2023.3.0.tar.gz", hash = "sha256:24e635549a590d74c6c18274ddd3ffab4753341753e923408b1904eaabafe04d"}, +] + +[package.dependencies] +aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""} +requests = {version = "*", optional = true, markers = "extra == \"http\""} + +[package.extras] +abfs = ["adlfs"] +adl = ["adlfs"] +arrow = ["pyarrow (>=1)"] +dask = ["dask", "distributed"] +dropbox = ["dropbox", "dropboxdrivefs", "requests"] +fuse = ["fusepy"] +gcs = ["gcsfs"] +git = ["pygit2"] +github = ["requests"] +gs = ["gcsfs"] +gui = ["panel"] +hdfs = ["pyarrow (>=1)"] +http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"] +libarchive = ["libarchive-c"] +oci = ["ocifs"] +s3 = ["s3fs"] +sftp = ["paramiko"] +smb = ["smbprotocol"] +ssh = ["paramiko"] +tqdm = ["tqdm"] + @@ -543,0 +964,28 @@ requirements-deprecated-finder = ["pip-api", "pipreqs"] +[[package]] +name = "joblib" +version = "1.2.0" +description = "Lightweight pipelining with Python functions" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "joblib-1.2.0-py3-none-any.whl", hash = "sha256:091138ed78f800342968c523bdde947e7a305b8594b910a0fea2ab83c3c6d385"}, + {file = "joblib-1.2.0.tar.gz", hash = "sha256:e1cee4a79e4af22881164f218d4311f60074197fb707e082e803b61f6d137018"}, +] + +[[package]] +name = "lazy-loader" +version = "0.2" +description = "lazy_loader" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "lazy_loader-0.2-py3-none-any.whl", hash = "sha256:c35875f815c340f823ce3271ed645045397213f961b40ad0c0d395c3f5218eeb"}, + {file = "lazy_loader-0.2.tar.gz", hash = "sha256:0edc7a5175c400acb108f283749951fefdadedeb00adcec6e88b974a9254f18a"}, +] + +[package.extras] +lint = ["pre-commit (>=3.1)"] +test = ["pytest (>=7.2)", "pytest-cov (>=4.0)"] + @@ -555,0 +1004 @@ appdirs = "^1.4.4" +datasets = {version = "^2.11.0", extras = ["audio", "vision"]} @@ -560,0 +1010 @@ networkx = "^3.0" +numba = "0.56.4" @@ -562,0 +1013 @@ psutil = "^5.9.4" +pydub = "^0.25.1" @@ -565,0 +1017 @@ requests = "^2.28.2" +soundfile = ">=0.12.1" @@ -570,0 +1023,70 @@ url = "../../libs/libcommon" +[[package]] +name = "librosa" +version = "0.10.0.post2" +description = "Python module for audio and music processing" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"}, + {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"}, +] + +[package.dependencies] +audioread = ">=2.1.9" +decorator = ">=4.3.0" +joblib = ">=0.14" +lazy-loader = ">=0.1" +msgpack = ">=1.0" +numba = ">=0.51.0" +numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2" +pooch = ">=1.0,<1.7" +scikit-learn = ">=0.20.0" +scipy = ">=1.2.0" +soundfile = ">=0.12.1" +soxr = ">=0.3.2" +typing-extensions = ">=4.1.1" + +[package.extras] +display = ["matplotlib (>=3.3.0)"] +docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"] +tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"] + +[[package]] +name = "llvmlite" +version = "0.39.1" +description = "lightweight wrapper around basic LLVM functionality" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"}, + {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"}, + {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"}, + {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"}, + {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"}, + {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"}, + {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"}, + {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"}, + {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"}, + {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"}, +] + @@ -684 +1206 @@ description = "MessagePack serializer" -category = "dev" +category = "main" @@ -741,0 +1264,111 @@ files = [ +[[package]] +name = "multidict" +version = "6.0.4" +description = "multidict implementation" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"}, + {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"}, + {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"}, + {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"}, + {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"}, + {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"}, + {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"}, + {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"}, + {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"}, + {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"}, + {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"}, + {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"}, + {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"}, +] + +[[package]] +name = "multiprocess" +version = "0.70.14" +description = "better multiprocessing and multithreading in python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"}, + {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"}, + {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"}, + {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"}, + {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"}, + {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"}, +] + +[package.dependencies] +dill = ">=0.3.6" + @@ -819,0 +1453,81 @@ test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] +[[package]] +name = "numba" +version = "0.56.4" +description = "compiling Python code using LLVM" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"}, + {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"}, + {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"}, + {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"}, + {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"}, + {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"}, + {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"}, + {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"}, + {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"}, + {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"}, + {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"}, + {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"}, + {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"}, + {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"}, + {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"}, + {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"}, +] + +[package.dependencies] +llvmlite = ">=0.39.0dev0,<0.40" +numpy = ">=1.18,<1.24" +setuptools = "*" + +[[package]] +name = "numpy" +version = "1.23.5" +description = "NumPy is the fundamental package for array computing with Python." +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"}, + {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"}, + {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"}, + {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"}, + {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"}, + {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"}, + {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"}, + {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"}, + {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"}, + {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"}, +] + @@ -901,0 +1616,64 @@ files = [ +[[package]] +name = "pandas" +version = "2.0.0" +description = "Powerful data structures for data analysis, time series, and statistics" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "pandas-2.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:bbb2c5e94d6aa4e632646a3bacd05c2a871c3aa3e85c9bec9be99cb1267279f2"}, + {file = "pandas-2.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b5337c87c4e963f97becb1217965b6b75c6fe5f54c4cf09b9a5ac52fc0bd03d3"}, + {file = "pandas-2.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6ded51f7e3dd9b4f8b87f2ceb7bd1a8df2491f7ee72f7074c6927a512607199e"}, + {file = "pandas-2.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:52c858de9e9fc422d25e67e1592a6e6135d7bcf9a19fcaf4d0831a0be496bf21"}, + {file = "pandas-2.0.0-cp310-cp310-win32.whl", hash = "sha256:2d1d138848dd71b37e3cbe7cd952ff84e2ab04d8988972166e18567dcc811245"}, + {file = "pandas-2.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:d08e41d96bc4de6f500afe80936c68fce6099d5a434e2af7c7fd8e7c72a3265d"}, + {file = "pandas-2.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:24472cfc7ced511ac90608728b88312be56edc8f19b9ed885a7d2e47ffaf69c0"}, + {file = "pandas-2.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4ffb14f50c74ee541610668137830bb93e9dfa319b1bef2cedf2814cd5ac9c70"}, + {file = "pandas-2.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c24c7d12d033a372a9daf9ff2c80f8b0af6f98d14664dbb0a4f6a029094928a7"}, + {file = "pandas-2.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8318de0f886e4dcb8f9f36e45a3d6a6c3d1cfdc508354da85e739090f0222991"}, + {file = "pandas-2.0.0-cp311-cp311-win32.whl", hash = "sha256:57c34b79c13249505e850d0377b722961b99140f81dafbe6f19ef10239f6284a"}, + {file = "pandas-2.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:8f987ec26e96a8490909bc5d98c514147236e49830cba7df8690f6087c12bbae"}, + {file = "pandas-2.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b3ba8f5dd470d8bfbc4259829589f4a32881151c49e36384d9eb982b35a12020"}, + {file = "pandas-2.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:fcd471c9d9f60926ab2f15c6c29164112f458acb42280365fbefa542d0c2fc74"}, + {file = "pandas-2.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9253edfd015520ce77a9343eb7097429479c039cd3ebe81d7810ea11b4b24695"}, + {file = "pandas-2.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:977326039bd1ded620001a1889e2ed4798460a6bc5a24fbaebb5f07a41c32a55"}, + {file = "pandas-2.0.0-cp38-cp38-win32.whl", hash = "sha256:78425ca12314b23356c28b16765639db10ebb7d8983f705d6759ff7fe41357fa"}, + {file = "pandas-2.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:d93b7fcfd9f3328072b250d6d001dcfeec5d3bb66c1b9c8941e109a46c0c01a8"}, + {file = "pandas-2.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:425705cee8be54db2504e8dd2a730684790b15e5904b750c367611ede49098ab"}, + {file = "pandas-2.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a4f789b7c012a608c08cda4ff0872fd979cb18907a37982abe884e6f529b8793"}, + {file = "pandas-2.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3bb9d840bf15656805f6a3d87eea9dcb7efdf1314a82adcf7f00b820427c5570"}, + {file = "pandas-2.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0778ab54c8f399d83d98ffb674d11ec716449956bc6f6821891ab835848687f2"}, + {file = "pandas-2.0.0-cp39-cp39-win32.whl", hash = "sha256:70db5c278bbec0306d32bf78751ff56b9594c05a5098386f6c8a563659124f91"}, + {file = "pandas-2.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:4f3320bb55f34af4193020158ef8118ee0fb9aec7cc47d2084dbfdd868a0a24f"}, + {file = "pandas-2.0.0.tar.gz", hash = "sha256:cda9789e61b44463c1c4fe17ef755de77bcd13b09ba31c940d20f193d63a5dc8"}, +] + +[package.dependencies] +numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""} +python-dateutil = ">=2.8.2" +pytz = ">=2020.1" +tzdata = ">=2022.1" + +[package.extras] +all = ["PyQt5 (>=5.15.1)", "SQLAlchemy (>=1.4.16)", "beautifulsoup4 (>=4.9.3)", "bottleneck (>=1.3.2)", "brotlipy (>=0.7.0)", "fastparquet (>=0.6.3)", "fsspec (>=2021.07.0)", "gcsfs (>=2021.07.0)", "html5lib (>=1.1)", "hypothesis (>=6.34.2)", "jinja2 (>=3.0.0)", "lxml (>=4.6.3)", "matplotlib (>=3.6.1)", "numba (>=0.53.1)", "numexpr (>=2.7.3)", "odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pandas-gbq (>=0.15.0)", "psycopg2 (>=2.8.6)", "pyarrow (>=7.0.0)", "pymysql (>=1.0.2)", "pyreadstat (>=1.1.2)", "pytest (>=7.0.0)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)", "python-snappy (>=0.6.0)", "pyxlsb (>=1.0.8)", "qtpy (>=2.2.0)", "s3fs (>=2021.08.0)", "scipy (>=1.7.1)", "tables (>=3.6.1)", "tabulate (>=0.8.9)", "xarray (>=0.21.0)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)", "zstandard (>=0.15.2)"] +aws = ["s3fs (>=2021.08.0)"] +clipboard = ["PyQt5 (>=5.15.1)", "qtpy (>=2.2.0)"] +compression = ["brotlipy (>=0.7.0)", "python-snappy (>=0.6.0)", "zstandard (>=0.15.2)"] +computation = ["scipy (>=1.7.1)", "xarray (>=0.21.0)"] +excel = ["odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pyxlsb (>=1.0.8)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)"] +feather = ["pyarrow (>=7.0.0)"] +fss = ["fsspec (>=2021.07.0)"] +gcp = ["gcsfs (>=2021.07.0)", "pandas-gbq (>=0.15.0)"] +hdf5 = ["tables (>=3.6.1)"] +html = ["beautifulsoup4 (>=4.9.3)", "html5lib (>=1.1)", "lxml (>=4.6.3)"] +mysql = ["SQLAlchemy (>=1.4.16)", "pymysql (>=1.0.2)"] +output-formatting = ["jinja2 (>=3.0.0)", "tabulate (>=0.8.9)"] +parquet = ["pyarrow (>=7.0.0)"] +performance = ["bottleneck (>=1.3.2)", "numba (>=0.53.1)", "numexpr (>=2.7.1)"] +plot = ["matplotlib (>=3.6.1)"] +postgresql = ["SQLAlchemy (>=1.4.16)", "psycopg2 (>=2.8.6)"] +spss = ["pyreadstat (>=1.1.2)"] +sql-other = ["SQLAlchemy (>=1.4.16)"] +test = ["hypothesis (>=6.34.2)", "pytest (>=7.0.0)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)"] +xml = ["lxml (>=4.6.3)"] + @@ -925,0 +1704,80 @@ files = [ +[[package]] +name = "pillow" +version = "9.5.0" +description = "Python Imaging Library (Fork)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pillow-9.5.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:ace6ca218308447b9077c14ea4ef381ba0b67ee78d64046b3f19cf4e1139ad16"}, + {file = "Pillow-9.5.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d3d403753c9d5adc04d4694d35cf0391f0f3d57c8e0030aac09d7678fa8030aa"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5ba1b81ee69573fe7124881762bb4cd2e4b6ed9dd28c9c60a632902fe8db8b38"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe7e1c262d3392afcf5071df9afa574544f28eac825284596ac6db56e6d11062"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f36397bf3f7d7c6a3abdea815ecf6fd14e7fcd4418ab24bae01008d8d8ca15e"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:252a03f1bdddce077eff2354c3861bf437c892fb1832f75ce813ee94347aa9b5"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:85ec677246533e27770b0de5cf0f9d6e4ec0c212a1f89dfc941b64b21226009d"}, + {file = "Pillow-9.5.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:b416f03d37d27290cb93597335a2f85ed446731200705b22bb927405320de903"}, + {file = "Pillow-9.5.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:1781a624c229cb35a2ac31cc4a77e28cafc8900733a864870c49bfeedacd106a"}, + {file = "Pillow-9.5.0-cp310-cp310-win32.whl", hash = "sha256:8507eda3cd0608a1f94f58c64817e83ec12fa93a9436938b191b80d9e4c0fc44"}, + {file = "Pillow-9.5.0-cp310-cp310-win_amd64.whl", hash = "sha256:d3c6b54e304c60c4181da1c9dadf83e4a54fd266a99c70ba646a9baa626819eb"}, + {file = "Pillow-9.5.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:7ec6f6ce99dab90b52da21cf0dc519e21095e332ff3b399a357c187b1a5eee32"}, + {file = "Pillow-9.5.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:560737e70cb9c6255d6dcba3de6578a9e2ec4b573659943a5e7e4af13f298f5c"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:96e88745a55b88a7c64fa49bceff363a1a27d9a64e04019c2281049444a571e3"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9c206c29b46cfd343ea7cdfe1232443072bbb270d6a46f59c259460db76779a"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cfcc2c53c06f2ccb8976fb5c71d448bdd0a07d26d8e07e321c103416444c7ad1"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:a0f9bb6c80e6efcde93ffc51256d5cfb2155ff8f78292f074f60f9e70b942d99"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:8d935f924bbab8f0a9a28404422da8af4904e36d5c33fc6f677e4c4485515625"}, + {file = "Pillow-9.5.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:fed1e1cf6a42577953abbe8e6cf2fe2f566daebde7c34724ec8803c4c0cda579"}, + {file = "Pillow-9.5.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:c1170d6b195555644f0616fd6ed929dfcf6333b8675fcca044ae5ab110ded296"}, + {file = "Pillow-9.5.0-cp311-cp311-win32.whl", hash = "sha256:54f7102ad31a3de5666827526e248c3530b3a33539dbda27c6843d19d72644ec"}, + {file = "Pillow-9.5.0-cp311-cp311-win_amd64.whl", hash = "sha256:cfa4561277f677ecf651e2b22dc43e8f5368b74a25a8f7d1d4a3a243e573f2d4"}, + {file = "Pillow-9.5.0-cp311-cp311-win_arm64.whl", hash = "sha256:965e4a05ef364e7b973dd17fc765f42233415974d773e82144c9bbaaaea5d089"}, + {file = "Pillow-9.5.0-cp312-cp312-win32.whl", hash = "sha256:22baf0c3cf0c7f26e82d6e1adf118027afb325e703922c8dfc1d5d0156bb2eeb"}, + {file = "Pillow-9.5.0-cp312-cp312-win_amd64.whl", hash = "sha256:432b975c009cf649420615388561c0ce7cc31ce9b2e374db659ee4f7d57a1f8b"}, + {file = "Pillow-9.5.0-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:5d4ebf8e1db4441a55c509c4baa7a0587a0210f7cd25fcfe74dbbce7a4bd1906"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:375f6e5ee9620a271acb6820b3d1e94ffa8e741c0601db4c0c4d3cb0a9c224bf"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:99eb6cafb6ba90e436684e08dad8be1637efb71c4f2180ee6b8f940739406e78"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2dfaaf10b6172697b9bceb9a3bd7b951819d1ca339a5ef294d1f1ac6d7f63270"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_28_aarch64.whl", hash = "sha256:763782b2e03e45e2c77d7779875f4432e25121ef002a41829d8868700d119392"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:35f6e77122a0c0762268216315bf239cf52b88865bba522999dc38f1c52b9b47"}, + {file = "Pillow-9.5.0-cp37-cp37m-win32.whl", hash = "sha256:aca1c196f407ec7cf04dcbb15d19a43c507a81f7ffc45b690899d6a76ac9fda7"}, + {file = "Pillow-9.5.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322724c0032af6692456cd6ed554bb85f8149214d97398bb80613b04e33769f6"}, + {file = "Pillow-9.5.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:a0aa9417994d91301056f3d0038af1199eb7adc86e646a36b9e050b06f526597"}, + {file = "Pillow-9.5.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:f8286396b351785801a976b1e85ea88e937712ee2c3ac653710a4a57a8da5d9c"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c830a02caeb789633863b466b9de10c015bded434deb3ec87c768e53752ad22a"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbd359831c1657d69bb81f0db962905ee05e5e9451913b18b831febfe0519082"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f8fc330c3370a81bbf3f88557097d1ea26cd8b019d6433aa59f71195f5ddebbf"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:7002d0797a3e4193c7cdee3198d7c14f92c0836d6b4a3f3046a64bd1ce8df2bf"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:229e2c79c00e85989a34b5981a2b67aa079fd08c903f0aaead522a1d68d79e51"}, + {file = "Pillow-9.5.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:9adf58f5d64e474bed00d69bcd86ec4bcaa4123bfa70a65ce72e424bfb88ed96"}, + {file = "Pillow-9.5.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:662da1f3f89a302cc22faa9f14a262c2e3951f9dbc9617609a47521c69dd9f8f"}, + {file = "Pillow-9.5.0-cp38-cp38-win32.whl", hash = "sha256:6608ff3bf781eee0cd14d0901a2b9cc3d3834516532e3bd673a0a204dc8615fc"}, + {file = "Pillow-9.5.0-cp38-cp38-win_amd64.whl", hash = "sha256:e49eb4e95ff6fd7c0c402508894b1ef0e01b99a44320ba7d8ecbabefddcc5569"}, + {file = "Pillow-9.5.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:482877592e927fd263028c105b36272398e3e1be3269efda09f6ba21fd83ec66"}, + {file = "Pillow-9.5.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3ded42b9ad70e5f1754fb7c2e2d6465a9c842e41d178f262e08b8c85ed8a1d8e"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c446d2245ba29820d405315083d55299a796695d747efceb5717a8b450324115"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8aca1152d93dcc27dc55395604dcfc55bed5f25ef4c98716a928bacba90d33a3"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:608488bdcbdb4ba7837461442b90ea6f3079397ddc968c31265c1e056964f1ef"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:60037a8db8750e474af7ffc9faa9b5859e6c6d0a50e55c45576bf28be7419705"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:07999f5834bdc404c442146942a2ecadd1cb6292f5229f4ed3b31e0a108746b1"}, + {file = "Pillow-9.5.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:a127ae76092974abfbfa38ca2d12cbeddcdeac0fb71f9627cc1135bedaf9d51a"}, + {file = "Pillow-9.5.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:489f8389261e5ed43ac8ff7b453162af39c3e8abd730af8363587ba64bb2e865"}, + {file = "Pillow-9.5.0-cp39-cp39-win32.whl", hash = "sha256:9b1af95c3a967bf1da94f253e56b6286b50af23392a886720f563c547e48e964"}, + {file = "Pillow-9.5.0-cp39-cp39-win_amd64.whl", hash = "sha256:77165c4a5e7d5a284f10a6efaa39a0ae8ba839da344f20b111d62cc932fa4e5d"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:833b86a98e0ede388fa29363159c9b1a294b0905b5128baf01db683672f230f5"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:aaf305d6d40bd9632198c766fb64f0c1a83ca5b667f16c1e79e1661ab5060140"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0852ddb76d85f127c135b6dd1f0bb88dbb9ee990d2cd9aa9e28526c93e794fba"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:91ec6fe47b5eb5a9968c79ad9ed78c342b1f97a091677ba0e012701add857829"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cb841572862f629b99725ebaec3287fc6d275be9b14443ea746c1dd325053cbd"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:c380b27d041209b849ed246b111b7c166ba36d7933ec6e41175fd15ab9eb1572"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7c9af5a3b406a50e313467e3565fc99929717f780164fe6fbb7704edba0cebbe"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5671583eab84af046a397d6d0ba25343c00cd50bce03787948e0fff01d4fd9b1"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:84a6f19ce086c1bf894644b43cd129702f781ba5751ca8572f08aa40ef0ab7b7"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:1e7723bd90ef94eda669a3c2c19d549874dd5badaeefabefd26053304abe5799"}, + {file = "Pillow-9.5.0.tar.gz", hash = "sha256:bf548479d336726d7a0eceb6e767e179fbde37833ae42794602631a070d630f1"}, +] + +[package.extras] +docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-removed-in", "sphinxext-opengraph"] +tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"] + @@ -955 +1813 @@ name = "pip-audit" -version = "2.4.14" +version = "2.5.4" @@ -961,2 +1819,2 @@ files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, + {file = "pip_audit-2.5.4-py3-none-any.whl", hash = "sha256:bee3748030c895488b4dd8a6196fa44f484da000cfd52d8fc64dfa3a2e121624"}, + {file = "pip_audit-2.5.4.tar.gz", hash = "sha256:a4cb03f9e2896d626f5b153973d3ac0d32fdb18594d78d393b153c83bb8089b6"}, @@ -967 +1825 @@ CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +cyclonedx-python-lib = ">=2.0,<2.5.0 || >2.5.0,<3.0" @@ -972 +1829,0 @@ pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" @@ -979 +1836 @@ doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.260)", "types-html5lib", "types-requests", "types-toml"] @@ -1033,0 +1891,22 @@ testing = ["pytest", "pytest-benchmark"] +[[package]] +name = "pooch" +version = "1.6.0" +description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\"" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"}, + {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"}, +] + +[package.dependencies] +appdirs = ">=1.3.0" +packaging = ">=20.0" +requests = ">=2.19.0" + +[package.extras] +progress = ["tqdm (>=4.41.0,<5.0.0)"] +sftp = ["paramiko (>=2.7.0)"] +xxhash = ["xxhash (>=1.4.3)"] + @@ -1060,0 +1940,38 @@ test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] +[[package]] +name = "pyarrow" +version = "11.0.0" +description = "Python library for Apache Arrow" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"}, + {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"}, + {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"}, + {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"}, + {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"}, + {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"}, + {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"}, + {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"}, +] + +[package.dependencies] +numpy = ">=1.16.6" + @@ -1072,0 +1990,24 @@ files = [ +[[package]] +name = "pycparser" +version = "2.21" +description = "C parser in Python" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, + {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, +] + +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + @@ -1290,0 +2232,15 @@ testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtuale +[[package]] +name = "python-dateutil" +version = "2.8.2" +description = "Extensions to the standard Python datetime module" +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" +files = [ + {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"}, + {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"}, +] + +[package.dependencies] +six = ">=1.5" + @@ -1391,4 +2347,4 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -name = "resolvelib" -version = "0.9.0" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" +name = "responses" +version = "0.18.0" +description = "A utility library for mocking out the `requests` Python library." +category = "main" @@ -1396 +2352 @@ optional = false -python-versions = "*" +python-versions = ">=3.7" @@ -1398,2 +2354,2 @@ files = [ - {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"}, - {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"}, + {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, + {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, @@ -1401,0 +2358,4 @@ files = [ +[package.dependencies] +requests = ">=2.0,<3.0" +urllib3 = ">=1.25.10" + @@ -1403,4 +2363 @@ files = [ -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] +tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] @@ -1426,0 +2384,82 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] +[[package]] +name = "scikit-learn" +version = "1.2.2" +description = "A set of python modules for machine learning and data mining" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "scikit-learn-1.2.2.tar.gz", hash = "sha256:8429aea30ec24e7a8c7ed8a3fa6213adf3814a6efbea09e16e0a0c71e1a1a3d7"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:99cc01184e347de485bf253d19fcb3b1a3fb0ee4cea5ee3c43ec0cc429b6d29f"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:e6e574db9914afcb4e11ade84fab084536a895ca60aadea3041e85b8ac963edb"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6fe83b676f407f00afa388dd1fdd49e5c6612e551ed84f3b1b182858f09e987d"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e2642baa0ad1e8f8188917423dd73994bf25429f8893ddbe115be3ca3183584"}, + {file = "scikit_learn-1.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:ad66c3848c0a1ec13464b2a95d0a484fd5b02ce74268eaa7e0c697b904f31d6c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:dfeaf8be72117eb61a164ea6fc8afb6dfe08c6f90365bde2dc16456e4bc8e45f"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:fe0aa1a7029ed3e1dcbf4a5bc675aa3b1bc468d9012ecf6c6f081251ca47f590"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:065e9673e24e0dc5113e2dd2b4ca30c9d8aa2fa90f4c0597241c93b63130d233"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf036ea7ef66115e0d49655f16febfa547886deba20149555a41d28f56fd6d3c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:8b0670d4224a3c2d596fd572fb4fa673b2a0ccfb07152688ebd2ea0b8c61025c"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9c710ff9f9936ba8a3b74a455ccf0dcf59b230caa1e9ba0223773c490cab1e51"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:2dd3ffd3950e3d6c0c0ef9033a9b9b32d910c61bd06cb8206303fb4514b88a49"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:44b47a305190c28dd8dd73fc9445f802b6ea716669cfc22ab1eb97b335d238b1"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:953236889928d104c2ef14027539f5f2609a47ebf716b8cbe4437e85dce42744"}, + {file = "scikit_learn-1.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:7f69313884e8eb311460cc2f28676d5e400bd929841a2c8eb8742ae78ebf7c20"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8156db41e1c39c69aa2d8599ab7577af53e9e5e7a57b0504e116cc73c39138dd"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:fe175ee1dab589d2e1033657c5b6bec92a8a3b69103e3dd361b58014729975c3"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d5312d9674bed14f73773d2acf15a3272639b981e60b72c9b190a0cffed5bad"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ea061bf0283bf9a9f36ea3c5d3231ba2176221bbd430abd2603b1c3b2ed85c89"}, + {file = "scikit_learn-1.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:6477eed40dbce190f9f9e9d0d37e020815825b300121307942ec2110302b66a3"}, +] + +[package.dependencies] +joblib = ">=1.1.1" +numpy = ">=1.17.3" +scipy = ">=1.3.2" +threadpoolctl = ">=2.0.0" + +[package.extras] +benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"] +docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=4.0.1)", "sphinx-gallery (>=0.7.0)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"] +examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"] +tests = ["black (>=22.3.0)", "flake8 (>=3.8.2)", "matplotlib (>=3.1.3)", "mypy (>=0.961)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=5.3.1)", "pytest-cov (>=2.9.0)", "scikit-image (>=0.16.2)"] + +[[package]] +name = "scipy" +version = "1.10.1" +description = "Fundamental algorithms for scientific computing in Python" +category = "main" +optional = false +python-versions = "<3.12,>=3.8" +files = [ + {file = "scipy-1.10.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e7354fd7527a4b0377ce55f286805b34e8c54b91be865bac273f527e1b839019"}, + {file = "scipy-1.10.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:4b3f429188c66603a1a5c549fb414e4d3bdc2a24792e061ffbd607d3d75fd84e"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1553b5dcddd64ba9a0d95355e63fe6c3fc303a8fd77c7bc91e77d61363f7433f"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c0ff64b06b10e35215abce517252b375e580a6125fd5fdf6421b98efbefb2d2"}, + {file = "scipy-1.10.1-cp310-cp310-win_amd64.whl", hash = "sha256:fae8a7b898c42dffe3f7361c40d5952b6bf32d10c4569098d276b4c547905ee1"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:0f1564ea217e82c1bbe75ddf7285ba0709ecd503f048cb1236ae9995f64217bd"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d925fa1c81b772882aa55bcc10bf88324dadb66ff85d548c71515f6689c6dac5"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaea0a6be54462ec027de54fca511540980d1e9eea68b2d5c1dbfe084797be35"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15a35c4242ec5f292c3dd364a7c71a61be87a3d4ddcc693372813c0b73c9af1d"}, + {file = "scipy-1.10.1-cp311-cp311-win_amd64.whl", hash = "sha256:43b8e0bcb877faf0abfb613d51026cd5cc78918e9530e375727bf0625c82788f"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5678f88c68ea866ed9ebe3a989091088553ba12c6090244fdae3e467b1139c35"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:39becb03541f9e58243f4197584286e339029e8908c46f7221abeea4b749fa88"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bce5869c8d68cf383ce240e44c1d9ae7c06078a9396df68ce88a1230f93a30c1"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:07c3457ce0b3ad5124f98a86533106b643dd811dd61b548e78cf4c8786652f6f"}, + {file = "scipy-1.10.1-cp38-cp38-win_amd64.whl", hash = "sha256:049a8bbf0ad95277ffba9b3b7d23e5369cc39e66406d60422c8cfef40ccc8415"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cd9f1027ff30d90618914a64ca9b1a77a431159df0e2a195d8a9e8a04c78abf9"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:79c8e5a6c6ffaf3a2262ef1be1e108a035cf4f05c14df56057b64acc5bebffb6"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:51af417a000d2dbe1ec6c372dfe688e041a7084da4fdd350aeb139bd3fb55353"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b4735d6c28aad3cdcf52117e0e91d6b39acd4272f3f5cd9907c24ee931ad601"}, + {file = "scipy-1.10.1-cp39-cp39-win_amd64.whl", hash = "sha256:7ff7f37b1bf4417baca958d254e8e2875d0cc23aaadbe65b3d5b3077b0eb23ea"}, + {file = "scipy-1.10.1.tar.gz", hash = "sha256:2cf9dfb80a7b4589ba4c40ce7588986d6d5cebc5457cad2c2880f6bc2d42f3a5"}, +] + +[package.dependencies] +numpy = ">=1.19.5,<1.27.0" + +[package.extras] +dev = ["click", "doit (>=0.36.0)", "flake8", "mypy", "pycodestyle", "pydevtool", "rich-click", "typing_extensions"] +doc = ["matplotlib (>2)", "numpydoc", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"] +test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"] + @@ -1431 +2470 @@ description = "Easily download, build, install, upgrade, and uninstall Python pa -category = "dev" +category = "main" @@ -1448 +2487 @@ description = "Python 2 and 3 compatibility utilities" -category = "dev" +category = "main" @@ -1479,0 +2519,68 @@ files = [ +[[package]] +name = "soundfile" +version = "0.12.1" +description = "An audio library based on libsndfile, CFFI and NumPy" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"}, + {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"}, + {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"}, + {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"}, +] + +[package.dependencies] +cffi = ">=1.0" + +[package.extras] +numpy = ["numpy"] + +[[package]] +name = "soxr" +version = "0.3.4" +description = "High quality, one-dimensional sample-rate conversion library" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "soxr-0.3.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b7b84126643c063d5daa203f7f9137e21734dabbd7e68c097607b2ef457e2f2e"}, + {file = "soxr-0.3.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:380d2d43871a68e8b1ef1702a0abe6f9e48ddb3933c7a303c45d67e121503e7c"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4a1b4019c9972f57612482c4f85523d6e832e3d10935e2f070a9dcd334a4dcb"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e613cee023b7c3f162b9da3f6b169cd7f58de345275be1fde9f19adc9cf144df"}, + {file = "soxr-0.3.4-cp310-cp310-win32.whl", hash = "sha256:182c02a7ba45a159a0dbb0a297335df2381ead03a65377b19663ea0ff720ecb7"}, + {file = "soxr-0.3.4-cp310-cp310-win_amd64.whl", hash = "sha256:1e95c96ce94524fae453b4331c9910d33f97506f99bae06d76a9c0649710619e"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2678d2719e7496803983584d661eb5fddc7017154a8dda4a774407c56ff07973"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:11bd1396052049e6d389225a0e96a9df15f706da501c619b35d3c72ac6bc7257"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7e23de4dfe54ac30e583bbc9cc3feda1cd776fedce13206bc4b3115b75ecab82"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3e7396498a5f5b7d8f23b656f65c24517a6ff5bdc3ee0623ccd491036a43ea08"}, + {file = "soxr-0.3.4-cp311-cp311-win32.whl", hash = "sha256:e57e9703c2bff834cabc06800d3c11a259544891d2c24a78949f3cf2f5492cc5"}, + {file = "soxr-0.3.4-cp311-cp311-win_amd64.whl", hash = "sha256:7c8350acd7150f74261a0569b47ccb1bb4aa39b2d575860bc97cfa69aab8aead"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:941f7355116fe77fe6a82938fa7799a0e466a494ebc093f676969ce32b2815b1"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:00fdbf24f64d8c3fb800425c383048cb24c32defac80901cde4a57fb6ce5d431"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7bb6d4dc807d04c536674429e2b05ae08a1efac9815c4595e41ffd6b57c2c662"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ff15853895b54f1b627799c6600be1ce5f7286724e7a93e4b7b9d79e5d4166f5"}, + {file = "soxr-0.3.4-cp38-cp38-win32.whl", hash = "sha256:d858becbc1fcc7b38c3436d3276290fae09403cdcbdf1d5986a18dab7023a6c3"}, + {file = "soxr-0.3.4-cp38-cp38-win_amd64.whl", hash = "sha256:068ab4df549df5783cc1eb4eb6c94f53823b164dc27134fc621fc9f5097f38cd"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:20130329985f9767c8417bbd125fe138790a71802b000481c386a800e2ad2bca"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:78090e97abfb326b7cf14ef37d08a17252b07d438388dcbbd82a6836a9d551b1"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84e590e75b7e5dca12bf68bfb090276f34a88fbcd793781c62d47f5d7dbe525e"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3479d265574b960e12bca0878baba0862c43278915e0319d84679bb4d4fcd33"}, + {file = "soxr-0.3.4-cp39-cp39-win32.whl", hash = "sha256:83de825d6a713c7b2e76d9ec3f229a58a9ed290237e7adc05d80e8b39be995a6"}, + {file = "soxr-0.3.4-cp39-cp39-win_amd64.whl", hash = "sha256:2082f88cae89de854c3e0d62f55d0cb31eb11764f5c2a28299121fb642a22472"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:fe8b5f92c802f1e7793c40344f5368dc6163718c9ffa82e79ee6ad779d318ac5"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d0063d5f9a4e1a367084f4705301e9da131cf4d2d32aa3fe0072a1245e18088f"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:a680bab57adae462cdc86abcc7330beb5daa3ba5101165583eedcda88b7ba551"}, + {file = "soxr-0.3.4.tar.gz", hash = "sha256:fe68daf00e8f020977b187699903d219f9e39b9fb3d915f3f923eed8ba431449"}, +] + +[package.dependencies] +numpy = "*" + +[package.extras] +docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"] +test = ["pytest"] + @@ -1494,0 +2602,12 @@ pbr = ">=2.0.0,<2.1.0 || >2.1.0" +[[package]] +name = "threadpoolctl" +version = "3.1.0" +description = "threadpoolctl" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, + {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"}, +] + @@ -1551,0 +2671,12 @@ files = [ +[[package]] +name = "tzdata" +version = "2023.3" +description = "Provider of IANA time zone data" +category = "main" +optional = false +python-versions = ">=2" +files = [ + {file = "tzdata-2023.3-py2.py3-none-any.whl", hash = "sha256:7e65763eef3120314099b6939b5546db7adce1e7d6f2e179e3df563c70511eda"}, + {file = "tzdata-2023.3.tar.gz", hash = "sha256:11ef1e08e54acb0d4f95bdb1be05da659673de4acbd21bf9c69e94cc5e907a3a"}, +] + @@ -1580,0 +2712,196 @@ files = [ +[[package]] +name = "xxhash" +version = "3.2.0" +description = "Python binding for xxHash" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"}, + {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"}, + {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"}, + {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"}, + {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"}, + {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"}, + {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"}, + {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"}, + {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"}, + {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"}, + {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"}, + {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"}, + {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"}, + {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"}, + {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"}, + {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"}, + {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"}, +] + +[[package]] +name = "yarl" +version = "1.8.2" +description = "Yet another URL library" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:bb81f753c815f6b8e2ddd2eef3c855cf7da193b82396ac013c661aaa6cc6b0a5"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:47d49ac96156f0928f002e2424299b2c91d9db73e08c4cd6742923a086f1c863"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3fc056e35fa6fba63248d93ff6e672c096f95f7836938241ebc8260e062832fe"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58a3c13d1c3005dbbac5c9f0d3210b60220a65a999b1833aa46bd6677c69b08e"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:10b08293cda921157f1e7c2790999d903b3fd28cd5c208cf8826b3b508026996"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:de986979bbd87272fe557e0a8fcb66fd40ae2ddfe28a8b1ce4eae22681728fef"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c4fcfa71e2c6a3cb568cf81aadc12768b9995323186a10827beccf5fa23d4f8"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae4d7ff1049f36accde9e1ef7301912a751e5bae0a9d142459646114c70ecba6"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:bf071f797aec5b96abfc735ab97da9fd8f8768b43ce2abd85356a3127909d146"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:74dece2bfc60f0f70907c34b857ee98f2c6dd0f75185db133770cd67300d505f"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:df60a94d332158b444301c7f569659c926168e4d4aad2cfbf4bce0e8fb8be826"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:63243b21c6e28ec2375f932a10ce7eda65139b5b854c0f6b82ed945ba526bff3"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:cfa2bbca929aa742b5084fd4663dd4b87c191c844326fcb21c3afd2d11497f80"}, + {file = "yarl-1.8.2-cp310-cp310-win32.whl", hash = "sha256:b05df9ea7496df11b710081bd90ecc3a3db6adb4fee36f6a411e7bc91a18aa42"}, + {file = "yarl-1.8.2-cp310-cp310-win_amd64.whl", hash = "sha256:24ad1d10c9db1953291f56b5fe76203977f1ed05f82d09ec97acb623a7976574"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:2a1fca9588f360036242f379bfea2b8b44cae2721859b1c56d033adfd5893634"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:f37db05c6051eff17bc832914fe46869f8849de5b92dc4a3466cd63095d23dfd"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:77e913b846a6b9c5f767b14dc1e759e5aff05502fe73079f6f4176359d832581"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0978f29222e649c351b173da2b9b4665ad1feb8d1daa9d971eb90df08702668a"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:388a45dc77198b2460eac0aca1efd6a7c09e976ee768b0d5109173e521a19daf"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2305517e332a862ef75be8fad3606ea10108662bc6fe08509d5ca99503ac2aee"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:42430ff511571940d51e75cf42f1e4dbdded477e71c1b7a17f4da76c1da8ea76"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3150078118f62371375e1e69b13b48288e44f6691c1069340081c3fd12c94d5b"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:c15163b6125db87c8f53c98baa5e785782078fbd2dbeaa04c6141935eb6dab7a"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:4d04acba75c72e6eb90745447d69f84e6c9056390f7a9724605ca9c56b4afcc6"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:e7fd20d6576c10306dea2d6a5765f46f0ac5d6f53436217913e952d19237efc4"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:75c16b2a900b3536dfc7014905a128a2bea8fb01f9ee26d2d7d8db0a08e7cb2c"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:6d88056a04860a98341a0cf53e950e3ac9f4e51d1b6f61a53b0609df342cc8b2"}, + {file = "yarl-1.8.2-cp311-cp311-win32.whl", hash = "sha256:fb742dcdd5eec9f26b61224c23baea46c9055cf16f62475e11b9b15dfd5c117b"}, + {file = "yarl-1.8.2-cp311-cp311-win_amd64.whl", hash = "sha256:8c46d3d89902c393a1d1e243ac847e0442d0196bbd81aecc94fcebbc2fd5857c"}, + {file = "yarl-1.8.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:ceff9722e0df2e0a9e8a79c610842004fa54e5b309fe6d218e47cd52f791d7ef"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3f6b4aca43b602ba0f1459de647af954769919c4714706be36af670a5f44c9c1"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1684a9bd9077e922300ecd48003ddae7a7474e0412bea38d4631443a91d61077"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ebb78745273e51b9832ef90c0898501006670d6e059f2cdb0e999494eb1450c2"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3adeef150d528ded2a8e734ebf9ae2e658f4c49bf413f5f157a470e17a4a2e89"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:57a7c87927a468e5a1dc60c17caf9597161d66457a34273ab1760219953f7f4c"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:efff27bd8cbe1f9bd127e7894942ccc20c857aa8b5a0327874f30201e5ce83d0"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:a783cd344113cb88c5ff7ca32f1f16532a6f2142185147822187913eb989f739"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:705227dccbe96ab02c7cb2c43e1228e2826e7ead880bb19ec94ef279e9555b5b"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:34c09b43bd538bf6c4b891ecce94b6fa4f1f10663a8d4ca589a079a5018f6ed7"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:a48f4f7fea9a51098b02209d90297ac324241bf37ff6be6d2b0149ab2bd51b37"}, + {file = "yarl-1.8.2-cp37-cp37m-win32.whl", hash = "sha256:0414fd91ce0b763d4eadb4456795b307a71524dbacd015c657bb2a39db2eab89"}, + {file = "yarl-1.8.2-cp37-cp37m-win_amd64.whl", hash = "sha256:d881d152ae0007809c2c02e22aa534e702f12071e6b285e90945aa3c376463c5"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5df5e3d04101c1e5c3b1d69710b0574171cc02fddc4b23d1b2813e75f35a30b1"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7a66c506ec67eb3159eea5096acd05f5e788ceec7b96087d30c7d2865a243918"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2b4fa2606adf392051d990c3b3877d768771adc3faf2e117b9de7eb977741229"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1e21fb44e1eff06dd6ef971d4bdc611807d6bd3691223d9c01a18cec3677939e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:93202666046d9edadfe9f2e7bf5e0782ea0d497b6d63da322e541665d65a044e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fc77086ce244453e074e445104f0ecb27530d6fd3a46698e33f6c38951d5a0f1"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:64dd68a92cab699a233641f5929a40f02a4ede8c009068ca8aa1fe87b8c20ae3"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1b372aad2b5f81db66ee7ec085cbad72c4da660d994e8e590c997e9b01e44901"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e6f3515aafe0209dd17fb9bdd3b4e892963370b3de781f53e1746a521fb39fc0"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:dfef7350ee369197106805e193d420b75467b6cceac646ea5ed3049fcc950a05"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:728be34f70a190566d20aa13dc1f01dc44b6aa74580e10a3fb159691bc76909d"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ff205b58dc2929191f68162633d5e10e8044398d7a45265f90a0f1d51f85f72c"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:baf211dcad448a87a0d9047dc8282d7de59473ade7d7fdf22150b1d23859f946"}, + {file = "yarl-1.8.2-cp38-cp38-win32.whl", hash = "sha256:272b4f1599f1b621bf2aabe4e5b54f39a933971f4e7c9aa311d6d7dc06965165"}, + {file = "yarl-1.8.2-cp38-cp38-win_amd64.whl", hash = "sha256:326dd1d3caf910cd26a26ccbfb84c03b608ba32499b5d6eeb09252c920bcbe4f"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:f8ca8ad414c85bbc50f49c0a106f951613dfa5f948ab69c10ce9b128d368baf8"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:418857f837347e8aaef682679f41e36c24250097f9e2f315d39bae3a99a34cbf"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:ae0eec05ab49e91a78700761777f284c2df119376e391db42c38ab46fd662b77"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:009a028127e0a1755c38b03244c0bea9d5565630db9c4cf9572496e947137a87"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3edac5d74bb3209c418805bda77f973117836e1de7c000e9755e572c1f7850d0"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:da65c3f263729e47351261351b8679c6429151ef9649bba08ef2528ff2c423b2"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ef8fb25e52663a1c85d608f6dd72e19bd390e2ecaf29c17fb08f730226e3a08"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bcd7bb1e5c45274af9a1dd7494d3c52b2be5e6bd8d7e49c612705fd45420b12d"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:44ceac0450e648de86da8e42674f9b7077d763ea80c8ceb9d1c3e41f0f0a9951"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:97209cc91189b48e7cfe777237c04af8e7cc51eb369004e061809bcdf4e55220"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:48dd18adcf98ea9cd721a25313aef49d70d413a999d7d89df44f469edfb38a06"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:e59399dda559688461762800d7fb34d9e8a6a7444fd76ec33220a926c8be1516"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d617c241c8c3ad5c4e78a08429fa49e4b04bedfc507b34b4d8dceb83b4af3588"}, + {file = "yarl-1.8.2-cp39-cp39-win32.whl", hash = "sha256:cb6d48d80a41f68de41212f3dfd1a9d9898d7841c8f7ce6696cf2fd9cb57ef83"}, + {file = "yarl-1.8.2-cp39-cp39-win_amd64.whl", hash = "sha256:6604711362f2dbf7160df21c416f81fac0de6dbcf0b5445a2ef25478ecc4c778"}, + {file = "yarl-1.8.2.tar.gz", hash = "sha256:49d43402c6e3013ad0978602bf6bf5328535c48d192304b91b97a3c6790b1562"}, +] + +[package.dependencies] +idna = ">=2.0" +multidict = ">=4.0" + @@ -1584 +2911 @@ python-versions = "3.9.15" -content-hash = "67ae1eb17c3921f944b323ddcaa58919c1deff1e7cc25e63cf3a82ec23340c62" +content-hash = "f916a260ab2e57ba4b18f783a88b5fbb1631a59795bed892690777d2c7ee0cd6" diff --git a/jobs/mongodb_migration/pyproject.toml b/jobs/mongodb_migration/pyproject.toml index 8f80140e..ee7a4f9c 100644 --- a/jobs/mongodb_migration/pyproject.toml +++ b/jobs/mongodb_migration/pyproject.toml @@ -19 +19 @@ mypy = "^1.0.0" -pip-audit = "^2.4.14" +pip-audit = "^2.5.4" diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock index c1bc5c87..f82da90e 100644 --- a/libs/libcommon/poetry.lock +++ b/libs/libcommon/poetry.lock @@ -1 +1,125 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. + +[[package]] +name = "aiohttp" +version = "3.8.4" +description = "Async http client/server framework (asyncio)" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"}, + {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"}, + {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"}, + {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"}, + {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"}, + {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"}, + {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"}, + {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"}, + {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"}, + {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"}, + {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"}, + {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"}, +] + +[package.dependencies] +aiosignal = ">=1.1.2" +async-timeout = ">=4.0.0a3,<5.0" +attrs = ">=17.3.0" +charset-normalizer = ">=2.0,<4.0" +frozenlist = ">=1.1.1" +multidict = ">=4.5,<7.0" +yarl = ">=1.0,<2.0" + +[package.extras] +speedups = ["Brotli", "aiodns", "cchardet"] + +[[package]] +name = "aiosignal" +version = "1.3.1" +description = "aiosignal: a list of registered asynchronous callbacks" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, + {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, +] + +[package.dependencies] +frozenlist = ">=1.1.0" @@ -14,0 +139,12 @@ files = [ +[[package]] +name = "async-timeout" +version = "4.0.2" +description = "Timeout context manager for asyncio programs" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"}, + {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"}, +] + @@ -19 +155 @@ description = "Classes Without Boilerplate" -category = "dev" +category = "main" @@ -33,0 +170,11 @@ tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy +[[package]] +name = "audioread" +version = "3.0.0" +description = "multi-library, cross-platform audio decoding" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"}, +] + @@ -36 +183 @@ name = "bandit" -version = "1.7.4" +version = "1.7.5" @@ -42,2 +189,2 @@ files = [ - {file = "bandit-1.7.4-py3-none-any.whl", hash = "sha256:412d3f259dab4077d0e7f0c11f50f650cc7d10db905d98f6520a95a18049658a"}, - {file = "bandit-1.7.4.tar.gz", hash = "sha256:2d63a8c573417bae338962d4b9b06fbc6080f74ecd955a092849e1e65c717bd2"}, + {file = "bandit-1.7.5-py3-none-any.whl", hash = "sha256:75665181dc1e0096369112541a056c59d1c5f66f9bb74a8d686c3c362b83f549"}, + {file = "bandit-1.7.5.tar.gz", hash = "sha256:bdfc739baa03b880c2d15d0431b31c658ffc348e907fe197e54e0389dd59e11e"}, @@ -49,0 +197 @@ PyYAML = ">=5.3.1" +rich = "*" @@ -53,2 +201,2 @@ stevedore = ">=1.20.0" -test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "toml"] -toml = ["toml"] +test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "tomli (>=1.1.0)"] +toml = ["tomli (>=1.1.0)"] @@ -125,0 +274,77 @@ files = [ +[[package]] +name = "cffi" +version = "1.15.1" +description = "Foreign Function Interface for Python calling C code." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"}, + {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"}, + {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"}, + {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"}, + {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"}, + {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"}, + {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"}, + {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"}, + {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"}, + {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"}, + {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"}, + {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"}, + {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"}, + {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"}, + {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"}, + {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"}, + {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"}, + {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"}, + {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"}, +] + +[package.dependencies] +pycparser = "*" + @@ -128 +353 @@ name = "charset-normalizer" -version = "3.0.1" +version = "3.1.0" @@ -132 +357 @@ optional = false -python-versions = "*" +python-versions = ">=3.7.0" @@ -134,88 +359,75 @@ files = [ - {file = "charset-normalizer-3.0.1.tar.gz", hash = "sha256:ebea339af930f8ca5d7a699b921106c6e29c617fe9606fa7baa043c1cdae326f"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:88600c72ef7587fe1708fd242b385b6ed4b8904976d5da0893e31df8b3480cb6"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c75ffc45f25324e68ab238cb4b5c0a38cd1c3d7f1fb1f72b5541de469e2247db"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:db72b07027db150f468fbada4d85b3b2729a3db39178abf5c543b784c1254539"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:62595ab75873d50d57323a91dd03e6966eb79c41fa834b7a1661ed043b2d404d"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ff6f3db31555657f3163b15a6b7c6938d08df7adbfc9dd13d9d19edad678f1e8"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:772b87914ff1152b92a197ef4ea40efe27a378606c39446ded52c8f80f79702e"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:70990b9c51340e4044cfc394a81f614f3f90d41397104d226f21e66de668730d"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:292d5e8ba896bbfd6334b096e34bffb56161c81408d6d036a7dfa6929cff8783"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:2edb64ee7bf1ed524a1da60cdcd2e1f6e2b4f66ef7c077680739f1641f62f555"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:31a9ddf4718d10ae04d9b18801bd776693487cbb57d74cc3458a7673f6f34639"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:44ba614de5361b3e5278e1241fda3dc1838deed864b50a10d7ce92983797fa76"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:12db3b2c533c23ab812c2b25934f60383361f8a376ae272665f8e48b88e8e1c6"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c512accbd6ff0270939b9ac214b84fb5ada5f0409c44298361b2f5e13f9aed9e"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-win32.whl", hash = "sha256:502218f52498a36d6bf5ea77081844017bf7982cdbe521ad85e64cabee1b608b"}, - {file = "charset_normalizer-3.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:601f36512f9e28f029d9481bdaf8e89e5148ac5d89cffd3b05cd533eeb423b59"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0298eafff88c99982a4cf66ba2efa1128e4ddaca0b05eec4c456bbc7db691d8d"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:a8d0fc946c784ff7f7c3742310cc8a57c5c6dc31631269876a88b809dbeff3d3"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:87701167f2a5c930b403e9756fab1d31d4d4da52856143b609e30a1ce7160f3c"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:14e76c0f23218b8f46c4d87018ca2e441535aed3632ca134b10239dfb6dadd6b"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0c0a590235ccd933d9892c627dec5bc7511ce6ad6c1011fdf5b11363022746c1"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8c7fe7afa480e3e82eed58e0ca89f751cd14d767638e2550c77a92a9e749c317"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:79909e27e8e4fcc9db4addea88aa63f6423ebb171db091fb4373e3312cb6d603"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8ac7b6a045b814cf0c47f3623d21ebd88b3e8cf216a14790b455ea7ff0135d18"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:72966d1b297c741541ca8cf1223ff262a6febe52481af742036a0b296e35fa5a"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f9d0c5c045a3ca9bedfc35dca8526798eb91a07aa7a2c0fee134c6c6f321cbd7"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:5995f0164fa7df59db4746112fec3f49c461dd6b31b841873443bdb077c13cfc"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:4a8fcf28c05c1f6d7e177a9a46a1c52798bfe2ad80681d275b10dcf317deaf0b"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:761e8904c07ad053d285670f36dd94e1b6ab7f16ce62b9805c475b7aa1cffde6"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-win32.whl", hash = "sha256:71140351489970dfe5e60fc621ada3e0f41104a5eddaca47a7acb3c1b851d6d3"}, - {file = "charset_normalizer-3.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:9ab77acb98eba3fd2a85cd160851816bfce6871d944d885febf012713f06659c"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:84c3990934bae40ea69a82034912ffe5a62c60bbf6ec5bc9691419641d7d5c9a"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:74292fc76c905c0ef095fe11e188a32ebd03bc38f3f3e9bcb85e4e6db177b7ea"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c95a03c79bbe30eec3ec2b7f076074f4281526724c8685a42872974ef4d36b72"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f4c39b0e3eac288fedc2b43055cfc2ca7a60362d0e5e87a637beac5d801ef478"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:df2c707231459e8a4028eabcd3cfc827befd635b3ef72eada84ab13b52e1574d"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:93ad6d87ac18e2a90b0fe89df7c65263b9a99a0eb98f0a3d2e079f12a0735837"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:59e5686dd847347e55dffcc191a96622f016bc0ad89105e24c14e0d6305acbc6"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:cd6056167405314a4dc3c173943f11249fa0f1b204f8b51ed4bde1a9cd1834dc"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:083c8d17153ecb403e5e1eb76a7ef4babfc2c48d58899c98fcaa04833e7a2f9a"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:f5057856d21e7586765171eac8b9fc3f7d44ef39425f85dbcccb13b3ebea806c"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:7eb33a30d75562222b64f569c642ff3dc6689e09adda43a082208397f016c39a"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-win32.whl", hash = "sha256:95dea361dd73757c6f1c0a1480ac499952c16ac83f7f5f4f84f0658a01b8ef41"}, - {file = "charset_normalizer-3.0.1-cp36-cp36m-win_amd64.whl", hash = "sha256:eaa379fcd227ca235d04152ca6704c7cb55564116f8bc52545ff357628e10602"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:3e45867f1f2ab0711d60c6c71746ac53537f1684baa699f4f668d4c6f6ce8e14"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cadaeaba78750d58d3cc6ac4d1fd867da6fc73c88156b7a3212a3cd4819d679d"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:911d8a40b2bef5b8bbae2e36a0b103f142ac53557ab421dc16ac4aafee6f53dc"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:503e65837c71b875ecdd733877d852adbc465bd82c768a067badd953bf1bc5a3"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a60332922359f920193b1d4826953c507a877b523b2395ad7bc716ddd386d866"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16a8663d6e281208d78806dbe14ee9903715361cf81f6d4309944e4d1e59ac5b"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:a16418ecf1329f71df119e8a65f3aa68004a3f9383821edcb20f0702934d8087"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:9d9153257a3f70d5f69edf2325357251ed20f772b12e593f3b3377b5f78e7ef8"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:02a51034802cbf38db3f89c66fb5d2ec57e6fe7ef2f4a44d070a593c3688667b"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:2e396d70bc4ef5325b72b593a72c8979999aa52fb8bcf03f701c1b03e1166918"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:11b53acf2411c3b09e6af37e4b9005cba376c872503c8f28218c7243582df45d"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-win32.whl", hash = "sha256:0bf2dae5291758b6f84cf923bfaa285632816007db0330002fa1de38bfcb7154"}, - {file = "charset_normalizer-3.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:2c03cc56021a4bd59be889c2b9257dae13bf55041a3372d3295416f86b295fb5"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:024e606be3ed92216e2b6952ed859d86b4cfa52cd5bc5f050e7dc28f9b43ec42"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:4b0d02d7102dd0f997580b51edc4cebcf2ab6397a7edf89f1c73b586c614272c"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:358a7c4cb8ba9b46c453b1dd8d9e431452d5249072e4f56cfda3149f6ab1405e"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81d6741ab457d14fdedc215516665050f3822d3e56508921cc7239f8c8e66a58"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8b8af03d2e37866d023ad0ddea594edefc31e827fee64f8de5611a1dbc373174"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9cf4e8ad252f7c38dd1f676b46514f92dc0ebeb0db5552f5f403509705e24753"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e696f0dd336161fca9adbb846875d40752e6eba585843c768935ba5c9960722b"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c22d3fe05ce11d3671297dc8973267daa0f938b93ec716e12e0f6dee81591dc1"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:109487860ef6a328f3eec66f2bf78b0b72400280d8f8ea05f69c51644ba6521a"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:37f8febc8ec50c14f3ec9637505f28e58d4f66752207ea177c1d67df25da5aed"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:f97e83fa6c25693c7a35de154681fcc257c1c41b38beb0304b9c4d2d9e164479"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:a152f5f33d64a6be73f1d30c9cc82dfc73cec6477ec268e7c6e4c7d23c2d2291"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:39049da0ffb96c8cbb65cbf5c5f3ca3168990adf3551bd1dee10c48fce8ae820"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-win32.whl", hash = "sha256:4457ea6774b5611f4bed5eaa5df55f70abde42364d498c5134b7ef4c6958e20e"}, - {file = "charset_normalizer-3.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:e62164b50f84e20601c1ff8eb55620d2ad25fb81b59e3cd776a1902527a788af"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:8eade758719add78ec36dc13201483f8e9b5d940329285edcd5f70c0a9edbd7f"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8499ca8f4502af841f68135133d8258f7b32a53a1d594aa98cc52013fff55678"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3fc1c4a2ffd64890aebdb3f97e1278b0cc72579a08ca4de8cd2c04799a3a22be"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:00d3ffdaafe92a5dc603cb9bd5111aaa36dfa187c8285c543be562e61b755f6b"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c2ac1b08635a8cd4e0cbeaf6f5e922085908d48eb05d44c5ae9eabab148512ca"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f6f45710b4459401609ebebdbcfb34515da4fc2aa886f95107f556ac69a9147e"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3ae1de54a77dc0d6d5fcf623290af4266412a7c4be0b1ff7444394f03f5c54e3"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3b590df687e3c5ee0deef9fc8c547d81986d9a1b56073d82de008744452d6541"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ab5de034a886f616a5668aa5d098af2b5385ed70142090e2a31bcbd0af0fdb3d"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:9cb3032517f1627cc012dbc80a8ec976ae76d93ea2b5feaa9d2a5b8882597579"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:608862a7bf6957f2333fc54ab4399e405baad0163dc9f8d99cb236816db169d4"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0f438ae3532723fb6ead77e7c604be7c8374094ef4ee2c5e03a3a17f1fca256c"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:356541bf4381fa35856dafa6a965916e54bed415ad8a24ee6de6e37deccf2786"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-win32.whl", hash = "sha256:39cf9ed17fe3b1bc81f33c9ceb6ce67683ee7526e65fde1447c772afc54a1bb8"}, - {file = "charset_normalizer-3.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:0a11e971ed097d24c534c037d298ad32c6ce81a45736d31e0ff0ad37ab437d59"}, - {file = "charset_normalizer-3.0.1-py3-none-any.whl", hash = "sha256:7e189e2e1d3ed2f4aebabd2d5b0f931e883676e51c7624826e0a4e5fe8a0bf24"}, + {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"}, + {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"}, @@ -253 +465 @@ name = "coverage" -version = "7.2.1" +version = "7.2.2" @@ -259,51 +471,51 @@ files = [ - {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"}, - {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"}, - {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"}, - {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"}, - {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"}, - {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"}, - {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"}, - {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"}, - {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"}, - {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"}, - {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"}, - {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"}, - {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"}, - {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"}, - {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"}, - {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"}, - {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"}, - {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"}, - {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"}, - {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"}, - {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"}, - {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"}, - {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"}, - {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"}, - {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"}, - {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"}, - {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"}, - {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"}, - {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"}, - {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"}, - {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"}, - {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"}, - {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"}, - {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"}, - {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"}, - {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"}, - {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"}, - {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"}, - {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"}, - {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"}, - {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"}, - {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"}, - {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"}, - {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"}, - {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"}, - {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"}, - {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"}, - {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"}, - {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"}, - {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"}, - {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"}, + {file = "coverage-7.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c90e73bdecb7b0d1cea65a08cb41e9d672ac6d7995603d6465ed4914b98b9ad7"}, + {file = "coverage-7.2.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e2926b8abedf750c2ecf5035c07515770944acf02e1c46ab08f6348d24c5f94d"}, + {file = "coverage-7.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57b77b9099f172804e695a40ebaa374f79e4fb8b92f3e167f66facbf92e8e7f5"}, + {file = "coverage-7.2.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:efe1c0adad110bf0ad7fb59f833880e489a61e39d699d37249bdf42f80590169"}, + {file = "coverage-7.2.2-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2199988e0bc8325d941b209f4fd1c6fa007024b1442c5576f1a32ca2e48941e6"}, + {file = "coverage-7.2.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:81f63e0fb74effd5be736cfe07d710307cc0a3ccb8f4741f7f053c057615a137"}, + {file = "coverage-7.2.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:186e0fc9cf497365036d51d4d2ab76113fb74f729bd25da0975daab2e107fd90"}, + {file = "coverage-7.2.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:420f94a35e3e00a2b43ad5740f935358e24478354ce41c99407cddd283be00d2"}, + {file = "coverage-7.2.2-cp310-cp310-win32.whl", hash = "sha256:38004671848b5745bb05d4d621526fca30cee164db42a1f185615f39dc997292"}, + {file = "coverage-7.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:0ce383d5f56d0729d2dd40e53fe3afeb8f2237244b0975e1427bfb2cf0d32bab"}, + {file = "coverage-7.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3eb55b7b26389dd4f8ae911ba9bc8c027411163839dea4c8b8be54c4ee9ae10b"}, + {file = "coverage-7.2.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d2b96123a453a2d7f3995ddb9f28d01fd112319a7a4d5ca99796a7ff43f02af5"}, + {file = "coverage-7.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:299bc75cb2a41e6741b5e470b8c9fb78d931edbd0cd009c58e5c84de57c06731"}, + {file = "coverage-7.2.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5e1df45c23d4230e3d56d04414f9057eba501f78db60d4eeecfcb940501b08fd"}, + {file = "coverage-7.2.2-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:006ed5582e9cbc8115d2e22d6d2144a0725db542f654d9d4fda86793832f873d"}, + {file = "coverage-7.2.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d683d230b5774816e7d784d7ed8444f2a40e7a450e5720d58af593cb0b94a212"}, + {file = "coverage-7.2.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:8efb48fa743d1c1a65ee8787b5b552681610f06c40a40b7ef94a5b517d885c54"}, + {file = "coverage-7.2.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:4c752d5264053a7cf2fe81c9e14f8a4fb261370a7bb344c2a011836a96fb3f57"}, + {file = "coverage-7.2.2-cp311-cp311-win32.whl", hash = "sha256:55272f33da9a5d7cccd3774aeca7a01e500a614eaea2a77091e9be000ecd401d"}, + {file = "coverage-7.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:92ebc1619650409da324d001b3a36f14f63644c7f0a588e331f3b0f67491f512"}, + {file = "coverage-7.2.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5afdad4cc4cc199fdf3e18088812edcf8f4c5a3c8e6cb69127513ad4cb7471a9"}, + {file = "coverage-7.2.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0484d9dd1e6f481b24070c87561c8d7151bdd8b044c93ac99faafd01f695c78e"}, + {file = "coverage-7.2.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d530191aa9c66ab4f190be8ac8cc7cfd8f4f3217da379606f3dd4e3d83feba69"}, + {file = "coverage-7.2.2-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ac0f522c3b6109c4b764ffec71bf04ebc0523e926ca7cbe6c5ac88f84faced0"}, + {file = "coverage-7.2.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:ba279aae162b20444881fc3ed4e4f934c1cf8620f3dab3b531480cf602c76b7f"}, + {file = "coverage-7.2.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:53d0fd4c17175aded9c633e319360d41a1f3c6e352ba94edcb0fa5167e2bad67"}, + {file = "coverage-7.2.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8c99cb7c26a3039a8a4ee3ca1efdde471e61b4837108847fb7d5be7789ed8fd9"}, + {file = "coverage-7.2.2-cp37-cp37m-win32.whl", hash = "sha256:5cc0783844c84af2522e3a99b9b761a979a3ef10fb87fc4048d1ee174e18a7d8"}, + {file = "coverage-7.2.2-cp37-cp37m-win_amd64.whl", hash = "sha256:817295f06eacdc8623dc4df7d8b49cea65925030d4e1e2a7c7218380c0072c25"}, + {file = "coverage-7.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:6146910231ece63facfc5984234ad1b06a36cecc9fd0c028e59ac7c9b18c38c6"}, + {file = "coverage-7.2.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:387fb46cb8e53ba7304d80aadca5dca84a2fbf6fe3faf6951d8cf2d46485d1e5"}, + {file = "coverage-7.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:046936ab032a2810dcaafd39cc4ef6dd295df1a7cbead08fe996d4765fca9fe4"}, + {file = "coverage-7.2.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e627dee428a176ffb13697a2c4318d3f60b2ccdde3acdc9b3f304206ec130ccd"}, + {file = "coverage-7.2.2-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4fa54fb483decc45f94011898727802309a109d89446a3c76387d016057d2c84"}, + {file = "coverage-7.2.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:3668291b50b69a0c1ef9f462c7df2c235da3c4073f49543b01e7eb1dee7dd540"}, + {file = "coverage-7.2.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:7c20b731211261dc9739bbe080c579a1835b0c2d9b274e5fcd903c3a7821cf88"}, + {file = "coverage-7.2.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:5764e1f7471cb8f64b8cda0554f3d4c4085ae4b417bfeab236799863703e5de2"}, + {file = "coverage-7.2.2-cp38-cp38-win32.whl", hash = "sha256:4f01911c010122f49a3e9bdc730eccc66f9b72bd410a3a9d3cb8448bb50d65d3"}, + {file = "coverage-7.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:c448b5c9e3df5448a362208b8d4b9ed85305528313fca1b479f14f9fe0d873b8"}, + {file = "coverage-7.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bfe7085783cda55e53510482fa7b5efc761fad1abe4d653b32710eb548ebdd2d"}, + {file = "coverage-7.2.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:9d22e94e6dc86de981b1b684b342bec5e331401599ce652900ec59db52940005"}, + {file = "coverage-7.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:507e4720791977934bba016101579b8c500fb21c5fa3cd4cf256477331ddd988"}, + {file = "coverage-7.2.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bc4803779f0e4b06a2361f666e76f5c2e3715e8e379889d02251ec911befd149"}, + {file = "coverage-7.2.2-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db8c2c5ace167fd25ab5dd732714c51d4633f58bac21fb0ff63b0349f62755a8"}, + {file = "coverage-7.2.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:4f68ee32d7c4164f1e2c8797535a6d0a3733355f5861e0f667e37df2d4b07140"}, + {file = "coverage-7.2.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:d52f0a114b6a58305b11a5cdecd42b2e7f1ec77eb20e2b33969d702feafdd016"}, + {file = "coverage-7.2.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:797aad79e7b6182cb49c08cc5d2f7aa7b2128133b0926060d0a8889ac43843be"}, + {file = "coverage-7.2.2-cp39-cp39-win32.whl", hash = "sha256:db45eec1dfccdadb179b0f9ca616872c6f700d23945ecc8f21bb105d74b1c5fc"}, + {file = "coverage-7.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:8dbe2647bf58d2c5a6c5bcc685f23b5f371909a5624e9f5cd51436d6a9f6c6ef"}, + {file = "coverage-7.2.2-pp37.pp38.pp39-none-any.whl", hash = "sha256:872d6ce1f5be73f05bea4df498c140b9e7ee5418bfa2cc8204e7f9b817caa968"}, + {file = "coverage-7.2.2.tar.gz", hash = "sha256:36dd42da34fe94ed98c39887b86db9d06777b1c8f860520e21126a75507024f2"}, @@ -317 +529 @@ name = "cyclonedx-python-lib" -version = "3.1.5" +version = "2.7.1" @@ -323,2 +535,2 @@ files = [ - {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, - {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, + {file = "cyclonedx-python-lib-2.7.1.tar.gz", hash = "sha256:493bf2f30e26c48f305f745ed8580ce10d05a8d68d62a598fe95f05a0d9007dc"}, + {file = "cyclonedx_python_lib-2.7.1-py3-none-any.whl", hash = "sha256:fabc4c8baf722faeea01c3bbca83730e3489dfb37d85c6036baa67a9a7519d40"}, @@ -332,0 +545,74 @@ toml = ">=0.10.0,<0.11.0" +[[package]] +name = "datasets" +version = "2.11.0" +description = "HuggingFace community-driven open-source library of datasets" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "datasets-2.11.0-py3-none-any.whl", hash = "sha256:d946cdb8c4885d3016a2ab3129c9403dd3358fe9107e8ab5e549ceab672774af"}, + {file = "datasets-2.11.0.tar.gz", hash = "sha256:1ca53b9cd6ece7a3fdb81176dadd5b9e646420e52e68e85307b27db3a36ca18c"}, +] + +[package.dependencies] +aiohttp = "*" +dill = ">=0.3.0,<0.3.7" +fsspec = {version = ">=2021.11.1", extras = ["http"]} +huggingface-hub = ">=0.11.0,<1.0.0" +librosa = {version = "*", optional = true, markers = "extra == \"audio\""} +multiprocess = "*" +numpy = ">=1.17" +packaging = "*" +pandas = "*" +Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} +pyarrow = ">=8.0.0" +pyyaml = ">=5.1" +requests = ">=2.19.0" +responses = "<0.19" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} +tqdm = ">=4.62.1" +xxhash = "*" + +[package.extras] +apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] +audio = ["librosa", "soundfile (>=0.12.1)"] +benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +docs = ["s3fs"] +jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"] +s3 = ["s3fs"] +tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"] +tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +torch = ["torch"] +vision = ["Pillow (>=6.2.1)"] + +[[package]] +name = "decorator" +version = "5.1.1" +description = "Decorators for Humans" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"}, + {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"}, +] + +[[package]] +name = "dill" +version = "0.3.6" +description = "serialize all of python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"}, + {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"}, +] + +[package.extras] +graph = ["objgraph (>=1.7.2)"] + @@ -373 +659 @@ name = "exceptiongroup" -version = "1.1.0" +version = "1.1.1" @@ -379,2 +665,2 @@ files = [ - {file = "exceptiongroup-1.1.0-py3-none-any.whl", hash = "sha256:327cbda3da756e2de031a3107b81ab7b3770a602c4d16ca618298c526f4bec1e"}, - {file = "exceptiongroup-1.1.0.tar.gz", hash = "sha256:bcb67d800a4497e1b404c2dd44fca47d3b7a5e5433dbab67f96c1a685cdfdf23"}, + {file = "exceptiongroup-1.1.1-py3-none-any.whl", hash = "sha256:232c37c63e4f682982c8b6459f33a8981039e5fb8756b2074364e5055c498c9e"}, + {file = "exceptiongroup-1.1.1.tar.gz", hash = "sha256:d484c3090ba2889ae2928419117447a14daf3c1231d5e30d0aae34f354f01785"}, @@ -388 +674 @@ name = "filelock" -version = "3.9.0" +version = "3.10.7" @@ -394,2 +680,2 @@ files = [ - {file = "filelock-3.9.0-py3-none-any.whl", hash = "sha256:f58d535af89bb9ad5cd4df046f741f8553a418c01a7856bf0d173bbc9f6bd16d"}, - {file = "filelock-3.9.0.tar.gz", hash = "sha256:7b319f24340b51f55a2bf7a12ac0755a9b03e718311dac567a0f4f7fabd2f5de"}, + {file = "filelock-3.10.7-py3-none-any.whl", hash = "sha256:bde48477b15fde2c7e5a0713cbe72721cb5a5ad32ee0b8f419907960b9d75536"}, + {file = "filelock-3.10.7.tar.gz", hash = "sha256:892be14aa8efc01673b5ed6589dbccb95f9a8596f0507e232626155495c18105"}, @@ -399,2 +685,2 @@ files = [ -docs = ["furo (>=2022.12.7)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"] -testing = ["covdefaults (>=2.2.2)", "coverage (>=7.0.1)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-timeout (>=2.1)"] +docs = ["furo (>=2022.12.7)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"] +testing = ["covdefaults (>=2.3)", "coverage (>=7.2.2)", "diff-cover (>=7.5)", "pytest (>=7.2.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)", "pytest-timeout (>=2.1)"] @@ -418,0 +705,122 @@ pyflakes = ">=2.3.0,<2.4.0" +[[package]] +name = "frozenlist" +version = "1.3.3" +description = "A list-like structure which implements collections.abc.MutableSequence" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"}, + {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"}, + {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"}, + {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"}, + {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"}, + {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"}, + {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"}, + {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"}, + {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"}, + {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"}, + {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"}, +] + +[[package]] +name = "fsspec" +version = "2023.3.0" +description = "File-system specification" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "fsspec-2023.3.0-py3-none-any.whl", hash = "sha256:bf57215e19dbfa4fe7edae53040cc1deef825e3b1605cca9a8d2c2fadd2328a0"}, + {file = "fsspec-2023.3.0.tar.gz", hash = "sha256:24e635549a590d74c6c18274ddd3ffab4753341753e923408b1904eaabafe04d"}, +] + +[package.dependencies] +aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""} +requests = {version = "*", optional = true, markers = "extra == \"http\""} + +[package.extras] +abfs = ["adlfs"] +adl = ["adlfs"] +arrow = ["pyarrow (>=1)"] +dask = ["dask", "distributed"] +dropbox = ["dropbox", "dropboxdrivefs", "requests"] +fuse = ["fusepy"] +gcs = ["gcsfs"] +git = ["pygit2"] +github = ["requests"] +gs = ["gcsfs"] +gui = ["panel"] +hdfs = ["pyarrow (>=1)"] +http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"] +libarchive = ["libarchive-c"] +oci = ["ocifs"] +s3 = ["s3fs"] +sftp = ["paramiko"] +smb = ["smbprotocol"] +ssh = ["paramiko"] +tqdm = ["tqdm"] + @@ -543,0 +952,98 @@ requirements-deprecated-finder = ["pip-api", "pipreqs"] +[[package]] +name = "joblib" +version = "1.2.0" +description = "Lightweight pipelining with Python functions" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "joblib-1.2.0-py3-none-any.whl", hash = "sha256:091138ed78f800342968c523bdde947e7a305b8594b910a0fea2ab83c3c6d385"}, + {file = "joblib-1.2.0.tar.gz", hash = "sha256:e1cee4a79e4af22881164f218d4311f60074197fb707e082e803b61f6d137018"}, +] + +[[package]] +name = "lazy-loader" +version = "0.2" +description = "lazy_loader" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "lazy_loader-0.2-py3-none-any.whl", hash = "sha256:c35875f815c340f823ce3271ed645045397213f961b40ad0c0d395c3f5218eeb"}, + {file = "lazy_loader-0.2.tar.gz", hash = "sha256:0edc7a5175c400acb108f283749951fefdadedeb00adcec6e88b974a9254f18a"}, +] + +[package.extras] +lint = ["pre-commit (>=3.1)"] +test = ["pytest (>=7.2)", "pytest-cov (>=4.0)"] + +[[package]] +name = "librosa" +version = "0.10.0.post2" +description = "Python module for audio and music processing" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"}, + {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"}, +] + +[package.dependencies] +audioread = ">=2.1.9" +decorator = ">=4.3.0" +joblib = ">=0.14" +lazy-loader = ">=0.1" +msgpack = ">=1.0" +numba = ">=0.51.0" +numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2" +pooch = ">=1.0,<1.7" +scikit-learn = ">=0.20.0" +scipy = ">=1.2.0" +soundfile = ">=0.12.1" +soxr = ">=0.3.2" +typing-extensions = ">=4.1.1" + +[package.extras] +display = ["matplotlib (>=3.3.0)"] +docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"] +tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"] + +[[package]] +name = "llvmlite" +version = "0.39.1" +description = "lightweight wrapper around basic LLVM functionality" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"}, + {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"}, + {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"}, + {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"}, + {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"}, + {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"}, + {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"}, + {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"}, + {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"}, + {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"}, +] + @@ -655 +1161 @@ name = "msgpack" -version = "1.0.4" +version = "1.0.5" @@ -657 +1163 @@ description = "MessagePack serializer" -category = "dev" +category = "main" @@ -661,52 +1167,171 @@ files = [ - {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:4ab251d229d10498e9a2f3b1e68ef64cb393394ec477e3370c457f9430ce9250"}, - {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:112b0f93202d7c0fef0b7810d465fde23c746a2d482e1e2de2aafd2ce1492c88"}, - {file = "msgpack-1.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:002b5c72b6cd9b4bafd790f364b8480e859b4712e91f43014fe01e4f957b8467"}, - {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:35bc0faa494b0f1d851fd29129b2575b2e26d41d177caacd4206d81502d4c6a6"}, - {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4733359808c56d5d7756628736061c432ded018e7a1dff2d35a02439043321aa"}, - {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:eb514ad14edf07a1dbe63761fd30f89ae79b42625731e1ccf5e1f1092950eaa6"}, - {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c23080fdeec4716aede32b4e0ef7e213c7b1093eede9ee010949f2a418ced6ba"}, - {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:49565b0e3d7896d9ea71d9095df15b7f75a035c49be733051c34762ca95bbf7e"}, - {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:aca0f1644d6b5a73eb3e74d4d64d5d8c6c3d577e753a04c9e9c87d07692c58db"}, - {file = "msgpack-1.0.4-cp310-cp310-win32.whl", hash = "sha256:0dfe3947db5fb9ce52aaea6ca28112a170db9eae75adf9339a1aec434dc954ef"}, - {file = "msgpack-1.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dea20515f660aa6b7e964433b1808d098dcfcabbebeaaad240d11f909298075"}, - {file = "msgpack-1.0.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:e83f80a7fec1a62cf4e6c9a660e39c7f878f603737a0cdac8c13131d11d97f52"}, - {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3c11a48cf5e59026ad7cb0dc29e29a01b5a66a3e333dc11c04f7e991fc5510a9"}, - {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1276e8f34e139aeff1c77a3cefb295598b504ac5314d32c8c3d54d24fadb94c9"}, - {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6c9566f2c39ccced0a38d37c26cc3570983b97833c365a6044edef3574a00c08"}, - {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:fcb8a47f43acc113e24e910399376f7277cf8508b27e5b88499f053de6b115a8"}, - {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:76ee788122de3a68a02ed6f3a16bbcd97bc7c2e39bd4d94be2f1821e7c4a64e6"}, - {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0a68d3ac0104e2d3510de90a1091720157c319ceeb90d74f7b5295a6bee51bae"}, - {file = "msgpack-1.0.4-cp36-cp36m-win32.whl", hash = "sha256:85f279d88d8e833ec015650fd15ae5eddce0791e1e8a59165318f371158efec6"}, - {file = "msgpack-1.0.4-cp36-cp36m-win_amd64.whl", hash = "sha256:c1683841cd4fa45ac427c18854c3ec3cd9b681694caf5bff04edb9387602d661"}, - {file = "msgpack-1.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:a75dfb03f8b06f4ab093dafe3ddcc2d633259e6c3f74bb1b01996f5d8aa5868c"}, - {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9667bdfdf523c40d2511f0e98a6c9d3603be6b371ae9a238b7ef2dc4e7a427b0"}, - {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11184bc7e56fd74c00ead4f9cc9a3091d62ecb96e97653add7a879a14b003227"}, - {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ac5bd7901487c4a1dd51a8c58f2632b15d838d07ceedaa5e4c080f7190925bff"}, - {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1e91d641d2bfe91ba4c52039adc5bccf27c335356055825c7f88742c8bb900dd"}, - {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:2a2df1b55a78eb5f5b7d2a4bb221cd8363913830145fad05374a80bf0877cb1e"}, - {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:545e3cf0cf74f3e48b470f68ed19551ae6f9722814ea969305794645da091236"}, - {file = "msgpack-1.0.4-cp37-cp37m-win32.whl", hash = "sha256:2cc5ca2712ac0003bcb625c96368fd08a0f86bbc1a5578802512d87bc592fe44"}, - {file = "msgpack-1.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:eba96145051ccec0ec86611fe9cf693ce55f2a3ce89c06ed307de0e085730ec1"}, - {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:7760f85956c415578c17edb39eed99f9181a48375b0d4a94076d84148cf67b2d"}, - {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:449e57cc1ff18d3b444eb554e44613cffcccb32805d16726a5494038c3b93dab"}, - {file = "msgpack-1.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d603de2b8d2ea3f3bcb2efe286849aa7a81531abc52d8454da12f46235092bcb"}, - {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:48f5d88c99f64c456413d74a975bd605a9b0526293218a3b77220a2c15458ba9"}, - {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6916c78f33602ecf0509cc40379271ba0f9ab572b066bd4bdafd7434dee4bc6e"}, - {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:81fc7ba725464651190b196f3cd848e8553d4d510114a954681fd0b9c479d7e1"}, - {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:d5b5b962221fa2c5d3a7f8133f9abffc114fe218eb4365e40f17732ade576c8e"}, - {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:77ccd2af37f3db0ea59fb280fa2165bf1b096510ba9fe0cc2bf8fa92a22fdb43"}, - {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b17be2478b622939e39b816e0aa8242611cc8d3583d1cd8ec31b249f04623243"}, - {file = "msgpack-1.0.4-cp38-cp38-win32.whl", hash = "sha256:2bb8cdf50dd623392fa75525cce44a65a12a00c98e1e37bf0fb08ddce2ff60d2"}, - {file = "msgpack-1.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:26b8feaca40a90cbe031b03d82b2898bf560027160d3eae1423f4a67654ec5d6"}, - {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:462497af5fd4e0edbb1559c352ad84f6c577ffbbb708566a0abaaa84acd9f3ae"}, - {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2999623886c5c02deefe156e8f869c3b0aaeba14bfc50aa2486a0415178fce55"}, - {file = "msgpack-1.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f0029245c51fd9473dc1aede1160b0a29f4a912e6b1dd353fa6d317085b219da"}, - {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ed6f7b854a823ea44cf94919ba3f727e230da29feb4a99711433f25800cf747f"}, - {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0df96d6eaf45ceca04b3f3b4b111b86b33785683d682c655063ef8057d61fd92"}, - {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6a4192b1ab40f8dca3f2877b70e63799d95c62c068c84dc028b40a6cb03ccd0f"}, - {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:0e3590f9fb9f7fbc36df366267870e77269c03172d086fa76bb4eba8b2b46624"}, - {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:1576bd97527a93c44fa856770197dec00d223b0b9f36ef03f65bac60197cedf8"}, - {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:63e29d6e8c9ca22b21846234913c3466b7e4ee6e422f205a2988083de3b08cae"}, - {file = "msgpack-1.0.4-cp39-cp39-win32.whl", hash = "sha256:fb62ea4b62bfcb0b380d5680f9a4b3f9a2d166d9394e9bbd9666c0ee09a3645c"}, - {file = "msgpack-1.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:4d5834a2a48965a349da1c5a79760d94a1a0172fbb5ab6b5b33cbf8447e109ce"}, - {file = "msgpack-1.0.4.tar.gz", hash = "sha256:f5d869c18f030202eb412f08b28d2afeea553d6613aee89e200d7aca7ef01f5f"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:525228efd79bb831cf6830a732e2e80bc1b05436b086d4264814b4b2955b2fa9"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4f8d8b3bf1ff2672567d6b5c725a1b347fe838b912772aa8ae2bf70338d5a198"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cdc793c50be3f01106245a61b739328f7dccc2c648b501e237f0699fe1395b81"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb47c21a8a65b165ce29f2bec852790cbc04936f502966768e4aae9fa763cb7"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e42b9594cc3bf4d838d67d6ed62b9e59e201862a25e9a157019e171fbe672dd3"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:55b56a24893105dc52c1253649b60f475f36b3aa0fc66115bffafb624d7cb30b"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:1967f6129fc50a43bfe0951c35acbb729be89a55d849fab7686004da85103f1c"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:20a97bf595a232c3ee6d57ddaadd5453d174a52594bf9c21d10407e2a2d9b3bd"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d25dd59bbbbb996eacf7be6b4ad082ed7eacc4e8f3d2df1ba43822da9bfa122a"}, + {file = "msgpack-1.0.5-cp310-cp310-win32.whl", hash = "sha256:382b2c77589331f2cb80b67cc058c00f225e19827dbc818d700f61513ab47bea"}, + {file = "msgpack-1.0.5-cp310-cp310-win_amd64.whl", hash = "sha256:4867aa2df9e2a5fa5f76d7d5565d25ec76e84c106b55509e78c1ede0f152659a"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9f5ae84c5c8a857ec44dc180a8b0cc08238e021f57abdf51a8182e915e6299f0"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:9e6ca5d5699bcd89ae605c150aee83b5321f2115695e741b99618f4856c50898"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5494ea30d517a3576749cad32fa27f7585c65f5f38309c88c6d137877fa28a5a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ab2f3331cb1b54165976a9d976cb251a83183631c88076613c6c780f0d6e45a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:28592e20bbb1620848256ebc105fc420436af59515793ed27d5c77a217477705"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe5c63197c55bce6385d9aee16c4d0641684628f63ace85f73571e65ad1c1e8d"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:ed40e926fa2f297e8a653c954b732f125ef97bdd4c889f243182299de27e2aa9"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b2de4c1c0538dcb7010902a2b97f4e00fc4ddf2c8cda9749af0e594d3b7fa3d7"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf22a83f973b50f9d38e55c6aade04c41ddda19b00c4ebc558930d78eecc64ed"}, + {file = "msgpack-1.0.5-cp311-cp311-win32.whl", hash = "sha256:c396e2cc213d12ce017b686e0f53497f94f8ba2b24799c25d913d46c08ec422c"}, + {file = "msgpack-1.0.5-cp311-cp311-win_amd64.whl", hash = "sha256:6c4c68d87497f66f96d50142a2b73b97972130d93677ce930718f68828b382e2"}, + {file = "msgpack-1.0.5-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:a2b031c2e9b9af485d5e3c4520f4220d74f4d222a5b8dc8c1a3ab9448ca79c57"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f837b93669ce4336e24d08286c38761132bc7ab29782727f8557e1eb21b2080"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1d46dfe3832660f53b13b925d4e0fa1432b00f5f7210eb3ad3bb9a13c6204a6"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:366c9a7b9057e1547f4ad51d8facad8b406bab69c7d72c0eb6f529cf76d4b85f"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4c075728a1095efd0634a7dccb06204919a2f67d1893b6aa8e00497258bf926c"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:f933bbda5a3ee63b8834179096923b094b76f0c7a73c1cfe8f07ad608c58844b"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:36961b0568c36027c76e2ae3ca1132e35123dcec0706c4b7992683cc26c1320c"}, + {file = "msgpack-1.0.5-cp36-cp36m-win32.whl", hash = "sha256:b5ef2f015b95f912c2fcab19c36814963b5463f1fb9049846994b007962743e9"}, + {file = "msgpack-1.0.5-cp36-cp36m-win_amd64.whl", hash = "sha256:288e32b47e67f7b171f86b030e527e302c91bd3f40fd9033483f2cacc37f327a"}, + {file = "msgpack-1.0.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:137850656634abddfb88236008339fdaba3178f4751b28f270d2ebe77a563b6c"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c05a4a96585525916b109bb85f8cb6511db1c6f5b9d9cbcbc940dc6b4be944b"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:56a62ec00b636583e5cb6ad313bbed36bb7ead5fa3a3e38938503142c72cba4f"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef8108f8dedf204bb7b42994abf93882da1159728a2d4c5e82012edd92c9da9f"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1835c84d65f46900920b3708f5ba829fb19b1096c1800ad60bae8418652a951d"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:e57916ef1bd0fee4f21c4600e9d1da352d8816b52a599c46460e93a6e9f17086"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:17358523b85973e5f242ad74aa4712b7ee560715562554aa2134d96e7aa4cbbf"}, + {file = "msgpack-1.0.5-cp37-cp37m-win32.whl", hash = "sha256:cb5aaa8c17760909ec6cb15e744c3ebc2ca8918e727216e79607b7bbce9c8f77"}, + {file = "msgpack-1.0.5-cp37-cp37m-win_amd64.whl", hash = "sha256:ab31e908d8424d55601ad7075e471b7d0140d4d3dd3272daf39c5c19d936bd82"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:b72d0698f86e8d9ddf9442bdedec15b71df3598199ba33322d9711a19f08145c"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:379026812e49258016dd84ad79ac8446922234d498058ae1d415f04b522d5b2d"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:332360ff25469c346a1c5e47cbe2a725517919892eda5cfaffe6046656f0b7bb"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:476a8fe8fae289fdf273d6d2a6cb6e35b5a58541693e8f9f019bfe990a51e4ba"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a9985b214f33311df47e274eb788a5893a761d025e2b92c723ba4c63936b69b1"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:48296af57cdb1d885843afd73c4656be5c76c0c6328db3440c9601a98f303d87"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:addab7e2e1fcc04bd08e4eb631c2a90960c340e40dfc4a5e24d2ff0d5a3b3edb"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:916723458c25dfb77ff07f4c66aed34e47503b2eb3188b3adbec8d8aa6e00f48"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:821c7e677cc6acf0fd3f7ac664c98803827ae6de594a9f99563e48c5a2f27eb0"}, + {file = "msgpack-1.0.5-cp38-cp38-win32.whl", hash = "sha256:1c0f7c47f0087ffda62961d425e4407961a7ffd2aa004c81b9c07d9269512f6e"}, + {file = "msgpack-1.0.5-cp38-cp38-win_amd64.whl", hash = "sha256:bae7de2026cbfe3782c8b78b0db9cbfc5455e079f1937cb0ab8d133496ac55e1"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:20c784e66b613c7f16f632e7b5e8a1651aa5702463d61394671ba07b2fc9e025"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:266fa4202c0eb94d26822d9bfd7af25d1e2c088927fe8de9033d929dd5ba24c5"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:18334484eafc2b1aa47a6d42427da7fa8f2ab3d60b674120bce7a895a0a85bdd"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57e1f3528bd95cc44684beda696f74d3aaa8a5e58c816214b9046512240ef437"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:586d0d636f9a628ddc6a17bfd45aa5b5efaf1606d2b60fa5d87b8986326e933f"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a740fa0e4087a734455f0fc3abf5e746004c9da72fbd541e9b113013c8dc3282"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3055b0455e45810820db1f29d900bf39466df96ddca11dfa6d074fa47054376d"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:a61215eac016f391129a013c9e46f3ab308db5f5ec9f25811e811f96962599a8"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:362d9655cd369b08fda06b6657a303eb7172d5279997abe094512e919cf74b11"}, + {file = "msgpack-1.0.5-cp39-cp39-win32.whl", hash = "sha256:ac9dd47af78cae935901a9a500104e2dea2e253207c924cc95de149606dc43cc"}, + {file = "msgpack-1.0.5-cp39-cp39-win_amd64.whl", hash = "sha256:06f5174b5f8ed0ed919da0e62cbd4ffde676a374aba4020034da05fab67b9164"}, + {file = "msgpack-1.0.5.tar.gz", hash = "sha256:c075544284eadc5cddc70f4757331d99dcbc16b2bbd4849d15f8aae4cf36d31c"}, +] + +[[package]] +name = "multidict" +version = "6.0.4" +description = "multidict implementation" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"}, + {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"}, + {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"}, + {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"}, + {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"}, + {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"}, + {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"}, + {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"}, + {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"}, + {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"}, + {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"}, + {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"}, + {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"}, +] + +[[package]] +name = "multiprocess" +version = "0.70.14" +description = "better multiprocessing and multithreading in python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"}, + {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"}, + {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"}, + {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"}, + {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"}, + {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"}, @@ -714,0 +1340,3 @@ files = [ +[package.dependencies] +dill = ">=0.3.6" + @@ -717 +1345 @@ name = "mypy" -version = "1.0.1" +version = "1.1.1" @@ -723,26 +1351,26 @@ files = [ - {file = "mypy-1.0.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:71a808334d3f41ef011faa5a5cd8153606df5fc0b56de5b2e89566c8093a0c9a"}, - {file = "mypy-1.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:920169f0184215eef19294fa86ea49ffd4635dedfdea2b57e45cb4ee85d5ccaf"}, - {file = "mypy-1.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:27a0f74a298769d9fdc8498fcb4f2beb86f0564bcdb1a37b58cbbe78e55cf8c0"}, - {file = "mypy-1.0.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:65b122a993d9c81ea0bfde7689b3365318a88bde952e4dfa1b3a8b4ac05d168b"}, - {file = "mypy-1.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:5deb252fd42a77add936b463033a59b8e48eb2eaec2976d76b6878d031933fe4"}, - {file = "mypy-1.0.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2013226d17f20468f34feddd6aae4635a55f79626549099354ce641bc7d40262"}, - {file = "mypy-1.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:48525aec92b47baed9b3380371ab8ab6e63a5aab317347dfe9e55e02aaad22e8"}, - {file = "mypy-1.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c96b8a0c019fe29040d520d9257d8c8f122a7343a8307bf8d6d4a43f5c5bfcc8"}, - {file = "mypy-1.0.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:448de661536d270ce04f2d7dddaa49b2fdba6e3bd8a83212164d4174ff43aa65"}, - {file = "mypy-1.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:d42a98e76070a365a1d1c220fcac8aa4ada12ae0db679cb4d910fabefc88b994"}, - {file = "mypy-1.0.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e64f48c6176e243ad015e995de05af7f22bbe370dbb5b32bd6988438ec873919"}, - {file = "mypy-1.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5fdd63e4f50e3538617887e9aee91855368d9fc1dea30da743837b0df7373bc4"}, - {file = "mypy-1.0.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:dbeb24514c4acbc78d205f85dd0e800f34062efcc1f4a4857c57e4b4b8712bff"}, - {file = "mypy-1.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a2948c40a7dd46c1c33765718936669dc1f628f134013b02ff5ac6c7ef6942bf"}, - {file = "mypy-1.0.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5bc8d6bd3b274dd3846597855d96d38d947aedba18776aa998a8d46fabdaed76"}, - {file = "mypy-1.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:17455cda53eeee0a4adb6371a21dd3dbf465897de82843751cf822605d152c8c"}, - {file = "mypy-1.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e831662208055b006eef68392a768ff83596035ffd6d846786578ba1714ba8f6"}, - {file = "mypy-1.0.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:e60d0b09f62ae97a94605c3f73fd952395286cf3e3b9e7b97f60b01ddfbbda88"}, - {file = "mypy-1.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:0af4f0e20706aadf4e6f8f8dc5ab739089146b83fd53cb4a7e0e850ef3de0bb6"}, - {file = "mypy-1.0.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:24189f23dc66f83b839bd1cce2dfc356020dfc9a8bae03978477b15be61b062e"}, - {file = "mypy-1.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:93a85495fb13dc484251b4c1fd7a5ac370cd0d812bbfc3b39c1bafefe95275d5"}, - {file = "mypy-1.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f546ac34093c6ce33f6278f7c88f0f147a4849386d3bf3ae193702f4fe31407"}, - {file = "mypy-1.0.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c6c2ccb7af7154673c591189c3687b013122c5a891bb5651eca3db8e6c6c55bd"}, - {file = "mypy-1.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:15b5a824b58c7c822c51bc66308e759243c32631896743f030daf449fe3677f3"}, - {file = "mypy-1.0.1-py3-none-any.whl", hash = "sha256:eda5c8b9949ed411ff752b9a01adda31afe7eae1e53e946dbdf9db23865e66c4"}, - {file = "mypy-1.0.1.tar.gz", hash = "sha256:28cea5a6392bb43d266782983b5a4216c25544cd7d80be681a155ddcdafd152d"}, + {file = "mypy-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39c7119335be05630611ee798cc982623b9e8f0cff04a0b48dfc26100e0b97af"}, + {file = "mypy-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:61bf08362e93b6b12fad3eab68c4ea903a077b87c90ac06c11e3d7a09b56b9c1"}, + {file = "mypy-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dbb19c9f662e41e474e0cff502b7064a7edc6764f5262b6cd91d698163196799"}, + {file = "mypy-1.1.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:315ac73cc1cce4771c27d426b7ea558fb4e2836f89cb0296cbe056894e3a1f78"}, + {file = "mypy-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:5cb14ff9919b7df3538590fc4d4c49a0f84392237cbf5f7a816b4161c061829e"}, + {file = "mypy-1.1.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:26cdd6a22b9b40b2fd71881a8a4f34b4d7914c679f154f43385ca878a8297389"}, + {file = "mypy-1.1.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5b5f81b40d94c785f288948c16e1f2da37203c6006546c5d947aab6f90aefef2"}, + {file = "mypy-1.1.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21b437be1c02712a605591e1ed1d858aba681757a1e55fe678a15c2244cd68a5"}, + {file = "mypy-1.1.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d809f88734f44a0d44959d795b1e6f64b2bbe0ea4d9cc4776aa588bb4229fc1c"}, + {file = "mypy-1.1.1-cp311-cp311-win_amd64.whl", hash = "sha256:a380c041db500e1410bb5b16b3c1c35e61e773a5c3517926b81dfdab7582be54"}, + {file = "mypy-1.1.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:b7c7b708fe9a871a96626d61912e3f4ddd365bf7f39128362bc50cbd74a634d5"}, + {file = "mypy-1.1.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c1c10fa12df1232c936830839e2e935d090fc9ee315744ac33b8a32216b93707"}, + {file = "mypy-1.1.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:0a28a76785bf57655a8ea5eb0540a15b0e781c807b5aa798bd463779988fa1d5"}, + {file = "mypy-1.1.1-cp37-cp37m-win_amd64.whl", hash = "sha256:ef6a01e563ec6a4940784c574d33f6ac1943864634517984471642908b30b6f7"}, + {file = "mypy-1.1.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d64c28e03ce40d5303450f547e07418c64c241669ab20610f273c9e6290b4b0b"}, + {file = "mypy-1.1.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:64cc3afb3e9e71a79d06e3ed24bb508a6d66f782aff7e56f628bf35ba2e0ba51"}, + {file = "mypy-1.1.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce61663faf7a8e5ec6f456857bfbcec2901fbdb3ad958b778403f63b9e606a1b"}, + {file = "mypy-1.1.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:2b0c373d071593deefbcdd87ec8db91ea13bd8f1328d44947e88beae21e8d5e9"}, + {file = "mypy-1.1.1-cp38-cp38-win_amd64.whl", hash = "sha256:2888ce4fe5aae5a673386fa232473014056967f3904f5abfcf6367b5af1f612a"}, + {file = "mypy-1.1.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:19ba15f9627a5723e522d007fe708007bae52b93faab00f95d72f03e1afa9598"}, + {file = "mypy-1.1.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:59bbd71e5c58eed2e992ce6523180e03c221dcd92b52f0e792f291d67b15a71c"}, + {file = "mypy-1.1.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9401e33814cec6aec8c03a9548e9385e0e228fc1b8b0a37b9ea21038e64cdd8a"}, + {file = "mypy-1.1.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4b398d8b1f4fba0e3c6463e02f8ad3346f71956b92287af22c9b12c3ec965a9f"}, + {file = "mypy-1.1.1-cp39-cp39-win_amd64.whl", hash = "sha256:69b35d1dcb5707382810765ed34da9db47e7f95b3528334a3c999b0c90fe523f"}, + {file = "mypy-1.1.1-py3-none-any.whl", hash = "sha256:4e4e8b362cdf99ba00c2b218036002bdcdf1e0de085cdb296a49df03fb31dfc4"}, + {file = "mypy-1.1.1.tar.gz", hash = "sha256:ae9ceae0f5b9059f33dbc62dea087e942c0ccab4b7a003719cb70f9b8abfa32f"}, @@ -752 +1380 @@ files = [ -mypy-extensions = ">=0.4.3" +mypy-extensions = ">=1.0.0" @@ -792,0 +1421,81 @@ test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] +[[package]] +name = "numba" +version = "0.56.4" +description = "compiling Python code using LLVM" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"}, + {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"}, + {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"}, + {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"}, + {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"}, + {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"}, + {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"}, + {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"}, + {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"}, + {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"}, + {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"}, + {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"}, + {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"}, + {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"}, + {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"}, + {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"}, +] + +[package.dependencies] +llvmlite = ">=0.39.0dev0,<0.40" +numpy = ">=1.18,<1.24" +setuptools = "*" + +[[package]] +name = "numpy" +version = "1.23.5" +description = "NumPy is the fundamental package for array computing with Python." +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"}, + {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"}, + {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"}, + {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"}, + {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"}, + {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"}, + {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"}, + {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"}, + {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"}, + {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"}, +] + @@ -795 +1504 @@ name = "orjson" -version = "3.8.6" +version = "3.8.8" @@ -801,44 +1510,49 @@ files = [ - {file = "orjson-3.8.6-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:062a9a74c10c439acc35cf67f31ac88d9464a11025700bab421e6cdf54a54a35"}, - {file = "orjson-3.8.6-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:692c255109867cc8211267f4416d2915845273bf4f403bbca5419f5b15ac9175"}, - {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4a20905c7a5ebc280343704c4dd19343ef966c9dea5a38ade6e0461a6deb8eda"}, - {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:34ce4a8b8f0fea483bce6985c015953f475540b7d756efd48a571b1803c318ee"}, - {file = "orjson-3.8.6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e57ecad7616ec842d8c382ed42a778cdcdadc67cfb46b804b43079f937b63b31"}, - {file = "orjson-3.8.6-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:323065cf14fdd4096dbf93ea1634e7e030044af8c1000803bcdc132fbfd395f5"}, - {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:4cb4f37fca8cf8309de421634447437f229bc03b240cec8ad4ac241fd4b1bcf4"}, - {file = "orjson-3.8.6-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:32353b14c5e0b55b6a8759e993482a2d8c44d492489840718b74658de67671e2"}, - {file = "orjson-3.8.6-cp310-none-win_amd64.whl", hash = "sha256:3e44f78db3a15902b5e8386119979691ed3dd61d1ded10bad2c7106fd50641ef"}, - {file = "orjson-3.8.6-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:c59ec129d523abd4f2d65c0733d0e57af7dd09c69142f1aa564b04358f04ace3"}, - {file = "orjson-3.8.6-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:d44d89314a66e98e690ce64c8771d963eb64ae6cea662d0a1d077ed024627228"}, - {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:865ef341c4d310ac2689bf811dbc0930b2f13272f8eade1511dc40b186f6d562"}, - {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:52809a37a0daa6992835ee0625aca22b4c0693dba3cb465948e6c9796de927b0"}, - {file = "orjson-3.8.6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d7402121d06d11fafcaed7d06f9d68b11bbe39868e0e1bc19239ee5b6b98b2b"}, - {file = "orjson-3.8.6-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:583338b7dabb509ca4c3b4f160f58a5228bf6c6e0f8a2981663f683791f39d45"}, - {file = "orjson-3.8.6-cp311-none-win_amd64.whl", hash = "sha256:4a6c0a0ef2f535ba7a5d01f014b53d05eeb372d43556edb25c75a4d52690a123"}, - {file = "orjson-3.8.6-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:9d35573e7f5817a26d8ce1134c3463d31bc3b39aad3ad7ae06bb67d6078fa9c0"}, - {file = "orjson-3.8.6-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:94d8fdc12adc0450994931d722cb38be5e4caa273219881abb96c15a9e9f151f"}, - {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8fc43bfb73d394b9bf12062cd6dab72abf728ac7869f972e4bb7327fd3330b8"}, - {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a38387387139695a7e52b9f568e39c1632b22eb34939afc5efed265fa8277b84"}, - {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8e048c6df7453c3da4de10fa5c44f6c655b157b712628888ce880cd5bbf30013"}, - {file = "orjson-3.8.6-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:d3b0950d792b25c0aa52505faf09237fd98136d09616a0837f7cdb0fde9e2730"}, - {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:38bc8a388080d8fd297388bcff4939e350ffafe4a006567e0dd81cdb8c7b86fa"}, - {file = "orjson-3.8.6-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5b3251ab7113f2400d76f2b4a2d6592e7d5a5cf45fa948c894340553671ef8f1"}, - {file = "orjson-3.8.6-cp37-none-win_amd64.whl", hash = "sha256:2c83a33cf389fd286bd9ef0befc406307444b9553d2e9ba14b90b9332524cfa6"}, - {file = "orjson-3.8.6-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:53f51c23398cfe818d9bb09079d31a60c6cd77e7eee1d555cfcc735460db4190"}, - {file = "orjson-3.8.6-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:6190e23a2fb9fc78228b289b3ec295094671ca0299319c8c72727aa9e7dbe06f"}, - {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:61fff8a8b4cd4e489b291fe5105b6138b1831490f1a0dc726d5e17ebe811d595"}, - {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:c192813f527f886bd85abc5a9e8d9dde16ffa06d7305de526a7c4657730dbf4e"}, - {file = "orjson-3.8.6-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aae1487fba9d955b2679f0a697665ed8fc32563b3252acc240e097184c184e29"}, - {file = "orjson-3.8.6-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:cd2bd48e9a14f2130790a3c2dcb897bd93c2e5c244919799430a6d9b8212cb50"}, - {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:006178fd654a0a4f14f5912b8320ba9a26ab9c0ae7ce1c7eeb4b5249d6cada29"}, - {file = "orjson-3.8.6-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:9d5ad2fddccc89ab64b6333823b250ce8430fc51f014954e5a2d4c933f5deb9f"}, - {file = "orjson-3.8.6-cp38-none-win_amd64.whl", hash = "sha256:aef3d558f5bd809733ebf2cbce7e1338ce62812db317478427236b97036aba0f"}, - {file = "orjson-3.8.6-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:7d216a5f3d23eac2c7c654e7bd30280c27cdf5edc32325e6ad8e880d36c265b7"}, - {file = "orjson-3.8.6-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:004122c95e08db7201b80224de3a8f2ad79b9717040e6884c6015f27b010127d"}, - {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:006c492577ad046cb7e50237a8d8935131a35f7e7f8320fbc3514da6fbc0b436"}, - {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67554103b415349b6ee2db82d2422da1c8f4c2d280d20772217f6d1d227410b6"}, - {file = "orjson-3.8.6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:caa5053f19584816f063c887d94385db481fc01d995d6a717ce4fbb929653ec2"}, - {file = "orjson-3.8.6-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:2bdd64566870a8a0bdcf8c7df2f4452391dd55070f5cd98cc581914e8c263d85"}, - {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:550a4dec128d1adfd0262ef9ad7878d62d1cc0bddaaa05e41d8ca28414dc86bc"}, - {file = "orjson-3.8.6-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3f5ad9442e8a99fb436279a8614a00aca272ea8dabb692cadee70a4874d6e03"}, - {file = "orjson-3.8.6-cp39-none-win_amd64.whl", hash = "sha256:aa7b112e3273d1744f7bc983ffd3dd0d004062c69dfa68e119515a7e115c46c8"}, - {file = "orjson-3.8.6.tar.gz", hash = "sha256:91ef8a554d33fbc5bb61c3972f3e8baa994f72c4967671e64e7dac1cc06f50e1"}, + {file = "orjson-3.8.8-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:18fcdea75d8b571dc9b185652b81397b62878ae7934fd62e6a0103a5b8448e34"}, + {file = "orjson-3.8.8-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:306618884929b596e2e083f82b5617da812df25b0c467542371f1d51f0c5a6f5"}, + {file = "orjson-3.8.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:edc65ddb6ae6f8fbb2bbf78ac98f75b729c9eeb0776d5508dd76d3a948dda1dd"}, + {file = "orjson-3.8.8-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:e6a6d55e01bce74516dff15302627a13b1f4edcb1c3942dd660978dee423ccf2"}, + {file = "orjson-3.8.8-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:28075c4b502d792fb6703e983d456b2a30d5d6f332d26092eb312dc782e64c64"}, + {file = "orjson-3.8.8-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eda4c37e48ff549763183a1549c10eec6ea40439520b17d09359cd74a425069"}, + {file = "orjson-3.8.8-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a3eac485a15493164867729f44e1e1247b3094ff19d37708e8cdc9c88a93c623"}, + {file = "orjson-3.8.8-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:88bf40e5468444c04374d1b8f1877cebbaef6bb7406cb6b4a34a570c5cbb87bc"}, + {file = "orjson-3.8.8-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:747bd4e09d8aa61e1ff677a7dd1cffd28a5d13c22f3769123c58ec988bf1b83d"}, + {file = "orjson-3.8.8-cp310-none-win_amd64.whl", hash = "sha256:dd7d86c5f5f820ac9d4783477e86eb984b63bdb32359935609eb33cf65049c54"}, + {file = "orjson-3.8.8-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:52293a6097750c2d434737966fe6e2a1ed489ac70cc8e584f5944af83de0b787"}, + {file = "orjson-3.8.8-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:9322450f392dceb49810d2f820b1932af22d66f67f1d45c31f160067dd06359f"}, + {file = "orjson-3.8.8-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:68d59e3ae84a9b6f14b45a89f7fde4a08a87ea5eb76bfc854b354640de8156f5"}, + {file = "orjson-3.8.8-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:022347dad2253081eaa25366834bb8b06a5aceb0e83b39c6b0aa865759e49d69"}, + {file = "orjson-3.8.8-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddfcc54793e266056fe1c257d0804c336bca1c5c1ee7979d674e1fc19cfb0a6a"}, + {file = "orjson-3.8.8-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:449d8ed1e0e6b24e9df5a06b59fd66ea7f7293e141257069601ae8ff9fad705c"}, + {file = "orjson-3.8.8-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:0204bc414bc6f7a595211569840b422d96649fd8686efa1fbbcb12eed5dd9521"}, + {file = "orjson-3.8.8-cp311-none-win_amd64.whl", hash = "sha256:e991a5c2c5f2f299c77e1d07ef2812ff5b68e1d97a2aab01aca29cf756473aa3"}, + {file = "orjson-3.8.8-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:2006d9c046bbf335c951f61e016a27bd4f17323dd116f601e4a8a11739cd0a62"}, + {file = "orjson-3.8.8-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:4553d85bad4cbd634a40b7b5d36daaa197a6025f9ce3e2165b371e528759093d"}, + {file = "orjson-3.8.8-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57ee45d2cc6c11c50afb5a0c09d7cd559aea76c77250dbe996be6a03464d4a50"}, + {file = "orjson-3.8.8-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:02f5b5db1e424706eb9f70f1c25699ff4cef16fadfc64af5b70f8628eafe4771"}, + {file = "orjson-3.8.8-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4d7c9f3b1598a1ccd806ef02257a76a00c7ede09662ddb54eec2b4bd92874254"}, + {file = "orjson-3.8.8-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b90d171932b6a9d50e79fa2762cb303e3556bbf25c08bb316fe346ec58af9c19"}, + {file = "orjson-3.8.8-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:28dfe774c345130f1117c4d023644ec52d9d50e3eaadb9bd1c668d91dc109bb5"}, + {file = "orjson-3.8.8-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:8f84116fcc3714e7ba3cbeb1b11ac5e4549e7d2726c50142f8299fff9dea7d53"}, + {file = "orjson-3.8.8-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:f989f8580db86166aaaa938ccd1597ba1817e3f5df14c047baafe783e3d24173"}, + {file = "orjson-3.8.8-cp37-none-win_amd64.whl", hash = "sha256:66045850f286090800a18662d81d44f88c3fcb60ea3a9947d5caeab5d1efc92e"}, + {file = "orjson-3.8.8-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:2c2c5f3d3bbd61dba646e2b9c54a0dd7941b03fba49726bd31c1c23fedf0b9aa"}, + {file = "orjson-3.8.8-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:9cb36d4a14f3a911369219d5abc19b907bc41ed2730f7bfe0847b0fd3e834c87"}, + {file = "orjson-3.8.8-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:343124f84da0a33c83ee106a98b3e3c42767c88323d4a2809683cbe83816e8be"}, + {file = "orjson-3.8.8-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:24ad122d8dd057acf2a9965a2ffc1bc12fb310ae1cfe2912db930cbb9ef7eaba"}, + {file = "orjson-3.8.8-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c2f28a92a9bcb4e8635524b20db1b539bda8613872f306b36cdfd9d3577d03ac"}, + {file = "orjson-3.8.8-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:81d3c5b253eebfc4a61cea1f255a576cb2b889afa99f4510f30ec13201d4f457"}, + {file = "orjson-3.8.8-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:317164f7d4c0540a6eb8b0a0faeec84ef011d359da05188423db762b65f84e1d"}, + {file = "orjson-3.8.8-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:5e7e39357371d4ae5649f33c01886508a4c8e5fa5c7344554af041dc0f004c01"}, + {file = "orjson-3.8.8-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:60fefd4bbd796b4296f478e705fe2c2c7defd28da98d3017743eb87c3238a380"}, + {file = "orjson-3.8.8-cp38-none-win_amd64.whl", hash = "sha256:0dc4a52f1087baeec6b58248fd6b01f17c124fb99f6f770596851ea434a7be0b"}, + {file = "orjson-3.8.8-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:a6bcb449537a99f55c5f05187bac00b4549a795e89c10dcca0d7629548852357"}, + {file = "orjson-3.8.8-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:9c98dc791aa44268ba7f6e21124cf885c813b155316c6bf257560571d243fe15"}, + {file = "orjson-3.8.8-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1b2abf93b727a6af7c5ec8816168cbdff39c716af18ced425dd50ae46d69765c"}, + {file = "orjson-3.8.8-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:23447d38375a19d57975d4e32d9ce9f533803c197fd4292e10d3234c052037a8"}, + {file = "orjson-3.8.8-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4c2e19d2b46cc93c7218bf8180807bf922ff61dc9883458a06edc66d22970fff"}, + {file = "orjson-3.8.8-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e480d74d7bf415e6548a364669404119a85dbe0e3c6cd5f7cb4c7003eac20164"}, + {file = "orjson-3.8.8-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:8e0bff5656b99dd975cae2e5230b39e5909d06c0692fd1f6f06dc46f1fe705d0"}, + {file = "orjson-3.8.8-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:56bb6eb7a254eec3b15feba9b20f4172ccbe6ea50a54cf66cbc8e1e4a19585c2"}, + {file = "orjson-3.8.8-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:1af1cfad5d90b68e15fd625c889c4f9f91d7a88f49512cdb89f01c3881e0c9d9"}, + {file = "orjson-3.8.8-cp39-none-win_amd64.whl", hash = "sha256:d5514dfe200356a1d5a6039e00dca78d87d063f3da1eb6a371253e5a8b7ab5b0"}, + {file = "orjson-3.8.8.tar.gz", hash = "sha256:c096d7a523bae6ffb9c4a228ba4691d66113f0f2231579dc945523fbef09c6da"}, @@ -849 +1563 @@ name = "packageurl-python" -version = "0.10.4" +version = "0.11.1" @@ -853 +1567 @@ optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" @@ -855,2 +1569,2 @@ files = [ - {file = "packageurl-python-0.10.4.tar.gz", hash = "sha256:5c91334f942cd55d45eb0c67dd339a535ef90e25f05b9ec016ad188ed0ef9048"}, - {file = "packageurl_python-0.10.4-py3-none-any.whl", hash = "sha256:bf8a1ffe755634776f6563904d792fb0aa13b377fc86115c36fe17f69b6e59db"}, + {file = "packageurl-python-0.11.1.tar.gz", hash = "sha256:bbcc53d2cb5920c815c1626c75992f319bfc450b73893fa7bd8aac5869aa49fe"}, + {file = "packageurl_python-0.11.1-py3-none-any.whl", hash = "sha256:4bad1d3ea4feb5e7a1db5ca8fb690ac9c82ab18e08d500755947b853df68817d"}, @@ -861 +1575,2 @@ build = ["wheel"] -test = ["black", "isort", "pytest"] +lint = ["black", "isort", "mypy"] +test = ["pytest"] @@ -874,0 +1590,60 @@ files = [ +[[package]] +name = "pandas" +version = "1.5.3" +description = "Powerful data structures for data analysis, time series, and statistics" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "pandas-1.5.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:3749077d86e3a2f0ed51367f30bf5b82e131cc0f14260c4d3e499186fccc4406"}, + {file = "pandas-1.5.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:972d8a45395f2a2d26733eb8d0f629b2f90bebe8e8eddbb8829b180c09639572"}, + {file = "pandas-1.5.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:50869a35cbb0f2e0cd5ec04b191e7b12ed688874bd05dd777c19b28cbea90996"}, + {file = "pandas-1.5.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c3ac844a0fe00bfaeb2c9b51ab1424e5c8744f89860b138434a363b1f620f354"}, + {file = "pandas-1.5.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7a0a56cef15fd1586726dace5616db75ebcfec9179a3a55e78f72c5639fa2a23"}, + {file = "pandas-1.5.3-cp310-cp310-win_amd64.whl", hash = "sha256:478ff646ca42b20376e4ed3fa2e8d7341e8a63105586efe54fa2508ee087f328"}, + {file = "pandas-1.5.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:6973549c01ca91ec96199e940495219c887ea815b2083722821f1d7abfa2b4dc"}, + {file = "pandas-1.5.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:c39a8da13cede5adcd3be1182883aea1c925476f4e84b2807a46e2775306305d"}, + {file = "pandas-1.5.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f76d097d12c82a535fda9dfe5e8dd4127952b45fea9b0276cb30cca5ea313fbc"}, + {file = "pandas-1.5.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e474390e60ed609cec869b0da796ad94f420bb057d86784191eefc62b65819ae"}, + {file = "pandas-1.5.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f2b952406a1588ad4cad5b3f55f520e82e902388a6d5a4a91baa8d38d23c7f6"}, + {file = "pandas-1.5.3-cp311-cp311-win_amd64.whl", hash = "sha256:bc4c368f42b551bf72fac35c5128963a171b40dce866fb066540eeaf46faa003"}, + {file = "pandas-1.5.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:14e45300521902689a81f3f41386dc86f19b8ba8dd5ac5a3c7010ef8d2932813"}, + {file = "pandas-1.5.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9842b6f4b8479e41968eced654487258ed81df7d1c9b7b870ceea24ed9459b31"}, + {file = "pandas-1.5.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:26d9c71772c7afb9d5046e6e9cf42d83dd147b5cf5bcb9d97252077118543792"}, + {file = "pandas-1.5.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5fbcb19d6fceb9e946b3e23258757c7b225ba450990d9ed63ccceeb8cae609f7"}, + {file = "pandas-1.5.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:565fa34a5434d38e9d250af3c12ff931abaf88050551d9fbcdfafca50d62babf"}, + {file = "pandas-1.5.3-cp38-cp38-win32.whl", hash = "sha256:87bd9c03da1ac870a6d2c8902a0e1fd4267ca00f13bc494c9e5a9020920e1d51"}, + {file = "pandas-1.5.3-cp38-cp38-win_amd64.whl", hash = "sha256:41179ce559943d83a9b4bbacb736b04c928b095b5f25dd2b7389eda08f46f373"}, + {file = "pandas-1.5.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:c74a62747864ed568f5a82a49a23a8d7fe171d0c69038b38cedf0976831296fa"}, + {file = "pandas-1.5.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c4c00e0b0597c8e4f59e8d461f797e5d70b4d025880516a8261b2817c47759ee"}, + {file = "pandas-1.5.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a50d9a4336a9621cab7b8eb3fb11adb82de58f9b91d84c2cd526576b881a0c5a"}, + {file = "pandas-1.5.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:dd05f7783b3274aa206a1af06f0ceed3f9b412cf665b7247eacd83be41cf7bf0"}, + {file = "pandas-1.5.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f69c4029613de47816b1bb30ff5ac778686688751a5e9c99ad8c7031f6508e5"}, + {file = "pandas-1.5.3-cp39-cp39-win32.whl", hash = "sha256:7cec0bee9f294e5de5bbfc14d0573f65526071029d036b753ee6507d2a21480a"}, + {file = "pandas-1.5.3-cp39-cp39-win_amd64.whl", hash = "sha256:dfd681c5dc216037e0b0a2c821f5ed99ba9f03ebcf119c7dac0e9a7b960b9ec9"}, + {file = "pandas-1.5.3.tar.gz", hash = "sha256:74a3fd7e5a7ec052f183273dc7b0acd3a863edf7520f5d3a1765c04ffdb3b0b1"}, +] + +[package.dependencies] +numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""} +python-dateutil = ">=2.8.1" +pytz = ">=2020.1" + +[package.extras] +test = ["hypothesis (>=5.5.3)", "pytest (>=6.0)", "pytest-xdist (>=1.31)"] + +[[package]] +name = "pandas-stubs" +version = "1.5.3.230321" +description = "Type annotations for pandas" +category = "dev" +optional = false +python-versions = ">=3.8,<3.12" +files = [ + {file = "pandas_stubs-1.5.3.230321-py3-none-any.whl", hash = "sha256:4bf36b3071dd55f0e558ac8efe07676a120f2ed89e7a3df0fb78ddf2733bf247"}, + {file = "pandas_stubs-1.5.3.230321.tar.gz", hash = "sha256:2fa860df9e6058e9f0d2c09bc711c09abb8f0516eee7f0b9f9950d29b835fc6f"}, +] + +[package.dependencies] +types-pytz = ">=2022.1.1" + @@ -877 +1652 @@ name = "pathspec" -version = "0.11.0" +version = "0.11.1" @@ -883,2 +1658,2 @@ files = [ - {file = "pathspec-0.11.0-py3-none-any.whl", hash = "sha256:3a66eb970cbac598f9e5ccb5b2cf58930cd8e3ed86d393d541eaf2d8b1705229"}, - {file = "pathspec-0.11.0.tar.gz", hash = "sha256:64d338d4e0914e91c1792321e6907b5a593f1ab1851de7fc269557a21b30ebbc"}, + {file = "pathspec-0.11.1-py3-none-any.whl", hash = "sha256:d8af70af76652554bd134c22b3e8a1cc46ed7d91edcdd721ef1a0c51a84a5293"}, + {file = "pathspec-0.11.1.tar.gz", hash = "sha256:2798de800fa92780e33acca925945e9a19a133b715067cf165b8866c15a31687"}, @@ -898,0 +1674,91 @@ files = [ +[[package]] +name = "pillow" +version = "9.4.0" +description = "Python Imaging Library (Fork)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pillow-9.4.0-1-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:1b4b4e9dda4f4e4c4e6896f93e84a8f0bcca3b059de9ddf67dac3c334b1195e1"}, + {file = "Pillow-9.4.0-1-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:fb5c1ad6bad98c57482236a21bf985ab0ef42bd51f7ad4e4538e89a997624e12"}, + {file = "Pillow-9.4.0-1-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:f0caf4a5dcf610d96c3bd32932bfac8aee61c96e60481c2a0ea58da435e25acd"}, + {file = "Pillow-9.4.0-1-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:3f4cc516e0b264c8d4ccd6b6cbc69a07c6d582d8337df79be1e15a5056b258c9"}, + {file = "Pillow-9.4.0-1-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:b8c2f6eb0df979ee99433d8b3f6d193d9590f735cf12274c108bd954e30ca858"}, + {file = "Pillow-9.4.0-1-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:b70756ec9417c34e097f987b4d8c510975216ad26ba6e57ccb53bc758f490dab"}, + {file = "Pillow-9.4.0-1-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:43521ce2c4b865d385e78579a082b6ad1166ebed2b1a2293c3be1d68dd7ca3b9"}, + {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"}, + {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"}, + {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"}, + {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"}, + {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"}, + {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"}, + {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"}, + {file = "Pillow-9.4.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:2968c58feca624bb6c8502f9564dd187d0e1389964898f5e9e1fbc8533169157"}, + {file = "Pillow-9.4.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c5c1362c14aee73f50143d74389b2c158707b4abce2cb055b7ad37ce60738d47"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bd752c5ff1b4a870b7661234694f24b1d2b9076b8bf337321a814c612665f343"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:9a3049a10261d7f2b6514d35bbb7a4dfc3ece4c4de14ef5876c4b7a23a0e566d"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16a8df99701f9095bea8a6c4b3197da105df6f74e6176c5b410bc2df2fd29a57"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:94cdff45173b1919350601f82d61365e792895e3c3a3443cf99819e6fbf717a5"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:ed3e4b4e1e6de75fdc16d3259098de7c6571b1a6cc863b1a49e7d3d53e036070"}, + {file = "Pillow-9.4.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d5b2f8a31bd43e0f18172d8ac82347c8f37ef3e0b414431157718aa234991b28"}, + {file = "Pillow-9.4.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:09b89ddc95c248ee788328528e6a2996e09eaccddeeb82a5356e92645733be35"}, + {file = "Pillow-9.4.0-cp310-cp310-win32.whl", hash = "sha256:f09598b416ba39a8f489c124447b007fe865f786a89dbfa48bb5cf395693132a"}, + {file = "Pillow-9.4.0-cp310-cp310-win_amd64.whl", hash = "sha256:f6e78171be3fb7941f9910ea15b4b14ec27725865a73c15277bc39f5ca4f8391"}, + {file = "Pillow-9.4.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:3fa1284762aacca6dc97474ee9c16f83990b8eeb6697f2ba17140d54b453e133"}, + {file = "Pillow-9.4.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:eaef5d2de3c7e9b21f1e762f289d17b726c2239a42b11e25446abf82b26ac132"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4dfdae195335abb4e89cc9762b2edc524f3c6e80d647a9a81bf81e17e3fb6f0"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6abfb51a82e919e3933eb137e17c4ae9c0475a25508ea88993bb59faf82f3b35"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:451f10ef963918e65b8869e17d67db5e2f4ab40e716ee6ce7129b0cde2876eab"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:6663977496d616b618b6cfa43ec86e479ee62b942e1da76a2c3daa1c75933ef4"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:60e7da3a3ad1812c128750fc1bc14a7ceeb8d29f77e0a2356a8fb2aa8925287d"}, + {file = "Pillow-9.4.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:19005a8e58b7c1796bc0167862b1f54a64d3b44ee5d48152b06bb861458bc0f8"}, + {file = "Pillow-9.4.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f715c32e774a60a337b2bb8ad9839b4abf75b267a0f18806f6f4f5f1688c4b5a"}, + {file = "Pillow-9.4.0-cp311-cp311-win32.whl", hash = "sha256:b222090c455d6d1a64e6b7bb5f4035c4dff479e22455c9eaa1bdd4c75b52c80c"}, + {file = "Pillow-9.4.0-cp311-cp311-win_amd64.whl", hash = "sha256:ba6612b6548220ff5e9df85261bddc811a057b0b465a1226b39bfb8550616aee"}, + {file = "Pillow-9.4.0-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:5f532a2ad4d174eb73494e7397988e22bf427f91acc8e6ebf5bb10597b49c493"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5dd5a9c3091a0f414a963d427f920368e2b6a4c2f7527fdd82cde8ef0bc7a327"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef21af928e807f10bf4141cad4746eee692a0dd3ff56cfb25fce076ec3cc8abe"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:847b114580c5cc9ebaf216dd8c8dbc6b00a3b7ab0131e173d7120e6deade1f57"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_28_aarch64.whl", hash = "sha256:653d7fb2df65efefbcbf81ef5fe5e5be931f1ee4332c2893ca638c9b11a409c4"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:46f39cab8bbf4a384ba7cb0bc8bae7b7062b6a11cfac1ca4bc144dea90d4a9f5"}, + {file = "Pillow-9.4.0-cp37-cp37m-win32.whl", hash = "sha256:7ac7594397698f77bce84382929747130765f66406dc2cd8b4ab4da68ade4c6e"}, + {file = "Pillow-9.4.0-cp37-cp37m-win_amd64.whl", hash = "sha256:46c259e87199041583658457372a183636ae8cd56dbf3f0755e0f376a7f9d0e6"}, + {file = "Pillow-9.4.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:0e51f608da093e5d9038c592b5b575cadc12fd748af1479b5e858045fff955a9"}, + {file = "Pillow-9.4.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:765cb54c0b8724a7c12c55146ae4647e0274a839fb6de7bcba841e04298e1011"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:519e14e2c49fcf7616d6d2cfc5c70adae95682ae20f0395e9280db85e8d6c4df"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d197df5489004db87d90b918033edbeee0bd6df3848a204bca3ff0a903bef837"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0845adc64fe9886db00f5ab68c4a8cd933ab749a87747555cec1c95acea64b0b"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:e1339790c083c5a4de48f688b4841f18df839eb3c9584a770cbd818b33e26d5d"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:a96e6e23f2b79433390273eaf8cc94fec9c6370842e577ab10dabdcc7ea0a66b"}, + {file = "Pillow-9.4.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:7cfc287da09f9d2a7ec146ee4d72d6ea1342e770d975e49a8621bf54eaa8f30f"}, + {file = "Pillow-9.4.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:d7081c084ceb58278dd3cf81f836bc818978c0ccc770cbbb202125ddabec6628"}, + {file = "Pillow-9.4.0-cp38-cp38-win32.whl", hash = "sha256:df41112ccce5d47770a0c13651479fbcd8793f34232a2dd9faeccb75eb5d0d0d"}, + {file = "Pillow-9.4.0-cp38-cp38-win_amd64.whl", hash = "sha256:7a21222644ab69ddd9967cfe6f2bb420b460dae4289c9d40ff9a4896e7c35c9a"}, + {file = "Pillow-9.4.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:0f3269304c1a7ce82f1759c12ce731ef9b6e95b6df829dccd9fe42912cc48569"}, + {file = "Pillow-9.4.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:cb362e3b0976dc994857391b776ddaa8c13c28a16f80ac6522c23d5257156bed"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a2e0f87144fcbbe54297cae708c5e7f9da21a4646523456b00cc956bd4c65815"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:28676836c7796805914b76b1837a40f76827ee0d5398f72f7dcc634bae7c6264"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0884ba7b515163a1a05440a138adeb722b8a6ae2c2b33aea93ea3118dd3a899e"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:53dcb50fbdc3fb2c55431a9b30caeb2f7027fcd2aeb501459464f0214200a503"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:e8c5cf126889a4de385c02a2c3d3aba4b00f70234bfddae82a5eaa3ee6d5e3e6"}, + {file = "Pillow-9.4.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:6c6b1389ed66cdd174d040105123a5a1bc91d0aa7059c7261d20e583b6d8cbd2"}, + {file = "Pillow-9.4.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:0dd4c681b82214b36273c18ca7ee87065a50e013112eea7d78c7a1b89a739153"}, + {file = "Pillow-9.4.0-cp39-cp39-win32.whl", hash = "sha256:6d9dfb9959a3b0039ee06c1a1a90dc23bac3b430842dcb97908ddde05870601c"}, + {file = "Pillow-9.4.0-cp39-cp39-win_amd64.whl", hash = "sha256:54614444887e0d3043557d9dbc697dbb16cfb5a35d672b7a0fcc1ed0cf1c600b"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:b9b752ab91e78234941e44abdecc07f1f0d8f51fb62941d32995b8161f68cfe5"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d3b56206244dc8711f7e8b7d6cad4663917cd5b2d950799425076681e8766286"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aabdab8ec1e7ca7f1434d042bf8b1e92056245fb179790dc97ed040361f16bfd"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:db74f5562c09953b2c5f8ec4b7dfd3f5421f31811e97d1dbc0a7c93d6e3a24df"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:e9d7747847c53a16a729b6ee5e737cf170f7a16611c143d95aa60a109a59c336"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:b52ff4f4e002f828ea6483faf4c4e8deea8d743cf801b74910243c58acc6eda3"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:575d8912dca808edd9acd6f7795199332696d3469665ef26163cd090fa1f8bfa"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c3c4ed2ff6760e98d262e0cc9c9a7f7b8a9f61aa4d47c58835cdaf7b0b8811bb"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:e621b0246192d3b9cb1dc62c78cfa4c6f6d2ddc0ec207d43c0dedecb914f152a"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:8f127e7b028900421cad64f51f75c051b628db17fb00e099eb148761eed598c9"}, + {file = "Pillow-9.4.0.tar.gz", hash = "sha256:a1c2d7780448eb93fbcc3789bf3916aa5720d942e37945f4056680317f1cd23e"}, +] + +[package.extras] +docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-issues (>=3.0.1)", "sphinx-removed-in", "sphinxext-opengraph"] +tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"] + @@ -928 +1794 @@ name = "pip-audit" -version = "2.4.14" +version = "2.5.4" @@ -934,2 +1800,2 @@ files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, + {file = "pip_audit-2.5.4-py3-none-any.whl", hash = "sha256:bee3748030c895488b4dd8a6196fa44f484da000cfd52d8fc64dfa3a2e121624"}, + {file = "pip_audit-2.5.4.tar.gz", hash = "sha256:a4cb03f9e2896d626f5b153973d3ac0d32fdb18594d78d393b153c83bb8089b6"}, @@ -940 +1806 @@ CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +cyclonedx-python-lib = ">=2.0,<2.5.0 || >2.5.0,<3.0" @@ -945 +1810,0 @@ pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" @@ -952 +1817 @@ doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.260)", "types-html5lib", "types-requests", "types-toml"] @@ -977 +1842 @@ name = "platformdirs" -version = "3.0.0" +version = "3.2.0" @@ -983,2 +1848,2 @@ files = [ - {file = "platformdirs-3.0.0-py3-none-any.whl", hash = "sha256:b1d5eb14f221506f50d6604a561f4c5786d9e80355219694a1b244bcd96f4567"}, - {file = "platformdirs-3.0.0.tar.gz", hash = "sha256:8a1228abb1ef82d788f74139988b137e78692984ec7b08eaa6c65f1723af28f9"}, + {file = "platformdirs-3.2.0-py3-none-any.whl", hash = "sha256:ebe11c0d7a805086e99506aa331612429a72ca7cd52a1f0d277dc4adc20cb10e"}, + {file = "platformdirs-3.2.0.tar.gz", hash = "sha256:d5b638ca397f25f979350ff789db335903d7ea010ab28903f57b27e1b16c2b08"}, @@ -989 +1854 @@ docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx- -test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"] +test = ["appdirs (==1.4.4)", "covdefaults (>=2.3)", "pytest (>=7.2.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"] @@ -1006,0 +1872,22 @@ testing = ["pytest", "pytest-benchmark"] +[[package]] +name = "pooch" +version = "1.6.0" +description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\"" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"}, + {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"}, +] + +[package.dependencies] +appdirs = ">=1.3.0" +packaging = ">=20.0" +requests = ">=2.19.0" + +[package.extras] +progress = ["tqdm (>=4.41.0,<5.0.0)"] +sftp = ["paramiko (>=2.7.0)"] +xxhash = ["xxhash (>=1.4.3)"] + @@ -1033,0 +1921,38 @@ test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] +[[package]] +name = "pyarrow" +version = "11.0.0" +description = "Python library for Apache Arrow" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"}, + {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"}, + {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"}, + {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"}, + {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"}, + {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"}, + {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"}, + {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"}, +] + +[package.dependencies] +numpy = ">=1.16.6" + @@ -1045,0 +1971,24 @@ files = [ +[[package]] +name = "pycparser" +version = "2.21" +description = "C parser in Python" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, + {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, +] + +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + @@ -1222 +2171 @@ name = "pytest" -version = "7.2.1" +version = "7.2.2" @@ -1228,2 +2177,2 @@ files = [ - {file = "pytest-7.2.1-py3-none-any.whl", hash = "sha256:c7c6ca206e93355074ae32f7403e8ea12163b1163c976fee7d4d84027c162be5"}, - {file = "pytest-7.2.1.tar.gz", hash = "sha256:d45e0952f3727241918b8fd0f376f5ff6b301cc0777c6f9a556935c92d8a7d42"}, + {file = "pytest-7.2.2-py3-none-any.whl", hash = "sha256:130328f552dcfac0b1cec75c12e3f005619dc5f874f0a06e8ff7263f0ee6225e"}, + {file = "pytest-7.2.2.tar.gz", hash = "sha256:c99ab0c73aceb050f68929bc93af19ab6db0558791c6a0715723abe9d0ade9d4"}, @@ -1263,0 +2213,15 @@ testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtuale +[[package]] +name = "python-dateutil" +version = "2.8.2" +description = "Extensions to the standard Python datetime module" +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" +files = [ + {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"}, + {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"}, +] + +[package.dependencies] +six = ">=1.5" + @@ -1364,4 +2328,4 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -name = "resolvelib" -version = "0.9.0" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" +name = "responses" +version = "0.18.0" +description = "A utility library for mocking out the `requests` Python library." +category = "main" @@ -1369 +2333 @@ optional = false -python-versions = "*" +python-versions = ">=3.7" @@ -1371,2 +2335,2 @@ files = [ - {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"}, - {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"}, + {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, + {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, @@ -1374,0 +2339,4 @@ files = [ +[package.dependencies] +requests = ">=2.0,<3.0" +urllib3 = ">=1.25.10" + @@ -1376,4 +2344 @@ files = [ -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] +tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] @@ -1383 +2348 @@ name = "rich" -version = "13.3.1" +version = "13.3.3" @@ -1389,2 +2354,2 @@ files = [ - {file = "rich-13.3.1-py3-none-any.whl", hash = "sha256:8aa57747f3fc3e977684f0176a88e789be314a99f99b43b75d1e9cb5dc6db9e9"}, - {file = "rich-13.3.1.tar.gz", hash = "sha256:125d96d20c92b946b983d0d392b84ff945461e5a06d3867e9f9e575f8697b67f"}, + {file = "rich-13.3.3-py3-none-any.whl", hash = "sha256:540c7d6d26a1178e8e8b37e9ba44573a3cd1464ff6348b99ee7061b95d1c6333"}, + {file = "rich-13.3.3.tar.gz", hash = "sha256:dc84400a9d842b3a9c5ff74addd8eb798d155f36c1c91303888e0a66850d2a15"}, @@ -1394,2 +2359,2 @@ files = [ -markdown-it-py = ">=2.1.0,<3.0.0" -pygments = ">=2.14.0,<3.0.0" +markdown-it-py = ">=2.2.0,<3.0.0" +pygments = ">=2.13.0,<3.0.0" @@ -1399,0 +2365,82 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] +[[package]] +name = "scikit-learn" +version = "1.2.2" +description = "A set of python modules for machine learning and data mining" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "scikit-learn-1.2.2.tar.gz", hash = "sha256:8429aea30ec24e7a8c7ed8a3fa6213adf3814a6efbea09e16e0a0c71e1a1a3d7"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:99cc01184e347de485bf253d19fcb3b1a3fb0ee4cea5ee3c43ec0cc429b6d29f"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:e6e574db9914afcb4e11ade84fab084536a895ca60aadea3041e85b8ac963edb"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6fe83b676f407f00afa388dd1fdd49e5c6612e551ed84f3b1b182858f09e987d"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e2642baa0ad1e8f8188917423dd73994bf25429f8893ddbe115be3ca3183584"}, + {file = "scikit_learn-1.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:ad66c3848c0a1ec13464b2a95d0a484fd5b02ce74268eaa7e0c697b904f31d6c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:dfeaf8be72117eb61a164ea6fc8afb6dfe08c6f90365bde2dc16456e4bc8e45f"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:fe0aa1a7029ed3e1dcbf4a5bc675aa3b1bc468d9012ecf6c6f081251ca47f590"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:065e9673e24e0dc5113e2dd2b4ca30c9d8aa2fa90f4c0597241c93b63130d233"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf036ea7ef66115e0d49655f16febfa547886deba20149555a41d28f56fd6d3c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:8b0670d4224a3c2d596fd572fb4fa673b2a0ccfb07152688ebd2ea0b8c61025c"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9c710ff9f9936ba8a3b74a455ccf0dcf59b230caa1e9ba0223773c490cab1e51"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:2dd3ffd3950e3d6c0c0ef9033a9b9b32d910c61bd06cb8206303fb4514b88a49"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:44b47a305190c28dd8dd73fc9445f802b6ea716669cfc22ab1eb97b335d238b1"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:953236889928d104c2ef14027539f5f2609a47ebf716b8cbe4437e85dce42744"}, + {file = "scikit_learn-1.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:7f69313884e8eb311460cc2f28676d5e400bd929841a2c8eb8742ae78ebf7c20"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8156db41e1c39c69aa2d8599ab7577af53e9e5e7a57b0504e116cc73c39138dd"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:fe175ee1dab589d2e1033657c5b6bec92a8a3b69103e3dd361b58014729975c3"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d5312d9674bed14f73773d2acf15a3272639b981e60b72c9b190a0cffed5bad"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ea061bf0283bf9a9f36ea3c5d3231ba2176221bbd430abd2603b1c3b2ed85c89"}, + {file = "scikit_learn-1.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:6477eed40dbce190f9f9e9d0d37e020815825b300121307942ec2110302b66a3"}, +] + +[package.dependencies] +joblib = ">=1.1.1" +numpy = ">=1.17.3" +scipy = ">=1.3.2" +threadpoolctl = ">=2.0.0" + +[package.extras] +benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"] +docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=4.0.1)", "sphinx-gallery (>=0.7.0)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"] +examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"] +tests = ["black (>=22.3.0)", "flake8 (>=3.8.2)", "matplotlib (>=3.1.3)", "mypy (>=0.961)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=5.3.1)", "pytest-cov (>=2.9.0)", "scikit-image (>=0.16.2)"] + +[[package]] +name = "scipy" +version = "1.10.1" +description = "Fundamental algorithms for scientific computing in Python" +category = "main" +optional = false +python-versions = "<3.12,>=3.8" +files = [ + {file = "scipy-1.10.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e7354fd7527a4b0377ce55f286805b34e8c54b91be865bac273f527e1b839019"}, + {file = "scipy-1.10.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:4b3f429188c66603a1a5c549fb414e4d3bdc2a24792e061ffbd607d3d75fd84e"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1553b5dcddd64ba9a0d95355e63fe6c3fc303a8fd77c7bc91e77d61363f7433f"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c0ff64b06b10e35215abce517252b375e580a6125fd5fdf6421b98efbefb2d2"}, + {file = "scipy-1.10.1-cp310-cp310-win_amd64.whl", hash = "sha256:fae8a7b898c42dffe3f7361c40d5952b6bf32d10c4569098d276b4c547905ee1"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:0f1564ea217e82c1bbe75ddf7285ba0709ecd503f048cb1236ae9995f64217bd"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d925fa1c81b772882aa55bcc10bf88324dadb66ff85d548c71515f6689c6dac5"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaea0a6be54462ec027de54fca511540980d1e9eea68b2d5c1dbfe084797be35"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15a35c4242ec5f292c3dd364a7c71a61be87a3d4ddcc693372813c0b73c9af1d"}, + {file = "scipy-1.10.1-cp311-cp311-win_amd64.whl", hash = "sha256:43b8e0bcb877faf0abfb613d51026cd5cc78918e9530e375727bf0625c82788f"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5678f88c68ea866ed9ebe3a989091088553ba12c6090244fdae3e467b1139c35"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:39becb03541f9e58243f4197584286e339029e8908c46f7221abeea4b749fa88"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bce5869c8d68cf383ce240e44c1d9ae7c06078a9396df68ce88a1230f93a30c1"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:07c3457ce0b3ad5124f98a86533106b643dd811dd61b548e78cf4c8786652f6f"}, + {file = "scipy-1.10.1-cp38-cp38-win_amd64.whl", hash = "sha256:049a8bbf0ad95277ffba9b3b7d23e5369cc39e66406d60422c8cfef40ccc8415"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cd9f1027ff30d90618914a64ca9b1a77a431159df0e2a195d8a9e8a04c78abf9"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:79c8e5a6c6ffaf3a2262ef1be1e108a035cf4f05c14df56057b64acc5bebffb6"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:51af417a000d2dbe1ec6c372dfe688e041a7084da4fdd350aeb139bd3fb55353"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b4735d6c28aad3cdcf52117e0e91d6b39acd4272f3f5cd9907c24ee931ad601"}, + {file = "scipy-1.10.1-cp39-cp39-win_amd64.whl", hash = "sha256:7ff7f37b1bf4417baca958d254e8e2875d0cc23aaadbe65b3d5b3077b0eb23ea"}, + {file = "scipy-1.10.1.tar.gz", hash = "sha256:2cf9dfb80a7b4589ba4c40ce7588986d6d5cebc5457cad2c2880f6bc2d42f3a5"}, +] + +[package.dependencies] +numpy = ">=1.19.5,<1.27.0" + +[package.extras] +dev = ["click", "doit (>=0.36.0)", "flake8", "mypy", "pycodestyle", "pydevtool", "rich-click", "typing_extensions"] +doc = ["matplotlib (>2)", "numpydoc", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"] +test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"] + @@ -1402 +2449 @@ name = "setuptools" -version = "67.4.0" +version = "67.6.1" @@ -1404 +2451 @@ description = "Easily download, build, install, upgrade, and uninstall Python pa -category = "dev" +category = "main" @@ -1408,2 +2455,2 @@ files = [ - {file = "setuptools-67.4.0-py3-none-any.whl", hash = "sha256:f106dee1b506dee5102cc3f3e9e68137bbad6d47b616be7991714b0c62204251"}, - {file = "setuptools-67.4.0.tar.gz", hash = "sha256:e5fd0a713141a4a105412233c63dc4e17ba0090c8e8334594ac790ec97792330"}, + {file = "setuptools-67.6.1-py3-none-any.whl", hash = "sha256:e728ca814a823bf7bf60162daf9db95b93d532948c4c0bea762ce62f60189078"}, + {file = "setuptools-67.6.1.tar.gz", hash = "sha256:257de92a9d50a60b8e22abfcbb771571fde0dbf3ec234463212027a4eeecbe9a"}, @@ -1421 +2468 @@ description = "Python 2 and 3 compatibility utilities" -category = "dev" +category = "main" @@ -1452,0 +2500,68 @@ files = [ +[[package]] +name = "soundfile" +version = "0.12.1" +description = "An audio library based on libsndfile, CFFI and NumPy" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"}, + {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"}, + {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"}, + {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"}, +] + +[package.dependencies] +cffi = ">=1.0" + +[package.extras] +numpy = ["numpy"] + +[[package]] +name = "soxr" +version = "0.3.4" +description = "High quality, one-dimensional sample-rate conversion library" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "soxr-0.3.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b7b84126643c063d5daa203f7f9137e21734dabbd7e68c097607b2ef457e2f2e"}, + {file = "soxr-0.3.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:380d2d43871a68e8b1ef1702a0abe6f9e48ddb3933c7a303c45d67e121503e7c"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4a1b4019c9972f57612482c4f85523d6e832e3d10935e2f070a9dcd334a4dcb"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e613cee023b7c3f162b9da3f6b169cd7f58de345275be1fde9f19adc9cf144df"}, + {file = "soxr-0.3.4-cp310-cp310-win32.whl", hash = "sha256:182c02a7ba45a159a0dbb0a297335df2381ead03a65377b19663ea0ff720ecb7"}, + {file = "soxr-0.3.4-cp310-cp310-win_amd64.whl", hash = "sha256:1e95c96ce94524fae453b4331c9910d33f97506f99bae06d76a9c0649710619e"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2678d2719e7496803983584d661eb5fddc7017154a8dda4a774407c56ff07973"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:11bd1396052049e6d389225a0e96a9df15f706da501c619b35d3c72ac6bc7257"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7e23de4dfe54ac30e583bbc9cc3feda1cd776fedce13206bc4b3115b75ecab82"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3e7396498a5f5b7d8f23b656f65c24517a6ff5bdc3ee0623ccd491036a43ea08"}, + {file = "soxr-0.3.4-cp311-cp311-win32.whl", hash = "sha256:e57e9703c2bff834cabc06800d3c11a259544891d2c24a78949f3cf2f5492cc5"}, + {file = "soxr-0.3.4-cp311-cp311-win_amd64.whl", hash = "sha256:7c8350acd7150f74261a0569b47ccb1bb4aa39b2d575860bc97cfa69aab8aead"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:941f7355116fe77fe6a82938fa7799a0e466a494ebc093f676969ce32b2815b1"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:00fdbf24f64d8c3fb800425c383048cb24c32defac80901cde4a57fb6ce5d431"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7bb6d4dc807d04c536674429e2b05ae08a1efac9815c4595e41ffd6b57c2c662"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ff15853895b54f1b627799c6600be1ce5f7286724e7a93e4b7b9d79e5d4166f5"}, + {file = "soxr-0.3.4-cp38-cp38-win32.whl", hash = "sha256:d858becbc1fcc7b38c3436d3276290fae09403cdcbdf1d5986a18dab7023a6c3"}, + {file = "soxr-0.3.4-cp38-cp38-win_amd64.whl", hash = "sha256:068ab4df549df5783cc1eb4eb6c94f53823b164dc27134fc621fc9f5097f38cd"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:20130329985f9767c8417bbd125fe138790a71802b000481c386a800e2ad2bca"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:78090e97abfb326b7cf14ef37d08a17252b07d438388dcbbd82a6836a9d551b1"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84e590e75b7e5dca12bf68bfb090276f34a88fbcd793781c62d47f5d7dbe525e"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3479d265574b960e12bca0878baba0862c43278915e0319d84679bb4d4fcd33"}, + {file = "soxr-0.3.4-cp39-cp39-win32.whl", hash = "sha256:83de825d6a713c7b2e76d9ec3f229a58a9ed290237e7adc05d80e8b39be995a6"}, + {file = "soxr-0.3.4-cp39-cp39-win_amd64.whl", hash = "sha256:2082f88cae89de854c3e0d62f55d0cb31eb11764f5c2a28299121fb642a22472"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:fe8b5f92c802f1e7793c40344f5368dc6163718c9ffa82e79ee6ad779d318ac5"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d0063d5f9a4e1a367084f4705301e9da131cf4d2d32aa3fe0072a1245e18088f"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:a680bab57adae462cdc86abcc7330beb5daa3ba5101165583eedcda88b7ba551"}, + {file = "soxr-0.3.4.tar.gz", hash = "sha256:fe68daf00e8f020977b187699903d219f9e39b9fb3d915f3f923eed8ba431449"}, +] + +[package.dependencies] +numpy = "*" + +[package.extras] +docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"] +test = ["pytest"] + @@ -1467,0 +2583,12 @@ pbr = ">=2.0.0,<2.1.0 || >2.1.0" +[[package]] +name = "threadpoolctl" +version = "3.1.0" +description = "threadpoolctl" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, + {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"}, +] + @@ -1494 +2621 @@ name = "tqdm" -version = "4.64.1" +version = "4.65.0" @@ -1498 +2625 @@ optional = false -python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,>=2.7" +python-versions = ">=3.7" @@ -1500,2 +2627,2 @@ files = [ - {file = "tqdm-4.64.1-py2.py3-none-any.whl", hash = "sha256:6fee160d6ffcd1b1c68c65f14c829c22832bc401726335ce92c52d395944a6a1"}, - {file = "tqdm-4.64.1.tar.gz", hash = "sha256:5f4f682a004951c1b450bc753c710e9280c5746ce6ffedee253ddbcbf54cf1e4"}, + {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"}, + {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"}, @@ -1515 +2642 @@ name = "types-psutil" -version = "5.9.5.9" +version = "5.9.5.11" @@ -1521,2 +2648,2 @@ files = [ - {file = "types-psutil-5.9.5.9.tar.gz", hash = "sha256:c6c7e6b41b6f7ebb87afd5650df360751d0e8acf0d276b7ac64d00f419beb922"}, - {file = "types_psutil-5.9.5.9-py3-none-any.whl", hash = "sha256:f7830fb3a7bc3288113ff08199330fced7851f5219f0af1b2abc24aaa3112236"}, + {file = "types-psutil-5.9.5.11.tar.gz", hash = "sha256:3d59da0758f056bfb59fef757366e538c5dd5473d81c35b38956624ae2484f31"}, + {file = "types_psutil-5.9.5.11-py3-none-any.whl", hash = "sha256:01cc541b187a11e758d336c4cc89abf71d0098627fa95d5cfaca536be31a7d1a"}, @@ -1539 +2666 @@ name = "types-requests" -version = "2.28.11.15" +version = "2.28.11.17" @@ -1545,2 +2672,2 @@ files = [ - {file = "types-requests-2.28.11.15.tar.gz", hash = "sha256:fc8eaa09cc014699c6b63c60c2e3add0c8b09a410c818b5ac6e65f92a26dde09"}, - {file = "types_requests-2.28.11.15-py3-none-any.whl", hash = "sha256:a05e4c7bc967518fba5789c341ea8b0c942776ee474c7873129a61161978e586"}, + {file = "types-requests-2.28.11.17.tar.gz", hash = "sha256:0d580652ce903f643f8c3b494dd01d29367ea57cea0c7ad7f65cf3169092edb0"}, + {file = "types_requests-2.28.11.17-py3-none-any.whl", hash = "sha256:cc1aba862575019306b2ed134eb1ea994cab1c887a22e18d3383e6dd42e9789b"}, @@ -1554 +2681 @@ name = "types-urllib3" -version = "1.26.25.8" +version = "1.26.25.9" @@ -1560,2 +2687,2 @@ files = [ - {file = "types-urllib3-1.26.25.8.tar.gz", hash = "sha256:ecf43c42d8ee439d732a1110b4901e9017a79a38daca26f08e42c8460069392c"}, - {file = "types_urllib3-1.26.25.8-py3-none-any.whl", hash = "sha256:95ea847fbf0bf675f50c8ae19a665baedcf07e6b4641662c4c3c72e7b2edf1a9"}, + {file = "types-urllib3-1.26.25.9.tar.gz", hash = "sha256:160727879bdbe52f11f5feeca092a473f38d68ed3be88abb461b59cda40fb9bc"}, + {file = "types_urllib3-1.26.25.9-py3-none-any.whl", hash = "sha256:b327d360ba4a9edd80ea82f5990ba19e76175a20b5b64be4b4813d9a1c424caa"}, @@ -1578 +2705 @@ name = "urllib3" -version = "1.26.14" +version = "1.26.15" @@ -1584,2 +2711,2 @@ files = [ - {file = "urllib3-1.26.14-py2.py3-none-any.whl", hash = "sha256:75edcdc2f7d85b137124a6c3c9fc3933cdeaa12ecb9a6a959f22797a0feca7e1"}, - {file = "urllib3-1.26.14.tar.gz", hash = "sha256:076907bf8fd355cde77728471316625a4d2f7e713c125f51953bb5b3eecf4f72"}, + {file = "urllib3-1.26.15-py2.py3-none-any.whl", hash = "sha256:aa751d169e23c7479ce47a0cb0da579e3ede798f994f5816a74e4f4500dcea42"}, + {file = "urllib3-1.26.15.tar.gz", hash = "sha256:8a388717b9476f934a21484e8c8e61875ab60644d29b9b39e11e4b9dc1c6b305"}, @@ -1604,0 +2732,196 @@ files = [ +[[package]] +name = "xxhash" +version = "3.2.0" +description = "Python binding for xxHash" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"}, + {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"}, + {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"}, + {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"}, + {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"}, + {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"}, + {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"}, + {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"}, + {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"}, + {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"}, + {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"}, + {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"}, + {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"}, + {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"}, + {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"}, + {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"}, + {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"}, +] + +[[package]] +name = "yarl" +version = "1.8.2" +description = "Yet another URL library" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:bb81f753c815f6b8e2ddd2eef3c855cf7da193b82396ac013c661aaa6cc6b0a5"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:47d49ac96156f0928f002e2424299b2c91d9db73e08c4cd6742923a086f1c863"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3fc056e35fa6fba63248d93ff6e672c096f95f7836938241ebc8260e062832fe"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58a3c13d1c3005dbbac5c9f0d3210b60220a65a999b1833aa46bd6677c69b08e"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:10b08293cda921157f1e7c2790999d903b3fd28cd5c208cf8826b3b508026996"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:de986979bbd87272fe557e0a8fcb66fd40ae2ddfe28a8b1ce4eae22681728fef"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c4fcfa71e2c6a3cb568cf81aadc12768b9995323186a10827beccf5fa23d4f8"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae4d7ff1049f36accde9e1ef7301912a751e5bae0a9d142459646114c70ecba6"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:bf071f797aec5b96abfc735ab97da9fd8f8768b43ce2abd85356a3127909d146"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:74dece2bfc60f0f70907c34b857ee98f2c6dd0f75185db133770cd67300d505f"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:df60a94d332158b444301c7f569659c926168e4d4aad2cfbf4bce0e8fb8be826"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:63243b21c6e28ec2375f932a10ce7eda65139b5b854c0f6b82ed945ba526bff3"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:cfa2bbca929aa742b5084fd4663dd4b87c191c844326fcb21c3afd2d11497f80"}, + {file = "yarl-1.8.2-cp310-cp310-win32.whl", hash = "sha256:b05df9ea7496df11b710081bd90ecc3a3db6adb4fee36f6a411e7bc91a18aa42"}, + {file = "yarl-1.8.2-cp310-cp310-win_amd64.whl", hash = "sha256:24ad1d10c9db1953291f56b5fe76203977f1ed05f82d09ec97acb623a7976574"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:2a1fca9588f360036242f379bfea2b8b44cae2721859b1c56d033adfd5893634"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:f37db05c6051eff17bc832914fe46869f8849de5b92dc4a3466cd63095d23dfd"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:77e913b846a6b9c5f767b14dc1e759e5aff05502fe73079f6f4176359d832581"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0978f29222e649c351b173da2b9b4665ad1feb8d1daa9d971eb90df08702668a"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:388a45dc77198b2460eac0aca1efd6a7c09e976ee768b0d5109173e521a19daf"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2305517e332a862ef75be8fad3606ea10108662bc6fe08509d5ca99503ac2aee"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:42430ff511571940d51e75cf42f1e4dbdded477e71c1b7a17f4da76c1da8ea76"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3150078118f62371375e1e69b13b48288e44f6691c1069340081c3fd12c94d5b"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:c15163b6125db87c8f53c98baa5e785782078fbd2dbeaa04c6141935eb6dab7a"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:4d04acba75c72e6eb90745447d69f84e6c9056390f7a9724605ca9c56b4afcc6"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:e7fd20d6576c10306dea2d6a5765f46f0ac5d6f53436217913e952d19237efc4"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:75c16b2a900b3536dfc7014905a128a2bea8fb01f9ee26d2d7d8db0a08e7cb2c"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:6d88056a04860a98341a0cf53e950e3ac9f4e51d1b6f61a53b0609df342cc8b2"}, + {file = "yarl-1.8.2-cp311-cp311-win32.whl", hash = "sha256:fb742dcdd5eec9f26b61224c23baea46c9055cf16f62475e11b9b15dfd5c117b"}, + {file = "yarl-1.8.2-cp311-cp311-win_amd64.whl", hash = "sha256:8c46d3d89902c393a1d1e243ac847e0442d0196bbd81aecc94fcebbc2fd5857c"}, + {file = "yarl-1.8.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:ceff9722e0df2e0a9e8a79c610842004fa54e5b309fe6d218e47cd52f791d7ef"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3f6b4aca43b602ba0f1459de647af954769919c4714706be36af670a5f44c9c1"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1684a9bd9077e922300ecd48003ddae7a7474e0412bea38d4631443a91d61077"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ebb78745273e51b9832ef90c0898501006670d6e059f2cdb0e999494eb1450c2"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3adeef150d528ded2a8e734ebf9ae2e658f4c49bf413f5f157a470e17a4a2e89"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:57a7c87927a468e5a1dc60c17caf9597161d66457a34273ab1760219953f7f4c"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:efff27bd8cbe1f9bd127e7894942ccc20c857aa8b5a0327874f30201e5ce83d0"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:a783cd344113cb88c5ff7ca32f1f16532a6f2142185147822187913eb989f739"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:705227dccbe96ab02c7cb2c43e1228e2826e7ead880bb19ec94ef279e9555b5b"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:34c09b43bd538bf6c4b891ecce94b6fa4f1f10663a8d4ca589a079a5018f6ed7"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:a48f4f7fea9a51098b02209d90297ac324241bf37ff6be6d2b0149ab2bd51b37"}, + {file = "yarl-1.8.2-cp37-cp37m-win32.whl", hash = "sha256:0414fd91ce0b763d4eadb4456795b307a71524dbacd015c657bb2a39db2eab89"}, + {file = "yarl-1.8.2-cp37-cp37m-win_amd64.whl", hash = "sha256:d881d152ae0007809c2c02e22aa534e702f12071e6b285e90945aa3c376463c5"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5df5e3d04101c1e5c3b1d69710b0574171cc02fddc4b23d1b2813e75f35a30b1"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7a66c506ec67eb3159eea5096acd05f5e788ceec7b96087d30c7d2865a243918"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2b4fa2606adf392051d990c3b3877d768771adc3faf2e117b9de7eb977741229"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1e21fb44e1eff06dd6ef971d4bdc611807d6bd3691223d9c01a18cec3677939e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:93202666046d9edadfe9f2e7bf5e0782ea0d497b6d63da322e541665d65a044e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fc77086ce244453e074e445104f0ecb27530d6fd3a46698e33f6c38951d5a0f1"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:64dd68a92cab699a233641f5929a40f02a4ede8c009068ca8aa1fe87b8c20ae3"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1b372aad2b5f81db66ee7ec085cbad72c4da660d994e8e590c997e9b01e44901"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e6f3515aafe0209dd17fb9bdd3b4e892963370b3de781f53e1746a521fb39fc0"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:dfef7350ee369197106805e193d420b75467b6cceac646ea5ed3049fcc950a05"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:728be34f70a190566d20aa13dc1f01dc44b6aa74580e10a3fb159691bc76909d"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ff205b58dc2929191f68162633d5e10e8044398d7a45265f90a0f1d51f85f72c"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:baf211dcad448a87a0d9047dc8282d7de59473ade7d7fdf22150b1d23859f946"}, + {file = "yarl-1.8.2-cp38-cp38-win32.whl", hash = "sha256:272b4f1599f1b621bf2aabe4e5b54f39a933971f4e7c9aa311d6d7dc06965165"}, + {file = "yarl-1.8.2-cp38-cp38-win_amd64.whl", hash = "sha256:326dd1d3caf910cd26a26ccbfb84c03b608ba32499b5d6eeb09252c920bcbe4f"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:f8ca8ad414c85bbc50f49c0a106f951613dfa5f948ab69c10ce9b128d368baf8"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:418857f837347e8aaef682679f41e36c24250097f9e2f315d39bae3a99a34cbf"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:ae0eec05ab49e91a78700761777f284c2df119376e391db42c38ab46fd662b77"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:009a028127e0a1755c38b03244c0bea9d5565630db9c4cf9572496e947137a87"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3edac5d74bb3209c418805bda77f973117836e1de7c000e9755e572c1f7850d0"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:da65c3f263729e47351261351b8679c6429151ef9649bba08ef2528ff2c423b2"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ef8fb25e52663a1c85d608f6dd72e19bd390e2ecaf29c17fb08f730226e3a08"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bcd7bb1e5c45274af9a1dd7494d3c52b2be5e6bd8d7e49c612705fd45420b12d"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:44ceac0450e648de86da8e42674f9b7077d763ea80c8ceb9d1c3e41f0f0a9951"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:97209cc91189b48e7cfe777237c04af8e7cc51eb369004e061809bcdf4e55220"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:48dd18adcf98ea9cd721a25313aef49d70d413a999d7d89df44f469edfb38a06"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:e59399dda559688461762800d7fb34d9e8a6a7444fd76ec33220a926c8be1516"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d617c241c8c3ad5c4e78a08429fa49e4b04bedfc507b34b4d8dceb83b4af3588"}, + {file = "yarl-1.8.2-cp39-cp39-win32.whl", hash = "sha256:cb6d48d80a41f68de41212f3dfd1a9d9898d7841c8f7ce6696cf2fd9cb57ef83"}, + {file = "yarl-1.8.2-cp39-cp39-win_amd64.whl", hash = "sha256:6604711362f2dbf7160df21c416f81fac0de6dbcf0b5445a2ef25478ecc4c778"}, + {file = "yarl-1.8.2.tar.gz", hash = "sha256:49d43402c6e3013ad0978602bf6bf5328535c48d192304b91b97a3c6790b1562"}, +] + +[package.dependencies] +idna = ">=2.0" +multidict = ">=4.0" + @@ -1608 +2931 @@ python-versions = "3.9.15" -content-hash = "7a54f8bf0afda6421a0b8af6d43de261f0bb88e4e6b2f41552d2bc64ee521f20" +content-hash = "75aa9df715db07857fb386e175284037aa6b3202c852476f339ea619f558358c" diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml index 5f64c2a9..11378269 100644 --- a/libs/libcommon/pyproject.toml +++ b/libs/libcommon/pyproject.toml @@ -9,0 +10 @@ appdirs = "^1.4.4" +datasets = { extras = ["audio", "vision"], version = "^2.11.0" } @@ -16,0 +18 @@ psutil = "^5.9.4" +pydub = "^0.25.1" @@ -20,0 +23,2 @@ requests = "^2.28.2" +soundfile = ">=0.12.1" +numba = "0.56.4" @@ -28 +32,2 @@ mypy = "^1.0.0" -pip-audit = "^2.4.14" +pandas-stubs = "^1.5.3" +pip-audit = "^2.5.4" @@ -59,0 +65 @@ module = [ + "datasets.*", diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 208da39d..ddd7ed31 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -33 +33 @@ ASSETS_BASE_URL = "assets" -ASSETS_STORE_DIRECTORY = None +ASSETS_STORAGE_DIRECTORY = None @@ -39 +39 @@ class AssetsConfig: - storage_directory: Optional[str] = ASSETS_STORE_DIRECTORY + storage_directory: Optional[str] = ASSETS_STORAGE_DIRECTORY @@ -47 +47,38 @@ class AssetsConfig: - storage_directory=env.str(name="STORAGE_DIRECTORY", default=ASSETS_STORE_DIRECTORY), + storage_directory=env.str(name="STORAGE_DIRECTORY", default=ASSETS_STORAGE_DIRECTORY), + ) + + +CACHED_ASSETS_BASE_URL = "cached-assets" +CACHED_ASSETS_STORAGE_DIRECTORY = None +CACHED_ASSETS_CLEAN_CACHE_PROBA = 0.05 +CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER = 100 +CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER = 200 +CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER = 10_000 + + +@dataclass(frozen=True) +class CachedAssetsConfig: + base_url: str = ASSETS_BASE_URL + storage_directory: Optional[str] = ASSETS_STORAGE_DIRECTORY + clean_cache_proba: float = CACHED_ASSETS_CLEAN_CACHE_PROBA + keep_first_rows_number: int = CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER + keep_most_recent_rows_number: int = CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER + max_cleaned_rows_number: int = CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER + + @classmethod + def from_env(cls) -> "CachedAssetsConfig": + env = Env(expand_vars=True) + with env.prefixed("CACHED_ASSETS_"): + return cls( + base_url=env.str(name="BASE_URL", default=CACHED_ASSETS_BASE_URL), + storage_directory=env.str(name="STORAGE_DIRECTORY", default=CACHED_ASSETS_STORAGE_DIRECTORY), + clean_cache_proba=env.float(name="CLEAN_CACHE_PROBA", default=CACHED_ASSETS_CLEAN_CACHE_PROBA), + keep_first_rows_number=env.float( + name="KEEP_FIRST_ROWS_NUMBER", default=CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER + ), + keep_most_recent_rows_number=env.float( + name="KEEP_MOST_RECENT_ROWS_NUMBER", default=CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER + ), + max_cleaned_rows_number=env.float( + name="MAX_CLEAN_SAMPLE_SIZE", default=CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER + ), diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 71f30795..83327cc6 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -4,0 +5 @@ ASSETS_CACHE_APPNAME = "datasets_server_assets" +CACHED_ASSETS_CACHE_APPNAME = "datasets_server_cached_assets" diff --git a/libs/libcommon/src/libcommon/storage.py b/libs/libcommon/src/libcommon/storage.py index a4197e8b..d462cbbd 100644 --- a/libs/libcommon/src/libcommon/storage.py +++ b/libs/libcommon/src/libcommon/storage.py @@ -12 +12 @@ from appdirs import user_cache_dir # type:ignore -from libcommon.constants import ASSETS_CACHE_APPNAME +from libcommon.constants import ASSETS_CACHE_APPNAME, CACHED_ASSETS_CACHE_APPNAME @@ -51,0 +52,14 @@ def init_assets_dir(directory: Optional[StrPath] = None) -> StrPath: +def init_cached_assets_dir(directory: Optional[StrPath] = None) -> StrPath: + """Initialize the cached assets directory. + + If directory is None, it will be set to the default cache location on the machine. + + Args: + directory (Optional[Union[str, PathLike[str]]], optional): The directory to initialize. Defaults to None. + + Returns: + Union[str, PathLike[str]]: The directory. + """ + return init_dir(directory, appname=CACHED_ASSETS_CACHE_APPNAME) + + diff --git a/libs/libcommon/src/libcommon/viewer_utils/__init__.py b/libs/libcommon/src/libcommon/viewer_utils/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/services/worker/src/worker/asset.py b/libs/libcommon/src/libcommon/viewer_utils/asset.py similarity index 53% rename from services/worker/src/worker/asset.py rename to libs/libcommon/src/libcommon/viewer_utils/asset.py index c26a8b57..28a8322c 100644 --- a/services/worker/src/worker/asset.py +++ b/libs/libcommon/src/libcommon/viewer_utils/asset.py @@ -3,0 +4 @@ +import os @@ -6 +7 @@ from pathlib import Path -from typing import List, Tuple, TypedDict +from typing import Generator, List, Tuple, TypedDict @@ -9 +9,0 @@ import soundfile # type:ignore -from libcommon.storage import StrPath @@ -13,0 +14,2 @@ from pydub import AudioSegment # type:ignore +from libcommon.storage import StrPath + @@ -15,0 +18 @@ ASSET_DIR_MODE = 0o755 +DATASETS_SERVER_MDATE_FILENAME = ".dss" @@ -26,0 +30,24 @@ def create_asset_dir( +def glob_rows_in_assets_dir( + dataset: str, + assets_directory: StrPath, +) -> Generator[Path, None, None]: + return Path(assets_directory).resolve().glob(os.path.join(dataset, DATASET_SEPARATOR, "*", "*", "*")) + + +def update_last_modified_date_of_rows_in_assets_dir( + dataset: str, + config: str, + split: str, + offset: int, + length: int, + assets_directory: StrPath, +) -> None: + row_dirs_path = Path(assets_directory).resolve() / dataset / DATASET_SEPARATOR / config / split + for row_idx in range(offset, offset + length): + if (row_dirs_path / str(row_idx)).is_dir(): + # update the directory's last modified date + if (row_dirs_path / str(row_idx) / DATASETS_SERVER_MDATE_FILENAME).is_file(): + (row_dirs_path / str(row_idx) / DATASETS_SERVER_MDATE_FILENAME).unlink() + (row_dirs_path / str(row_idx) / DATASETS_SERVER_MDATE_FILENAME).touch() + + @@ -42,0 +70 @@ def create_image_file( + overwrite: bool = True, @@ -45 +73,6 @@ def create_image_file( - dataset=dataset, config=config, split=split, row_idx=row_idx, column=column, assets_directory=assets_directory + dataset=dataset, + config=config, + split=split, + row_idx=row_idx, + column=column, + assets_directory=assets_directory, @@ -46,0 +80 @@ def create_image_file( + makedirs(dir_path, ASSET_DIR_MODE, exist_ok=True) @@ -48 +82,2 @@ def create_image_file( - image.save(file_path) + if overwrite or not file_path.exists(): + image.save(file_path) @@ -71,0 +107 @@ def create_audio_files( + overwrite: bool = True, @@ -76 +112,6 @@ def create_audio_files( - dataset=dataset, config=config, split=split, row_idx=row_idx, column=column, assets_directory=assets_directory + dataset=dataset, + config=config, + split=split, + row_idx=row_idx, + column=column, + assets_directory=assets_directory, @@ -77,0 +119 @@ def create_audio_files( + makedirs(dir_path, ASSET_DIR_MODE, exist_ok=True) @@ -80,3 +122,5 @@ def create_audio_files( - soundfile.write(wav_file_path, array, sampling_rate) - segment = AudioSegment.from_wav(wav_file_path) - segment.export(mp3_file_path, format="mp3") + if overwrite or not wav_file_path.exists(): + soundfile.write(wav_file_path, array, sampling_rate) + if overwrite or not mp3_file_path.exists(): + segment = AudioSegment.from_wav(wav_file_path) + segment.export(mp3_file_path, format="mp3") diff --git a/services/worker/src/worker/features.py b/libs/libcommon/src/libcommon/viewer_utils/features.py similarity index 87% rename from services/worker/src/worker/features.py rename to libs/libcommon/src/libcommon/viewer_utils/features.py index aacd6690..3bd2702f 100644 --- a/services/worker/src/worker/features.py +++ b/libs/libcommon/src/libcommon/viewer_utils/features.py @@ -4 +3,0 @@ -import io @@ -5,0 +5 @@ import json +from io import BytesIO @@ -9,2 +8,0 @@ from zlib import adler32 -import numpy -import soundfile # type:ignore @@ -24 +22 @@ from datasets import ( -from libcommon.storage import StrPath +from numpy import ndarray @@ -27 +25,2 @@ from PIL import Image as PILImage # type: ignore -from worker.asset import create_audio_files, create_image_file +from libcommon.storage import StrPath +from libcommon.viewer_utils.asset import create_audio_files, create_image_file @@ -56,0 +56 @@ def image( + overwrite: bool = True, @@ -59,0 +60,2 @@ def image( + if isinstance(value, dict) and value.get("bytes"): + value = PILImage.open(BytesIO(value["bytes"])) @@ -61,5 +63,4 @@ def image( - try: - image_bytes = value["bytes"] - value = PILImage.open(io.BytesIO(image_bytes)) - except Exception: - raise TypeError("image cell must be a PIL image") + raise TypeError( + "Image cell must be a PIL image or an encoded dict of an image, " + f"but got {str(value)[:300]}{'...' if len(str(value)) > 300 else ''}" + ) @@ -78,0 +80 @@ def image( + overwrite=overwrite, @@ -97,0 +100 @@ def audio( + overwrite: bool = True, @@ -100,0 +104,2 @@ def audio( + if isinstance(value, dict) and value.get("bytes"): + value = Audio().decode_example(value) @@ -105,6 +110,5 @@ def audio( - if "bytes" in value: - bytes_array, sampling_rate = soundfile.read(io.BytesIO(value["bytes"])) - array = numpy.array(bytes_array) - else: - raise TypeError("audio cell must contain 'array' and 'sampling_rate' fields") from e - if type(array) != numpy.ndarray: + raise TypeError( + "audio cell must contain 'array' and 'sampling_rate' fields, " + f"but got {str(value)[:300]}{'...' if len(str(value)) > 300 else ''}" + ) from e + if type(array) != ndarray: @@ -125,0 +130 @@ def audio( + overwrite=overwrite, @@ -139,0 +145 @@ def get_cell_value( + overwrite: bool = True, @@ -154,0 +161 @@ def get_cell_value( + overwrite=overwrite, @@ -166,0 +174 @@ def get_cell_value( + overwrite=overwrite, @@ -185,0 +194 @@ def get_cell_value( + overwrite=overwrite, @@ -204,0 +214 @@ def get_cell_value( + overwrite=overwrite, @@ -226,0 +237 @@ def get_cell_value( + overwrite=overwrite, @@ -248,0 +260 @@ def get_cell_value( + overwrite=overwrite, diff --git a/libs/libcommon/tests/conftest.py b/libs/libcommon/tests/conftest.py index 628df4bd..8e81561b 100644 --- a/libs/libcommon/tests/conftest.py +++ b/libs/libcommon/tests/conftest.py @@ -2,0 +3 @@ +from pathlib import Path @@ -6,0 +8,5 @@ from pytest import fixture +from libcommon.storage import StrPath, init_cached_assets_dir + +# Import fixture modules as plugins +pytest_plugins = ["tests.fixtures.datasets"] + @@ -32,0 +39,6 @@ def queue_mongo_host(env: Env) -> str: + + +@fixture +def cached_assets_directory(tmp_path: Path) -> StrPath: + cached_assets_directory = tmp_path / "cached-assets" + return init_cached_assets_dir(cached_assets_directory) diff --git a/libs/libcommon/tests/fixtures/__init__.py b/libs/libcommon/tests/fixtures/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/libs/libcommon/tests/fixtures/data/test_image_rgb.jpg b/libs/libcommon/tests/fixtures/data/test_image_rgb.jpg new file mode 100644 index 00000000..e131e8ec Binary files /dev/null and b/libs/libcommon/tests/fixtures/data/test_image_rgb.jpg differ diff --git a/libs/libcommon/tests/fixtures/datasets.py b/libs/libcommon/tests/fixtures/datasets.py new file mode 100644 index 00000000..ea07ce7f --- /dev/null +++ b/libs/libcommon/tests/fixtures/datasets.py @@ -0,0 +1,133 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import datetime +from pathlib import Path +from typing import Any, Mapping, Optional + +import numpy as np +import pandas as pd +import pytest +from datasets import ( + Array2D, + Array3D, + Array4D, + Array5D, + Audio, + ClassLabel, + Dataset, + Features, + Image, + Sequence, + Translation, + TranslationVariableLanguages, + Value, +) +from datasets.features.features import FeatureType + + +def value(content: Any, dtype: Any) -> Dataset: + return Dataset.from_pandas(pd.DataFrame({"col": [content]}, dtype=dtype)) + + +def other(content: Any, feature_type: Optional[FeatureType] = None) -> Dataset: + if feature_type: + features = Features({"col": feature_type}) + return Dataset.from_dict({"col": [content]}, features=features) + else: + return Dataset.from_dict({"col": [content]}) + + [email protected](scope="session") +def datasets() -> Mapping[str, Dataset]: + sampling_rate = 16_000 + return { + # Value feature + "null": value(None, None), + "bool": value(False, pd.BooleanDtype()), + "int8": value(-7, pd.Int8Dtype()), + "int16": value(-7, pd.Int16Dtype()), + "int32": value(-7, pd.Int32Dtype()), + "int64": value(-7, pd.Int64Dtype()), + "uint8": value(7, pd.UInt8Dtype()), + "uint16": value(7, pd.UInt16Dtype()), + "uint32": value(7, pd.UInt32Dtype()), + "uint64": value(7, pd.UInt64Dtype()), + "float16": value(-3.14, np.float16), + "float32": value(-3.14, np.float32), + "float64": value(-3.14, np.float64), + "time": value(datetime.time(1, 1, 1), None), + "timestamp_1": value(pd.Timestamp(2020, 1, 1), None), + "timestamp_2": value(pd.Timestamp(1513393355.5, unit="s"), None), + "timestamp_3": value(pd.Timestamp(1513393355500, unit="ms"), None), + "timestamp_tz": value(pd.Timestamp(year=2020, month=1, day=1, tz="US/Pacific"), None), + "string": value("a string", pd.StringDtype(storage="python")), + # other types of features + "class_label": other("positive", ClassLabel(names=["negative", "positive"])), + "dict": other({"a": 0}, None), + "list": other([{"a": 0}], None), + "sequence_simple": other([0], None), + "sequence": other([{"a": 0}], Sequence(feature={"a": Value(dtype="int64")})), + "array2d": other(np.zeros((2, 2), dtype="float32"), Array2D(shape=(2, 2), dtype="float32")), + "array3d": other(np.zeros((2, 2, 2), dtype="float32"), Array3D(shape=(2, 2, 2), dtype="float32")), + "array4d": other(np.zeros((2, 2, 2, 2), dtype="float32"), Array4D(shape=(2, 2, 2, 2), dtype="float32")), + "array5d": other(np.zeros((2, 2, 2, 2, 2), dtype="float32"), Array5D(shape=(2, 2, 2, 2, 2), dtype="float32")), + "audio": other({"array": [0.1, 0.2, 0.3], "sampling_rate": sampling_rate}, Audio(sampling_rate=sampling_rate)), + "image": other(str(Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"), Image()), + "translation": other({"en": "the cat", "fr": "le chat"}, Translation(languages=["en", "fr"])), + "translation_variable_languages": other( + {"en": "the cat", "fr": ["le chat", "la chatte"]}, + TranslationVariableLanguages(languages=["en", "fr"]), + ), + "images_list": other( + [ + str(Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"), + str(Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"), + ], + [Image()], + ), + "audios_list": other( + [ + {"array": [0.1, 0.2, 0.3], "sampling_rate": 16_000}, + {"array": [0.1, 0.2, 0.3], "sampling_rate": 16_000}, + ], + [Audio()], + ), + "images_sequence": other( + [ + str(Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"), + str(Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"), + ], + Sequence(feature=Image()), + ), + "audios_sequence": other( + [ + {"array": [0.1, 0.2, 0.3], "sampling_rate": 16_000}, + {"array": [0.1, 0.2, 0.3], "sampling_rate": 16_000}, + ], + Sequence(feature=Audio()), + ), + "dict_of_audios_and_images": other( + { + "a": 0, + "b": [ + str(Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"), + str(Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"), + ], + "c": { + "ca": [ + {"array": [0.1, 0.2, 0.3], "sampling_rate": 16_000}, + {"array": [0.1, 0.2, 0.3], "sampling_rate": 16_000}, + ] + }, + }, + {"a": Value(dtype="int64"), "b": [Image()], "c": {"ca": [Audio()]}}, + ), + "sequence_of_dicts": other( + [{"a": {"b": 0}}, {"a": {"b": 1}}], Sequence(feature={"a": {"b": Value(dtype="int64")}}) + ), + "none_value": other({"a": None}, {"a": Value(dtype="int64")}), + "big": Dataset.from_pandas( + pd.DataFrame({"col": ["a" * 1_234 for _ in range(4_567)]}, dtype=pd.StringDtype(storage="python")) + ), + } diff --git a/libs/libcommon/tests/viewer_utils/__init__.py b/libs/libcommon/tests/viewer_utils/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/services/worker/tests/test_features.py b/libs/libcommon/tests/viewer_utils/test_features.py similarity index 96% rename from services/worker/tests/test_features.py rename to libs/libcommon/tests/viewer_utils/test_features.py index ef6c076e..1d716bd3 100644 --- a/services/worker/tests/test_features.py +++ b/libs/libcommon/tests/viewer_utils/test_features.py @@ -11 +10,0 @@ from datasets import Audio, Dataset, Image, Value -from libcommon.storage import StrPath @@ -13,2 +12,2 @@ from libcommon.storage import StrPath -from worker.config import AppConfig -from worker.features import get_cell_value +from libcommon.storage import StrPath +from libcommon.viewer_utils.features import get_cell_value @@ -60,2 +59 @@ def test_value( - app_config: AppConfig, - assets_directory: StrPath, + cached_assets_directory: StrPath, @@ -75,2 +73,2 @@ def test_value( - assets_base_url=app_config.assets.base_url, - assets_directory=assets_directory, + assets_base_url="http://localhost/assets", + assets_directory=cached_assets_directory, @@ -302,2 +300 @@ def test_others( - app_config: AppConfig, - assets_directory: StrPath, + cached_assets_directory: StrPath, @@ -319,2 +316,2 @@ def test_others( - assets_base_url=app_config.assets.base_url, - assets_directory=assets_directory, + assets_base_url="http://localhost/assets", + assets_directory=cached_assets_directory, diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock index af90c2db..ab4087a0 100644 --- a/services/admin/poetry.lock +++ b/services/admin/poetry.lock @@ -1 +1,125 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. + +[[package]] +name = "aiohttp" +version = "3.8.4" +description = "Async http client/server framework (asyncio)" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"}, + {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"}, + {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"}, + {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"}, + {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"}, + {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"}, + {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"}, + {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"}, + {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"}, + {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"}, + {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"}, + {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"}, +] + +[package.dependencies] +aiosignal = ">=1.1.2" +async-timeout = ">=4.0.0a3,<5.0" +attrs = ">=17.3.0" +charset-normalizer = ">=2.0,<4.0" +frozenlist = ">=1.1.1" +multidict = ">=4.5,<7.0" +yarl = ">=1.0,<2.0" + +[package.extras] +speedups = ["Brotli", "aiodns", "cchardet"] + +[[package]] +name = "aiosignal" +version = "1.3.1" +description = "aiosignal: a list of registered asynchronous callbacks" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, + {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, +] + +[package.dependencies] +frozenlist = ">=1.1.0" @@ -35,0 +160,12 @@ files = [ +[[package]] +name = "async-timeout" +version = "4.0.2" +description = "Timeout context manager for asyncio programs" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"}, + {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"}, +] + @@ -40 +176 @@ description = "Classes Without Boilerplate" -category = "dev" +category = "main" @@ -54,0 +191,11 @@ tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy +[[package]] +name = "audioread" +version = "3.0.0" +description = "multi-library, cross-platform audio decoding" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"}, +] + @@ -146,0 +294,77 @@ files = [ +[[package]] +name = "cffi" +version = "1.15.1" +description = "Foreign Function Interface for Python calling C code." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"}, + {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"}, + {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"}, + {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"}, + {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"}, + {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"}, + {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"}, + {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"}, + {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"}, + {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"}, + {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"}, + {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"}, + {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"}, + {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"}, + {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"}, + {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"}, + {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"}, + {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"}, + {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"}, +] + +[package.dependencies] +pycparser = "*" + @@ -338 +562 @@ name = "cyclonedx-python-lib" -version = "3.1.5" +version = "2.7.1" @@ -344,2 +568,2 @@ files = [ - {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, - {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, + {file = "cyclonedx-python-lib-2.7.1.tar.gz", hash = "sha256:493bf2f30e26c48f305f745ed8580ce10d05a8d68d62a598fe95f05a0d9007dc"}, + {file = "cyclonedx_python_lib-2.7.1-py3-none-any.whl", hash = "sha256:fabc4c8baf722faeea01c3bbca83730e3489dfb37d85c6036baa67a9a7519d40"}, @@ -353,0 +578,74 @@ toml = ">=0.10.0,<0.11.0" +[[package]] +name = "datasets" +version = "2.11.0" +description = "HuggingFace community-driven open-source library of datasets" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "datasets-2.11.0-py3-none-any.whl", hash = "sha256:d946cdb8c4885d3016a2ab3129c9403dd3358fe9107e8ab5e549ceab672774af"}, + {file = "datasets-2.11.0.tar.gz", hash = "sha256:1ca53b9cd6ece7a3fdb81176dadd5b9e646420e52e68e85307b27db3a36ca18c"}, +] + +[package.dependencies] +aiohttp = "*" +dill = ">=0.3.0,<0.3.7" +fsspec = {version = ">=2021.11.1", extras = ["http"]} +huggingface-hub = ">=0.11.0,<1.0.0" +librosa = {version = "*", optional = true, markers = "extra == \"audio\""} +multiprocess = "*" +numpy = ">=1.17" +packaging = "*" +pandas = "*" +Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} +pyarrow = ">=8.0.0" +pyyaml = ">=5.1" +requests = ">=2.19.0" +responses = "<0.19" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} +tqdm = ">=4.62.1" +xxhash = "*" + +[package.extras] +apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] +audio = ["librosa", "soundfile (>=0.12.1)"] +benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +docs = ["s3fs"] +jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"] +s3 = ["s3fs"] +tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"] +tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +torch = ["torch"] +vision = ["Pillow (>=6.2.1)"] + +[[package]] +name = "decorator" +version = "5.1.1" +description = "Decorators for Humans" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"}, + {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"}, +] + +[[package]] +name = "dill" +version = "0.3.6" +description = "serialize all of python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"}, + {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"}, +] + +[package.extras] +graph = ["objgraph (>=1.7.2)"] + @@ -439,0 +738,122 @@ pyflakes = ">=2.3.0,<2.4.0" +[[package]] +name = "frozenlist" +version = "1.3.3" +description = "A list-like structure which implements collections.abc.MutableSequence" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"}, + {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"}, + {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"}, + {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"}, + {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"}, + {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"}, + {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"}, + {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"}, + {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"}, + {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"}, + {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"}, +] + +[[package]] +name = "fsspec" +version = "2023.3.0" +description = "File-system specification" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "fsspec-2023.3.0-py3-none-any.whl", hash = "sha256:bf57215e19dbfa4fe7edae53040cc1deef825e3b1605cca9a8d2c2fadd2328a0"}, + {file = "fsspec-2023.3.0.tar.gz", hash = "sha256:24e635549a590d74c6c18274ddd3ffab4753341753e923408b1904eaabafe04d"}, +] + +[package.dependencies] +aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""} +requests = {version = "*", optional = true, markers = "extra == \"http\""} + +[package.extras] +abfs = ["adlfs"] +adl = ["adlfs"] +arrow = ["pyarrow (>=1)"] +dask = ["dask", "distributed"] +dropbox = ["dropbox", "dropboxdrivefs", "requests"] +fuse = ["fusepy"] +gcs = ["gcsfs"] +git = ["pygit2"] +github = ["requests"] +gs = ["gcsfs"] +gui = ["panel"] +hdfs = ["pyarrow (>=1)"] +http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"] +libarchive = ["libarchive-c"] +oci = ["ocifs"] +s3 = ["s3fs"] +sftp = ["paramiko"] +smb = ["smbprotocol"] +ssh = ["paramiko"] +tqdm = ["tqdm"] + @@ -622,0 +1043,28 @@ requirements-deprecated-finder = ["pip-api", "pipreqs"] +[[package]] +name = "joblib" +version = "1.2.0" +description = "Lightweight pipelining with Python functions" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "joblib-1.2.0-py3-none-any.whl", hash = "sha256:091138ed78f800342968c523bdde947e7a305b8594b910a0fea2ab83c3c6d385"}, + {file = "joblib-1.2.0.tar.gz", hash = "sha256:e1cee4a79e4af22881164f218d4311f60074197fb707e082e803b61f6d137018"}, +] + +[[package]] +name = "lazy-loader" +version = "0.2" +description = "lazy_loader" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "lazy_loader-0.2-py3-none-any.whl", hash = "sha256:c35875f815c340f823ce3271ed645045397213f961b40ad0c0d395c3f5218eeb"}, + {file = "lazy_loader-0.2.tar.gz", hash = "sha256:0edc7a5175c400acb108f283749951fefdadedeb00adcec6e88b974a9254f18a"}, +] + +[package.extras] +lint = ["pre-commit (>=3.1)"] +test = ["pytest (>=7.2)", "pytest-cov (>=4.0)"] + @@ -634,0 +1083 @@ appdirs = "^1.4.4" +datasets = {version = "^2.11.0", extras = ["audio", "vision"]} @@ -639,0 +1089 @@ networkx = "^3.0" +numba = "0.56.4" @@ -641,0 +1092 @@ psutil = "^5.9.4" +pydub = "^0.25.1" @@ -644,0 +1096 @@ requests = "^2.28.2" +soundfile = ">=0.12.1" @@ -649,0 +1102,70 @@ url = "../../libs/libcommon" +[[package]] +name = "librosa" +version = "0.10.0.post2" +description = "Python module for audio and music processing" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"}, + {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"}, +] + +[package.dependencies] +audioread = ">=2.1.9" +decorator = ">=4.3.0" +joblib = ">=0.14" +lazy-loader = ">=0.1" +msgpack = ">=1.0" +numba = ">=0.51.0" +numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2" +pooch = ">=1.0,<1.7" +scikit-learn = ">=0.20.0" +scipy = ">=1.2.0" +soundfile = ">=0.12.1" +soxr = ">=0.3.2" +typing-extensions = ">=4.1.1" + +[package.extras] +display = ["matplotlib (>=3.3.0)"] +docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"] +tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"] + +[[package]] +name = "llvmlite" +version = "0.39.1" +description = "lightweight wrapper around basic LLVM functionality" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"}, + {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"}, + {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"}, + {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"}, + {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"}, + {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"}, + {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"}, + {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"}, + {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"}, + {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"}, +] + @@ -763 +1285 @@ description = "MessagePack serializer" -category = "dev" +category = "main" @@ -820,0 +1343,111 @@ files = [ +[[package]] +name = "multidict" +version = "6.0.4" +description = "multidict implementation" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"}, + {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"}, + {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"}, + {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"}, + {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"}, + {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"}, + {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"}, + {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"}, + {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"}, + {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"}, + {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"}, + {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"}, + {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"}, +] + +[[package]] +name = "multiprocess" +version = "0.70.14" +description = "better multiprocessing and multithreading in python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"}, + {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"}, + {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"}, + {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"}, + {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"}, + {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"}, +] + +[package.dependencies] +dill = ">=0.3.6" + @@ -898,0 +1532,81 @@ test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] +[[package]] +name = "numba" +version = "0.56.4" +description = "compiling Python code using LLVM" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"}, + {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"}, + {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"}, + {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"}, + {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"}, + {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"}, + {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"}, + {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"}, + {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"}, + {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"}, + {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"}, + {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"}, + {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"}, + {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"}, + {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"}, + {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"}, +] + +[package.dependencies] +llvmlite = ">=0.39.0dev0,<0.40" +numpy = ">=1.18,<1.24" +setuptools = "*" + +[[package]] +name = "numpy" +version = "1.23.5" +description = "NumPy is the fundamental package for array computing with Python." +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"}, + {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"}, + {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"}, + {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"}, + {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"}, + {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"}, + {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"}, + {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"}, + {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"}, + {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"}, +] + @@ -980,0 +1695,64 @@ files = [ +[[package]] +name = "pandas" +version = "2.0.0" +description = "Powerful data structures for data analysis, time series, and statistics" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "pandas-2.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:bbb2c5e94d6aa4e632646a3bacd05c2a871c3aa3e85c9bec9be99cb1267279f2"}, + {file = "pandas-2.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b5337c87c4e963f97becb1217965b6b75c6fe5f54c4cf09b9a5ac52fc0bd03d3"}, + {file = "pandas-2.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6ded51f7e3dd9b4f8b87f2ceb7bd1a8df2491f7ee72f7074c6927a512607199e"}, + {file = "pandas-2.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:52c858de9e9fc422d25e67e1592a6e6135d7bcf9a19fcaf4d0831a0be496bf21"}, + {file = "pandas-2.0.0-cp310-cp310-win32.whl", hash = "sha256:2d1d138848dd71b37e3cbe7cd952ff84e2ab04d8988972166e18567dcc811245"}, + {file = "pandas-2.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:d08e41d96bc4de6f500afe80936c68fce6099d5a434e2af7c7fd8e7c72a3265d"}, + {file = "pandas-2.0.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:24472cfc7ced511ac90608728b88312be56edc8f19b9ed885a7d2e47ffaf69c0"}, + {file = "pandas-2.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4ffb14f50c74ee541610668137830bb93e9dfa319b1bef2cedf2814cd5ac9c70"}, + {file = "pandas-2.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c24c7d12d033a372a9daf9ff2c80f8b0af6f98d14664dbb0a4f6a029094928a7"}, + {file = "pandas-2.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8318de0f886e4dcb8f9f36e45a3d6a6c3d1cfdc508354da85e739090f0222991"}, + {file = "pandas-2.0.0-cp311-cp311-win32.whl", hash = "sha256:57c34b79c13249505e850d0377b722961b99140f81dafbe6f19ef10239f6284a"}, + {file = "pandas-2.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:8f987ec26e96a8490909bc5d98c514147236e49830cba7df8690f6087c12bbae"}, + {file = "pandas-2.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b3ba8f5dd470d8bfbc4259829589f4a32881151c49e36384d9eb982b35a12020"}, + {file = "pandas-2.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:fcd471c9d9f60926ab2f15c6c29164112f458acb42280365fbefa542d0c2fc74"}, + {file = "pandas-2.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9253edfd015520ce77a9343eb7097429479c039cd3ebe81d7810ea11b4b24695"}, + {file = "pandas-2.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:977326039bd1ded620001a1889e2ed4798460a6bc5a24fbaebb5f07a41c32a55"}, + {file = "pandas-2.0.0-cp38-cp38-win32.whl", hash = "sha256:78425ca12314b23356c28b16765639db10ebb7d8983f705d6759ff7fe41357fa"}, + {file = "pandas-2.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:d93b7fcfd9f3328072b250d6d001dcfeec5d3bb66c1b9c8941e109a46c0c01a8"}, + {file = "pandas-2.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:425705cee8be54db2504e8dd2a730684790b15e5904b750c367611ede49098ab"}, + {file = "pandas-2.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a4f789b7c012a608c08cda4ff0872fd979cb18907a37982abe884e6f529b8793"}, + {file = "pandas-2.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3bb9d840bf15656805f6a3d87eea9dcb7efdf1314a82adcf7f00b820427c5570"}, + {file = "pandas-2.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0778ab54c8f399d83d98ffb674d11ec716449956bc6f6821891ab835848687f2"}, + {file = "pandas-2.0.0-cp39-cp39-win32.whl", hash = "sha256:70db5c278bbec0306d32bf78751ff56b9594c05a5098386f6c8a563659124f91"}, + {file = "pandas-2.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:4f3320bb55f34af4193020158ef8118ee0fb9aec7cc47d2084dbfdd868a0a24f"}, + {file = "pandas-2.0.0.tar.gz", hash = "sha256:cda9789e61b44463c1c4fe17ef755de77bcd13b09ba31c940d20f193d63a5dc8"}, +] + +[package.dependencies] +numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""} +python-dateutil = ">=2.8.2" +pytz = ">=2020.1" +tzdata = ">=2022.1" + +[package.extras] +all = ["PyQt5 (>=5.15.1)", "SQLAlchemy (>=1.4.16)", "beautifulsoup4 (>=4.9.3)", "bottleneck (>=1.3.2)", "brotlipy (>=0.7.0)", "fastparquet (>=0.6.3)", "fsspec (>=2021.07.0)", "gcsfs (>=2021.07.0)", "html5lib (>=1.1)", "hypothesis (>=6.34.2)", "jinja2 (>=3.0.0)", "lxml (>=4.6.3)", "matplotlib (>=3.6.1)", "numba (>=0.53.1)", "numexpr (>=2.7.3)", "odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pandas-gbq (>=0.15.0)", "psycopg2 (>=2.8.6)", "pyarrow (>=7.0.0)", "pymysql (>=1.0.2)", "pyreadstat (>=1.1.2)", "pytest (>=7.0.0)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)", "python-snappy (>=0.6.0)", "pyxlsb (>=1.0.8)", "qtpy (>=2.2.0)", "s3fs (>=2021.08.0)", "scipy (>=1.7.1)", "tables (>=3.6.1)", "tabulate (>=0.8.9)", "xarray (>=0.21.0)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)", "zstandard (>=0.15.2)"] +aws = ["s3fs (>=2021.08.0)"] +clipboard = ["PyQt5 (>=5.15.1)", "qtpy (>=2.2.0)"] +compression = ["brotlipy (>=0.7.0)", "python-snappy (>=0.6.0)", "zstandard (>=0.15.2)"] +computation = ["scipy (>=1.7.1)", "xarray (>=0.21.0)"] +excel = ["odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pyxlsb (>=1.0.8)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)"] +feather = ["pyarrow (>=7.0.0)"] +fss = ["fsspec (>=2021.07.0)"] +gcp = ["gcsfs (>=2021.07.0)", "pandas-gbq (>=0.15.0)"] +hdf5 = ["tables (>=3.6.1)"] +html = ["beautifulsoup4 (>=4.9.3)", "html5lib (>=1.1)", "lxml (>=4.6.3)"] +mysql = ["SQLAlchemy (>=1.4.16)", "pymysql (>=1.0.2)"] +output-formatting = ["jinja2 (>=3.0.0)", "tabulate (>=0.8.9)"] +parquet = ["pyarrow (>=7.0.0)"] +performance = ["bottleneck (>=1.3.2)", "numba (>=0.53.1)", "numexpr (>=2.7.1)"] +plot = ["matplotlib (>=3.6.1)"] +postgresql = ["SQLAlchemy (>=1.4.16)", "psycopg2 (>=2.8.6)"] +spss = ["pyreadstat (>=1.1.2)"] +sql-other = ["SQLAlchemy (>=1.4.16)"] +test = ["hypothesis (>=6.34.2)", "pytest (>=7.0.0)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)"] +xml = ["lxml (>=4.6.3)"] + @@ -1004,0 +1783,80 @@ files = [ +[[package]] +name = "pillow" +version = "9.5.0" +description = "Python Imaging Library (Fork)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pillow-9.5.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:ace6ca218308447b9077c14ea4ef381ba0b67ee78d64046b3f19cf4e1139ad16"}, + {file = "Pillow-9.5.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d3d403753c9d5adc04d4694d35cf0391f0f3d57c8e0030aac09d7678fa8030aa"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5ba1b81ee69573fe7124881762bb4cd2e4b6ed9dd28c9c60a632902fe8db8b38"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe7e1c262d3392afcf5071df9afa574544f28eac825284596ac6db56e6d11062"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f36397bf3f7d7c6a3abdea815ecf6fd14e7fcd4418ab24bae01008d8d8ca15e"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:252a03f1bdddce077eff2354c3861bf437c892fb1832f75ce813ee94347aa9b5"}, + {file = "Pillow-9.5.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:85ec677246533e27770b0de5cf0f9d6e4ec0c212a1f89dfc941b64b21226009d"}, + {file = "Pillow-9.5.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:b416f03d37d27290cb93597335a2f85ed446731200705b22bb927405320de903"}, + {file = "Pillow-9.5.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:1781a624c229cb35a2ac31cc4a77e28cafc8900733a864870c49bfeedacd106a"}, + {file = "Pillow-9.5.0-cp310-cp310-win32.whl", hash = "sha256:8507eda3cd0608a1f94f58c64817e83ec12fa93a9436938b191b80d9e4c0fc44"}, + {file = "Pillow-9.5.0-cp310-cp310-win_amd64.whl", hash = "sha256:d3c6b54e304c60c4181da1c9dadf83e4a54fd266a99c70ba646a9baa626819eb"}, + {file = "Pillow-9.5.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:7ec6f6ce99dab90b52da21cf0dc519e21095e332ff3b399a357c187b1a5eee32"}, + {file = "Pillow-9.5.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:560737e70cb9c6255d6dcba3de6578a9e2ec4b573659943a5e7e4af13f298f5c"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:96e88745a55b88a7c64fa49bceff363a1a27d9a64e04019c2281049444a571e3"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9c206c29b46cfd343ea7cdfe1232443072bbb270d6a46f59c259460db76779a"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cfcc2c53c06f2ccb8976fb5c71d448bdd0a07d26d8e07e321c103416444c7ad1"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:a0f9bb6c80e6efcde93ffc51256d5cfb2155ff8f78292f074f60f9e70b942d99"}, + {file = "Pillow-9.5.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:8d935f924bbab8f0a9a28404422da8af4904e36d5c33fc6f677e4c4485515625"}, + {file = "Pillow-9.5.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:fed1e1cf6a42577953abbe8e6cf2fe2f566daebde7c34724ec8803c4c0cda579"}, + {file = "Pillow-9.5.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:c1170d6b195555644f0616fd6ed929dfcf6333b8675fcca044ae5ab110ded296"}, + {file = "Pillow-9.5.0-cp311-cp311-win32.whl", hash = "sha256:54f7102ad31a3de5666827526e248c3530b3a33539dbda27c6843d19d72644ec"}, + {file = "Pillow-9.5.0-cp311-cp311-win_amd64.whl", hash = "sha256:cfa4561277f677ecf651e2b22dc43e8f5368b74a25a8f7d1d4a3a243e573f2d4"}, + {file = "Pillow-9.5.0-cp311-cp311-win_arm64.whl", hash = "sha256:965e4a05ef364e7b973dd17fc765f42233415974d773e82144c9bbaaaea5d089"}, + {file = "Pillow-9.5.0-cp312-cp312-win32.whl", hash = "sha256:22baf0c3cf0c7f26e82d6e1adf118027afb325e703922c8dfc1d5d0156bb2eeb"}, + {file = "Pillow-9.5.0-cp312-cp312-win_amd64.whl", hash = "sha256:432b975c009cf649420615388561c0ce7cc31ce9b2e374db659ee4f7d57a1f8b"}, + {file = "Pillow-9.5.0-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:5d4ebf8e1db4441a55c509c4baa7a0587a0210f7cd25fcfe74dbbce7a4bd1906"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:375f6e5ee9620a271acb6820b3d1e94ffa8e741c0601db4c0c4d3cb0a9c224bf"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:99eb6cafb6ba90e436684e08dad8be1637efb71c4f2180ee6b8f940739406e78"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2dfaaf10b6172697b9bceb9a3bd7b951819d1ca339a5ef294d1f1ac6d7f63270"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_28_aarch64.whl", hash = "sha256:763782b2e03e45e2c77d7779875f4432e25121ef002a41829d8868700d119392"}, + {file = "Pillow-9.5.0-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:35f6e77122a0c0762268216315bf239cf52b88865bba522999dc38f1c52b9b47"}, + {file = "Pillow-9.5.0-cp37-cp37m-win32.whl", hash = "sha256:aca1c196f407ec7cf04dcbb15d19a43c507a81f7ffc45b690899d6a76ac9fda7"}, + {file = "Pillow-9.5.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322724c0032af6692456cd6ed554bb85f8149214d97398bb80613b04e33769f6"}, + {file = "Pillow-9.5.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:a0aa9417994d91301056f3d0038af1199eb7adc86e646a36b9e050b06f526597"}, + {file = "Pillow-9.5.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:f8286396b351785801a976b1e85ea88e937712ee2c3ac653710a4a57a8da5d9c"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c830a02caeb789633863b466b9de10c015bded434deb3ec87c768e53752ad22a"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbd359831c1657d69bb81f0db962905ee05e5e9451913b18b831febfe0519082"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f8fc330c3370a81bbf3f88557097d1ea26cd8b019d6433aa59f71195f5ddebbf"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:7002d0797a3e4193c7cdee3198d7c14f92c0836d6b4a3f3046a64bd1ce8df2bf"}, + {file = "Pillow-9.5.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:229e2c79c00e85989a34b5981a2b67aa079fd08c903f0aaead522a1d68d79e51"}, + {file = "Pillow-9.5.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:9adf58f5d64e474bed00d69bcd86ec4bcaa4123bfa70a65ce72e424bfb88ed96"}, + {file = "Pillow-9.5.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:662da1f3f89a302cc22faa9f14a262c2e3951f9dbc9617609a47521c69dd9f8f"}, + {file = "Pillow-9.5.0-cp38-cp38-win32.whl", hash = "sha256:6608ff3bf781eee0cd14d0901a2b9cc3d3834516532e3bd673a0a204dc8615fc"}, + {file = "Pillow-9.5.0-cp38-cp38-win_amd64.whl", hash = "sha256:e49eb4e95ff6fd7c0c402508894b1ef0e01b99a44320ba7d8ecbabefddcc5569"}, + {file = "Pillow-9.5.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:482877592e927fd263028c105b36272398e3e1be3269efda09f6ba21fd83ec66"}, + {file = "Pillow-9.5.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3ded42b9ad70e5f1754fb7c2e2d6465a9c842e41d178f262e08b8c85ed8a1d8e"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c446d2245ba29820d405315083d55299a796695d747efceb5717a8b450324115"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8aca1152d93dcc27dc55395604dcfc55bed5f25ef4c98716a928bacba90d33a3"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:608488bdcbdb4ba7837461442b90ea6f3079397ddc968c31265c1e056964f1ef"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:60037a8db8750e474af7ffc9faa9b5859e6c6d0a50e55c45576bf28be7419705"}, + {file = "Pillow-9.5.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:07999f5834bdc404c442146942a2ecadd1cb6292f5229f4ed3b31e0a108746b1"}, + {file = "Pillow-9.5.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:a127ae76092974abfbfa38ca2d12cbeddcdeac0fb71f9627cc1135bedaf9d51a"}, + {file = "Pillow-9.5.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:489f8389261e5ed43ac8ff7b453162af39c3e8abd730af8363587ba64bb2e865"}, + {file = "Pillow-9.5.0-cp39-cp39-win32.whl", hash = "sha256:9b1af95c3a967bf1da94f253e56b6286b50af23392a886720f563c547e48e964"}, + {file = "Pillow-9.5.0-cp39-cp39-win_amd64.whl", hash = "sha256:77165c4a5e7d5a284f10a6efaa39a0ae8ba839da344f20b111d62cc932fa4e5d"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:833b86a98e0ede388fa29363159c9b1a294b0905b5128baf01db683672f230f5"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:aaf305d6d40bd9632198c766fb64f0c1a83ca5b667f16c1e79e1661ab5060140"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0852ddb76d85f127c135b6dd1f0bb88dbb9ee990d2cd9aa9e28526c93e794fba"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:91ec6fe47b5eb5a9968c79ad9ed78c342b1f97a091677ba0e012701add857829"}, + {file = "Pillow-9.5.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cb841572862f629b99725ebaec3287fc6d275be9b14443ea746c1dd325053cbd"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:c380b27d041209b849ed246b111b7c166ba36d7933ec6e41175fd15ab9eb1572"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7c9af5a3b406a50e313467e3565fc99929717f780164fe6fbb7704edba0cebbe"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5671583eab84af046a397d6d0ba25343c00cd50bce03787948e0fff01d4fd9b1"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:84a6f19ce086c1bf894644b43cd129702f781ba5751ca8572f08aa40ef0ab7b7"}, + {file = "Pillow-9.5.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:1e7723bd90ef94eda669a3c2c19d549874dd5badaeefabefd26053304abe5799"}, + {file = "Pillow-9.5.0.tar.gz", hash = "sha256:bf548479d336726d7a0eceb6e767e179fbde37833ae42794602631a070d630f1"}, +] + +[package.extras] +docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-removed-in", "sphinxext-opengraph"] +tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"] + @@ -1034 +1892 @@ name = "pip-audit" -version = "2.4.14" +version = "2.5.4" @@ -1040,2 +1898,2 @@ files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, + {file = "pip_audit-2.5.4-py3-none-any.whl", hash = "sha256:bee3748030c895488b4dd8a6196fa44f484da000cfd52d8fc64dfa3a2e121624"}, + {file = "pip_audit-2.5.4.tar.gz", hash = "sha256:a4cb03f9e2896d626f5b153973d3ac0d32fdb18594d78d393b153c83bb8089b6"}, @@ -1046 +1904 @@ CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +cyclonedx-python-lib = ">=2.0,<2.5.0 || >2.5.0,<3.0" @@ -1051 +1908,0 @@ pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" @@ -1058 +1915 @@ doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.260)", "types-html5lib", "types-requests", "types-toml"] @@ -1112,0 +1970,22 @@ testing = ["pytest", "pytest-benchmark"] +[[package]] +name = "pooch" +version = "1.6.0" +description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\"" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"}, + {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"}, +] + +[package.dependencies] +appdirs = ">=1.3.0" +packaging = ">=20.0" +requests = ">=2.19.0" + +[package.extras] +progress = ["tqdm (>=4.41.0,<5.0.0)"] +sftp = ["paramiko (>=2.7.0)"] +xxhash = ["xxhash (>=1.4.3)"] + @@ -1154,0 +2034,38 @@ test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] +[[package]] +name = "pyarrow" +version = "11.0.0" +description = "Python library for Apache Arrow" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"}, + {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"}, + {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"}, + {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"}, + {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"}, + {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"}, + {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"}, + {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"}, +] + +[package.dependencies] +numpy = ">=1.16.6" + @@ -1166,0 +2084,24 @@ files = [ +[[package]] +name = "pycparser" +version = "2.21" +description = "C parser in Python" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, + {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, +] + +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + @@ -1384,0 +2326,15 @@ testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtuale +[[package]] +name = "python-dateutil" +version = "2.8.2" +description = "Extensions to the standard Python datetime module" +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" +files = [ + {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"}, + {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"}, +] + +[package.dependencies] +six = ">=1.5" + @@ -1484,18 +2439,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "resolvelib" -version = "0.9.0" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"}, - {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"}, -] - -[package.extras] -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] - @@ -1504 +2442 @@ name = "responses" -version = "0.22.0" +version = "0.18.0" @@ -1506 +2444 @@ description = "A utility library for mocking out the `requests` Python library." -category = "dev" +category = "main" @@ -1510,2 +2448,2 @@ files = [ - {file = "responses-0.22.0-py3-none-any.whl", hash = "sha256:dcf294d204d14c436fddcc74caefdbc5764795a40ff4e6a7740ed8ddbf3294be"}, - {file = "responses-0.22.0.tar.gz", hash = "sha256:396acb2a13d25297789a5866b4881cf4e46ffd49cc26c43ab1117f40b973102e"}, + {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, + {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, @@ -1515,3 +2453 @@ files = [ -requests = ">=2.22.0,<3.0" -toml = "*" -types-toml = "*" +requests = ">=2.0,<3.0" @@ -1521 +2457 @@ urllib3 = ">=1.25.10" -tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=7.0.0)", "pytest-asyncio", "pytest-cov", "pytest-httpserver", "types-requests"] +tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] @@ -1559,0 +2496,82 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] +[[package]] +name = "scikit-learn" +version = "1.2.2" +description = "A set of python modules for machine learning and data mining" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "scikit-learn-1.2.2.tar.gz", hash = "sha256:8429aea30ec24e7a8c7ed8a3fa6213adf3814a6efbea09e16e0a0c71e1a1a3d7"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:99cc01184e347de485bf253d19fcb3b1a3fb0ee4cea5ee3c43ec0cc429b6d29f"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:e6e574db9914afcb4e11ade84fab084536a895ca60aadea3041e85b8ac963edb"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6fe83b676f407f00afa388dd1fdd49e5c6612e551ed84f3b1b182858f09e987d"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e2642baa0ad1e8f8188917423dd73994bf25429f8893ddbe115be3ca3183584"}, + {file = "scikit_learn-1.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:ad66c3848c0a1ec13464b2a95d0a484fd5b02ce74268eaa7e0c697b904f31d6c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:dfeaf8be72117eb61a164ea6fc8afb6dfe08c6f90365bde2dc16456e4bc8e45f"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:fe0aa1a7029ed3e1dcbf4a5bc675aa3b1bc468d9012ecf6c6f081251ca47f590"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:065e9673e24e0dc5113e2dd2b4ca30c9d8aa2fa90f4c0597241c93b63130d233"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf036ea7ef66115e0d49655f16febfa547886deba20149555a41d28f56fd6d3c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:8b0670d4224a3c2d596fd572fb4fa673b2a0ccfb07152688ebd2ea0b8c61025c"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9c710ff9f9936ba8a3b74a455ccf0dcf59b230caa1e9ba0223773c490cab1e51"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:2dd3ffd3950e3d6c0c0ef9033a9b9b32d910c61bd06cb8206303fb4514b88a49"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:44b47a305190c28dd8dd73fc9445f802b6ea716669cfc22ab1eb97b335d238b1"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:953236889928d104c2ef14027539f5f2609a47ebf716b8cbe4437e85dce42744"}, + {file = "scikit_learn-1.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:7f69313884e8eb311460cc2f28676d5e400bd929841a2c8eb8742ae78ebf7c20"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8156db41e1c39c69aa2d8599ab7577af53e9e5e7a57b0504e116cc73c39138dd"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:fe175ee1dab589d2e1033657c5b6bec92a8a3b69103e3dd361b58014729975c3"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d5312d9674bed14f73773d2acf15a3272639b981e60b72c9b190a0cffed5bad"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ea061bf0283bf9a9f36ea3c5d3231ba2176221bbd430abd2603b1c3b2ed85c89"}, + {file = "scikit_learn-1.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:6477eed40dbce190f9f9e9d0d37e020815825b300121307942ec2110302b66a3"}, +] + +[package.dependencies] +joblib = ">=1.1.1" +numpy = ">=1.17.3" +scipy = ">=1.3.2" +threadpoolctl = ">=2.0.0" + +[package.extras] +benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"] +docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=4.0.1)", "sphinx-gallery (>=0.7.0)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"] +examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"] +tests = ["black (>=22.3.0)", "flake8 (>=3.8.2)", "matplotlib (>=3.1.3)", "mypy (>=0.961)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=5.3.1)", "pytest-cov (>=2.9.0)", "scikit-image (>=0.16.2)"] + +[[package]] +name = "scipy" +version = "1.10.1" +description = "Fundamental algorithms for scientific computing in Python" +category = "main" +optional = false +python-versions = "<3.12,>=3.8" +files = [ + {file = "scipy-1.10.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e7354fd7527a4b0377ce55f286805b34e8c54b91be865bac273f527e1b839019"}, + {file = "scipy-1.10.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:4b3f429188c66603a1a5c549fb414e4d3bdc2a24792e061ffbd607d3d75fd84e"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1553b5dcddd64ba9a0d95355e63fe6c3fc303a8fd77c7bc91e77d61363f7433f"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c0ff64b06b10e35215abce517252b375e580a6125fd5fdf6421b98efbefb2d2"}, + {file = "scipy-1.10.1-cp310-cp310-win_amd64.whl", hash = "sha256:fae8a7b898c42dffe3f7361c40d5952b6bf32d10c4569098d276b4c547905ee1"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:0f1564ea217e82c1bbe75ddf7285ba0709ecd503f048cb1236ae9995f64217bd"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d925fa1c81b772882aa55bcc10bf88324dadb66ff85d548c71515f6689c6dac5"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaea0a6be54462ec027de54fca511540980d1e9eea68b2d5c1dbfe084797be35"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15a35c4242ec5f292c3dd364a7c71a61be87a3d4ddcc693372813c0b73c9af1d"}, + {file = "scipy-1.10.1-cp311-cp311-win_amd64.whl", hash = "sha256:43b8e0bcb877faf0abfb613d51026cd5cc78918e9530e375727bf0625c82788f"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5678f88c68ea866ed9ebe3a989091088553ba12c6090244fdae3e467b1139c35"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:39becb03541f9e58243f4197584286e339029e8908c46f7221abeea4b749fa88"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bce5869c8d68cf383ce240e44c1d9ae7c06078a9396df68ce88a1230f93a30c1"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:07c3457ce0b3ad5124f98a86533106b643dd811dd61b548e78cf4c8786652f6f"}, + {file = "scipy-1.10.1-cp38-cp38-win_amd64.whl", hash = "sha256:049a8bbf0ad95277ffba9b3b7d23e5369cc39e66406d60422c8cfef40ccc8415"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cd9f1027ff30d90618914a64ca9b1a77a431159df0e2a195d8a9e8a04c78abf9"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:79c8e5a6c6ffaf3a2262ef1be1e108a035cf4f05c14df56057b64acc5bebffb6"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:51af417a000d2dbe1ec6c372dfe688e041a7084da4fdd350aeb139bd3fb55353"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b4735d6c28aad3cdcf52117e0e91d6b39acd4272f3f5cd9907c24ee931ad601"}, + {file = "scipy-1.10.1-cp39-cp39-win_amd64.whl", hash = "sha256:7ff7f37b1bf4417baca958d254e8e2875d0cc23aaadbe65b3d5b3077b0eb23ea"}, + {file = "scipy-1.10.1.tar.gz", hash = "sha256:2cf9dfb80a7b4589ba4c40ce7588986d6d5cebc5457cad2c2880f6bc2d42f3a5"}, +] + +[package.dependencies] +numpy = ">=1.19.5,<1.27.0" + +[package.extras] +dev = ["click", "doit (>=0.36.0)", "flake8", "mypy", "pycodestyle", "pydevtool", "rich-click", "typing_extensions"] +doc = ["matplotlib (>2)", "numpydoc", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"] +test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"] + @@ -1564 +2582 @@ description = "Easily download, build, install, upgrade, and uninstall Python pa -category = "dev" +category = "main" @@ -1581 +2599 @@ description = "Python 2 and 3 compatibility utilities" -category = "dev" +category = "main" @@ -1624,0 +2643,68 @@ files = [ +[[package]] +name = "soundfile" +version = "0.12.1" +description = "An audio library based on libsndfile, CFFI and NumPy" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"}, + {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"}, + {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"}, + {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"}, +] + +[package.dependencies] +cffi = ">=1.0" + +[package.extras] +numpy = ["numpy"] + +[[package]] +name = "soxr" +version = "0.3.4" +description = "High quality, one-dimensional sample-rate conversion library" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "soxr-0.3.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b7b84126643c063d5daa203f7f9137e21734dabbd7e68c097607b2ef457e2f2e"}, + {file = "soxr-0.3.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:380d2d43871a68e8b1ef1702a0abe6f9e48ddb3933c7a303c45d67e121503e7c"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4a1b4019c9972f57612482c4f85523d6e832e3d10935e2f070a9dcd334a4dcb"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e613cee023b7c3f162b9da3f6b169cd7f58de345275be1fde9f19adc9cf144df"}, + {file = "soxr-0.3.4-cp310-cp310-win32.whl", hash = "sha256:182c02a7ba45a159a0dbb0a297335df2381ead03a65377b19663ea0ff720ecb7"}, + {file = "soxr-0.3.4-cp310-cp310-win_amd64.whl", hash = "sha256:1e95c96ce94524fae453b4331c9910d33f97506f99bae06d76a9c0649710619e"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2678d2719e7496803983584d661eb5fddc7017154a8dda4a774407c56ff07973"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:11bd1396052049e6d389225a0e96a9df15f706da501c619b35d3c72ac6bc7257"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7e23de4dfe54ac30e583bbc9cc3feda1cd776fedce13206bc4b3115b75ecab82"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3e7396498a5f5b7d8f23b656f65c24517a6ff5bdc3ee0623ccd491036a43ea08"}, + {file = "soxr-0.3.4-cp311-cp311-win32.whl", hash = "sha256:e57e9703c2bff834cabc06800d3c11a259544891d2c24a78949f3cf2f5492cc5"}, + {file = "soxr-0.3.4-cp311-cp311-win_amd64.whl", hash = "sha256:7c8350acd7150f74261a0569b47ccb1bb4aa39b2d575860bc97cfa69aab8aead"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:941f7355116fe77fe6a82938fa7799a0e466a494ebc093f676969ce32b2815b1"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:00fdbf24f64d8c3fb800425c383048cb24c32defac80901cde4a57fb6ce5d431"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7bb6d4dc807d04c536674429e2b05ae08a1efac9815c4595e41ffd6b57c2c662"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ff15853895b54f1b627799c6600be1ce5f7286724e7a93e4b7b9d79e5d4166f5"}, + {file = "soxr-0.3.4-cp38-cp38-win32.whl", hash = "sha256:d858becbc1fcc7b38c3436d3276290fae09403cdcbdf1d5986a18dab7023a6c3"}, + {file = "soxr-0.3.4-cp38-cp38-win_amd64.whl", hash = "sha256:068ab4df549df5783cc1eb4eb6c94f53823b164dc27134fc621fc9f5097f38cd"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:20130329985f9767c8417bbd125fe138790a71802b000481c386a800e2ad2bca"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:78090e97abfb326b7cf14ef37d08a17252b07d438388dcbbd82a6836a9d551b1"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84e590e75b7e5dca12bf68bfb090276f34a88fbcd793781c62d47f5d7dbe525e"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3479d265574b960e12bca0878baba0862c43278915e0319d84679bb4d4fcd33"}, + {file = "soxr-0.3.4-cp39-cp39-win32.whl", hash = "sha256:83de825d6a713c7b2e76d9ec3f229a58a9ed290237e7adc05d80e8b39be995a6"}, + {file = "soxr-0.3.4-cp39-cp39-win_amd64.whl", hash = "sha256:2082f88cae89de854c3e0d62f55d0cb31eb11764f5c2a28299121fb642a22472"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:fe8b5f92c802f1e7793c40344f5368dc6163718c9ffa82e79ee6ad779d318ac5"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d0063d5f9a4e1a367084f4705301e9da131cf4d2d32aa3fe0072a1245e18088f"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:a680bab57adae462cdc86abcc7330beb5daa3ba5101165583eedcda88b7ba551"}, + {file = "soxr-0.3.4.tar.gz", hash = "sha256:fe68daf00e8f020977b187699903d219f9e39b9fb3d915f3f923eed8ba431449"}, +] + +[package.dependencies] +numpy = "*" + +[package.extras] +docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"] +test = ["pytest"] + @@ -1674,0 +2761,12 @@ pbr = ">=2.0.0,<2.1.0 || >2.1.0" +[[package]] +name = "threadpoolctl" +version = "3.1.0" +description = "threadpoolctl" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, + {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"}, +] + @@ -1747,12 +2844,0 @@ types-urllib3 = "<1.27" -[[package]] -name = "types-toml" -version = "0.10.8.5" -description = "Typing stubs for toml" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "types-toml-0.10.8.5.tar.gz", hash = "sha256:bf80fce7d2d74be91148f47b88d9ae5adeb1024abef22aa2fdbabc036d6b8b3c"}, - {file = "types_toml-0.10.8.5-py3-none-any.whl", hash = "sha256:2432017febe43174af0f3c65f03116e3d3cf43e7e1406b8200e106da8cf98992"}, -] - @@ -1782,0 +2869,12 @@ files = [ +[[package]] +name = "tzdata" +version = "2023.3" +description = "Provider of IANA time zone data" +category = "main" +optional = false +python-versions = ">=2" +files = [ + {file = "tzdata-2023.3-py2.py3-none-any.whl", hash = "sha256:7e65763eef3120314099b6939b5546db7adce1e7d6f2e179e3df563c70511eda"}, + {file = "tzdata-2023.3.tar.gz", hash = "sha256:11ef1e08e54acb0d4f95bdb1be05da659673de4acbd21bf9c69e94cc5e907a3a"}, +] + @@ -1874,0 +2973,196 @@ files = [ +[[package]] +name = "xxhash" +version = "3.2.0" +description = "Python binding for xxHash" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"}, + {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"}, + {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"}, + {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"}, + {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"}, + {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"}, + {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"}, + {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"}, + {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"}, + {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"}, + {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"}, + {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"}, + {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"}, + {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"}, + {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"}, + {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"}, + {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"}, +] + +[[package]] +name = "yarl" +version = "1.8.2" +description = "Yet another URL library" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:bb81f753c815f6b8e2ddd2eef3c855cf7da193b82396ac013c661aaa6cc6b0a5"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:47d49ac96156f0928f002e2424299b2c91d9db73e08c4cd6742923a086f1c863"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3fc056e35fa6fba63248d93ff6e672c096f95f7836938241ebc8260e062832fe"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58a3c13d1c3005dbbac5c9f0d3210b60220a65a999b1833aa46bd6677c69b08e"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:10b08293cda921157f1e7c2790999d903b3fd28cd5c208cf8826b3b508026996"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:de986979bbd87272fe557e0a8fcb66fd40ae2ddfe28a8b1ce4eae22681728fef"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c4fcfa71e2c6a3cb568cf81aadc12768b9995323186a10827beccf5fa23d4f8"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae4d7ff1049f36accde9e1ef7301912a751e5bae0a9d142459646114c70ecba6"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:bf071f797aec5b96abfc735ab97da9fd8f8768b43ce2abd85356a3127909d146"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:74dece2bfc60f0f70907c34b857ee98f2c6dd0f75185db133770cd67300d505f"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:df60a94d332158b444301c7f569659c926168e4d4aad2cfbf4bce0e8fb8be826"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:63243b21c6e28ec2375f932a10ce7eda65139b5b854c0f6b82ed945ba526bff3"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:cfa2bbca929aa742b5084fd4663dd4b87c191c844326fcb21c3afd2d11497f80"}, + {file = "yarl-1.8.2-cp310-cp310-win32.whl", hash = "sha256:b05df9ea7496df11b710081bd90ecc3a3db6adb4fee36f6a411e7bc91a18aa42"}, + {file = "yarl-1.8.2-cp310-cp310-win_amd64.whl", hash = "sha256:24ad1d10c9db1953291f56b5fe76203977f1ed05f82d09ec97acb623a7976574"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:2a1fca9588f360036242f379bfea2b8b44cae2721859b1c56d033adfd5893634"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:f37db05c6051eff17bc832914fe46869f8849de5b92dc4a3466cd63095d23dfd"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:77e913b846a6b9c5f767b14dc1e759e5aff05502fe73079f6f4176359d832581"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0978f29222e649c351b173da2b9b4665ad1feb8d1daa9d971eb90df08702668a"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:388a45dc77198b2460eac0aca1efd6a7c09e976ee768b0d5109173e521a19daf"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2305517e332a862ef75be8fad3606ea10108662bc6fe08509d5ca99503ac2aee"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:42430ff511571940d51e75cf42f1e4dbdded477e71c1b7a17f4da76c1da8ea76"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3150078118f62371375e1e69b13b48288e44f6691c1069340081c3fd12c94d5b"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:c15163b6125db87c8f53c98baa5e785782078fbd2dbeaa04c6141935eb6dab7a"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:4d04acba75c72e6eb90745447d69f84e6c9056390f7a9724605ca9c56b4afcc6"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:e7fd20d6576c10306dea2d6a5765f46f0ac5d6f53436217913e952d19237efc4"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:75c16b2a900b3536dfc7014905a128a2bea8fb01f9ee26d2d7d8db0a08e7cb2c"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:6d88056a04860a98341a0cf53e950e3ac9f4e51d1b6f61a53b0609df342cc8b2"}, + {file = "yarl-1.8.2-cp311-cp311-win32.whl", hash = "sha256:fb742dcdd5eec9f26b61224c23baea46c9055cf16f62475e11b9b15dfd5c117b"}, + {file = "yarl-1.8.2-cp311-cp311-win_amd64.whl", hash = "sha256:8c46d3d89902c393a1d1e243ac847e0442d0196bbd81aecc94fcebbc2fd5857c"}, + {file = "yarl-1.8.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:ceff9722e0df2e0a9e8a79c610842004fa54e5b309fe6d218e47cd52f791d7ef"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3f6b4aca43b602ba0f1459de647af954769919c4714706be36af670a5f44c9c1"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1684a9bd9077e922300ecd48003ddae7a7474e0412bea38d4631443a91d61077"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ebb78745273e51b9832ef90c0898501006670d6e059f2cdb0e999494eb1450c2"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3adeef150d528ded2a8e734ebf9ae2e658f4c49bf413f5f157a470e17a4a2e89"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:57a7c87927a468e5a1dc60c17caf9597161d66457a34273ab1760219953f7f4c"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:efff27bd8cbe1f9bd127e7894942ccc20c857aa8b5a0327874f30201e5ce83d0"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:a783cd344113cb88c5ff7ca32f1f16532a6f2142185147822187913eb989f739"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:705227dccbe96ab02c7cb2c43e1228e2826e7ead880bb19ec94ef279e9555b5b"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:34c09b43bd538bf6c4b891ecce94b6fa4f1f10663a8d4ca589a079a5018f6ed7"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:a48f4f7fea9a51098b02209d90297ac324241bf37ff6be6d2b0149ab2bd51b37"}, + {file = "yarl-1.8.2-cp37-cp37m-win32.whl", hash = "sha256:0414fd91ce0b763d4eadb4456795b307a71524dbacd015c657bb2a39db2eab89"}, + {file = "yarl-1.8.2-cp37-cp37m-win_amd64.whl", hash = "sha256:d881d152ae0007809c2c02e22aa534e702f12071e6b285e90945aa3c376463c5"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5df5e3d04101c1e5c3b1d69710b0574171cc02fddc4b23d1b2813e75f35a30b1"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7a66c506ec67eb3159eea5096acd05f5e788ceec7b96087d30c7d2865a243918"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2b4fa2606adf392051d990c3b3877d768771adc3faf2e117b9de7eb977741229"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1e21fb44e1eff06dd6ef971d4bdc611807d6bd3691223d9c01a18cec3677939e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:93202666046d9edadfe9f2e7bf5e0782ea0d497b6d63da322e541665d65a044e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fc77086ce244453e074e445104f0ecb27530d6fd3a46698e33f6c38951d5a0f1"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:64dd68a92cab699a233641f5929a40f02a4ede8c009068ca8aa1fe87b8c20ae3"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1b372aad2b5f81db66ee7ec085cbad72c4da660d994e8e590c997e9b01e44901"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e6f3515aafe0209dd17fb9bdd3b4e892963370b3de781f53e1746a521fb39fc0"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:dfef7350ee369197106805e193d420b75467b6cceac646ea5ed3049fcc950a05"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:728be34f70a190566d20aa13dc1f01dc44b6aa74580e10a3fb159691bc76909d"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ff205b58dc2929191f68162633d5e10e8044398d7a45265f90a0f1d51f85f72c"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:baf211dcad448a87a0d9047dc8282d7de59473ade7d7fdf22150b1d23859f946"}, + {file = "yarl-1.8.2-cp38-cp38-win32.whl", hash = "sha256:272b4f1599f1b621bf2aabe4e5b54f39a933971f4e7c9aa311d6d7dc06965165"}, + {file = "yarl-1.8.2-cp38-cp38-win_amd64.whl", hash = "sha256:326dd1d3caf910cd26a26ccbfb84c03b608ba32499b5d6eeb09252c920bcbe4f"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:f8ca8ad414c85bbc50f49c0a106f951613dfa5f948ab69c10ce9b128d368baf8"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:418857f837347e8aaef682679f41e36c24250097f9e2f315d39bae3a99a34cbf"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:ae0eec05ab49e91a78700761777f284c2df119376e391db42c38ab46fd662b77"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:009a028127e0a1755c38b03244c0bea9d5565630db9c4cf9572496e947137a87"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3edac5d74bb3209c418805bda77f973117836e1de7c000e9755e572c1f7850d0"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:da65c3f263729e47351261351b8679c6429151ef9649bba08ef2528ff2c423b2"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ef8fb25e52663a1c85d608f6dd72e19bd390e2ecaf29c17fb08f730226e3a08"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bcd7bb1e5c45274af9a1dd7494d3c52b2be5e6bd8d7e49c612705fd45420b12d"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:44ceac0450e648de86da8e42674f9b7077d763ea80c8ceb9d1c3e41f0f0a9951"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:97209cc91189b48e7cfe777237c04af8e7cc51eb369004e061809bcdf4e55220"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:48dd18adcf98ea9cd721a25313aef49d70d413a999d7d89df44f469edfb38a06"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:e59399dda559688461762800d7fb34d9e8a6a7444fd76ec33220a926c8be1516"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d617c241c8c3ad5c4e78a08429fa49e4b04bedfc507b34b4d8dceb83b4af3588"}, + {file = "yarl-1.8.2-cp39-cp39-win32.whl", hash = "sha256:cb6d48d80a41f68de41212f3dfd1a9d9898d7841c8f7ce6696cf2fd9cb57ef83"}, + {file = "yarl-1.8.2-cp39-cp39-win_amd64.whl", hash = "sha256:6604711362f2dbf7160df21c416f81fac0de6dbcf0b5445a2ef25478ecc4c778"}, + {file = "yarl-1.8.2.tar.gz", hash = "sha256:49d43402c6e3013ad0978602bf6bf5328535c48d192304b91b97a3c6790b1562"}, +] + +[package.dependencies] +idna = ">=2.0" +multidict = ">=4.0" + @@ -1878 +3172 @@ python-versions = "3.9.15" -content-hash = "78aebf9305fa94fab4f0ac3d225c25480c48e8cf74a4eee1c63af22e4f76686d" +content-hash = "d91bad72ecda127472f1546ed66fe5f916c1da91b741e17a9951500756188a10" diff --git a/services/admin/pyproject.toml b/services/admin/pyproject.toml index 8efbf362..ad274b56 100644 --- a/services/admin/pyproject.toml +++ b/services/admin/pyproject.toml @@ -26 +26 @@ mypy = "^1.0.0" -pip-audit = "^2.4.14" +pip-audit = "^2.5.4" @@ -29 +29 @@ pytest-cov = "^2.12.1" -responses = "^0.22.0" +responses = "^0.18.0" diff --git a/services/admin/tests/utils.py b/services/admin/tests/utils.py index 0b875c1a..bd008b9b 100644 --- a/services/admin/tests/utils.py +++ b/services/admin/tests/utils.py @@ -3,0 +4 @@ +from io import BufferedReader @@ -7 +8,3 @@ from requests import PreparedRequest -from responses import _Body +from responses import Response + +_Body = Union[str, BaseException, Response, BufferedReader, bytes] diff --git a/services/api/Dockerfile b/services/api/Dockerfile index 3a77155e..9a42a9d5 100644 --- a/services/api/Dockerfile +++ b/services/api/Dockerfile @@ -19,0 +20 @@ RUN apt-get update \ + libicu-dev ffmpeg libavcodec-extra libsndfile1 llvm pkg-config \ diff --git a/services/api/dev.Dockerfile b/services/api/dev.Dockerfile index 340c5927..7af4bd54 100644 --- a/services/api/dev.Dockerfile +++ b/services/api/dev.Dockerfile @@ -19,0 +20 @@ RUN apt-get update \ + libicu-dev ffmpeg libavcodec-extra libsndfile1 llvm pkg-config \ diff --git a/services/api/poetry.lock b/services/api/poetry.lock index d5d7da5d..834ee07f 100644 --- a/services/api/poetry.lock +++ b/services/api/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -190,0 +191,11 @@ tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy +[[package]] +name = "audioread" +version = "3.0.0" +description = "multi-library, cross-platform audio decoding" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"}, +] + @@ -597 +608 @@ name = "cyclonedx-python-lib" -version = "3.1.5" +version = "2.7.1" @@ -603,2 +614,2 @@ files = [ - {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, - {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, + {file = "cyclonedx-python-lib-2.7.1.tar.gz", hash = "sha256:493bf2f30e26c48f305f745ed8580ce10d05a8d68d62a598fe95f05a0d9007dc"}, + {file = "cyclonedx_python_lib-2.7.1-py3-none-any.whl", hash = "sha256:fabc4c8baf722faeea01c3bbca83730e3489dfb37d85c6036baa67a9a7519d40"}, @@ -615 +626 @@ name = "datasets" -version = "2.10.1" +version = "2.11.0" @@ -621,2 +632,2 @@ files = [ - {file = "datasets-2.10.1-py3-none-any.whl", hash = "sha256:bfde7253b31abfd075f9ad55961b13c135a5a9295cea04ae86c47b9c0e0466fd"}, - {file = "datasets-2.10.1.tar.gz", hash = "sha256:e2764c90aa3af96450a9747a934b8893b121f79f58d89e123cb1a7046bb8e81e"}, + {file = "datasets-2.11.0-py3-none-any.whl", hash = "sha256:d946cdb8c4885d3016a2ab3129c9403dd3358fe9107e8ab5e549ceab672774af"}, + {file = "datasets-2.11.0.tar.gz", hash = "sha256:1ca53b9cd6ece7a3fdb81176dadd5b9e646420e52e68e85307b27db3a36ca18c"}, @@ -629 +640,2 @@ fsspec = {version = ">=2021.11.1", extras = ["http"]} -huggingface-hub = ">=0.2.0,<1.0.0" +huggingface-hub = ">=0.11.0,<1.0.0" +librosa = {version = "*", optional = true, markers = "extra == \"audio\""} @@ -634 +646,2 @@ pandas = "*" -pyarrow = ">=6.0.0" +Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} +pyarrow = ">=8.0.0" @@ -637,0 +651 @@ responses = "<0.19" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} @@ -643 +657 @@ apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] -audio = ["librosa"] +audio = ["librosa", "soundfile (>=0.12.1)"] @@ -645 +659 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -648 +662 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "sqlalchemy (<2.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -653 +667 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -656,0 +671,12 @@ vision = ["Pillow (>=6.2.1)"] +[[package]] +name = "decorator" +version = "5.1.1" +description = "Decorators for Humans" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"}, + {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"}, +] + @@ -1090,0 +1117,12 @@ requirements-deprecated-finder = ["pip-api", "pipreqs"] +[[package]] +name = "joblib" +version = "1.2.0" +description = "Lightweight pipelining with Python functions" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "joblib-1.2.0-py3-none-any.whl", hash = "sha256:091138ed78f800342968c523bdde947e7a305b8594b910a0fea2ab83c3c6d385"}, + {file = "joblib-1.2.0.tar.gz", hash = "sha256:e1cee4a79e4af22881164f218d4311f60074197fb707e082e803b61f6d137018"}, +] + @@ -1110,0 +1149,16 @@ format-nongpl = ["fqdn", "idna", "isoduration", "jsonpointer (>1.13)", "rfc3339- +[[package]] +name = "lazy-loader" +version = "0.2" +description = "lazy_loader" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "lazy_loader-0.2-py3-none-any.whl", hash = "sha256:c35875f815c340f823ce3271ed645045397213f961b40ad0c0d395c3f5218eeb"}, + {file = "lazy_loader-0.2.tar.gz", hash = "sha256:0edc7a5175c400acb108f283749951fefdadedeb00adcec6e88b974a9254f18a"}, +] + +[package.extras] +lint = ["pre-commit (>=3.1)"] +test = ["pytest (>=7.2)", "pytest-cov (>=4.0)"] + @@ -1122,0 +1177 @@ appdirs = "^1.4.4" +datasets = {version = "^2.11.0", extras = ["audio", "vision"]} @@ -1127,0 +1183 @@ networkx = "^3.0" +numba = "0.56.4" @@ -1129,0 +1186 @@ psutil = "^5.9.4" +pydub = "^0.25.1" @@ -1132,0 +1190 @@ requests = "^2.28.2" +soundfile = ">=0.12.1" @@ -1137,0 +1196,70 @@ url = "../../libs/libcommon" +[[package]] +name = "librosa" +version = "0.10.0.post2" +description = "Python module for audio and music processing" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"}, + {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"}, +] + +[package.dependencies] +audioread = ">=2.1.9" +decorator = ">=4.3.0" +joblib = ">=0.14" +lazy-loader = ">=0.1" +msgpack = ">=1.0" +numba = ">=0.51.0" +numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2" +pooch = ">=1.0,<1.7" +scikit-learn = ">=0.20.0" +scipy = ">=1.2.0" +soundfile = ">=0.12.1" +soxr = ">=0.3.2" +typing-extensions = ">=4.1.1" + +[package.extras] +display = ["matplotlib (>=3.3.0)"] +docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"] +tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"] + +[[package]] +name = "llvmlite" +version = "0.39.1" +description = "lightweight wrapper around basic LLVM functionality" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"}, + {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"}, + {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"}, + {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"}, + {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"}, + {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"}, + {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"}, + {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"}, + {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"}, + {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"}, +] + @@ -1311 +1439 @@ description = "MessagePack serializer" -category = "dev" +category = "main" @@ -1557,0 +1686,43 @@ test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] +[[package]] +name = "numba" +version = "0.56.4" +description = "compiling Python code using LLVM" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"}, + {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"}, + {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"}, + {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"}, + {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"}, + {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"}, + {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"}, + {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"}, + {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"}, + {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"}, + {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"}, + {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"}, + {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"}, + {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"}, + {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"}, + {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"}, +] + +[package.dependencies] +llvmlite = ">=0.39.0dev0,<0.40" +numpy = ">=1.18,<1.24" +setuptools = "*" + @@ -1560,2 +1731,2 @@ name = "numpy" -version = "1.24.2" -description = "Fundamental package for array computing in Python" +version = "1.23.5" +description = "NumPy is the fundamental package for array computing with Python." @@ -1566,28 +1737,28 @@ files = [ - {file = "numpy-1.24.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eef70b4fc1e872ebddc38cddacc87c19a3709c0e3e5d20bf3954c147b1dd941d"}, - {file = "numpy-1.24.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e8d2859428712785e8a8b7d2b3ef0a1d1565892367b32f915c4a4df44d0e64f5"}, - {file = "numpy-1.24.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6524630f71631be2dabe0c541e7675db82651eb998496bbe16bc4f77f0772253"}, - {file = "numpy-1.24.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a51725a815a6188c662fb66fb32077709a9ca38053f0274640293a14fdd22978"}, - {file = "numpy-1.24.2-cp310-cp310-win32.whl", hash = "sha256:2620e8592136e073bd12ee4536149380695fbe9ebeae845b81237f986479ffc9"}, - {file = "numpy-1.24.2-cp310-cp310-win_amd64.whl", hash = "sha256:97cf27e51fa078078c649a51d7ade3c92d9e709ba2bfb97493007103c741f1d0"}, - {file = "numpy-1.24.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7de8fdde0003f4294655aa5d5f0a89c26b9f22c0a58790c38fae1ed392d44a5a"}, - {file = "numpy-1.24.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4173bde9fa2a005c2c6e2ea8ac1618e2ed2c1c6ec8a7657237854d42094123a0"}, - {file = "numpy-1.24.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4cecaed30dc14123020f77b03601559fff3e6cd0c048f8b5289f4eeabb0eb281"}, - {file = "numpy-1.24.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9a23f8440561a633204a67fb44617ce2a299beecf3295f0d13c495518908e910"}, - {file = "numpy-1.24.2-cp311-cp311-win32.whl", hash = "sha256:e428c4fbfa085f947b536706a2fc349245d7baa8334f0c5723c56a10595f9b95"}, - {file = "numpy-1.24.2-cp311-cp311-win_amd64.whl", hash = "sha256:557d42778a6869c2162deb40ad82612645e21d79e11c1dc62c6e82a2220ffb04"}, - {file = "numpy-1.24.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d0a2db9d20117bf523dde15858398e7c0858aadca7c0f088ac0d6edd360e9ad2"}, - {file = "numpy-1.24.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:c72a6b2f4af1adfe193f7beb91ddf708ff867a3f977ef2ec53c0ffb8283ab9f5"}, - {file = "numpy-1.24.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c29e6bd0ec49a44d7690ecb623a8eac5ab8a923bce0bea6293953992edf3a76a"}, - {file = "numpy-1.24.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2eabd64ddb96a1239791da78fa5f4e1693ae2dadc82a76bc76a14cbb2b966e96"}, - {file = "numpy-1.24.2-cp38-cp38-win32.whl", hash = "sha256:e3ab5d32784e843fc0dd3ab6dcafc67ef806e6b6828dc6af2f689be0eb4d781d"}, - {file = "numpy-1.24.2-cp38-cp38-win_amd64.whl", hash = "sha256:76807b4063f0002c8532cfeac47a3068a69561e9c8715efdad3c642eb27c0756"}, - {file = "numpy-1.24.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4199e7cfc307a778f72d293372736223e39ec9ac096ff0a2e64853b866a8e18a"}, - {file = "numpy-1.24.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:adbdce121896fd3a17a77ab0b0b5eedf05a9834a18699db6829a64e1dfccca7f"}, - {file = "numpy-1.24.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:889b2cc88b837d86eda1b17008ebeb679d82875022200c6e8e4ce6cf549b7acb"}, - {file = "numpy-1.24.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f64bb98ac59b3ea3bf74b02f13836eb2e24e48e0ab0145bbda646295769bd780"}, - {file = "numpy-1.24.2-cp39-cp39-win32.whl", hash = "sha256:63e45511ee4d9d976637d11e6c9864eae50e12dc9598f531c035265991910468"}, - {file = "numpy-1.24.2-cp39-cp39-win_amd64.whl", hash = "sha256:a77d3e1163a7770164404607b7ba3967fb49b24782a6ef85d9b5f54126cc39e5"}, - {file = "numpy-1.24.2-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92011118955724465fb6853def593cf397b4a1367495e0b59a7e69d40c4eb71d"}, - {file = "numpy-1.24.2-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9006288bcf4895917d02583cf3411f98631275bc67cce355a7f39f8c14338fa"}, - {file = "numpy-1.24.2-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:150947adbdfeceec4e5926d956a06865c1c690f2fd902efede4ca6fe2e657c3f"}, - {file = "numpy-1.24.2.tar.gz", hash = "sha256:003a9f530e880cb2cd177cba1af7220b9aa42def9c4afc2a2fc3ee6be7eb2b22"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"}, + {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"}, + {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"}, + {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"}, + {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"}, + {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"}, + {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"}, + {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"}, + {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"}, + {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"}, @@ -1722,0 +1894,15 @@ test = ["hypothesis (>=5.5.3)", "pytest (>=6.0)", "pytest-xdist (>=1.31)"] +[[package]] +name = "pandas-stubs" +version = "1.5.3.230321" +description = "Type annotations for pandas" +category = "dev" +optional = false +python-versions = ">=3.8,<3.12" +files = [ + {file = "pandas_stubs-1.5.3.230321-py3-none-any.whl", hash = "sha256:4bf36b3071dd55f0e558ac8efe07676a120f2ed89e7a3df0fb78ddf2733bf247"}, + {file = "pandas_stubs-1.5.3.230321.tar.gz", hash = "sha256:2fa860df9e6058e9f0d2c09bc711c09abb8f0516eee7f0b9f9950d29b835fc6f"}, +] + +[package.dependencies] +types-pytz = ">=2022.1.1" + @@ -1746,0 +1933,91 @@ files = [ +[[package]] +name = "pillow" +version = "9.4.0" +description = "Python Imaging Library (Fork)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pillow-9.4.0-1-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:1b4b4e9dda4f4e4c4e6896f93e84a8f0bcca3b059de9ddf67dac3c334b1195e1"}, + {file = "Pillow-9.4.0-1-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:fb5c1ad6bad98c57482236a21bf985ab0ef42bd51f7ad4e4538e89a997624e12"}, + {file = "Pillow-9.4.0-1-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:f0caf4a5dcf610d96c3bd32932bfac8aee61c96e60481c2a0ea58da435e25acd"}, + {file = "Pillow-9.4.0-1-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:3f4cc516e0b264c8d4ccd6b6cbc69a07c6d582d8337df79be1e15a5056b258c9"}, + {file = "Pillow-9.4.0-1-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:b8c2f6eb0df979ee99433d8b3f6d193d9590f735cf12274c108bd954e30ca858"}, + {file = "Pillow-9.4.0-1-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:b70756ec9417c34e097f987b4d8c510975216ad26ba6e57ccb53bc758f490dab"}, + {file = "Pillow-9.4.0-1-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:43521ce2c4b865d385e78579a082b6ad1166ebed2b1a2293c3be1d68dd7ca3b9"}, + {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"}, + {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"}, + {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"}, + {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"}, + {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"}, + {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"}, + {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"}, + {file = "Pillow-9.4.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:2968c58feca624bb6c8502f9564dd187d0e1389964898f5e9e1fbc8533169157"}, + {file = "Pillow-9.4.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c5c1362c14aee73f50143d74389b2c158707b4abce2cb055b7ad37ce60738d47"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bd752c5ff1b4a870b7661234694f24b1d2b9076b8bf337321a814c612665f343"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:9a3049a10261d7f2b6514d35bbb7a4dfc3ece4c4de14ef5876c4b7a23a0e566d"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16a8df99701f9095bea8a6c4b3197da105df6f74e6176c5b410bc2df2fd29a57"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:94cdff45173b1919350601f82d61365e792895e3c3a3443cf99819e6fbf717a5"}, + {file = "Pillow-9.4.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:ed3e4b4e1e6de75fdc16d3259098de7c6571b1a6cc863b1a49e7d3d53e036070"}, + {file = "Pillow-9.4.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d5b2f8a31bd43e0f18172d8ac82347c8f37ef3e0b414431157718aa234991b28"}, + {file = "Pillow-9.4.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:09b89ddc95c248ee788328528e6a2996e09eaccddeeb82a5356e92645733be35"}, + {file = "Pillow-9.4.0-cp310-cp310-win32.whl", hash = "sha256:f09598b416ba39a8f489c124447b007fe865f786a89dbfa48bb5cf395693132a"}, + {file = "Pillow-9.4.0-cp310-cp310-win_amd64.whl", hash = "sha256:f6e78171be3fb7941f9910ea15b4b14ec27725865a73c15277bc39f5ca4f8391"}, + {file = "Pillow-9.4.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:3fa1284762aacca6dc97474ee9c16f83990b8eeb6697f2ba17140d54b453e133"}, + {file = "Pillow-9.4.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:eaef5d2de3c7e9b21f1e762f289d17b726c2239a42b11e25446abf82b26ac132"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4dfdae195335abb4e89cc9762b2edc524f3c6e80d647a9a81bf81e17e3fb6f0"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6abfb51a82e919e3933eb137e17c4ae9c0475a25508ea88993bb59faf82f3b35"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:451f10ef963918e65b8869e17d67db5e2f4ab40e716ee6ce7129b0cde2876eab"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:6663977496d616b618b6cfa43ec86e479ee62b942e1da76a2c3daa1c75933ef4"}, + {file = "Pillow-9.4.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:60e7da3a3ad1812c128750fc1bc14a7ceeb8d29f77e0a2356a8fb2aa8925287d"}, + {file = "Pillow-9.4.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:19005a8e58b7c1796bc0167862b1f54a64d3b44ee5d48152b06bb861458bc0f8"}, + {file = "Pillow-9.4.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f715c32e774a60a337b2bb8ad9839b4abf75b267a0f18806f6f4f5f1688c4b5a"}, + {file = "Pillow-9.4.0-cp311-cp311-win32.whl", hash = "sha256:b222090c455d6d1a64e6b7bb5f4035c4dff479e22455c9eaa1bdd4c75b52c80c"}, + {file = "Pillow-9.4.0-cp311-cp311-win_amd64.whl", hash = "sha256:ba6612b6548220ff5e9df85261bddc811a057b0b465a1226b39bfb8550616aee"}, + {file = "Pillow-9.4.0-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:5f532a2ad4d174eb73494e7397988e22bf427f91acc8e6ebf5bb10597b49c493"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5dd5a9c3091a0f414a963d427f920368e2b6a4c2f7527fdd82cde8ef0bc7a327"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef21af928e807f10bf4141cad4746eee692a0dd3ff56cfb25fce076ec3cc8abe"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:847b114580c5cc9ebaf216dd8c8dbc6b00a3b7ab0131e173d7120e6deade1f57"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_28_aarch64.whl", hash = "sha256:653d7fb2df65efefbcbf81ef5fe5e5be931f1ee4332c2893ca638c9b11a409c4"}, + {file = "Pillow-9.4.0-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:46f39cab8bbf4a384ba7cb0bc8bae7b7062b6a11cfac1ca4bc144dea90d4a9f5"}, + {file = "Pillow-9.4.0-cp37-cp37m-win32.whl", hash = "sha256:7ac7594397698f77bce84382929747130765f66406dc2cd8b4ab4da68ade4c6e"}, + {file = "Pillow-9.4.0-cp37-cp37m-win_amd64.whl", hash = "sha256:46c259e87199041583658457372a183636ae8cd56dbf3f0755e0f376a7f9d0e6"}, + {file = "Pillow-9.4.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:0e51f608da093e5d9038c592b5b575cadc12fd748af1479b5e858045fff955a9"}, + {file = "Pillow-9.4.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:765cb54c0b8724a7c12c55146ae4647e0274a839fb6de7bcba841e04298e1011"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:519e14e2c49fcf7616d6d2cfc5c70adae95682ae20f0395e9280db85e8d6c4df"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d197df5489004db87d90b918033edbeee0bd6df3848a204bca3ff0a903bef837"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0845adc64fe9886db00f5ab68c4a8cd933ab749a87747555cec1c95acea64b0b"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:e1339790c083c5a4de48f688b4841f18df839eb3c9584a770cbd818b33e26d5d"}, + {file = "Pillow-9.4.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:a96e6e23f2b79433390273eaf8cc94fec9c6370842e577ab10dabdcc7ea0a66b"}, + {file = "Pillow-9.4.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:7cfc287da09f9d2a7ec146ee4d72d6ea1342e770d975e49a8621bf54eaa8f30f"}, + {file = "Pillow-9.4.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:d7081c084ceb58278dd3cf81f836bc818978c0ccc770cbbb202125ddabec6628"}, + {file = "Pillow-9.4.0-cp38-cp38-win32.whl", hash = "sha256:df41112ccce5d47770a0c13651479fbcd8793f34232a2dd9faeccb75eb5d0d0d"}, + {file = "Pillow-9.4.0-cp38-cp38-win_amd64.whl", hash = "sha256:7a21222644ab69ddd9967cfe6f2bb420b460dae4289c9d40ff9a4896e7c35c9a"}, + {file = "Pillow-9.4.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:0f3269304c1a7ce82f1759c12ce731ef9b6e95b6df829dccd9fe42912cc48569"}, + {file = "Pillow-9.4.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:cb362e3b0976dc994857391b776ddaa8c13c28a16f80ac6522c23d5257156bed"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a2e0f87144fcbbe54297cae708c5e7f9da21a4646523456b00cc956bd4c65815"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:28676836c7796805914b76b1837a40f76827ee0d5398f72f7dcc634bae7c6264"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0884ba7b515163a1a05440a138adeb722b8a6ae2c2b33aea93ea3118dd3a899e"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:53dcb50fbdc3fb2c55431a9b30caeb2f7027fcd2aeb501459464f0214200a503"}, + {file = "Pillow-9.4.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:e8c5cf126889a4de385c02a2c3d3aba4b00f70234bfddae82a5eaa3ee6d5e3e6"}, + {file = "Pillow-9.4.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:6c6b1389ed66cdd174d040105123a5a1bc91d0aa7059c7261d20e583b6d8cbd2"}, + {file = "Pillow-9.4.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:0dd4c681b82214b36273c18ca7ee87065a50e013112eea7d78c7a1b89a739153"}, + {file = "Pillow-9.4.0-cp39-cp39-win32.whl", hash = "sha256:6d9dfb9959a3b0039ee06c1a1a90dc23bac3b430842dcb97908ddde05870601c"}, + {file = "Pillow-9.4.0-cp39-cp39-win_amd64.whl", hash = "sha256:54614444887e0d3043557d9dbc697dbb16cfb5a35d672b7a0fcc1ed0cf1c600b"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:b9b752ab91e78234941e44abdecc07f1f0d8f51fb62941d32995b8161f68cfe5"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d3b56206244dc8711f7e8b7d6cad4663917cd5b2d950799425076681e8766286"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:aabdab8ec1e7ca7f1434d042bf8b1e92056245fb179790dc97ed040361f16bfd"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:db74f5562c09953b2c5f8ec4b7dfd3f5421f31811e97d1dbc0a7c93d6e3a24df"}, + {file = "Pillow-9.4.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:e9d7747847c53a16a729b6ee5e737cf170f7a16611c143d95aa60a109a59c336"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:b52ff4f4e002f828ea6483faf4c4e8deea8d743cf801b74910243c58acc6eda3"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:575d8912dca808edd9acd6f7795199332696d3469665ef26163cd090fa1f8bfa"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c3c4ed2ff6760e98d262e0cc9c9a7f7b8a9f61aa4d47c58835cdaf7b0b8811bb"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:e621b0246192d3b9cb1dc62c78cfa4c6f6d2ddc0ec207d43c0dedecb914f152a"}, + {file = "Pillow-9.4.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:8f127e7b028900421cad64f51f75c051b628db17fb00e099eb148761eed598c9"}, + {file = "Pillow-9.4.0.tar.gz", hash = "sha256:a1c2d7780448eb93fbcc3789bf3916aa5720d942e37945f4056680317f1cd23e"}, +] + +[package.extras] +docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-issues (>=3.0.1)", "sphinx-removed-in", "sphinxext-opengraph"] +tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"] + @@ -1776 +2053 @@ name = "pip-audit" -version = "2.4.14" +version = "2.5.4" @@ -1782,2 +2059,2 @@ files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, + {file = "pip_audit-2.5.4-py3-none-any.whl", hash = "sha256:bee3748030c895488b4dd8a6196fa44f484da000cfd52d8fc64dfa3a2e121624"}, + {file = "pip_audit-2.5.4.tar.gz", hash = "sha256:a4cb03f9e2896d626f5b153973d3ac0d32fdb18594d78d393b153c83bb8089b6"}, @@ -1788 +2065 @@ CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +cyclonedx-python-lib = ">=2.0,<2.5.0 || >2.5.0,<3.0" @@ -1793 +2069,0 @@ pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" @@ -1800 +2076 @@ doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.260)", "types-html5lib", "types-requests", "types-toml"] @@ -1854,0 +2131,22 @@ testing = ["pytest", "pytest-benchmark"] +[[package]] +name = "pooch" +version = "1.6.0" +description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\"" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"}, + {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"}, +] + +[package.dependencies] +appdirs = ">=1.3.0" +packaging = ">=20.0" +requests = ">=2.19.0" + +[package.extras] +progress = ["tqdm (>=4.41.0,<5.0.0)"] +sftp = ["paramiko (>=2.7.0)"] +xxhash = ["xxhash (>=1.4.3)"] + @@ -1958,0 +2257,12 @@ files = [ +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + @@ -2364,18 +2673,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "resolvelib" -version = "0.9.0" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"}, - {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"}, -] - -[package.extras] -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] - @@ -2437,0 +2730,82 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] +[[package]] +name = "scikit-learn" +version = "1.2.2" +description = "A set of python modules for machine learning and data mining" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "scikit-learn-1.2.2.tar.gz", hash = "sha256:8429aea30ec24e7a8c7ed8a3fa6213adf3814a6efbea09e16e0a0c71e1a1a3d7"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:99cc01184e347de485bf253d19fcb3b1a3fb0ee4cea5ee3c43ec0cc429b6d29f"}, + {file = "scikit_learn-1.2.2-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:e6e574db9914afcb4e11ade84fab084536a895ca60aadea3041e85b8ac963edb"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6fe83b676f407f00afa388dd1fdd49e5c6612e551ed84f3b1b182858f09e987d"}, + {file = "scikit_learn-1.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e2642baa0ad1e8f8188917423dd73994bf25429f8893ddbe115be3ca3183584"}, + {file = "scikit_learn-1.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:ad66c3848c0a1ec13464b2a95d0a484fd5b02ce74268eaa7e0c697b904f31d6c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:dfeaf8be72117eb61a164ea6fc8afb6dfe08c6f90365bde2dc16456e4bc8e45f"}, + {file = "scikit_learn-1.2.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:fe0aa1a7029ed3e1dcbf4a5bc675aa3b1bc468d9012ecf6c6f081251ca47f590"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:065e9673e24e0dc5113e2dd2b4ca30c9d8aa2fa90f4c0597241c93b63130d233"}, + {file = "scikit_learn-1.2.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf036ea7ef66115e0d49655f16febfa547886deba20149555a41d28f56fd6d3c"}, + {file = "scikit_learn-1.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:8b0670d4224a3c2d596fd572fb4fa673b2a0ccfb07152688ebd2ea0b8c61025c"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9c710ff9f9936ba8a3b74a455ccf0dcf59b230caa1e9ba0223773c490cab1e51"}, + {file = "scikit_learn-1.2.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:2dd3ffd3950e3d6c0c0ef9033a9b9b32d910c61bd06cb8206303fb4514b88a49"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:44b47a305190c28dd8dd73fc9445f802b6ea716669cfc22ab1eb97b335d238b1"}, + {file = "scikit_learn-1.2.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:953236889928d104c2ef14027539f5f2609a47ebf716b8cbe4437e85dce42744"}, + {file = "scikit_learn-1.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:7f69313884e8eb311460cc2f28676d5e400bd929841a2c8eb8742ae78ebf7c20"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8156db41e1c39c69aa2d8599ab7577af53e9e5e7a57b0504e116cc73c39138dd"}, + {file = "scikit_learn-1.2.2-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:fe175ee1dab589d2e1033657c5b6bec92a8a3b69103e3dd361b58014729975c3"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7d5312d9674bed14f73773d2acf15a3272639b981e60b72c9b190a0cffed5bad"}, + {file = "scikit_learn-1.2.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ea061bf0283bf9a9f36ea3c5d3231ba2176221bbd430abd2603b1c3b2ed85c89"}, + {file = "scikit_learn-1.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:6477eed40dbce190f9f9e9d0d37e020815825b300121307942ec2110302b66a3"}, +] + +[package.dependencies] +joblib = ">=1.1.1" +numpy = ">=1.17.3" +scipy = ">=1.3.2" +threadpoolctl = ">=2.0.0" + +[package.extras] +benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"] +docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=4.0.1)", "sphinx-gallery (>=0.7.0)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"] +examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.10.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"] +tests = ["black (>=22.3.0)", "flake8 (>=3.8.2)", "matplotlib (>=3.1.3)", "mypy (>=0.961)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=5.3.1)", "pytest-cov (>=2.9.0)", "scikit-image (>=0.16.2)"] + +[[package]] +name = "scipy" +version = "1.10.1" +description = "Fundamental algorithms for scientific computing in Python" +category = "main" +optional = false +python-versions = "<3.12,>=3.8" +files = [ + {file = "scipy-1.10.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e7354fd7527a4b0377ce55f286805b34e8c54b91be865bac273f527e1b839019"}, + {file = "scipy-1.10.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:4b3f429188c66603a1a5c549fb414e4d3bdc2a24792e061ffbd607d3d75fd84e"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1553b5dcddd64ba9a0d95355e63fe6c3fc303a8fd77c7bc91e77d61363f7433f"}, + {file = "scipy-1.10.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c0ff64b06b10e35215abce517252b375e580a6125fd5fdf6421b98efbefb2d2"}, + {file = "scipy-1.10.1-cp310-cp310-win_amd64.whl", hash = "sha256:fae8a7b898c42dffe3f7361c40d5952b6bf32d10c4569098d276b4c547905ee1"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:0f1564ea217e82c1bbe75ddf7285ba0709ecd503f048cb1236ae9995f64217bd"}, + {file = "scipy-1.10.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d925fa1c81b772882aa55bcc10bf88324dadb66ff85d548c71515f6689c6dac5"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaea0a6be54462ec027de54fca511540980d1e9eea68b2d5c1dbfe084797be35"}, + {file = "scipy-1.10.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15a35c4242ec5f292c3dd364a7c71a61be87a3d4ddcc693372813c0b73c9af1d"}, + {file = "scipy-1.10.1-cp311-cp311-win_amd64.whl", hash = "sha256:43b8e0bcb877faf0abfb613d51026cd5cc78918e9530e375727bf0625c82788f"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:5678f88c68ea866ed9ebe3a989091088553ba12c6090244fdae3e467b1139c35"}, + {file = "scipy-1.10.1-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:39becb03541f9e58243f4197584286e339029e8908c46f7221abeea4b749fa88"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bce5869c8d68cf383ce240e44c1d9ae7c06078a9396df68ce88a1230f93a30c1"}, + {file = "scipy-1.10.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:07c3457ce0b3ad5124f98a86533106b643dd811dd61b548e78cf4c8786652f6f"}, + {file = "scipy-1.10.1-cp38-cp38-win_amd64.whl", hash = "sha256:049a8bbf0ad95277ffba9b3b7d23e5369cc39e66406d60422c8cfef40ccc8415"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:cd9f1027ff30d90618914a64ca9b1a77a431159df0e2a195d8a9e8a04c78abf9"}, + {file = "scipy-1.10.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:79c8e5a6c6ffaf3a2262ef1be1e108a035cf4f05c14df56057b64acc5bebffb6"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:51af417a000d2dbe1ec6c372dfe688e041a7084da4fdd350aeb139bd3fb55353"}, + {file = "scipy-1.10.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b4735d6c28aad3cdcf52117e0e91d6b39acd4272f3f5cd9907c24ee931ad601"}, + {file = "scipy-1.10.1-cp39-cp39-win_amd64.whl", hash = "sha256:7ff7f37b1bf4417baca958d254e8e2875d0cc23aaadbe65b3d5b3077b0eb23ea"}, + {file = "scipy-1.10.1.tar.gz", hash = "sha256:2cf9dfb80a7b4589ba4c40ce7588986d6d5cebc5457cad2c2880f6bc2d42f3a5"}, +] + +[package.dependencies] +numpy = ">=1.19.5,<1.27.0" + +[package.extras] +dev = ["click", "doit (>=0.36.0)", "flake8", "mypy", "pycodestyle", "pydevtool", "rich-click", "typing_extensions"] +doc = ["matplotlib (>2)", "numpydoc", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"] +test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"] + @@ -2442 +2816 @@ description = "Easily download, build, install, upgrade, and uninstall Python pa -category = "dev" +category = "main" @@ -2502,0 +2877,68 @@ files = [ +[[package]] +name = "soundfile" +version = "0.12.1" +description = "An audio library based on libsndfile, CFFI and NumPy" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"}, + {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"}, + {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"}, + {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"}, +] + +[package.dependencies] +cffi = ">=1.0" + +[package.extras] +numpy = ["numpy"] + +[[package]] +name = "soxr" +version = "0.3.4" +description = "High quality, one-dimensional sample-rate conversion library" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "soxr-0.3.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b7b84126643c063d5daa203f7f9137e21734dabbd7e68c097607b2ef457e2f2e"}, + {file = "soxr-0.3.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:380d2d43871a68e8b1ef1702a0abe6f9e48ddb3933c7a303c45d67e121503e7c"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a4a1b4019c9972f57612482c4f85523d6e832e3d10935e2f070a9dcd334a4dcb"}, + {file = "soxr-0.3.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e613cee023b7c3f162b9da3f6b169cd7f58de345275be1fde9f19adc9cf144df"}, + {file = "soxr-0.3.4-cp310-cp310-win32.whl", hash = "sha256:182c02a7ba45a159a0dbb0a297335df2381ead03a65377b19663ea0ff720ecb7"}, + {file = "soxr-0.3.4-cp310-cp310-win_amd64.whl", hash = "sha256:1e95c96ce94524fae453b4331c9910d33f97506f99bae06d76a9c0649710619e"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2678d2719e7496803983584d661eb5fddc7017154a8dda4a774407c56ff07973"}, + {file = "soxr-0.3.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:11bd1396052049e6d389225a0e96a9df15f706da501c619b35d3c72ac6bc7257"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7e23de4dfe54ac30e583bbc9cc3feda1cd776fedce13206bc4b3115b75ecab82"}, + {file = "soxr-0.3.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3e7396498a5f5b7d8f23b656f65c24517a6ff5bdc3ee0623ccd491036a43ea08"}, + {file = "soxr-0.3.4-cp311-cp311-win32.whl", hash = "sha256:e57e9703c2bff834cabc06800d3c11a259544891d2c24a78949f3cf2f5492cc5"}, + {file = "soxr-0.3.4-cp311-cp311-win_amd64.whl", hash = "sha256:7c8350acd7150f74261a0569b47ccb1bb4aa39b2d575860bc97cfa69aab8aead"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:941f7355116fe77fe6a82938fa7799a0e466a494ebc093f676969ce32b2815b1"}, + {file = "soxr-0.3.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:00fdbf24f64d8c3fb800425c383048cb24c32defac80901cde4a57fb6ce5d431"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7bb6d4dc807d04c536674429e2b05ae08a1efac9815c4595e41ffd6b57c2c662"}, + {file = "soxr-0.3.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ff15853895b54f1b627799c6600be1ce5f7286724e7a93e4b7b9d79e5d4166f5"}, + {file = "soxr-0.3.4-cp38-cp38-win32.whl", hash = "sha256:d858becbc1fcc7b38c3436d3276290fae09403cdcbdf1d5986a18dab7023a6c3"}, + {file = "soxr-0.3.4-cp38-cp38-win_amd64.whl", hash = "sha256:068ab4df549df5783cc1eb4eb6c94f53823b164dc27134fc621fc9f5097f38cd"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:20130329985f9767c8417bbd125fe138790a71802b000481c386a800e2ad2bca"}, + {file = "soxr-0.3.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:78090e97abfb326b7cf14ef37d08a17252b07d438388dcbbd82a6836a9d551b1"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84e590e75b7e5dca12bf68bfb090276f34a88fbcd793781c62d47f5d7dbe525e"}, + {file = "soxr-0.3.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3479d265574b960e12bca0878baba0862c43278915e0319d84679bb4d4fcd33"}, + {file = "soxr-0.3.4-cp39-cp39-win32.whl", hash = "sha256:83de825d6a713c7b2e76d9ec3f229a58a9ed290237e7adc05d80e8b39be995a6"}, + {file = "soxr-0.3.4-cp39-cp39-win_amd64.whl", hash = "sha256:2082f88cae89de854c3e0d62f55d0cb31eb11764f5c2a28299121fb642a22472"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:fe8b5f92c802f1e7793c40344f5368dc6163718c9ffa82e79ee6ad779d318ac5"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d0063d5f9a4e1a367084f4705301e9da131cf4d2d32aa3fe0072a1245e18088f"}, + {file = "soxr-0.3.4-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:a680bab57adae462cdc86abcc7330beb5daa3ba5101165583eedcda88b7ba551"}, + {file = "soxr-0.3.4.tar.gz", hash = "sha256:fe68daf00e8f020977b187699903d219f9e39b9fb3d915f3f923eed8ba431449"}, +] + +[package.dependencies] +numpy = "*" + +[package.extras] +docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"] +test = ["pytest"] + @@ -2552,0 +2995,12 @@ pbr = ">=2.0.0,<2.1.0 || >2.1.0" +[[package]] +name = "threadpoolctl" +version = "3.1.0" +description = "threadpoolctl" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, + {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"}, +] + @@ -2621,0 +3076,12 @@ files = [ +[[package]] +name = "types-pytz" +version = "2023.3.0.0" +description = "Typing stubs for pytz" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-pytz-2023.3.0.0.tar.gz", hash = "sha256:ecdc70d543aaf3616a7e48631543a884f74205f284cefd6649ddf44c6a820aac"}, + {file = "types_pytz-2023.3.0.0-py3-none-any.whl", hash = "sha256:4fc2a7fbbc315f0b6630e0b899fd6c743705abe1094d007b0e612d10da15e0f3"}, +] + @@ -2970 +3436 @@ python-versions = "3.9.15" -content-hash = "37daa2dafb9854811634de242057022f58b4e57ad36346f87be27f57aebf0277" +content-hash = "c15c0ae96cd56833f965b36c6f3222108f92905459fff4a80112bd9b1e83f535" diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml index c96e59ad..62414ee8 100644 --- a/services/api/pyproject.toml +++ b/services/api/pyproject.toml @@ -9 +9 @@ license = "Apache-2.0" -datasets = "^2.10.1" +datasets = { extras = ["audio", "vision"], version = "^2.11.0" } @@ -17,0 +18 @@ requests = "^2.28.2" +soundfile = ">=0.12.1" @@ -30 +31,2 @@ mypy = "^1.0.0" -pip-audit = "^2.4.14" +pandas-stubs = "^1.5.3" +pip-audit = "^2.5.4" @@ -70 +72,2 @@ module = [ - "tqdm.*" + "tqdm.*", + "fsspec.*" diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index 10ab008a..b69b7b7d 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -7,0 +8 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource +from libcommon.storage import exists, init_cached_assets_dir @@ -33,0 +35,3 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + cached_assets_directory = init_cached_assets_dir(directory=app_config.cached_assets.storage_directory) + if not exists(cached_assets_directory): + raise RuntimeError("The assets storage directory could not be accessed. Exiting.") @@ -130,0 +135,2 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + cached_assets_base_url=app_config.cached_assets.base_url, + cached_assets_directory=cached_assets_directory, diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index e2986d72..5f5f7af6 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -9,0 +10 @@ from libcommon.config import ( + CachedAssetsConfig, @@ -80,0 +82 @@ class AppConfig: + cached_assets: CachedAssetsConfig = field(default_factory=CachedAssetsConfig) @@ -91,0 +94 @@ class AppConfig: + cached_assets=CachedAssetsConfig.from_env(), diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 9eb26fd7..03cb5aa5 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -4,0 +5,3 @@ import logging +import os +import random +import shutil @@ -6 +9,3 @@ from functools import lru_cache, partial -from typing import Any, List, Mapping, Optional, TypedDict +from itertools import islice +from os import PathLike +from typing import Any, Callable, List, Mapping, Optional, TypedDict, Union @@ -13,0 +19,5 @@ from libcommon.processing_graph import ProcessingStep +from libcommon.viewer_utils.asset import ( + glob_rows_in_assets_dir, + update_last_modified_date_of_rows_in_assets_dir, +) +from libcommon.viewer_utils.features import get_cell_value @@ -31,0 +42,2 @@ from api.utils import ( +logger = logging.getLogger(__name__) + @@ -36,0 +49,3 @@ PARQUET_REVISION = "refs/convert/parquet" +StrPath = Union[str, PathLike[str]] + + @@ -48,0 +64,4 @@ class ParquetResponseEmptyError(Exception): +class ParquetDataProcessingError(Exception): + pass + + @@ -66 +85 @@ def get_parquet_fs(dataset: str, hf_token: Optional[str]) -> HfFileSystem: -UNSUPPORTED_FEATURES_MAGIC_STRINGS = ["Image(", "Audio(", "'binary'"] +UNSUPPORTED_FEATURES_MAGIC_STRINGS = ["'binary'"] @@ -161 +180 @@ class RowsIndex: - self.row_group_readers = [ + self.row_group_readers: List[Callable[[], pa.Table]] = [ @@ -227,0 +247,3 @@ class Indexer: +Row = Mapping[str, Any] + + @@ -231 +253 @@ class FeatureItem(TypedDict): - type: Mapping[str, Any] + type: Row @@ -259 +281,11 @@ class RowItem(TypedDict): -def to_rows_list(pa_table: pa.Table, offset: int, features: Features, unsupported_columns: List[str]) -> List[RowItem]: +def to_rows_list( + pa_table: pa.Table, + dataset: str, + config: str, + split: str, + cached_assets_base_url: str, + cached_assets_directory: StrPath, + offset: int, + features: Features, + unsupported_columns: List[str], +) -> List[RowItem]: @@ -263,0 +296,16 @@ def to_rows_list(pa_table: pa.Table, offset: int, features: Features, unsupporte + # transform the rows, if needed (e.g. save the images or audio to the assets, and return their URL) + try: + transformed_rows = transform_rows( + dataset=dataset, + config=config, + split=split, + rows=pa_table.to_pylist(), + features=features, + cached_assets_base_url=cached_assets_base_url, + cached_assets_directory=cached_assets_directory, + offset=offset, + ) + except Exception as err: + raise ParquetDataProcessingError( + "Server error while post-processing the split rows. Please report the issue." + ) from err @@ -270 +318,30 @@ def to_rows_list(pa_table: pa.Table, offset: int, features: Features, unsupporte - for idx, row in enumerate(pa_table.to_pylist()) + for idx, row in enumerate(transformed_rows) + ] + + +def transform_rows( + dataset: str, + config: str, + split: str, + rows: List[Row], + features: Features, + cached_assets_base_url: str, + cached_assets_directory: StrPath, + offset: int, +) -> List[Row]: + return [ + { + featureName: get_cell_value( + dataset=dataset, + config=config, + split=split, + row_idx=offset + row_idx, + cell=row[featureName] if featureName in row else None, + featureName=featureName, + fieldType=fieldType, + assets_base_url=cached_assets_base_url, + assets_directory=cached_assets_directory, + ) + for (featureName, fieldType) in features.items() + } + for row_idx, row in enumerate(rows) @@ -274 +351,74 @@ def to_rows_list(pa_table: pa.Table, offset: int, features: Features, unsupporte -def create_response(pa_table: pa.Table, offset: int, features: Features, unsupported_columns: List[str]) -> Any: +def _greater_or_equal(row_dir_name: str, row_idx: int, on_error: bool) -> bool: + try: + return int(row_dir_name) >= row_idx + except ValueError: + return on_error + + +def clean_cached_assets( + dataset: str, + cached_assets_directory: StrPath, + keep_first_rows_number: int, + keep_most_recent_rows_number: int, + max_cleaned_rows_number: int, +) -> None: + """ + The cached assets directory is cleaned to save disk space using this simple (?) heuristic: + + 1. it takes a big sample of rows from the cache using glob (max `max_cleaned_rows_number`) + 2. it keeps the most recent ones (max `keep_most_recent_rows_number`) + 3. it keeps the rows below a certain index (max `keep_first_rows_number`) + 4. it discards the rest + + To check for the most recent rows, it looks at the "last modified time" of rows directories. + This time is updated every time a row is accessed using `update_last_modified_date_of_rows_in_assets_dir()`. + + Args: + dataset (`str`): + Dataset name e.g `squad` or `lhoestq/demo1`. + Rows are cleaned in any dataset configuration or split of this dataset. + cached_assets_directory (`str`): + Directory containing the cached image and audio files + keep_first_rows_number (`int`): + Keep the rows with an index below a certain number + keep_most_recent_rows_number (`int`): + Keep the most recently accessed rows. + max_cleaned_rows_number (`int`): + Maximum number of rows to discard. + """ + if keep_first_rows_number < 0 or keep_most_recent_rows_number < 0 or max_cleaned_rows_number < 0: + raise ValueError( + "Failed to run cached assets cleaning. Make sure all of keep_first_rows_number," + f" keep_most_recent_rows_number and max_cleaned_rows_number are set (got {keep_first_rows_number}," + f" {keep_most_recent_rows_number} and {max_cleaned_rows_number})" + ) + row_directories = glob_rows_in_assets_dir(dataset, cached_assets_directory) + row_directories_sample = list( + islice( + ( + row_dir + for row_dir in row_directories + if _greater_or_equal(row_dir.name, keep_first_rows_number, on_error=True) + ), + max_cleaned_rows_number + keep_most_recent_rows_number, + ) + ) + if len(row_directories_sample) > keep_most_recent_rows_number: + row_dirs_to_delete = sorted(row_directories_sample, key=os.path.getmtime, reverse=True)[ + keep_most_recent_rows_number: + ] + for row_dir_to_delete in row_dirs_to_delete: + shutil.rmtree(row_dir_to_delete, ignore_errors=True) + + +def create_response( + dataset: str, + config: str, + split: str, + cached_assets_base_url: str, + cached_assets_directory: StrPath, + pa_table: pa.Table, + offset: int, + features: Features, + unsupported_columns: List[str], +) -> Any: @@ -281 +431,11 @@ def create_response(pa_table: pa.Table, offset: int, features: Features, unsuppo - "rows": to_rows_list(pa_table, offset, features, unsupported_columns), + "rows": to_rows_list( + pa_table, + dataset, + config, + split, + cached_assets_base_url, + cached_assets_directory, + offset, + features, + unsupported_columns, + ), @@ -287,0 +448,2 @@ def create_rows_endpoint( + cached_assets_base_url: str, + cached_assets_directory: StrPath, @@ -295,0 +458,4 @@ def create_rows_endpoint( + clean_cache_proba: float = 0.0, + keep_first_rows_number: int = -1, + keep_most_recent_rows_number: int = -1, + max_cleaned_rows_number: int = -1, @@ -327 +493 @@ def create_rows_endpoint( - dataset, + dataset=dataset, @@ -337,0 +504,20 @@ def create_rows_endpoint( + with StepProfiler(method="rows_endpoint", step="clean cache"): + # no need to do it every time + if random.random() < clean_cache_proba: # nosec + if ( + keep_first_rows_number < 0 + and keep_most_recent_rows_number < 0 + and max_cleaned_rows_number < 0 + ): + logger.debug( + "Params keep_first_rows_number, keep_most_recent_rows_number and" + " max_cleaned_rows_number are not set. Skipping cached assets cleaning." + ) + else: + clean_cached_assets( + dataset=dataset, + cached_assets_directory=cached_assets_directory, + keep_first_rows_number=keep_first_rows_number, + keep_most_recent_rows_number=keep_most_recent_rows_number, + max_cleaned_rows_number=max_cleaned_rows_number, + ) @@ -339 +525,20 @@ def create_rows_endpoint( - response = create_response(pa_table, offset, rows_index.features, rows_index.unsupported_columns) + response = create_response( + dataset=dataset, + config=config, + split=split, + cached_assets_base_url=cached_assets_base_url, + cached_assets_directory=cached_assets_directory, + pa_table=pa_table, + offset=offset, + features=rows_index.features, + unsupported_columns=rows_index.unsupported_columns, + ) + with StepProfiler(method="rows_endpoint", step="update last modified time of rows in asset dir"): + update_last_modified_date_of_rows_in_assets_dir( + dataset=dataset, + config=config, + split=split, + offset=offset, + length=length, + assets_directory=cached_assets_directory, + ) diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py index 04eb497e..4d8c0813 100644 --- a/services/api/tests/conftest.py +++ b/services/api/tests/conftest.py @@ -3,0 +4 @@ +from pathlib import Path @@ -9,0 +11 @@ from libcommon.simple_cache import _clean_cache_database +from libcommon.storage import StrPath, init_cached_assets_dir @@ -14,0 +17,3 @@ from api.routes.endpoint import EndpointsDefinition, StepsByInputTypeAndEndpoint +# Import fixture modules as plugins +pytest_plugins = ["tests.fixtures.fsspec"] + @@ -21,0 +27 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]: + monkeypatch_session.setenv("CACHED_ASSETS_BASE_URL", "http://localhost/cached-assets") @@ -118,0 +125,12 @@ def hf_auth_path(app_config: AppConfig) -> str: + + +@fixture +def cached_assets_directory(app_config: AppConfig) -> StrPath: + return init_cached_assets_dir(app_config.cached_assets.storage_directory) + + +@fixture +def image_path() -> str: + image_path = Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg" + assert image_path.is_file() + return str(image_path) diff --git a/services/api/tests/data/test_image_rgb.jpg b/services/api/tests/data/test_image_rgb.jpg new file mode 100644 index 00000000..e131e8ec Binary files /dev/null and b/services/api/tests/data/test_image_rgb.jpg differ diff --git a/services/api/tests/fixtures/fsspec.py b/services/api/tests/fixtures/fsspec.py new file mode 100644 index 00000000..c0891023 --- /dev/null +++ b/services/api/tests/fixtures/fsspec.py @@ -0,0 +1,117 @@ +# type: ignore +import posixpath +from pathlib import Path +from unittest.mock import patch + +import fsspec +import pytest +from fsspec.implementations.local import ( + AbstractFileSystem, + LocalFileSystem, + stringify_path, +) + + +class MockFileSystem(AbstractFileSystem): + protocol = "mock" + + def __init__(self, *args, local_root_dir, **kwargs): + super().__init__() + self._fs = LocalFileSystem(*args, **kwargs) + self.local_root_dir = Path(local_root_dir).resolve().as_posix() + "/" + + def mkdir(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.mkdir(path, *args, **kwargs) + + def makedirs(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.makedirs(path, *args, **kwargs) + + def rmdir(self, path): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.rmdir(path) + + def ls(self, path, detail=True, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + out = self._fs.ls(path, detail=detail, *args, **kwargs) + if detail: + return [{**info, "name": info["name"][len(self.local_root_dir) :]} for info in out] # noqa: E203 + else: + return [name[len(self.local_root_dir) :] for name in out] # noqa: E203 + + def info(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + out = dict(self._fs.info(path, *args, **kwargs)) + out["name"] = out["name"][len(self.local_root_dir) :] # noqa: E203 + return out + + def cp_file(self, path1, path2, *args, **kwargs): + path1 = posixpath.join(self.local_root_dir, self._strip_protocol(path1)) + path2 = posixpath.join(self.local_root_dir, self._strip_protocol(path2)) + return self._fs.cp_file(path1, path2, *args, **kwargs) + + def rm_file(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.rm_file(path, *args, **kwargs) + + def rm(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.rm(path, *args, **kwargs) + + def _open(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs._open(path, *args, **kwargs) + + def created(self, path): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.created(path) + + def modified(self, path): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.modified(path) + + @classmethod + def _strip_protocol(cls, path): + path = stringify_path(path) + if path.startswith("mock://"): + path = path[7:] + return path + + +class TmpDirFileSystem(MockFileSystem): + protocol = "tmp" + tmp_dir = None + + def __init__(self, *args, **kwargs): + assert self.tmp_dir is not None, "TmpDirFileSystem.tmp_dir is not set" + super().__init__(*args, **kwargs, local_root_dir=self.tmp_dir, auto_mkdir=True) + + @classmethod + def _strip_protocol(cls, path): + path = stringify_path(path) + if path.startswith("tmp://"): + path = path[6:] + return path + + [email protected] +def mock_fsspec(): + original_registry = fsspec.registry.copy() + fsspec.register_implementation("mock", MockFileSystem) + fsspec.register_implementation("tmp", TmpDirFileSystem) + yield + fsspec.registry = original_registry + + [email protected] +def mockfs(tmp_path_factory, mock_fsspec): + local_fs_dir = tmp_path_factory.mktemp("mockfs") + return MockFileSystem(local_root_dir=local_fs_dir, auto_mkdir=True) + + [email protected] +def tmpfs(tmp_path_factory, mock_fsspec): + tmp_fs_dir = tmp_path_factory.mktemp("tmpfs") + with patch.object(TmpDirFileSystem, "tmp_dir", tmp_fs_dir): + yield TmpDirFileSystem() diff --git a/services/api/tests/routes/test_rows.py b/services/api/tests/routes/test_rows.py new file mode 100644 index 00000000..b22fe7b8 --- /dev/null +++ b/services/api/tests/routes/test_rows.py @@ -0,0 +1,327 @@ +import os +import time +from http import HTTPStatus +from pathlib import Path +from typing import Any, Generator, List +from unittest.mock import patch + +import numpy as np +import pytest +from datasets import Dataset, Image, concatenate_datasets +from datasets.table import embed_table_storage +from fsspec import AbstractFileSystem +from libcommon.processing_graph import ProcessingStep +from libcommon.simple_cache import _clean_cache_database, upsert_response +from libcommon.storage import StrPath +from libcommon.viewer_utils.asset import update_last_modified_date_of_rows_in_assets_dir + +from api.config import AppConfig +from api.routes.endpoint import StepsByInputTypeAndEndpoint +from api.routes.rows import Indexer, RowsIndex, clean_cached_assets, create_response + + [email protected](autouse=True) +def clean_mongo_databases(app_config: AppConfig) -> None: + _clean_cache_database() + + [email protected] +def ds() -> Dataset: + return Dataset.from_dict({"text": ["Hello there", "General Kenobi"]}) + + [email protected] +def ds_fs(ds: Dataset, tmpfs: AbstractFileSystem) -> Generator[AbstractFileSystem, None, None]: + with tmpfs.open("plain_text/ds-train.parquet", "wb") as f: + ds.to_parquet(f) + yield tmpfs + + [email protected] +def ds_sharded(ds: Dataset) -> Dataset: + return concatenate_datasets([ds] * 4) + + [email protected] +def ds_sharded_fs(ds: Dataset, tmpfs: AbstractFileSystem) -> Generator[AbstractFileSystem, None, None]: + num_shards = 4 + for shard_idx in range(num_shards): + with tmpfs.open(f"plain_text/ds_sharded-train-{shard_idx:05d}-of-{num_shards:05d}.parquet", "wb") as f: + ds.to_parquet(f) + yield tmpfs + + [email protected] +def ds_image(image_path: str) -> Dataset: + ds = Dataset.from_dict({"image": [image_path]}).cast_column("image", Image()) + return Dataset(embed_table_storage(ds.data)) + + [email protected] +def ds_image_fs(ds_image: Dataset, tmpfs: AbstractFileSystem) -> Generator[AbstractFileSystem, None, None]: + with tmpfs.open("plain_text/ds_image-train.parquet", "wb") as f: + ds_image.to_parquet(f) + yield tmpfs + + [email protected] +def dataset_with_config_parquet() -> dict[str, Any]: + config_parquet_content = { + "parquet_files": [ + { + "dataset": "ds", + "config": "plain_text", + "split": "train", + "url": "https://fake.huggingface.co/datasets/ds/resolve/refs%2Fconvert%2Fparquet/plain_text/ds-train.parquet", # noqa: E501 + "filename": "ds-train.parquet", + "size": 128, + } + ] + } + upsert_response( + kind="config-parquet", + dataset="ds", + config="plain_text", + content=config_parquet_content, + http_status=HTTPStatus.OK, + progress=1.0, + ) + return config_parquet_content + + [email protected] +def dataset_sharded_with_config_parquet() -> dict[str, Any]: + num_shards = 4 + config_parquet_content = { + "parquet_files": [ + { + "dataset": "ds_sharded", + "config": "plain_text", + "split": "train", + "url": f"https://fake.huggingface.co/datasets/ds/resolve/refs%2Fconvert%2Fparquet/plain_text/ds_sharded-train-{shard_idx:05d}-of-{num_shards:05d}.parquet", # noqa: E501 + "filename": f"ds_sharded-train-{shard_idx:05d}-of-{num_shards:05d}.parquet", + "size": 128, + } + for shard_idx in range(num_shards) + ] + } + upsert_response( + kind="config-parquet", + dataset="ds_sharded", + config="plain_text", + content=config_parquet_content, + http_status=HTTPStatus.OK, + progress=1.0, + ) + return config_parquet_content + + [email protected] +def dataset_image_with_config_parquet() -> dict[str, Any]: + config_parquet_content = { + "parquet_files": [ + { + "dataset": "ds_image", + "config": "plain_text", + "split": "train", + "url": "https://fake.huggingface.co/datasets/ds/resolve/refs%2Fconvert%2Fparquet/plain_text/ds_image-train.parquet", # noqa: E501 + "filename": "ds_image-train.parquet", + "size": 11128, + } + ] + } + upsert_response( + kind="config-parquet", + dataset="ds_image", + config="plain_text", + content=config_parquet_content, + http_status=HTTPStatus.OK, + progress=1.0, + ) + return config_parquet_content + + [email protected] +def config_parquet_processing_steps(endpoint_definition: StepsByInputTypeAndEndpoint) -> List[ProcessingStep]: + parquet_processing_steps_by_input_type = endpoint_definition.get("/parquet") + if not parquet_processing_steps_by_input_type or not parquet_processing_steps_by_input_type["config"]: + raise RuntimeError("The parquet endpoint is not configured. Exiting.") + return parquet_processing_steps_by_input_type["config"] + + [email protected] +def indexer(app_config: AppConfig, config_parquet_processing_steps: List[ProcessingStep]) -> Indexer: + return Indexer( + config_parquet_processing_steps=config_parquet_processing_steps, + init_processing_steps=[], + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + ) + + [email protected] +def rows_index( + indexer: Indexer, + ds_sharded: Dataset, + ds_sharded_fs: AbstractFileSystem, + dataset_sharded_with_config_parquet: dict[str, Any], +) -> Generator[RowsIndex, None, None]: + with patch("api.routes.rows.get_parquet_fs", return_value=ds_sharded_fs): + yield indexer.get_rows_index("ds_sharded", "plain_text", "train") + + +def test_indexer_get_rows_index( + indexer: Indexer, ds: Dataset, ds_fs: AbstractFileSystem, dataset_with_config_parquet: dict[str, Any] +) -> None: + with patch("api.routes.rows.get_parquet_fs", return_value=ds_fs): + index = indexer.get_rows_index("ds", "plain_text", "train") + assert index.features == ds.features + assert index.row_group_offsets.tolist() == [len(ds)] + assert len(index.row_group_readers) == 1 + row_group_reader = index.row_group_readers[0] + pa_table = row_group_reader() + assert pa_table.to_pydict() == ds.to_dict() + + +def test_indexer_get_rows_index_sharded( + indexer: Indexer, + ds: Dataset, + ds_sharded: Dataset, + ds_sharded_fs: AbstractFileSystem, + dataset_sharded_with_config_parquet: dict[str, Any], +) -> None: + with patch("api.routes.rows.get_parquet_fs", return_value=ds_sharded_fs): + index = indexer.get_rows_index("ds_sharded", "plain_text", "train") + assert index.features == ds_sharded.features + assert index.row_group_offsets.tolist() == np.cumsum([len(ds)] * 4).tolist() + assert len(index.row_group_readers) == 4 + row_group_reader = index.row_group_readers[0] + pa_table = row_group_reader() + assert pa_table.to_pydict() == ds.to_dict() + + +def test_rows_index_query(rows_index: RowsIndex, ds_sharded: Dataset) -> None: + assert rows_index.query(offset=1, length=3).to_pydict() == ds_sharded[1:4] + assert rows_index.query(offset=1, length=-1).to_pydict() == ds_sharded[:0] + assert rows_index.query(offset=1, length=0).to_pydict() == ds_sharded[:0] + assert rows_index.query(offset=999999, length=1).to_pydict() == ds_sharded[:0] + assert rows_index.query(offset=1, length=99999999).to_pydict() == ds_sharded[1:] + with pytest.raises(IndexError): + rows_index.query(offset=-1, length=2) + + +def test_create_response(ds: Dataset, app_config: AppConfig, cached_assets_directory: StrPath) -> None: + response = create_response( + dataset="ds", + config="plain_text", + split="train", + cached_assets_base_url=app_config.cached_assets.base_url, + cached_assets_directory=cached_assets_directory, + pa_table=ds.data, + offset=0, + features=ds.features, + unsupported_columns=[], + ) + assert response["features"] == [{"feature_idx": 0, "name": "text", "type": {"dtype": "string", "_type": "Value"}}] + assert response["rows"] == [ + {"row_idx": 0, "row": {"text": "Hello there"}, "truncated_cells": []}, + {"row_idx": 1, "row": {"text": "General Kenobi"}, "truncated_cells": []}, + ] + + +def test_create_response_with_image( + ds_image: Dataset, app_config: AppConfig, cached_assets_directory: StrPath +) -> None: + response = create_response( + dataset="ds_image", + config="plain_text", + split="train", + cached_assets_base_url=app_config.cached_assets.base_url, + cached_assets_directory=cached_assets_directory, + pa_table=ds_image.data, + offset=0, + features=ds_image.features, + unsupported_columns=[], + ) + assert response["features"] == [{"feature_idx": 0, "name": "image", "type": {"_type": "Image"}}] + assert response["rows"] == [ + { + "row_idx": 0, + "row": { + "image": { + "src": "http://localhost/cached-assets/ds_image/--/plain_text/train/0/image/image.jpg", + "height": 480, + "width": 640, + } + }, + "truncated_cells": [], + } + ] + cached_image_path = Path(cached_assets_directory) / "ds_image/--/plain_text/train/0/image/image.jpg" + assert cached_image_path.is_file() + + [email protected]( + "n_rows,keep_most_recent_rows_number,keep_first_rows_number,max_cleaned_rows_number,expected_remaining_rows", + [ + (8, 1, 1, 100, [0, 7]), + (8, 2, 2, 100, [0, 1, 6, 7]), + (8, 1, 1, 3, [0, 4, 5, 6, 7]), + ], +) +def test_clean_cached_assets( + tmp_path: Path, + n_rows: int, + keep_most_recent_rows_number: int, + keep_first_rows_number: int, + max_cleaned_rows_number: int, + expected_remaining_rows: list[int], +) -> None: + cached_assets_directory = tmp_path / "cached-assets" + split_dir = cached_assets_directory / "ds/--/plain_text/train" + split_dir.mkdir(parents=True) + for i in range(n_rows): + (split_dir / str(i)).mkdir() + time.sleep(0.01) + + def deterministic_glob_rows_in_assets_dir( + dataset: str, + assets_directory: StrPath, + ) -> List[Path]: + return sorted( + list(Path(assets_directory).resolve().glob(os.path.join(dataset, "--", "*", "*", "*"))), + key=lambda p: int(p.name), + ) + + with patch("api.routes.rows.glob_rows_in_assets_dir", deterministic_glob_rows_in_assets_dir): + clean_cached_assets( + "ds", + cached_assets_directory, + keep_most_recent_rows_number=keep_most_recent_rows_number, + keep_first_rows_number=keep_first_rows_number, + max_cleaned_rows_number=max_cleaned_rows_number, + ) + remaining_rows = sorted(int(row_dir.name) for row_dir in split_dir.glob("*")) + assert remaining_rows == expected_remaining_rows + + +def test_update_last_modified_date_of_rows_in_assets_dir(tmp_path: Path) -> None: + cached_assets_directory = tmp_path / "cached-assets" + split_dir = cached_assets_directory / "ds/--/plain_text/train" + split_dir.mkdir(parents=True) + n_rows = 8 + for i in range(n_rows): + (split_dir / str(i)).mkdir() + time.sleep(0.01) + update_last_modified_date_of_rows_in_assets_dir( + dataset="ds", + config="plain_text", + split="train", + offset=2, + length=3, + assets_directory=cached_assets_directory, + ) + most_recent_rows_dirs = sorted([row_dir for row_dir in split_dir.glob("*")], key=os.path.getmtime, reverse=True) + most_recent_rows = [int(row_dir.name) for row_dir in most_recent_rows_dirs] + assert sorted(most_recent_rows[:3]) == [2, 3, 4] + assert most_recent_rows[3:] == [7, 6, 5, 1, 0] diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 06d634c6..a8e1a47b 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -848 +848 @@ name = "cyclonedx-python-lib" -version = "3.1.5" +version = "2.7.1" @@ -854,2 +854,2 @@ files = [ - {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, - {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, + {file = "cyclonedx-python-lib-2.7.1.tar.gz", hash = "sha256:493bf2f30e26c48f305f745ed8580ce10d05a8d68d62a598fe95f05a0d9007dc"}, + {file = "cyclonedx_python_lib-2.7.1-py3-none-any.whl", hash = "sha256:fabc4c8baf722faeea01c3bbca83730e3489dfb37d85c6036baa67a9a7519d40"}, @@ -1702 +1701,0 @@ files = [ - {file = "libclang-15.0.6.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:0bf192c48a8d2992fc5034393ddc99e772ac30e105df84927d62fc88ef8a659f"}, @@ -1723,0 +1723 @@ appdirs = "^1.4.4" +datasets = {version = "^2.11.0", extras = ["audio", "vision"]} @@ -1728,0 +1729 @@ networkx = "^3.0" +numba = "0.56.4" @@ -1730,0 +1732 @@ psutil = "^5.9.4" +pydub = "^0.25.1" @@ -1733,0 +1736 @@ requests = "^2.28.2" +soundfile = ">=0.12.1" @@ -2748 +2750,0 @@ files = [ - {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -2874 +2876 @@ name = "pip-audit" -version = "2.4.14" +version = "2.5.4" @@ -2880,2 +2882,2 @@ files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, + {file = "pip_audit-2.5.4-py3-none-any.whl", hash = "sha256:bee3748030c895488b4dd8a6196fa44f484da000cfd52d8fc64dfa3a2e121624"}, + {file = "pip_audit-2.5.4.tar.gz", hash = "sha256:a4cb03f9e2896d626f5b153973d3ac0d32fdb18594d78d393b153c83bb8089b6"}, @@ -2886 +2888 @@ CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +cyclonedx-python-lib = ">=2.0,<2.5.0 || >2.5.0,<3.0" @@ -2891 +2892,0 @@ pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" @@ -2898 +2899 @@ doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.260)", "types-html5lib", "types-requests", "types-toml"] @@ -4015,18 +4015,0 @@ rsa = ["oauthlib[signedtoken] (>=3.0.0)"] -[[package]] -name = "resolvelib" -version = "1.0.1" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "resolvelib-1.0.1-py2.py3-none-any.whl", hash = "sha256:d2da45d1a8dfee81bdd591647783e340ef3bcb104b54c383f70d422ef5cc7dbf"}, - {file = "resolvelib-1.0.1.tar.gz", hash = "sha256:04ce76cbd63fded2078ce224785da6ecd42b9564b1390793f64ddecbe997b309"}, -] - -[package.extras] -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] - @@ -4232 +4214,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -5413 +5395 @@ python-versions = "3.9.15" -content-hash = "90df140fdcf886f4fae9406c7d5c0c5c3e18278e31d84674f9ae9444569d056d" +content-hash = "02e169574090bd1891985027473ac6278439746c700ceb5a3136b0b99f6223f1" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index 79322c36..bd348c1e 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -58 +58 @@ pandas-stubs = "^1.5.3" -pip-audit = "^2.4.14" +pip-audit = "^2.5.4" diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index cf3c21df..d9a292a0 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -22,0 +23 @@ from libcommon.storage import StrPath +from libcommon.viewer_utils.features import get_cell_value @@ -27 +27,0 @@ from worker.config import AppConfig, FirstRowsConfig -from worker.features import get_cell_value diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 454578f4..d4194e36 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -28,0 +29 @@ from libcommon.storage import StrPath +from libcommon.viewer_utils.features import get_cell_value @@ -31 +31,0 @@ from worker.config import AppConfig, FirstRowsConfig -from worker.features import get_cell_value diff --git a/tools/Python.mk b/tools/Python.mk index 18d1e236..c4ef52db 100644 --- a/tools/Python.mk +++ b/tools/Python.mk @@ -31 +31 @@ pip-audit: - bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^pymongo==/,+109 d' | sed '/^requests==2.28.2 ;/,+2 d' | sed '/^kenlm @/d' | sed '/^fsspec==/,+2 d' | sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" + bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' |sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml index d2efc006..485c15eb 100644 --- a/tools/docker-compose-datasets-server.yml +++ b/tools/docker-compose-datasets-server.yml @@ -8,0 +9 @@ services: + - cached-assets:${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets}:ro @@ -13,0 +15 @@ services: + CACHED_ASSETS_DIRECTORY: ${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets} @@ -53,0 +56,2 @@ services: + volumes: + - cached-assets:${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets} @@ -54,0 +59,6 @@ services: + CACHED_ASSETS_BASE_URL: "http://localhost:${PORT_REVERSE_PROXY-8000}/cached-assets" # hard-coded to work with the reverse-proxy + CACHED_ASSETS_STORAGE_DIRECTORY: ${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets} + CACHED_ASSETS_CLEAN_CACHE_PROBA: ${CACHED_ASSETS_CLEAN_CACHE_PROBA-0.05} + CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER: ${CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER-100} + CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER: ${CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER-200} + CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER: ${CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER-10000} @@ -112,0 +123 @@ volumes: + cached-assets: diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index 1c7e753e..cbb6a092 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -13 +13 @@ services: - CACHE_MONGO_URL: ${CACHE_MONGO_URL-mongodb://localhost:${MONGO_PORT-27017}} # use mongo container by default + CACHE_MONGO_URL: ${CACHE_MONGO_URL-mongodb://host.docker.internal:${MONGO_PORT-27017}} # use mongo container by default @@ -17 +17 @@ services: - QUEUE_MONGO_URL: ${QUEUE_MONGO_URL-mongodb://localhost:${MONGO_PORT-27017}} # use mongo container by default + QUEUE_MONGO_URL: ${QUEUE_MONGO_URL-mongodb://host.docker.internal:${MONGO_PORT-27017}} # use mongo container by default diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index b4be1f51..ffcd1dec 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -8,0 +9 @@ services: + - cached-assets:${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets}:ro @@ -13,0 +15 @@ services: + CACHED_ASSETS_DIRECTORY: ${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets} @@ -46 +48,2 @@ services: - network_mode: host + ports: + - ${ADMIN_UVICORN_PORT-8081}:${ADMIN_UVICORN_PORT-8081} @@ -53,0 +57,2 @@ services: + volumes: + - cached-assets:${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets} @@ -54,0 +60,6 @@ services: + CACHED_ASSETS_BASE_URL: "http://localhost:${PORT_REVERSE_PROXY-8000}/cached-assets" # hard-coded to work with the reverse-proxy + CACHED_ASSETS_STORAGE_DIRECTORY: ${CACHED_ASSETS_STORAGE_DIRECTORY-/cached-assets} + CACHED_ASSETS_CLEAN_CACHE_PROBA: ${CACHED_ASSETS_CLEAN_CACHE_PROBA-0.05} + CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER: ${CACHED_ASSETS_KEEP_FIRST_ROWS_NUMBER-100} + CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER: ${CACHED_ASSETS_KEEP_MOST_RECENT_ROWS_NUMBER-200} + CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER: ${CACHED_ASSETS_MAX_CLEANED_ROWS_NUMBER-10000} @@ -59 +70 @@ services: - API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-0.2} + API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-1.0} @@ -68 +79,2 @@ services: - network_mode: host + ports: + - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080} @@ -77 +89 @@ services: - replicas: 4 + replicas: ${DEV_WORKER_REPLICAS-4} @@ -101 +112,0 @@ services: - network_mode: host @@ -110,0 +122 @@ volumes: + cached-assets:
571d62650ae6249899afc90d9b2df4227a0c9620
Steven Liu
2023-04-11T16:57:07
[docs] Process Parquet files (#987)
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml index 67b9bc5d..cf5709b9 100644 --- a/docs/source/_toctree.yml +++ b/docs/source/_toctree.yml @@ -16 +16,3 @@ - title: List Parquet files + title: List Parquet file + - local: parquet_process + title: Query datasets from Datasets Server diff --git a/docs/source/parquet_process.mdx b/docs/source/parquet_process.mdx new file mode 100644 index 00000000..829465e4 --- /dev/null +++ b/docs/source/parquet_process.mdx @@ -0,0 +1,235 @@ +# Query datasets from Datasets Server + +Datasets Server automatically converts and publishes datasets on the Hub as Parquet files. [Parquet](https://parquet.apache.org/docs/) files are column-based and they shine when you're working with big data. There are several ways you can work with Parquet files, and this guide will show you how to: + +- read and query Parquet files with Pandas and Polars +- connect, read and query Parquet files with DuckDB and DuckDB-Wasm + +## Polars + +[Polars](https://pola-rs.github.io/polars-book/user-guide/introduction.html) is a fast DataFrame library written in Rust with Arrow as its foundation. + +<Tip> + +💡 Learn more about how to get the dataset URLs in the [List Parquet files](parquet) guide. + +</Tip> + +Let's start by grabbing the URLs to the `train` split of the [`blog_authorship_corpus`](https://huggingface.co/datasets/blog_authorship_corpus) dataset from Datasets Server: + +```py +r = requests.get("https://datasets-server.huggingface.co/parquet?dataset=blog_authorship_corpus") +j = r.json() +urls = [f['url'] for f in j['parquet_files'] if f['split'] == 'train'] +urls +['https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet', + 'https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00001-of-00002.parquet'] +``` + +To read from a single Parquet file, use the [`read_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.read_parquet.html) function to read it into a DataFrame and then execute your query: + +```py +import polars as pl + +df = ( + pl.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") + .groupby("horoscope") + .agg( + [ + pl.count(), + pl.col("text").str.n_chars().mean().alias("avg_blog_length") + ] + ) + .sort("avg_blog_length", descending=True) + .limit(5) +) +print(df) +shape: (5, 3) +┌───────────┬───────┬─────────────────┐ +│ horoscope ┆ count ┆ avg_blog_length │ +│ --- ┆ --- ┆ --- │ +│ str ┆ u32 ┆ f64 │ +╞═══════════╪═══════╪═════════════════╡ +│ Aquarius ┆ 34062 ┆ 1129.218836 │ +│ Cancer ┆ 41509 ┆ 1098.366812 │ +│ Capricorn ┆ 33961 ┆ 1073.2002 │ +│ Libra ┆ 40302 ┆ 1072.071833 │ +│ Leo ┆ 40587 ┆ 1064.053687 │ +└───────────┴───────┴─────────────────┘ +``` + +To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.concat.html) function to concatenate the files into a single DataFrame: + +```py +import polars as pl +df = ( + pl.concat([pl.read_parquet(url) for url in urls]) + .groupby("horoscope") + .agg( + [ + pl.count(), + pl.col("text").str.n_chars().mean().alias("avg_blog_length") + ] + ) + .sort("avg_blog_length", descending=True) + .limit(5) +) +print(df) +shape: (5, 3) +┌─────────────┬───────┬─────────────────┐ +│ horoscope ┆ count ┆ avg_blog_length │ +│ --- ┆ --- ┆ --- │ +│ str ┆ u32 ┆ f64 │ +╞═════════════╪═══════╪═════════════════╡ +│ Aquarius ┆ 49568 ┆ 1125.830677 │ +│ Cancer ┆ 63512 ┆ 1097.956087 │ +│ Libra ┆ 60304 ┆ 1060.611054 │ +│ Capricorn ┆ 49402 ┆ 1059.555261 │ +│ Sagittarius ┆ 50431 ┆ 1057.458984 │ +└─────────────┴───────┴─────────────────┘ +``` + +### Lazy API + +Polars offers a [lazy API](https://pola-rs.github.io/polars-book/user-guide/lazy-api/intro.html) that is more performant and memory-efficient for large Parquet files. The LazyFrame API keeps track of what you want to do, and it'll only execute the entire query when you're ready. This way, the lazy API doesn't load everything into RAM beforehand, and it allows you to work with datasets larger than your available RAM. + +To lazily read a Parquet file, use the [`scan_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.scan_parquet.html) function instead. Then, execute the entire query with the [`collect`](https://pola-rs.github.io/polars/py-polars/html/reference/lazyframe/api/polars.LazyFrame.collect.html) function: + +```py +import polars as pl + +q = ( + pl.scan_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") + .groupby("horoscope") + .agg( + [ + pl.count(), + pl.col("text").str.n_chars().mean().alias("avg_blog_length") + ] + ) + .sort("avg_blog_length", descending=True) + .limit(5) +) +df = q.collect() +``` + +## Pandas + +You can also use the popular Pandas DataFrame library to read Parquet files. + +To read from a single Parquet file, use the [`read_parquet`](https://pandas.pydata.org/docs/reference/api/pandas.read_parquet.html) function to read it into a DataFrame: + +```py +import pandas as pd + +df = ( + pd.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") + .groupby('horoscope')['text'] + .apply(lambda x: x.str.len().mean()) + .sort_values(ascending=False) + .head(5) +) +``` + +To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pandas.pydata.org/docs/reference/api/pandas.concat.html) function to concatenate the files into a single DataFrame: + +```py +df = ( + pd.concat([pd.read_parquet(url) for url in urls]) + .groupby('horoscope')['text'] + .apply(lambda x: x.str.len().mean()) + .sort_values(ascending=False) + .head(5) +) +``` + +## DuckDB + +[DuckDB](https://duckdb.org/docs/) is a database that supports reading and querying Parquet files really fast. Begin by creating a connection to DuckDB, and then install and load the [`httpfs`](https://duckdb.org/docs/extensions/httpfs.html) extension to read and write remote files: + +<inferencesnippet> +<python> +```py +import duckdb + +url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet" + +con = duckdb.connect() +con.execute("INSTALL httpfs;") +con.execute("LOAD httpfs;") +``` +</python> +<js> +```js +var duckdb = require('duckdb'); +var db = new duckdb.Database(':memory:'); +var con = db.connect(); +con.exec('INSTALL httpfs'); +con.exec('LOAD httpfs'); + +const url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet" +``` +</js> +</inferencesnippet> + +Now you can write and execute your SQL query on the Parquet file: + +<inferencesnippet> +<python> +```py +con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '{url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)") +┌───────────┬──────────────┬────────────────────┐ +│ horoscope │ count_star() │ avg_blog_length │ +│ varchar │ int64 │ double │ +├───────────┼──────────────┼────────────────────┤ +│ Aquarius │ 34062 │ 1129.218836239798 │ +│ Cancer │ 41509 │ 1098.366812016671 │ +│ Capricorn │ 33961 │ 1073.2002002296751 │ +│ Libra │ 40302 │ 1072.0718326633914 │ +│ Leo │ 40587 │ 1064.0536871412028 │ +└───────────┴──────────────┴────────────────────┘ +``` +</python> +<js> +```js +con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '${url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) { + if (err) { + throw err; + } + console.log(res) +}); +``` +</js> +</inferencesnippet> + +To query multiple files - for example, if the dataset is sharded: + +<inferencesnippet> +<python> +```py +con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet({urls[:2]}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)") +┌─────────────┬──────────────┬────────────────────┐ +│ horoscope │ count_star() │ avg_blog_length │ +│ varchar │ int64 │ double │ +├─────────────┼──────────────┼────────────────────┤ +│ Aquarius │ 49568 │ 1125.8306770497095 │ +│ Cancer │ 63512 │ 1097.95608703867 │ +│ Libra │ 60304 │ 1060.6110539931017 │ +│ Capricorn │ 49402 │ 1059.5552609206104 │ +│ Sagittarius │ 50431 │ 1057.4589835616982 │ +└─────────────┴──────────────┴────────────────────┘ +``` +</python> +<js> +```js +con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet(${JSON.stringify(urls)}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) { + if (err) { + throw err; + } + console.log(res) +}); +``` +</js> +</inferencesnippet> + +[DuckDB-Wasm](https://duckdb.org/docs/api/wasm), a package powered by , is also availabe for running DuckDB in a browser. This could be useful, for instance, if you want to create a web app to query Parquet files from the browser! \ No newline at end of file
af1a46caf0dd93a00b2a234ba937b231158350d5
Polina Kazakova
2023-04-11T14:35:13
Config-level parquet-and-dataset-info (#985)
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml index e4febb35..25e0b578 100644 --- a/.github/workflows/_e2e_tests.yml +++ b/.github/workflows/_e2e_tests.yml @@ -41 +41 @@ jobs: - PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: "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" + PARQUET_AND_INFO_COMMITTER_HF_TOKEN: "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" @@ -80 +80 @@ jobs: - PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: "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" + PARQUET_AND_INFO_COMMITTER_HF_TOKEN: "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" diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 302923c0..96ef959c 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -85 +85 @@ log: -parquetAndDatasetInfo: +parquetAndInfo: diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 32732890..6229aae0 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -86 +86 @@ firstRows: -parquetAndDatasetInfo: +parquetAndInfo: diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index fc3cd97c..1a34a7c5 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -39,6 +39,6 @@ -# specific to the /parquet-and-dataset-info job runner -- name: PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS - value: {{ .Values.parquetAndDatasetInfo.blockedDatasets | quote }} -- name: PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE - value: {{ .Values.parquetAndDatasetInfo.commitMessage | quote }} -- name: PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN +# specific to the /parquet-and-dataset-info and config-parquet-and-info job runners +- name: PARQUET_AND_INFO_BLOCKED_DATASETS + value: {{ .Values.parquetAndInfo.blockedDatasets | quote }} +- name: PARQUET_AND_INFO_COMMIT_MESSAGE + value: {{ .Values.parquetAndInfo.commitMessage | quote }} +- name: PARQUET_AND_INFO_COMMITTER_HF_TOKEN @@ -54,10 +54,10 @@ -- name: PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE - value: {{ .Values.parquetAndDatasetInfo.maxDatasetSize | quote }} -- name: PARQUET_AND_DATASET_INFO_SOURCE_REVISION - value: {{ .Values.parquetAndDatasetInfo.sourceRevision | quote }} -- name: PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS - value: {{ .Values.parquetAndDatasetInfo.supportedDatasets | quote }} -- name: PARQUET_AND_DATASET_INFO_TARGET_REVISION - value: {{ .Values.parquetAndDatasetInfo.targetRevision | quote }} -- name: PARQUET_AND_DATASET_INFO_URL_TEMPLATE - value: {{ .Values.parquetAndDatasetInfo.urlTemplate | quote }} +- name: PARQUET_AND_INFO_MAX_DATASET_SIZE + value: {{ .Values.parquetAndInfo.maxDatasetSize | quote }} +- name: PARQUET_AND_INFO_SOURCE_REVISION + value: {{ .Values.parquetAndInfo.sourceRevision | quote }} +- name: PARQUET_AND_INFO_SUPPORTED_DATASETS + value: {{ .Values.parquetAndInfo.supportedDatasets | quote }} +- name: PARQUET_AND_INFO_TARGET_REVISION + value: {{ .Values.parquetAndInfo.targetRevision | quote }} +- name: PARQUET_AND_INFO_URL_TEMPLATE + value: {{ .Values.parquetAndInfo.urlTemplate | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index 269f31c1..a88727f4 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -148 +148 @@ firstRows: -parquetAndDatasetInfo: +parquetAndInfo: diff --git a/e2e/Makefile b/e2e/Makefile index eca2c0f8..88fff331 100644 --- a/e2e/Makefile +++ b/e2e/Makefile @@ -17 +17 @@ export MONGO_PORT := 27050 -export PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN := 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 +export PARQUET_AND_INFO_COMMITTER_HF_TOKEN := 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 diff --git a/e2e/tests/test_11_auth.py b/e2e/tests/test_11_auth.py index dc21d724..7c018cc6 100644 --- a/e2e/tests/test_11_auth.py +++ b/e2e/tests/test_11_auth.py @@ -41 +41 @@ def test_auth_e2e( - f"/parquet-and-dataset-info?dataset={dataset}", + f"/parquet-and-dataset-info?dataset={dataset}&config={config}", @@ -42,0 +43 @@ def test_auth_e2e( + f"/parquet?dataset={dataset}&config={config}", @@ -43,0 +45 @@ def test_auth_e2e( + f"/dataset-info?dataset={dataset}&config={config}", @@ -44,0 +47 @@ def test_auth_e2e( + f"/size?dataset={dataset}&config={config}", diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 86bbf204..208da39d 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -13,0 +14 @@ from libcommon.constants import ( + PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, @@ -141,0 +143,5 @@ class ProcessingGraphConfig: + "config-parquet-and-info": { + "input_type": "config", + "requires": "/config-names", + "job_runner_version": PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION, + }, @@ -148 +154 @@ class ProcessingGraphConfig: - "requires": "/parquet-and-dataset-info", + "requires": "config-parquet-and-info", @@ -163 +169 @@ class ProcessingGraphConfig: - "requires": "/parquet-and-dataset-info", + "requires": "config-parquet-and-info", @@ -178 +184 @@ class ProcessingGraphConfig: - "requires": "/parquet-and-dataset-info", + "requires": "config-parquet-and-info", diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index db3c31cb..71f30795 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -20,0 +21 @@ PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION = 1 +PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 1 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 59c7cb84..0b4355f8 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -68 +68,2 @@ def graph() -> ProcessingGraph: - ("/config-names", ["/split-names-from-streaming"], []), + ("/config-names", ["/split-names-from-streaming", "config-parquet-and-info"], []), + ("config-parquet-and-info", ["config-parquet", "config-info", "config-size"], ["/config-names"]), @@ -72 +73 @@ def graph() -> ProcessingGraph: - ["/parquet-and-dataset-info", "config-info"], + ["/config-names", "config-parquet-and-info", "config-info"], @@ -82 +83 @@ def graph() -> ProcessingGraph: - ["/parquet-and-dataset-info", "config-info", "/split-names-from-dataset-info"], + ["/config-names", "config-parquet-and-info", "config-info", "/split-names-from-dataset-info"], @@ -89 +90,2 @@ def graph() -> ProcessingGraph: - "/parquet-and-dataset-info", + "/config-names", + "config-parquet-and-info", @@ -92 +93,0 @@ def graph() -> ProcessingGraph: - "/config-names", @@ -96 +97,5 @@ def graph() -> ProcessingGraph: - ("split-first-rows-from-parquet", ["dataset-is-valid"], ["config-parquet", "/parquet-and-dataset-info"]), + ( + "split-first-rows-from-parquet", + ["dataset-is-valid"], + ["config-parquet", "/config-names", "config-parquet-and-info"], + ), @@ -104 +109 @@ def graph() -> ProcessingGraph: - "/parquet-and-dataset-info", + "config-parquet-and-info", @@ -108,7 +113,15 @@ def graph() -> ProcessingGraph: - ("/parquet-and-dataset-info", ["config-parquet", "config-info", "config-size"], []), - ("config-parquet", ["split-first-rows-from-parquet", "dataset-parquet"], ["/parquet-and-dataset-info"]), - ("dataset-parquet", [], ["/parquet-and-dataset-info", "config-parquet"]), - ("config-info", ["dataset-info", "/split-names-from-dataset-info"], ["/parquet-and-dataset-info"]), - ("dataset-info", [], ["/parquet-and-dataset-info", "config-info"]), - ("config-size", ["dataset-size"], ["/parquet-and-dataset-info"]), - ("dataset-size", [], ["/parquet-and-dataset-info", "config-size"]), + ("/parquet-and-dataset-info", [], []), + ( + "config-parquet", + ["split-first-rows-from-parquet", "dataset-parquet"], + ["/config-names", "config-parquet-and-info"], + ), + ("dataset-parquet", [], ["/config-names", "config-parquet-and-info", "config-parquet"]), + ( + "config-info", + ["dataset-info", "/split-names-from-dataset-info"], + ["/config-names", "config-parquet-and-info"], + ), + ("dataset-info", [], ["/config-names", "config-parquet-and-info", "config-info"]), + ("config-size", ["dataset-size"], ["/config-names", "config-parquet-and-info"]), + ("dataset-size", [], ["/config-names", "config-parquet-and-info", "config-size"]), @@ -120 +133 @@ def graph() -> ProcessingGraph: - "/parquet-and-dataset-info", + "config-parquet-and-info", diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 96eb6363..e2986d72 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -127 +127,4 @@ class EndpointConfig: - "/parquet-and-dataset-info": {"dataset": ["/parquet-and-dataset-info"]}, + "/parquet-and-dataset-info": { + "dataset": ["/parquet-and-dataset-info"], + "config": ["config-parquet-and-info"], + }, diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index c2964873..ec72f741 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -48,5 +48,5 @@ def test_endpoints_definition() -> None: - parquet_and_dataset_info = definition["/parquet-and-dataset-info"] - assert parquet_and_dataset_info is not None - assert sorted(list(parquet_and_dataset_info)) == ["dataset"] - assert parquet_and_dataset_info["dataset"] is not None - assert len(parquet_and_dataset_info["dataset"]) == 1 # Only has one processing step + parquet_and_info = definition["/parquet-and-dataset-info"] + assert parquet_and_info is not None + assert sorted(list(parquet_and_info)) == ["config", "dataset"] + assert parquet_and_info["config"] is not None + assert len(parquet_and_info["config"]) == 1 # Only has one processing step diff --git a/services/worker/README.md b/services/worker/README.md index 8ce3ce69..a2436173 100644 --- a/services/worker/README.md +++ b/services/worker/README.md @@ -66,10 +66,10 @@ Also, set the assets-related configuration for the first-rows worker. See [../.. -Set environment variables to configure the parquet worker (`PARQUET_AND_DATASET_INFO_` prefix): - -- `PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS`: comma-separated list of the blocked datasets. If empty, no dataset is blocked. Defaults to empty. -- `PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE`: the git commit message when the worker uploads the parquet files to the Hub. Defaults to `Update parquet files`. -- `PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN`: the user token (https://huggingface.co/settings/tokens) to commit the parquet files to the Hub. The user must be allowed to create the `refs/convert/parquet` branch (see `PARQUET_AND_DATASET_INFO_TARGET_REVISION`) ([Hugging Face organization](https://huggingface.co/huggingface) members have this right). It must also have the right to push to the `refs/convert/parquet` branch ([Datasets maintainers](https://huggingface.co/datasets-maintainers) members have this right). It must have permission to write. If not set, the worker will fail. Defaults to None. -- `PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE`: the maximum size in bytes of the dataset to pre-compute the parquet files. Bigger datasets, or datasets without that information, are ignored. Defaults to `100_000_000`. -- `PARQUET_AND_DATASET_INFO_SOURCE_REVISION`: the git revision of the dataset to use to prepare the parquet files. Defaults to `main`. -- `PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS`: comma-separated list of the supported datasets. The worker does not test the size of supported datasets against the maximum dataset size. Defaults to empty. -- `PARQUET_AND_DATASET_INFO_TARGET_REVISION`: the git revision of the dataset where to store the parquet files. Make sure the committer token (`PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN`) has the permission to write there. Defaults to `refs/convert/parquet`. -- `PARQUET_AND_DATASET_INFO_URL_TEMPLATE`: the URL template to build the parquet file URLs. Defaults to `/datasets/%s/resolve/%s/%s`. +Set environment variables to configure the parquet worker (`PARQUET_AND_INFO_` prefix): + +- `PARQUET_AND_INFO_BLOCKED_DATASETS`: comma-separated list of the blocked datasets. If empty, no dataset is blocked. Defaults to empty. +- `PARQUET_AND_INFO_COMMIT_MESSAGE`: the git commit message when the worker uploads the parquet files to the Hub. Defaults to `Update parquet files`. +- `PARQUET_AND_INFO_COMMITTER_HF_TOKEN`: the user token (https://huggingface.co/settings/tokens) to commit the parquet files to the Hub. The user must be allowed to create the `refs/convert/parquet` branch (see `PARQUET_AND_INFO_TARGET_REVISION`) ([Hugging Face organization](https://huggingface.co/huggingface) members have this right). It must also have the right to push to the `refs/convert/parquet` branch ([Datasets maintainers](https://huggingface.co/datasets-maintainers) members have this right). It must have permission to write. If not set, the worker will fail. Defaults to None. +- `PARQUET_AND_INFO_MAX_DATASET_SIZE`: the maximum size in bytes of the dataset to pre-compute the parquet files. Bigger datasets, or datasets without that information, are ignored. Defaults to `100_000_000`. +- `PARQUET_AND_INFO_SOURCE_REVISION`: the git revision of the dataset to use to prepare the parquet files. Defaults to `main`. +- `PARQUET_AND_INFO_SUPPORTED_DATASETS`: comma-separated list of the supported datasets. The worker does not test the size of supported datasets against the maximum dataset size. Defaults to empty. +- `PARQUET_AND_INFO_TARGET_REVISION`: the git revision of the dataset where to store the parquet files. Make sure the committer token (`PARQUET_AND_INFO_COMMITTER_HF_TOKEN`) has the permission to write there. Defaults to `refs/convert/parquet`. +- `PARQUET_AND_INFO_URL_TEMPLATE`: the URL template to build the parquet file URLs. Defaults to `/datasets/%s/resolve/%s/%s`. diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index 586b4064..716b8476 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -126,7 +126,7 @@ class FirstRowsConfig: -PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE = "Update parquet files" -PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN = None -PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE = 100_000_000 -PARQUET_AND_DATASET_INFO_MAX_EXTERNAL_DATA_FILES = 10_000 -PARQUET_AND_DATASET_INFO_SOURCE_REVISION = "main" -PARQUET_AND_DATASET_INFO_TARGET_REVISION = "refs/convert/parquet" -PARQUET_AND_DATASET_INFO_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s" +PARQUET_AND_INFO_COMMIT_MESSAGE = "Update parquet files" +PARQUET_AND_INFO_COMMITTER_HF_TOKEN = None +PARQUET_AND_INFO_MAX_DATASET_SIZE = 100_000_000 +PARQUET_AND_INFO_MAX_EXTERNAL_DATA_FILES = 10_000 +PARQUET_AND_INFO_SOURCE_REVISION = "main" +PARQUET_AND_INFO_TARGET_REVISION = "refs/convert/parquet" +PARQUET_AND_INFO_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s" @@ -136 +136 @@ PARQUET_AND_DATASET_INFO_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s" -class ParquetAndDatasetInfoConfig: +class ParquetAndInfoConfig: @@ -139,7 +139,7 @@ class ParquetAndDatasetInfoConfig: - commit_message: str = PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE - committer_hf_token: Optional[str] = PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN - max_dataset_size: int = PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE - max_external_data_files: int = PARQUET_AND_DATASET_INFO_MAX_EXTERNAL_DATA_FILES - source_revision: str = PARQUET_AND_DATASET_INFO_SOURCE_REVISION - target_revision: str = PARQUET_AND_DATASET_INFO_TARGET_REVISION - url_template: str = PARQUET_AND_DATASET_INFO_URL_TEMPLATE + commit_message: str = PARQUET_AND_INFO_COMMIT_MESSAGE + committer_hf_token: Optional[str] = PARQUET_AND_INFO_COMMITTER_HF_TOKEN + max_dataset_size: int = PARQUET_AND_INFO_MAX_DATASET_SIZE + max_external_data_files: int = PARQUET_AND_INFO_MAX_EXTERNAL_DATA_FILES + source_revision: str = PARQUET_AND_INFO_SOURCE_REVISION + target_revision: str = PARQUET_AND_INFO_TARGET_REVISION + url_template: str = PARQUET_AND_INFO_URL_TEMPLATE @@ -148 +148 @@ class ParquetAndDatasetInfoConfig: - def from_env(cls) -> "ParquetAndDatasetInfoConfig": + def from_env(cls) -> "ParquetAndInfoConfig": @@ -150 +150 @@ class ParquetAndDatasetInfoConfig: - with env.prefixed("PARQUET_AND_DATASET_INFO_"): + with env.prefixed("PARQUET_AND_INFO_"): @@ -154,8 +154,6 @@ class ParquetAndDatasetInfoConfig: - commit_message=env.str(name="COMMIT_MESSAGE", default=PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE), - committer_hf_token=env.str( - name="COMMITTER_HF_TOKEN", default=PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN - ), - max_dataset_size=env.int(name="MAX_DATASET_SIZE", default=PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE), - source_revision=env.str(name="SOURCE_REVISION", default=PARQUET_AND_DATASET_INFO_SOURCE_REVISION), - target_revision=env.str(name="TARGET_REVISION", default=PARQUET_AND_DATASET_INFO_TARGET_REVISION), - url_template=env.str(name="URL_TEMPLATE", default=PARQUET_AND_DATASET_INFO_URL_TEMPLATE), + commit_message=env.str(name="COMMIT_MESSAGE", default=PARQUET_AND_INFO_COMMIT_MESSAGE), + committer_hf_token=env.str(name="COMMITTER_HF_TOKEN", default=PARQUET_AND_INFO_COMMITTER_HF_TOKEN), + max_dataset_size=env.int(name="MAX_DATASET_SIZE", default=PARQUET_AND_INFO_MAX_DATASET_SIZE), + source_revision=env.str(name="SOURCE_REVISION", default=PARQUET_AND_INFO_SOURCE_REVISION), + target_revision=env.str(name="TARGET_REVISION", default=PARQUET_AND_INFO_TARGET_REVISION), + url_template=env.str(name="URL_TEMPLATE", default=PARQUET_AND_INFO_URL_TEMPLATE), diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index 6ced35a3..7c963ca3 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -29,0 +30 @@ GeneralJobRunnerErrorCode = Literal[ + "ParameterMissingError", @@ -116,0 +118,13 @@ class SplitNotFoundError(GeneralJobRunnerError): +class ParameterMissingError(GeneralJobRunnerError): + """Raised when request is missing some parameter.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__( + message=message, + status_code=HTTPStatus.BAD_REQUEST, + code="ParameterMissingError", + cause=cause, + disclose_cause=False, + ) + + diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index b035dba8..0a3ea836 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -12 +12 @@ from libcommon.storage import StrPath -from worker.config import AppConfig, FirstRowsConfig, ParquetAndDatasetInfoConfig +from worker.config import AppConfig, FirstRowsConfig, ParquetAndInfoConfig @@ -15,0 +16 @@ from worker.job_runners.config.parquet import ConfigParquetJobRunner +from worker.job_runners.config.parquet_and_info import ConfigParquetAndInfoJobRunner @@ -107 +108,9 @@ class JobRunnerFactory(BaseJobRunnerFactory): - parquet_and_dataset_info_config=ParquetAndDatasetInfoConfig.from_env(), + parquet_and_dataset_info_config=ParquetAndInfoConfig.from_env(), + ) + if job_type == ConfigParquetAndInfoJobRunner.get_job_type(): + return ConfigParquetAndInfoJobRunner( + job_info=job_info, + app_config=self.app_config, + processing_step=processing_step, + hf_datasets_cache=self.hf_datasets_cache, + parquet_and_info_config=ParquetAndInfoConfig.from_env(), @@ -200,0 +210 @@ class JobRunnerFactory(BaseJobRunnerFactory): + ConfigParquetAndInfoJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py index fe913f81..a0944827 100644 --- a/services/worker/src/worker/job_runners/config/info.py +++ b/services/worker/src/worker/job_runners/config/info.py @@ -9,2 +9,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError -from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoResponse +from worker.job_runner import ( + CompleteJobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, +) @@ -66,0 +71,2 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons + config (`str`): + Dataset configuration name @@ -73 +79 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - - [`~job_runners.config_info.PreviousStepFormatError`] + - [`~job_runners.config.info.PreviousStepFormatError`] @@ -75 +81 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - - [`~job_runners.config_info.MissingInfoForConfigError`] + - [`~job_runners.config.info.MissingInfoForConfigError`] @@ -84 +90 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - + previous_step = "config-parquet-and-info" @@ -86 +92 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - response = get_response(kind="/parquet-and-dataset-info", dataset=dataset) + response = get_response(kind=previous_step, dataset=dataset, config=config) @@ -88,3 +94 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - raise DatasetNotFoundError( - "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e - ) from e + raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e @@ -93 +97 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - raise PreviousStepStatusError(f"Previous step raised an error: {response['http_status']}..") + raise PreviousStepStatusError(f"Previous step '{previous_step}' raised an error: {response['http_status']}..") @@ -94,0 +99 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons + content = response["content"] @@ -96,3 +101 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - content = ParquetAndDatasetInfoResponse( - parquet_files=response["content"]["parquet_files"], dataset_info=response["content"]["dataset_info"] - ) + config_info = content["dataset_info"] @@ -100,17 +103,3 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - raise PreviousStepFormatError("Previous step did not return the expected content: 'dataset_info'.", e) from e - - if config not in content["dataset_info"]: - if not isinstance(content["dataset_info"], dict): - raise PreviousStepFormatError( - "Previous step did not return the expected content.", - TypeError(f"dataset_info should be a dict, but got {type(content['dataset_info'])}"), - ) - raise MissingInfoForConfigError( - f"Dataset configuration '{config}' is missing in the dataset info from the parquet export. " - f"Available configurations: {', '.join(list(content['dataset_info'])[:10])}" - + f"... ({len(content['dataset_info']) - 10})" - if len(content["dataset_info"]) > 10 - else "" - ) - try: - config_info = content["dataset_info"][config] + raise PreviousStepFormatError( + f"Previous step '{previous_step}' did not return the expected content: 'dataset_info'.", e + ) from e @@ -118,2 +107,5 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - except Exception as e: - raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + if not isinstance(config_info, dict): + raise PreviousStepFormatError( + "Previous step did not return the expected content.", + TypeError(f"dataset_info should be a dict, but got {type(config_info)}"), + ) @@ -135 +127 @@ class ConfigInfoJobRunner(JobRunner): - raise ValueError("dataset is required") + raise ParameterMissingError("'dataset' parameter is required") @@ -137 +129 @@ class ConfigInfoJobRunner(JobRunner): - raise ValueError("config is required") + raise ParameterMissingError("'config' parameter is required") diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py index 6f3a2803..aca97cdc 100644 --- a/services/worker/src/worker/job_runners/config/parquet.py +++ b/services/worker/src/worker/job_runners/config/parquet.py @@ -12,4 +12,5 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError -from worker.job_runners.parquet_and_dataset_info import ( - ParquetAndDatasetInfoResponse, - ParquetFileItem, +from worker.job_runner import ( + CompleteJobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, @@ -16,0 +18 @@ from worker.job_runners.parquet_and_dataset_info import ( +from worker.job_runners.config.parquet_and_info import ParquetFileItem @@ -21 +22,0 @@ ConfigParquetJobRunnerErrorCode = Literal[ - "MissingInfoForConfigError", @@ -59,7 +59,0 @@ class PreviousStepFormatError(ConfigParquetJobRunnerError): -class MissingInfoForConfigError(ConfigParquetJobRunnerError): - """Raised when the dataset info from the parquet export is missing the requested dataset configuration.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "MissingInfoForConfigError", cause, False) - - @@ -83,2 +76,0 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - - [`~job_runners.config_size.MissingInfoForConfigError`] - If the dataset info from the parquet export is missing the requested dataset configuration @@ -91,0 +84 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse + previous_step = "config-parquet-and-info" @@ -93 +86 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - response = get_response(kind="/parquet-and-dataset-info", dataset=dataset) + response = get_response(kind=previous_step, dataset=dataset, config=config) @@ -95,3 +88 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - raise DatasetNotFoundError( - "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e - ) from e + raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e @@ -100,20 +90,0 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - try: - content = ParquetAndDatasetInfoResponse( - parquet_files=response["content"]["parquet_files"], dataset_info=response["content"]["dataset_info"] - ) - except Exception as e: - raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e - - if config not in content["dataset_info"]: - if not isinstance(content["dataset_info"], dict): - raise PreviousStepFormatError( - "Previous step did not return the expected content.", - TypeError(f"dataset_info should be a dict, but got {type(content['dataset_info'])}"), - ) - raise MissingInfoForConfigError( - f"Dataset configuration '{config}' is missing in the dataset info from the parquet export. " - f"Available configurations: {', '.join(list(content['dataset_info'])[:10])}" - + f"... ({len(content['dataset_info']) - 10})" - if len(content["dataset_info"]) > 10 - else "" - ) @@ -120,0 +92 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse + content = response["content"] @@ -122,0 +95 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse + @@ -136,0 +110,2 @@ class ConfigParquetJobRunner(JobRunner): + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") @@ -138 +113 @@ class ConfigParquetJobRunner(JobRunner): - raise ValueError("config is required") + raise ParameterMissingError("'config' parameter is required") @@ -144 +119 @@ class ConfigParquetJobRunner(JobRunner): - SplitFullName(dataset=parquet_file["dataset"], config=parquet_file["config"], split=parquet_file["split"]) + SplitFullName(dataset=self.dataset, config=self.config, split=parquet_file["split"]) diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py new file mode 100644 index 00000000..bae66220 --- /dev/null +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -0,0 +1,969 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import contextlib +import glob +import logging +import re +from functools import partial +from http import HTTPStatus +from multiprocessing.pool import ThreadPool +from pathlib import Path +from typing import Any, Dict, List, Literal, Mapping, Optional, Set, Tuple, TypedDict +from urllib.parse import quote + +import datasets +import datasets.config +import numpy as np +import requests +from datasets import DownloadConfig, get_dataset_config_info, load_dataset_builder +from datasets.builder import DatasetBuilder +from datasets.data_files import EmptyDatasetError as _EmptyDatasetError +from datasets.download import StreamingDownloadManager +from datasets.utils.file_utils import ( + get_authentication_headers_for_url, + http_head, + is_relative_path, + url_or_path_join, +) +from datasets.utils.py_utils import asdict, map_nested +from huggingface_hub._commit_api import ( + CommitOperation, + CommitOperationAdd, + CommitOperationDelete, +) +from huggingface_hub.hf_api import DatasetInfo, HfApi, RepoFile +from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFoundError +from libcommon.constants import PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION +from libcommon.dataset import DatasetNotFoundError, ask_access +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import JobInfo +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response + +from worker.config import AppConfig, ParquetAndInfoConfig +from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError +from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.job_runners.config_names import ConfigNamesError + +ConfigParquetAndInfoJobRunnerErrorCode = Literal[ + "DatasetRevisionNotFoundError", + "EmptyDatasetError", + "DatasetInBlockListError", + "DatasetTooBigFromHubError", + "DatasetTooBigFromDatasetsError", + "UnsupportedExternalFilesError", + "DatasetWithTooManyExternalFilesError", + "DatasetWithTooBigExternalFilesError", + "ExternalFilesSizeRequestHTTPError", + "ExternalFilesSizeRequestConnectionError", + "ExternalFilesSizeRequestTimeoutError", + "ExternalFilesSizeRequestError", + "PreviousStepStatusError", + "PreviousStepFormatError", +] + + +class ConfigParquetAndInfoJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: ConfigParquetAndInfoJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class DatasetRevisionNotFoundError(ConfigParquetAndInfoJobRunnerError): + """Raised when the revision of a dataset repository does not exist.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_FOUND, "DatasetRevisionNotFoundError", cause, False) + + +class EmptyDatasetError(ConfigParquetAndInfoJobRunnerError): + """Raised when the dataset has no data.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "EmptyDatasetError", cause, True) + + +class DatasetInBlockListError(ConfigParquetAndInfoJobRunnerError): + """Raised when the dataset is in the list of blocked datasets.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetInBlockListError", cause, False) + + +class DatasetTooBigFromHubError(ConfigParquetAndInfoJobRunnerError): + """Raised when the dataset size (sum of files on the Hub) is too big.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromHubError", cause, False) + + +class DatasetTooBigFromDatasetsError(ConfigParquetAndInfoJobRunnerError): + """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromDatasetsError", cause, False) + + +class PreviousStepStatusError(ConfigParquetAndInfoJobRunnerError): + """Raised when the previous step gave an error. The job should not have been created.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) + + +class PreviousStepFormatError(ConfigParquetAndInfoJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +class ParquetFileItem(TypedDict): + dataset: str + config: str + split: str + url: str + filename: str + size: int + + +class ConfigParquetAndInfoResponse(TypedDict): + parquet_files: List[ParquetFileItem] + dataset_info: Dict[str, Any] + + +DATASET_TYPE = "dataset" + + +class ParquetFile: + def __init__(self, local_file: str, local_dir: str, config: str): + if not local_file.startswith(local_dir): + raise ValueError(f"{local_file} is not in {local_dir}") + self.local_file = local_file + self.local_dir = local_dir + self.config = config + + def repo_file(self) -> str: + return f'{self.config}/{self.local_file.removeprefix(f"{self.local_dir}/")}' + + +# TODO: use huggingface_hub's hf_hub_url after +# https://github.com/huggingface/huggingface_hub/issues/1082 +def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url_template: str) -> str: + return (hf_endpoint + url_template) % (repo_id, quote(revision, safe=""), filename) + + +p = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>\w+(\.\w+)*?)(-[0-9]{5}-of-[0-9]{5})?.parquet") + + +def parse_repo_filename(filename: str) -> Tuple[str, str]: + parts = filename.split("/") + if len(parts) != 2: + raise ValueError(f"Invalid filename: {filename}") + config, fname = parts + m = p.match(fname) + if not m: + raise ValueError(f"Cannot parse {filename}") + split = m.group("split") + return config, split + + +def create_parquet_file_item( + repo_file: RepoFile, + dataset: str, + config: str, + hf_endpoint: str, + target_revision: str, + url_template: str, +) -> ParquetFileItem: + if repo_file.size is None: + raise ValueError(f"Cannot get size of {repo_file.rfilename}") + _, split = parse_repo_filename(repo_file.rfilename) + return { + "dataset": dataset, + "config": config, + "split": split, + "url": hf_hub_url( + repo_id=dataset, + filename=repo_file.rfilename, + hf_endpoint=hf_endpoint, + revision=target_revision, + url_template=url_template, + ), + "filename": Path(repo_file.rfilename).name, + "size": repo_file.size, + } + + +def raise_if_blocked( + dataset: str, + blocked_datasets: List[str], +) -> None: + """ + Raise an error if the dataset is in the list of blocked datasets + + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + blocked_datasets (`List[str]`): + The list of blocked datasets. If empty, no dataset is blocked. + Returns: + `None` + <Tip> + Raises the following errors: + - [`~job_runners.config.parquet_and_info.DatasetInBlockListError`] + If the dataset is in the list of blocked datasets. + </Tip> + """ + if dataset in blocked_datasets: + raise DatasetInBlockListError( + "The parquet conversion has been disabled for this dataset for now. Please open an issue in" + " https://github.com/huggingface/datasets-server if you want this dataset to be supported." + ) + + +def get_dataset_info_or_raise( + dataset: str, + hf_endpoint: str, + hf_token: Optional[str], + revision: str, +) -> DatasetInfo: + """ + Return the dataset info if possible. + Raise an error if the dataset cannot be accessed (does not exist, gated with extra fields, private) + + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + hf_endpoint (`str`): + The Hub endpoint (for example: "https://huggingface.co") + hf_token (`str`, `optional`): + An app authentication token with read access to all the datasets. + revision (`str`): + The git revision (e.g. "main" or sha) of the dataset + Returns: + `DatasetInfo`: The dataset info + <Tip> + Raises the following errors: + - [`~.job_runner.DatasetNotFoundError`] + If the repository to download from cannot be found. This may be because it doesn't exist, + or because it is set to `private` and you do not have access. + - [`~job_runners.config.parquet_and_info.DatasetRevisionNotFoundError`] + If the revision does not exist or cannot be accessed using the token. + </Tip> + """ + try: + dataset_info = HfApi(endpoint=hf_endpoint, token=hf_token).dataset_info( + repo_id=dataset, revision=revision, files_metadata=True + ) + except RepositoryNotFoundError as err: + raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err + except RevisionNotFoundError as err: + raise DatasetRevisionNotFoundError("The dataset revision does not exist on the Hub.") from err + if dataset_info.private: + raise DatasetNotFoundError("The dataset does not exist on the Hub.") + return dataset_info + + +def raise_if_too_big_from_hub( + dataset_info: DatasetInfo, + max_dataset_size: int, +) -> None: + """ + Raise an error if the dataset is too big to be converted to parquet + + Args: + dataset_info (`DatasetInfo`): + The dataset info + max_dataset_size (`int`): + The maximum size of the dataset in bytes + Returns: + `None` + <Tip> + Raises the following errors: + - [`~job_runners.config.parquet_and_info.DatasetTooBigFromHubError`] + If the dataset is too big to be converted to parquet + </Tip> + """ + dataset_size: int = sum(sibling.size for sibling in dataset_info.siblings if sibling.size is not None) + if dataset_size > max_dataset_size: + raise DatasetTooBigFromHubError( + f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes. " + f"Current size of files on the hub is {dataset_size} bytes." + ) + + +def raise_if_too_big_from_datasets( + dataset: str, + config: str, + hf_endpoint: str, + hf_token: Optional[str], + revision: str, + max_dataset_size: int, +) -> None: + """ + Raise an error if the dataset is too big to be converted to parquet, as measured by the sum of the configs + sizes given by the datasets library + + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + config (`str`): + Dataset configuration name + hf_endpoint (`str`): + The Hub endpoint (for example: "https://huggingface.co") + hf_token (`str`, `optional`): + An app authentication token with read access to all the datasets. + revision (`str`): + The git revision (e.g. "main" or sha) of the dataset + max_dataset_size (`int`): + The maximum size of the dataset in bytes + Returns: + `None` + <Tip> + Raises the following errors: + - [`ValueError`] + If the datasets.config.HF_ENDPOINT is not set to the expected value + - [`~job_runners.config.parquet_and_info.DatasetTooBigFromDatasetsError`] + If the dataset is too big to be converted to parquet + </Tip> + """ + if datasets.config.HF_ENDPOINT != hf_endpoint: + raise ValueError( + f"Invalid datasets.config.HF_ENDPOINT value: '{datasets.config.HF_ENDPOINT}'. Please set it to:" + f" '{hf_endpoint}'." + ) + dataset_size = 0 + with contextlib.suppress(Exception): + info = get_dataset_config_info(path=dataset, config_name=config, revision=revision, use_auth_token=hf_token) + dataset_size = info.dataset_size if info.dataset_size is not None else 0 + if dataset_size > max_dataset_size: + raise DatasetTooBigFromDatasetsError( + f"The dataset is too big to be converted to Parquet. The size of the dataset ({dataset_size} B, as given" + f" per the datasets library) exceeds the maximum supported size ({max_dataset_size} B). Please report the" + " issue." + ) + + +def raise_if_not_supported( + dataset: str, + config: str, + hf_endpoint: str, + hf_token: Optional[str], + committer_hf_token: Optional[str], + revision: str, + supported_datasets: List[str], + blocked_datasets: List[str], + max_dataset_size: int, +) -> None: + """ + Raise an error if the dataset is not supported: + - if the dataset is in the list of blocked datasets + - if the dataset cannot be accessed (does not exist, gated with extra fields, private) + - if the dataset is too big, and not in the list of supported datasets + + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + config (`str`): + Dataset configuration name + hf_endpoint (`str`): + The Hub endpoint (for example: "https://huggingface.co") + hf_token (`str`, `optional`): + An app authentication token with read access to all the datasets. + committer_hf_token (`str`, `optional`): + A user authentication token (See https://huggingface.co/settings/token) with write access. It must: + - be part of the `huggingface` organization (to create the ref/convert/parquet "branch") + - be part of the `datasets-maintainers` organization (to push to the ref/convert/parquet "branch") + revision (`str`): + The git revision (e.g. "main" or sha) of the dataset + supported_datasets (`List[str]`): + The list of supported datasets, saving the blocked datasets. If empty, all datasets are supported + (saving the blocked datasets). + blocked_datasets (`List[str]`): + The list of blocked datasets. If empty, no dataset is blocked. + max_dataset_size (`int`): + The maximum size of a dataset in bytes. If the dataset is under the limit (which means that the size + can be fetched), it will be allowed. + Returns: + `ParquetResponseResult`: An object with the parquet_response + (dataset and list of parquet files) and the dataset_git_revision (sha) if any. + <Tip> + Raises the following errors: + - [`~job_runners.config.parquet_and_info.DatasetInBlockListError`] + If the dataset is in the list of blocked datasets. + - [`~libcommon.dataset.GatedExtraFieldsError`]: if the dataset is gated, with extra fields. + Programmatic access is not implemented for this type of dataset because there is no easy + way to get the list of extra fields. + - [`~libcommon.dataset.GatedDisabledError`]: if the dataset is gated, but disabled. + - [`~libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the + token does not give the sufficient access to the dataset, or if the dataset is private + (private datasets are not supported by the datasets server) + - ['~requests.exceptions.HTTPError']: any other error when asking access + - [`~job_runners.config.parquet_and_info.DatasetRevisionNotFoundError`] + If the revision does not exist or cannot be accessed using the token. + - [`~job_runners.config.parquet_and_info.DatasetTooBigFromHubError`] + If the dataset is too big to be converted to parquet + - [`ValueError`] + If the datasets.config.HF_ENDPOINT is not set to the expected value + - [`~job_runners.config.parquet_and_info.DatasetTooBigFromDatasetsError`] + If the dataset is too big to be converted to parquet + </Tip> + """ + raise_if_blocked(dataset=dataset, blocked_datasets=blocked_datasets) + ask_access(dataset=dataset, hf_endpoint=hf_endpoint, hf_token=committer_hf_token) + dataset_info = get_dataset_info_or_raise( + dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, revision=revision + ) + if dataset in supported_datasets: + return + raise_if_too_big_from_datasets( + dataset=dataset, + config=config, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + revision=revision, + max_dataset_size=max_dataset_size, + ) + raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size) + + +class EmptySplitsError(Exception): + pass + + +class SplitInfoFormatError(Exception): + pass + + +class EmptyConfigNameError(Exception): + pass + + +class EmptyDownloadSizeError(Exception): + pass + + +class EmptyFeaturesError(Exception): + pass + + +class DatasetWithTooManyExternalFilesError(ConfigParquetAndInfoJobRunnerError): + """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyExternalFilesError", cause, True) + + +class DatasetWithTooBigExternalFilesError(ConfigParquetAndInfoJobRunnerError): + """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooBigExternalFilesError", cause, True) + + +class UnsupportedExternalFilesError(ConfigParquetAndInfoJobRunnerError): + """Raised when we failed to get the size of the external files.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "UnsupportedExternalFilesError", cause, True) + + +class ExternalFilesSizeRequestHTTPError(ConfigParquetAndInfoJobRunnerError): + """Raised when we failed to get the size of the external files.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestHTTPError", cause, True) + + +class ExternalFilesSizeRequestConnectionError(ConfigParquetAndInfoJobRunnerError): + """Raised when we failed to get the size of the external files.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestConnectionError", cause, True) + + +class ExternalFilesSizeRequestTimeoutError(ConfigParquetAndInfoJobRunnerError): + """Raised when we failed to get the size of the external files.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestTimeoutError", cause, True) + + +class ExternalFilesSizeRequestError(ConfigParquetAndInfoJobRunnerError): + """Raised when we failed to get the size of the external files.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "ExternalFilesSizeRequestError", cause, True) + + +def _request_size(url: str, hf_token: Optional[str] = None) -> Optional[int]: + headers = get_authentication_headers_for_url(url, use_auth_token=hf_token) + response = http_head(url, headers=headers, max_retries=3) + response.raise_for_status() + size = response.headers.get("Content-Length") if response.ok else None + return int(size) if size is not None else size + + +class _MockStreamingDownloadManager(StreamingDownloadManager): # type: ignore + def __init__(self, *args: Any, **kwargs: Any) -> None: + super().__init__(*args, **kwargs) + self.ext_data_files: List[str] = [] + + def download(self, url_or_urls: Any) -> Any: + url_or_urls = map_nested( + self._download, + url_or_urls, + map_tuple=True, + parallel_min_length=np.inf, + # ^ parallel_min_length has int type, but is currently used in datasets for a comparison only + # and it works with np.inf. No conversion is involved + # (would raise: OverflowError: cannot convert float infinity to integer) + ) + return url_or_urls + + def _download(self, urlpath: Any) -> str: + urlpath_str = str(urlpath) + if is_relative_path(urlpath_str): + # append the relative path to the base_path + urlpath_str = url_or_path_join(self._base_path, urlpath_str) + elif not urlpath_str.startswith(self._base_path): + # it's an external file + self.ext_data_files.append(urlpath_str) + return urlpath_str + + +def raise_if_too_big_from_external_data_files( + builder: DatasetBuilder, max_dataset_size: int, max_external_data_files: int, hf_token: Optional[str] +) -> None: + # Packaged dataset modules only load data files that are inside the dataset repository. + # No need to check them since they're already caught by `raise_if_too_big_from_hub` + if type(builder).__module__.startswith("datasets."): + return + # For datasets with a loading script however, we need to check the downloaded files + mock_dl_manager = _MockStreamingDownloadManager( + base_path=builder.base_path, download_config=DownloadConfig(use_auth_token=hf_token) + ) + try: + builder._split_generators(mock_dl_manager) + except (requests.exceptions.RequestException, NotImplementedError) as error: + if isinstance(error, NotImplementedError): + # we can ignore the errors from functions not implemented in streaming mode like `.extract()` on TAR files + if "is not implemented in streaming mode." not in str(error): + raise UnsupportedExternalFilesError( + ( + "Couldn't get the list of external files in `_split_generators` because it doesn't support" + f" streaming:\n{error}" + ), + error, + ) from error + elif isinstance(error, requests.exceptions.HTTPError): + raise ExternalFilesSizeRequestHTTPError( + ( + "Couldn't get the list of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + elif isinstance(error, requests.exceptions.ConnectionError): + raise ExternalFilesSizeRequestConnectionError( + ( + "Couldn't get the list of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + elif isinstance(error, requests.exceptions.Timeout): + raise ExternalFilesSizeRequestTimeoutError( + ( + "Couldn't get the list of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + else: + raise ExternalFilesSizeRequestError( + ( + "Couldn't get the list of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + ext_data_files = mock_dl_manager.ext_data_files + if len(ext_data_files) > max_external_data_files: + raise DatasetWithTooManyExternalFilesError( + f"The conversion to parquet is limited to datasets with less than {max_external_data_files} files. " + f"However it uses {len(ext_data_files)} data files." + ) + elif ext_data_files: + try: + with ThreadPool(16) as pool: + total_size = 0 + get_size = partial(_request_size, hf_token=hf_token) + for i, size in enumerate(pool.imap_unordered(get_size, ext_data_files)): + if size is not None: + total_size += size + if total_size > max_dataset_size: + raise DatasetWithTooBigExternalFilesError( + f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes." + f" However {i + 1} data files of {len(ext_data_files)} are already bigger than" + f" {total_size} bytes." + ) + except requests.exceptions.RequestException as error: + if isinstance(error, requests.exceptions.HTTPError): + raise ExternalFilesSizeRequestHTTPError( + ( + "Couldn't get the size of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + elif isinstance(error, requests.exceptions.ConnectionError): + raise ExternalFilesSizeRequestConnectionError( + ( + "Couldn't get the size of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + elif isinstance(error, requests.exceptions.Timeout): + raise ExternalFilesSizeRequestTimeoutError( + ( + "Couldn't get the size of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + else: + raise ExternalFilesSizeRequestError( + ( + "Couldn't get the size of external files in `_split_generators` because a request" + f" failed:\n{error}\nPlease consider moving your data files in this dataset repository instead" + " (e.g. inside a data/ folder)." + ), + error, + ) from error + + +def compute_config_parquet_and_info_response( + dataset: str, + config: str, + hf_endpoint: str, + hf_token: Optional[str], + committer_hf_token: Optional[str], + source_revision: str, + target_revision: str, + commit_message: str, + url_template: str, + supported_datasets: List[str], + blocked_datasets: List[str], + max_dataset_size: int, + max_external_data_files: int, +) -> ConfigParquetAndInfoResponse: + """ + Get the response of /parquet-and-dataset-info for one specific dataset on huggingface.co. + It is assumed that the dataset can be accessed with the token. + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + config (`str`): + Dataset configuration name + hf_endpoint (`str`): + The Hub endpoint (for example: "https://huggingface.co") + hf_token (`str`, `optional`): + An app authentication token with read access to all the datasets. + committer_hf_token (`str`, `optional`): + A user authentication token (See https://huggingface.co/settings/token) with write access. It must: + - be part of the `huggingface` organization (to create the ref/convert/parquet "branch") + - be part of the `datasets-maintainers` organization (to push to the ref/convert/parquet "branch") + source_revision (`str`): + The git revision (e.g. "main" or sha) of the dataset used to prepare the parquet files + target_revision (`str`): + The target git revision (e.g. "ref/convert/parquet") of the dataset where to store the parquet files + commit_message (`str`): + The commit message to use when storing the parquet files + url_template (`str`): + The template to use to build the parquet file url + supported_datasets (`List[str]`): + The list of supported datasets, saving the blocked datasets. If empty, all datasets are supported + (saving the blocked datasets). + blocked_datasets (`List[str]`): + The list of blocked datasets. If empty, no dataset is blocked. + max_dataset_size (`int`): + The maximum size of a dataset in bytes. If the dataset is under the limit (which means that the size + can be fetched), it will be allowed. + max_external_data_files (`int`): + The maximum number of external data files of a dataset. This is for datasets with loading scripts only. + Returns: + `ConfigParquetAndInfoResponse`: An object with the config_parquet_and_info_response + (dataset info and list of parquet files). + <Tip> + Raises the following errors: + - [`~job_runners.config.parquet_and_info.DatasetInBlockListError`] + If the dataset is in the list of blocked datasets. + - [`libcommon.dataset.GatedExtraFieldsError`]: if the dataset is gated, with extra fields. + Programmatic access is not implemented for this type of dataset because there is no easy + way to get the list of extra fields. + - [`libcommon.dataset.GatedDisabledError`]: if the dataset is gated, but disabled. + - [`libcommon.dataset.DatasetNotFoundError`]: if the dataset does not exist, or if the + token does not give the sufficient access to the dataset, or if the dataset is private + (private datasets are not supported by the datasets server) + - ['HTTPError'](https://requests.readthedocs.io/en/latest/api/#requests.HTTPError): any other error when + asking access + - [`~job_runners.config.parquet_and_info.DatasetRevisionNotFoundError`] + If the revision does not exist or cannot be accessed using the token. + - [`~job_runners.config.parquet_and_info.DatasetTooBigFromHubError`] + If the dataset is too big to be converted to parquet + - [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) + If the datasets.config.HF_ENDPOINT is not set to the expected value + - [`~job_runners.config.parquet_and_info.DatasetTooBigFromDatasetsError`] + If the dataset is too big to be converted to parquet + - [`~job_runners.config.parquet_and_info.EmptyDatasetError`] + The dataset is empty. + - [`~job_runners.config.parquet_and_info.ConfigNamesError`] + If the list of configurations could not be obtained using the datasets library. + - [`~job_runners.config.parquet_and_info.DatasetInBlockListError`] + If the dataset is in the list of blocked datasets. + - [`~job_runners.config.parquet_and_info.DatasetWithTooManyExternalFilesError`] + If the dataset has too many external files to be converted to parquet + - [`~job_runners.config.parquet_and_info.DatasetWithTooBigExternalFilesError`] + If the dataset is too big external files be converted to parquet + - [`~job_runners.config.parquet_and_info.UnsupportedExternalFilesError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`~job_runners.config.parquet_and_info.ExternalFilesSizeRequestHTTPError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`~job_runners.config.parquet_and_info.ExternalFilesSizeRequestConnectionError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`~job_runners.config.parquet_and_info.ExternalFilesSizeRequestTimeoutError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`~job_runners.config.parquet_and_info.ExternalFilesSizeRequestError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`~job_runners.config.parquet_and_info.PreviousStepStatusError`] + If the previous step gave an error. + - [`~job_runners.config.parquet_and_info.PreviousStepFormatError`] + If the content of the previous step has not the expected format + + </Tip> + """ + logging.info(f"get parquet files and dataset info for {dataset=} {config=}") + + raise_if_not_supported( + dataset=dataset, + config=config, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + committer_hf_token=committer_hf_token, + revision=source_revision, + supported_datasets=supported_datasets, + blocked_datasets=blocked_datasets, + max_dataset_size=max_dataset_size, + ) + + logging.info(f"get config names for {dataset=}") + previous_step = "/config-names" + try: + response = get_response(kind=previous_step, dataset=dataset) + except DoesNotExist as e: + raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e + if response["http_status"] != HTTPStatus.OK: + raise PreviousStepStatusError(f"Previous step {previous_step} gave an error: {response['http_status']}..") + + config_names_content = response["content"] + if "config_names" not in config_names_content: + raise PreviousStepFormatError("Previous step did not return the expected content: 'config_names'.") + + if not isinstance(config_names_content["config_names"], list): + raise PreviousStepFormatError( + "Previous step did not return the expected content.", + TypeError(f"config_names should be a list, but got {type(config_names_content['config_names'])}"), + ) + + config_names = {config_name_item["config"] for config_name_item in config_names_content["config_names"]} + if config not in config_names: + raise ConfigNamesError(f"{config=} does not exist in {dataset=}") + + hf_api = HfApi(endpoint=hf_endpoint, token=hf_token) + committer_hf_api = HfApi(endpoint=hf_endpoint, token=committer_hf_token) + + # prepare the parquet files locally + local_parquet_files: List[ParquetFile] = [] + download_config = DownloadConfig(delete_extracted=True) + try: + builder = load_dataset_builder( + path=dataset, + name=config, + revision=source_revision, + use_auth_token=hf_token, + download_config=download_config, + ) + except _EmptyDatasetError as err: + raise EmptyDatasetError(f"{dataset=} is empty.", cause=err) from err + raise_if_too_big_from_external_data_files( + builder=builder, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + hf_token=hf_token, + ) + builder.download_and_prepare(file_format="parquet") # the parquet files are stored in the cache dir + dataset_info = asdict(builder.info) + local_parquet_files.extend( + ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=config) + for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet") + ) + + # create the target revision if it does not exist yet (clone from initial commit to avoid cloning all repo's files) + try: + refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE) + if all(ref.ref != target_revision for ref in refs.converts): + initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id + committer_hf_api.create_branch( + repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit + ) + except RepositoryNotFoundError as err: + raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err + + target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False) + # - get repo parquet files + all_repo_files: Set[str] = {f.rfilename for f in target_dataset_info.siblings} + repo_parquet_files: Set[str] = {file for file in all_repo_files if file.endswith(".parquet")} + # - get parquet files for current config + config_files_to_add: Dict[str, str] = { + parquet_file.repo_file(): parquet_file.local_file for parquet_file in local_parquet_files + } + # - get files that will be preserved in repo: files belonging to other configs and .gitattributes + # we exclude files of current config because otherwise outdated files might be preserved + files_to_ignore: Set[str] = { + file + for other_config in config_names.difference({config}) + for file in repo_parquet_files + if file.startswith(f"{other_config}/") + }.union({".gitattributes"}) + # - get files to be deleted - all files except for: + # - parquet files obtained for current config at this processing step, + # - parquet files belonging to other existing configs + # - .gitattributes + files_to_delete = all_repo_files - set(config_files_to_add).union(files_to_ignore) + delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete] + logging.debug(f"{delete_operations=}") + + # send the files to the target revision + add_operations: List[CommitOperation] = [ + CommitOperationAdd(path_in_repo=file, path_or_fileobj=local_file) + for file, local_file in config_files_to_add.items() + ] + logging.debug(f"{add_operations=}") + + committer_hf_api.create_commit( + repo_id=dataset, + repo_type=DATASET_TYPE, + revision=target_revision, + operations=delete_operations + add_operations, + commit_message=commit_message, + parent_commit=target_dataset_info.sha, + ) + + # call the API again to get the list of parquet files + target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True) + repo_files = [ + repo_file + for repo_file in target_dataset_info.siblings + if repo_file.rfilename.startswith(f"{config}/") and repo_file.rfilename.endswith(".parquet") + ] + # we might want to check if the sha of the parquet files is the same as the one we just uploaded + # we could also check that the list of parquet files is exactly what we expect + # let's not over engineer this for now. After all, what is on the Hub is the source of truth + # and the /parquet response is more a helper to get the list of parquet files + return ConfigParquetAndInfoResponse( + parquet_files=[ + create_parquet_file_item( + repo_file=repo_file, + dataset=dataset, + config=config, + hf_endpoint=hf_endpoint, + target_revision=target_revision, + url_template=url_template, + ) + for repo_file in repo_files + ], + dataset_info=dataset_info, + ) + + +class ConfigParquetAndInfoJobRunner(DatasetsBasedJobRunner): + parquet_and_info_config: ParquetAndInfoConfig + + @staticmethod + def get_job_type() -> str: + return "config-parquet-and-info" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION + + def __init__( + self, + job_info: JobInfo, + app_config: AppConfig, + processing_step: ProcessingStep, + hf_datasets_cache: Path, + parquet_and_info_config: ParquetAndInfoConfig, + ) -> None: + super().__init__( + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + hf_datasets_cache=hf_datasets_cache, + ) + self.parquet_and_info_config = parquet_and_info_config + + def compute(self) -> CompleteJobResult: + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") + if self.config is None: + raise ParameterMissingError("'config' parameter is required") + return CompleteJobResult( + compute_config_parquet_and_info_response( + dataset=self.dataset, + config=self.config, + hf_endpoint=self.common_config.hf_endpoint, + hf_token=self.common_config.hf_token, + committer_hf_token=self.parquet_and_info_config.committer_hf_token, + source_revision=self.parquet_and_info_config.source_revision, + target_revision=self.parquet_and_info_config.target_revision, + commit_message=self.parquet_and_info_config.commit_message, + url_template=self.parquet_and_info_config.url_template, + supported_datasets=self.parquet_and_info_config.supported_datasets, + blocked_datasets=self.parquet_and_info_config.blocked_datasets, + max_dataset_size=self.parquet_and_info_config.max_dataset_size, + max_external_data_files=self.parquet_and_info_config.max_external_data_files, + ) + ) + + def get_new_splits(self, content: Mapping[str, Any]) -> Set[SplitFullName]: + """Get the set of new splits, from the content created by the compute.""" + return { + SplitFullName(dataset=self.dataset, config=self.config, split=split) + for split in content["dataset_info"]["splits"] + } diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py index 95b23b3b..b9f4650d 100644 --- a/services/worker/src/worker/job_runners/config/size.py +++ b/services/worker/src/worker/job_runners/config/size.py @@ -12,2 +12,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError -from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoResponse +from worker.job_runner import ( + CompleteJobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, +) @@ -18 +21,0 @@ ConfigSizeJobRunnerErrorCode = Literal[ - "MissingInfoForConfigError", @@ -81,7 +83,0 @@ class PreviousStepFormatError(ConfigSizeJobRunnerError): -class MissingInfoForConfigError(ConfigSizeJobRunnerError): - """Raised when the dataset info from the parquet export is missing the requested dataset configuration.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "MissingInfoForConfigError", cause, False) - - @@ -101 +97 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - - [`~job_runners.config_size.PreviousStepStatusError`] + - [`~job_runners.config.size.PreviousStepStatusError`] @@ -103 +99 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - - [`~job_runners.config_size.PreviousStepFormatError`] + - [`~job_runners.config.size.PreviousStepFormatError`] @@ -105,2 +100,0 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - - [`~job_runners.config_size.MissingInfoForConfigError`] - If the dataset info from the parquet export is missing the requested dataset configuration @@ -113,0 +108 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons + previous_step = "config-parquet-and-info" @@ -115 +110 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - response = get_response(kind="/parquet-and-dataset-info", dataset=dataset) + response = get_response(kind=previous_step, dataset=dataset, config=config) @@ -117,3 +112 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - raise DatasetNotFoundError( - "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e - ) from e + raise DatasetNotFoundError(f"No response found in previous step '{previous_step}' for this dataset.", e) from e @@ -121 +114 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - raise PreviousStepStatusError(f"Previous step gave an error: {response['http_status']}..") + raise PreviousStepStatusError(f"Previous step {previous_step} gave an error: {response['http_status']}..") @@ -123,6 +116,3 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - try: - content = ParquetAndDatasetInfoResponse( - parquet_files=response["content"]["parquet_files"], dataset_info=response["content"]["dataset_info"] - ) - except Exception as e: - raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + content = response["content"] + if "dataset_info" not in content: + raise PreviousStepFormatError("Previous step did not return the expected content: 'dataset_info'.") @@ -130,12 +120,4 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - if config not in content["dataset_info"]: - if not isinstance(content["dataset_info"], dict): - raise PreviousStepFormatError( - "Previous step did not return the expected content.", - TypeError(f"dataset_info should be a dict, but got {type(content['dataset_info'])}"), - ) - raise MissingInfoForConfigError( - f"Dataset configuration '{config}' is missing in the dataset info from the parquet export. " - f"Available configurations: {', '.join(list(content['dataset_info'])[:10])}" - + f"... ({len(content['dataset_info']) - 10})" - if len(content["dataset_info"]) > 10 - else "" + if not isinstance(content["dataset_info"], dict): + raise PreviousStepFormatError( + "Previous step did not return the expected content.", + TypeError(f"dataset_info should be a dict, but got {type(content['dataset_info'])}"), @@ -142,0 +125 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons + @@ -144,2 +127,2 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - config_dataset_info = content["dataset_info"][config] - num_columns = len(config_dataset_info["features"]) + config_info = content["dataset_info"] + num_columns = len(config_info["features"]) @@ -160 +143 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - for split_info in config_dataset_info["splits"].values() + for split_info in config_info["splits"].values() @@ -166 +149 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - "num_bytes_original_files": config_dataset_info["download_size"], + "num_bytes_original_files": config_info["download_size"], @@ -197,0 +181,2 @@ class ConfigSizeJobRunner(JobRunner): + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") @@ -199 +184 @@ class ConfigSizeJobRunner(JobRunner): - raise ValueError("config is required") + raise ParameterMissingError("'config' parameter is required") diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index 3ff1fc3f..81283100 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -15 +15 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import CompleteJobResult, JobRunnerError +from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError @@ -57 +57 @@ class ResponseAlreadyComputedError(SplitNamesFromDatasetInfoJobRunnerError): - """Raised when reponse has been already computed by /split-names-from-streaming job runner.""" + """Raised when response has been already computed by /split-names-from-streaming job runner.""" @@ -81 +81 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - - [`~job_runners.split_names_from_dataset_info.PreviousStepStatusError`] + - [`~job_runners.config.split_names_from_dataset_info.PreviousStepStatusError`] @@ -83 +83 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - - [`~job_runners.split_names_from_dataset_info.PreviousStepFormatError`] + - [`~job_runners.config.split_names_from_dataset_info.PreviousStepFormatError`] @@ -87 +87 @@ def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> - - [`~job_runners.split_names_from_dataset_info.ResponseAlreadyComputedError`] + - [`~job_runners.config.split_names_from_dataset_info.ResponseAlreadyComputedError`] @@ -124 +124 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): - raise ValueError("dataset is required") + raise ParameterMissingError("'dataset' parameter is required") @@ -126 +126 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): - raise ValueError("config is required") + raise ParameterMissingError("'config' parameter is required") diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index e3964bf9..ca798ff9 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -16 +16 @@ from libcommon.simple_cache import SplitFullName -from worker.job_runner import CompleteJobResult, JobRunnerError +from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError @@ -58 +58 @@ class ResponseAlreadyComputedError(SplitNamesFromStreamingJobRunnerError): - """Raised when reponse has been already computed by /split-names-from-dataset-info job runner.""" + """Raised when response has been already computed by /split-names-from-dataset-info job runner.""" @@ -94 +94 @@ def compute_split_names_from_streaming_response( - - [`~job_runners.split_names_from_streaming.EmptyDatasetError`] + - [`~job_runners.config.split_names_from_streaming.EmptyDatasetError`] @@ -96 +96 @@ def compute_split_names_from_streaming_response( - - [`~job_runners.split_names_from_streaming.SplitsNamesError`] + - [`~job_runners.config.split_names_from_streaming.SplitsNamesError`] @@ -98 +98 @@ def compute_split_names_from_streaming_response( - - [`~job_runners.split_names_from_streaming.ResponseAlreadyComputedError`] + - [`~job_runners.config.split_names_from_streaming.ResponseAlreadyComputedError`] @@ -131 +131 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): - raise ValueError("dataset is required") + raise ParameterMissingError("'dataset' parameter is required") @@ -133 +133 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): - raise ValueError("config is required") + raise ParameterMissingError("'config' parameter is required") diff --git a/services/worker/src/worker/job_runners/config_names.py b/services/worker/src/worker/job_runners/config_names.py index df6442b4..8043c227 100644 --- a/services/worker/src/worker/job_runners/config_names.py +++ b/services/worker/src/worker/job_runners/config_names.py @@ -13 +13 @@ from libcommon.simple_cache import SplitFullName -from worker.job_runner import CompleteJobResult, JobRunnerError +from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError @@ -54 +54 @@ class ConfigNameItem(TypedDict): -class ConfigNamesResponseContent(TypedDict): +class ConfigNamesResponse(TypedDict): @@ -61 +61 @@ def compute_config_names_response( -) -> ConfigNamesResponseContent: +) -> ConfigNamesResponse: @@ -75 +75 @@ def compute_config_names_response( - `ConfigNamesResponseContent`: An object with the list of config names. + `ConfigNamesResponse`: An object with the list of config names. @@ -96 +96 @@ def compute_config_names_response( - return {"config_names": config_name_items} + return ConfigNamesResponse(config_names=config_name_items) @@ -108,0 +109,2 @@ class ConfigNamesJobRunner(DatasetsBasedJobRunner): + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py index 5d3d1b39..ab01427a 100644 --- a/services/worker/src/worker/job_runners/dataset/info.py +++ b/services/worker/src/worker/job_runners/dataset/info.py @@ -12 +12,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, +) @@ -71 +76 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - - [`~job_runners.dataset_info.PreviousStepStatusError`] + - [`~job_runners.dataset.info.PreviousStepStatusError`] @@ -73 +78 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - - [`~job_runners.dataset_info.PreviousStepFormatError`] + - [`~job_runners.dataset.info.PreviousStepFormatError`] @@ -145,0 +151,2 @@ class DatasetInfoJobRunner(JobRunner): + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index e7171db2..4163d564 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -12 +12,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, +) @@ -14 +19 @@ from worker.job_runners.config.parquet import ConfigParquetResponse -from worker.job_runners.parquet_and_dataset_info import ParquetFileItem +from worker.job_runners.config.parquet_and_info import ParquetFileItem @@ -70 +75 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - - [`~job_runners.dataset_size.PreviousStepStatusError`] + - [`~job_runners.dataset.parquet.PreviousStepStatusError`] @@ -72 +77 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - - [`~job_runners.dataset_size.PreviousStepFormatError`] + - [`~job_runners.dataset.parquet.PreviousStepFormatError`] @@ -152,0 +158,2 @@ class DatasetParquetJobRunner(JobRunner): + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py index 06753c55..958807bb 100644 --- a/services/worker/src/worker/job_runners/dataset/size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -12 +12,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, +) @@ -83 +88 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - - [`~job_runners.dataset_size.PreviousStepStatusError`] + - [`~job_runners.dataset.size.PreviousStepStatusError`] @@ -85 +90 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - - [`~job_runners.dataset_size.PreviousStepFormatError`] + - [`~job_runners.dataset.size.PreviousStepFormatError`] @@ -180,0 +186,2 @@ class DatasetSizeJobRunner(JobRunner): + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py index 9a8f69a3..a93d2798 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py @@ -14 +14,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, +) @@ -70 +75 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - - [`~job_runners.dataset_split_names_from_dataset_info.PreviousStepStatusError`] + - [`~job_runners.dataset.split_names_from_dataset_info.PreviousStepStatusError`] @@ -72 +77 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - - [`~job_runners.dataset_split_names_from_dataset_info.PreviousStepFormatError`] + - [`~job_runners.dataset.split_names_from_dataset_info.PreviousStepFormatError`] @@ -152 +157 @@ class DatasetSplitNamesFromDatasetInfoJobRunner(JobRunner): - raise ValueError("dataset is required") + raise ParameterMissingError("'dataset' parameter is required") diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py index 0d02287f..f486683a 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py @@ -14 +14,6 @@ from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response -from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.job_runner import ( + JobResult, + JobRunner, + JobRunnerError, + ParameterMissingError, +) @@ -84 +89 @@ def compute_dataset_split_names_from_streaming_response( - - [`~job_runners.dataset_split_names_from_streaming.PreviousStepStatusError`] + - [`~job_runners.dataset.split_names_from_streaming.PreviousStepStatusError`] @@ -86 +91 @@ def compute_dataset_split_names_from_streaming_response( - - [`~job_runners.dataset_split_names_from_streaming.PreviousStepFormatError`] + - [`~job_runners.dataset.split_names_from_streaming.PreviousStepFormatError`] @@ -165 +170 @@ class DatasetSplitNamesFromStreamingJobRunner(JobRunner): - raise ValueError("dataset is required") + raise ParameterMissingError("'dataset' parameter is required") diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py index 7ffae0cb..6702deff 100644 --- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py +++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py @@ -48,2 +48,2 @@ from libcommon.simple_cache import SplitFullName -from worker.config import AppConfig, ParquetAndDatasetInfoConfig -from worker.job_runner import CompleteJobResult, JobRunnerError +from worker.config import AppConfig, ParquetAndInfoConfig +from worker.job_runner import CompleteJobResult, JobRunnerError, ParameterMissingError @@ -50,0 +51 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.job_runners.config.parquet_and_info import ParquetFile, ParquetFileItem @@ -127,9 +127,0 @@ class DatasetTooBigFromDatasetsError(ParquetAndDatasetInfoJobRunnerError): -class ParquetFileItem(TypedDict): - dataset: str - config: str - split: str - url: str - filename: str - size: int - - @@ -144,12 +135,0 @@ DATASET_TYPE = "dataset" -class ParquetFile: - def __init__(self, local_file: str, local_dir: str, config: str): - if not local_file.startswith(local_dir): - raise ValueError(f"{local_file} is not in {local_dir}") - self.local_file = local_file - self.local_dir = local_dir - self.config = config - - def repo_file(self) -> str: - return f'{self.config}/{self.local_file.removeprefix(f"{self.local_dir}/")}' - - @@ -866 +846 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner): - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig + parquet_and_dataset_info_config: ParquetAndInfoConfig @@ -882 +862 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner): - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -892,0 +873,2 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner): + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index c045229a..454578f4 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -35,0 +36 @@ from worker.job_runner import ( + ParameterMissingError, @@ -58 +58,0 @@ SplitFirstRowsFromStreamingJobRunnerErrorCode = Literal[ - "ResponseAlreadyComputedError", @@ -301 +301 @@ def compute_first_rows_response( - - [`~job_runners.first_rows.InfoError`] + - [`~job_runners.split.first_rows.InfoError`] @@ -303 +303 @@ def compute_first_rows_response( - - [`~job_runners.first_rows.FeaturesError`] + - [`~job_runners.split.first_rows.FeaturesError`] @@ -305 +305 @@ def compute_first_rows_response( - - [`~job_runners.first_rows.StreamingRowsError`] + - [`~job_runners.split.first_rows.StreamingRowsError`] @@ -307 +307 @@ def compute_first_rows_response( - - [`~job_runners.first_rows.NormalRowsError`] + - [`~job_runners.split.first_rows.NormalRowsError`] @@ -309 +309 @@ def compute_first_rows_response( - - [`~job_runners.first_rows.RowsPostProcessingError`] + - [`~job_runners.split.first_rows.RowsPostProcessingError`] @@ -311 +311 @@ def compute_first_rows_response( - - [`~job_runners.first_rows.TooManyColumnsError`] + - [`~job_runners.split.first_rows.TooManyColumnsError`] @@ -313 +313 @@ def compute_first_rows_response( - - [`~job_runners.first_rows.TooBigContentError`] + - [`~job_runners.split.first_rows.TooBigContentError`] @@ -505,2 +505,6 @@ class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): - if self.config is None or self.split is None: - raise ValueError("config and split are required") + if self.dataset is None: + raise ParameterMissingError("'dataset' parameter is required") + if self.config is None: + raise ParameterMissingError("'config' parameter is required") + if self.split is None: + raise ParameterMissingError("'split' parameter is required") diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 088065f0..2c8bac4d 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -60,3 +60,3 @@ def set_env_vars( - mp.setenv("PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE", "10_000") - mp.setenv("PARQUET_AND_DATASET_INFO_MAX_EXTERNAL_DATA_FILES", "10") - mp.setenv("PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN", CI_USER_TOKEN) + mp.setenv("PARQUET_AND_INFO_MAX_DATASET_SIZE", "10_000") + mp.setenv("PARQUET_AND_INFO_MAX_EXTERNAL_DATA_FILES", "10") + mp.setenv("PARQUET_AND_INFO_COMMITTER_HF_TOKEN", CI_USER_TOKEN) diff --git a/services/worker/tests/fixtures/files.py b/services/worker/tests/fixtures/files.py index 50a3b7af..c67a8730 100644 --- a/services/worker/tests/fixtures/files.py +++ b/services/worker/tests/fixtures/files.py @@ -115,0 +116,34 @@ def dataset_script_with_external_files_path(tmp_path_factory: pytest.TempPathFac + + +DATASET_SCRIPT_WITH_TWO_CONFIGS = """ +import os + +import datasets +from datasets import DatasetInfo, BuilderConfig, Features, Split, SplitGenerator, Value + + +class DummyDataset(datasets.GeneratorBasedBuilder): + + BUILDER_CONFIGS = [BuilderConfig(name="first"), BuilderConfig(name="second")] + + def _info(self) -> DatasetInfo: + return DatasetInfo(features=Features({"text": Value("string")})) + + def _split_generators(self, dl_manager): + return [ + SplitGenerator(Split.TRAIN, gen_kwargs={"text": self.config.name}), + SplitGenerator(Split.TEST, gen_kwargs={"text": self.config.name}), + ] + + def _generate_examples(self, text, **kwargs): + for i in range(1000): + yield i, {"text": text} +""" + + [email protected](scope="session") +def dataset_script_with_two_configs_path(tmp_path_factory: pytest.TempPathFactory) -> str: + path = str(tmp_path_factory.mktemp("data") / "{dataset_name}.py") + with open(path, "w", newline="") as f: + f.write(DATASET_SCRIPT_WITH_TWO_CONFIGS) + return path diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 982209ae..3fa69d89 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -246,0 +247,7 @@ def external_files_dataset_builder(hub_public_external_files: str) -> DatasetBui [email protected](scope="session") +def hub_public_legacy_configs(dataset_script_with_two_configs_path: str) -> Iterator[str]: + repo_id = create_hub_dataset_repo(prefix="legacy_configs", file_paths=[dataset_script_with_two_configs_path]) + yield repo_id + delete_hub_dataset_repo(repo_id=repo_id) + + @@ -252,0 +260 @@ class HubDatasetTest(TypedDict): + parquet_and_info_response: Any @@ -350,0 +359 @@ def create_dataset_info_response_for_audio() -> Any: +# TODO: remove when "/parquet-and-dataset-info" step is removed @@ -377,0 +387,27 @@ def create_parquet_and_dataset_info_response(dataset: str, data_type: Literal["c +def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "audio"]) -> Any: + dataset, config, split = get_default_config_split(dataset) + + filename = "csv-train.parquet" if data_type == "csv" else "parquet-train.parquet" + size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE + info = ( + create_dataset_info_response_for_csv(dataset, config) + if data_type == "csv" + else create_dataset_info_response_for_audio() + ) + return { + "parquet_files": [ + { + "dataset": dataset, + "config": config, + "split": split, + "url": CI_URL_TEMPLATE.format( + repo_id=f"datasets/{dataset}", revision="refs%2Fconvert%2Fparquet", filename=f"{config}/{filename}" + ), + "filename": filename, + "size": size, + } + ], + "dataset_info": info, # we don't have config key here, only dataset_info dictionary + } + + @@ -517,0 +554 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -524,0 +562 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -531,0 +570 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -538,0 +578 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -547,0 +588 @@ def hub_datasets( + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_csv, data_type="csv"), @@ -556,0 +598 @@ def hub_datasets( + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_private_csv, data_type="csv"), @@ -565,0 +608 @@ def hub_datasets( + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_gated_csv, data_type="csv"), @@ -572,0 +616 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -581,0 +626,3 @@ def hub_datasets( + "parquet_and_info_response": create_parquet_and_info_response( + dataset=hub_gated_extra_fields_csv, data_type="csv" + ), @@ -592,0 +640 @@ def hub_datasets( + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_audio, data_type="audio"), @@ -601,0 +650 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -610,0 +660 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -617,0 +668 @@ def hub_datasets( + "parquet_and_info_response": None, @@ -624,0 +676 @@ def hub_datasets( + "parquet_and_info_response": None, diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py index b07e5702..ed0ddf7c 100644 --- a/services/worker/tests/job_runners/config/test_info.py +++ b/services/worker/tests/job_runners/config/test_info.py @@ -184 +184 @@ def get_job_runner( - "dataset_info": DATASET_INFO_OK, + "dataset_info": CONFIG_INFO_1, @@ -222 +222,5 @@ def test_compute( - kind="/parquet-and-dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status + kind="config-parquet-and-info", + dataset=dataset, + config=config, + content=upstream_content, + http_status=upstream_status, diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index ae29300c..7282747d 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -21,2 +21,2 @@ from worker.job_runners.config.parquet import ( -from worker.job_runners.parquet_and_dataset_info import ( - ParquetAndDatasetInfoResponse, +from worker.job_runners.config.parquet_and_info import ( + ConfigParquetAndInfoResponse, @@ -80 +80 @@ def get_job_runner( - ParquetAndDatasetInfoResponse( + ConfigParquetAndInfoResponse( @@ -86 +86 @@ def get_job_runner( - dataset="ok", config="config_2", split="train", url="url2", filename="filename2", size=0 + dataset="ok", config="config_1", split="train", url="url2", filename="filename2", size=0 @@ -89 +89 @@ def get_job_runner( - dataset_info={"config_1": "value", "config_2": "value"}, + dataset_info={"description": "value", "dataset_size": 10}, @@ -96 +96,4 @@ def get_job_runner( - ) + ), + ParquetFileItem( + dataset="ok", config="config_1", split="train", url="url2", filename="filename2", size=0 + ), @@ -133 +136,5 @@ def test_compute( - kind="/parquet-and-dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status + kind="config-parquet-and-info", + dataset=dataset, + config=config, + content=upstream_content, + http_status=upstream_status, @@ -145,2 +152 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - dataset = "doesnotexist" - config = "doesnotexist" + dataset = config = "doesnotexist" diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py new file mode 100644 index 00000000..5e76eb95 --- /dev/null +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -0,0 +1,641 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import io +from fnmatch import fnmatch +from http import HTTPStatus +from typing import Any, Callable, Iterator, List, Optional + +import datasets.builder +import pandas as pd +import pytest +import requests +from huggingface_hub.hf_api import HfApi +from libcommon.exceptions import CustomError +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import DoesNotExist, get_response, upsert_response + +from worker.config import AppConfig, ParquetAndInfoConfig +from worker.job_runners.config.parquet_and_info import ( + ConfigParquetAndInfoJobRunner, + DatasetInBlockListError, + DatasetTooBigFromDatasetsError, + DatasetTooBigFromHubError, + DatasetWithTooBigExternalFilesError, + DatasetWithTooManyExternalFilesError, + get_dataset_info_or_raise, + parse_repo_filename, + raise_if_blocked, + raise_if_not_supported, + raise_if_too_big_from_datasets, + raise_if_too_big_from_external_data_files, + raise_if_too_big_from_hub, +) +from worker.resources import LibrariesResource + +from ...constants import CI_HUB_ENDPOINT, CI_USER_TOKEN +from ...fixtures.hub import HubDatasets + + +# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200 [email protected](scope="module", autouse=True) +def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyPatch]: + mp = pytest.MonkeyPatch() + mp.setenv( + "PARQUET_AND_INFO_BLOCKED_DATASETS", + ",".join(value["name"] for value in hub_datasets.values() if "jsonl" in value["name"]), + ) + mp.setenv( + "PARQUET_AND_INFO_SUPPORTED_DATASETS", + ",".join(value["name"] for value in hub_datasets.values() if "big" not in value["name"]), + ) + yield mp + mp.undo() + + [email protected] +def parquet_and_info_config( + set_env_vars: pytest.MonkeyPatch, set_supported_datasets: pytest.MonkeyPatch +) -> ParquetAndInfoConfig: + return ParquetAndInfoConfig.from_env() + + +GetJobRunner = Callable[[str, str, AppConfig, ParquetAndInfoConfig, bool], ConfigParquetAndInfoJobRunner] + + [email protected] +def get_job_runner( + libraries_resource: LibrariesResource, + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, + force: bool = False, + ) -> ConfigParquetAndInfoJobRunner: + return ConfigParquetAndInfoJobRunner( + job_info={ + "type": ConfigParquetAndInfoJobRunner.get_job_type(), + "dataset": dataset, + "config": config, + "split": None, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + app_config=app_config, + processing_step=ProcessingStep( + name=ConfigParquetAndInfoJobRunner.get_job_type(), + input_type="config", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + job_runner_version=ConfigParquetAndInfoJobRunner.get_job_runner_version(), + ), + hf_datasets_cache=libraries_resource.hf_datasets_cache, + parquet_and_info_config=parquet_and_info_config, + ) + + return _get_job_runner + + +def assert_content_is_equal(content: Any, expected: Any) -> None: + print(content) + assert set(content) == {"parquet_files", "dataset_info"}, content + assert content["parquet_files"] == expected["parquet_files"], content + assert len(content["dataset_info"]) == len(expected["dataset_info"]), content + content_value = content["dataset_info"] + expected_value = expected["dataset_info"] + assert set(content_value.keys()) == set(expected_value.keys()), content + for key in content_value.keys(): + if key != "download_checksums": + assert content_value[key] == expected_value[key], content + assert len(content_value["download_checksums"]) == 1, content + content_checksum = list(content_value["download_checksums"].values())[0] + expected_checksum = list(expected_value["download_checksums"].values())[0] + assert content_checksum == expected_checksum, content + + +def test_compute( + app_config: AppConfig, + get_job_runner: GetJobRunner, + parquet_and_info_config: ParquetAndInfoConfig, + hub_datasets: HubDatasets, +) -> None: + dataset = hub_datasets["public"]["name"] + config = hub_datasets["public"]["config_names_response"]["config_names"][0]["config"] + upsert_response( + "/config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets["public"]["config_names_response"], + ) + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + assert job_runner.process() + cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config) + assert cached_response["http_status"] == HTTPStatus.OK + assert cached_response["error_code"] is None + assert cached_response["job_runner_version"] == job_runner.get_job_runner_version() + assert cached_response["dataset_git_revision"] is not None + content = cached_response["content"] + assert len(content["parquet_files"]) == 1 + assert_content_is_equal(content, hub_datasets["public"]["parquet_and_info_response"]) + + +def test_compute_legacy_configs( + app_config: AppConfig, + get_job_runner: GetJobRunner, + parquet_and_info_config: ParquetAndInfoConfig, + hub_public_legacy_configs: str, +) -> None: + dataset_name = hub_public_legacy_configs + original_configs = {"first", "second"} + upsert_response( + kind="/config-names", + dataset=hub_public_legacy_configs, + http_status=HTTPStatus.OK, + content={ + "config_names": [ + {"dataset": hub_public_legacy_configs, "config": "first"}, + {"dataset": hub_public_legacy_configs, "config": "second"}, + ], + }, + ) + # first compute and push parquet files for each config for dataset with script with two configs + for config in original_configs: + job_runner = get_job_runner(dataset_name, config, app_config, parquet_and_info_config, False) + assert job_runner.process() + hf_api = HfApi(endpoint=CI_HUB_ENDPOINT, token=CI_USER_TOKEN) + dataset_info = hf_api.dataset_info( + repo_id=hub_public_legacy_configs, revision=parquet_and_info_config.target_revision, files_metadata=False + ) + repo_files = {f.rfilename for f in dataset_info.siblings} + # assert that there are only parquet files for dataset's configs and ".gitattributes" in a repo + # (no files from 'main') + assert ".gitattributes" in repo_files + assert all( + fnmatch(file, "first/*.parquet") or fnmatch(file, "second/*.parquet") + for file in repo_files.difference({".gitattributes"}) + ) + orig_repo_configs = {f.rfilename.split("/")[0] for f in dataset_info.siblings if f.rfilename.endswith(".parquet")} + # assert that both configs are pushed (push of second config didn't delete first config's files) + assert len(orig_repo_configs) == 2 + assert orig_repo_configs == original_configs + # then change the set of dataset configs (remove "second") + upsert_response( + kind="/config-names", + dataset=hub_public_legacy_configs, + http_status=HTTPStatus.OK, + content={ + "config_names": [ + {"dataset": hub_public_legacy_configs, "config": "first"}, + ], + }, + ) + job_runner = get_job_runner(dataset_name, "first", app_config, parquet_and_info_config, False) + assert job_runner.process() + dataset_info = hf_api.dataset_info( + repo_id=hub_public_legacy_configs, revision=parquet_and_info_config.target_revision, files_metadata=False + ) + updated_repo_files = {f.rfilename for f in dataset_info.siblings} + # assert that legacy config is removed from the repo + # and there are only files for config that was just pushed and .gitattributes + assert ".gitattributes" in updated_repo_files + assert all(fnmatch(file, "first/*") for file in updated_repo_files.difference({".gitattributes"})) + updated_repo_configs = { + f.rfilename.split("/")[0] for f in dataset_info.siblings if f.rfilename.endswith(".parquet") + } + assert len(updated_repo_configs) == 1 + assert updated_repo_configs == {"first"} + + +def test_doesnotexist( + app_config: AppConfig, get_job_runner: GetJobRunner, parquet_and_info_config: ParquetAndInfoConfig +) -> None: + dataset, config = "doesnotexist", "nonexisting" + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + assert not job_runner.process() + with pytest.raises(DoesNotExist): + get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config) + + [email protected]( + "dataset,blocked,raises", + [ + ("public", ["public"], True), + ("public", ["public", "audio"], True), + ("public", ["audio"], False), + ("public", [], False), + ], +) +def test_raise_if_blocked(dataset: str, blocked: List[str], raises: bool) -> None: + if raises: + with pytest.raises(DatasetInBlockListError): + raise_if_blocked(dataset=dataset, blocked_datasets=blocked) + else: + raise_if_blocked(dataset=dataset, blocked_datasets=blocked) + + [email protected]( + "name,raises", + [("public", False), ("big", True)], +) +def test_raise_if_too_big_from_hub( + hub_datasets: HubDatasets, + name: str, + raises: bool, + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, +) -> None: + dataset = hub_datasets[name]["name"] + dataset_info = get_dataset_info_or_raise( + dataset=dataset, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + revision="main", + ) + if raises: + with pytest.raises(DatasetTooBigFromHubError): + raise_if_too_big_from_hub( + dataset_info=dataset_info, max_dataset_size=parquet_and_info_config.max_dataset_size + ) + else: + raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=parquet_and_info_config.max_dataset_size) + + [email protected]( + "name,raises", + [("public", False), ("big", True)], +) +def test_raise_if_too_big_from_datasets( + hub_datasets: HubDatasets, + name: str, + raises: bool, + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, +) -> None: + dataset = hub_datasets[name]["name"] + config = hub_datasets[name]["config_names_response"]["config_names"][0]["config"] + if raises: + with pytest.raises(DatasetTooBigFromDatasetsError): + raise_if_too_big_from_datasets( + dataset=dataset, + config=config, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + revision="main", + max_dataset_size=parquet_and_info_config.max_dataset_size, + ) + else: + raise_if_too_big_from_datasets( + dataset=dataset, + config=config, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + revision="main", + max_dataset_size=parquet_and_info_config.max_dataset_size, + ) + + [email protected]( + "max_dataset_size,max_external_data_files,raises", + [ + (None, None, False), + (10, None, True), + ], +) +def test_raise_if_too_big_external_files( + external_files_dataset_builder: "datasets.builder.DatasetBuilder", + raises: bool, + max_dataset_size: Optional[int], + max_external_data_files: Optional[int], + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, +) -> None: + max_dataset_size = max_dataset_size or parquet_and_info_config.max_dataset_size + max_external_data_files = max_external_data_files or parquet_and_info_config.max_external_data_files + if raises: + with pytest.raises(DatasetWithTooBigExternalFilesError): + raise_if_too_big_from_external_data_files( + builder=external_files_dataset_builder, + hf_token=app_config.common.hf_token, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + ) + else: + raise_if_too_big_from_external_data_files( + builder=external_files_dataset_builder, + hf_token=app_config.common.hf_token, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + ) + + [email protected]( + "max_dataset_size,max_external_data_files,raises", + [ + (None, None, False), + (None, 1, True), + ], +) +def test_raise_if_too_many_external_files( + external_files_dataset_builder: "datasets.builder.DatasetBuilder", + raises: bool, + max_dataset_size: Optional[int], + max_external_data_files: Optional[int], + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, +) -> None: + max_dataset_size = max_dataset_size or parquet_and_info_config.max_dataset_size + max_external_data_files = max_external_data_files or parquet_and_info_config.max_external_data_files + if raises: + with pytest.raises(DatasetWithTooManyExternalFilesError): + raise_if_too_big_from_external_data_files( + builder=external_files_dataset_builder, + hf_token=app_config.common.hf_token, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + ) + else: + raise_if_too_big_from_external_data_files( + builder=external_files_dataset_builder, + hf_token=app_config.common.hf_token, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + ) + + [email protected]( + "in_list,raises", + [ + (True, False), + (False, True), + ], +) +def test_raise_if_not_supported( + hub_datasets: HubDatasets, + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, + in_list: bool, + raises: bool, +) -> None: + dataset = hub_datasets["big"]["name"] + config = hub_datasets["big"]["config_names_response"]["config_names"][0]["config"] + if raises: + with pytest.raises(DatasetTooBigFromDatasetsError): + raise_if_not_supported( + dataset=dataset, + config=config, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + committer_hf_token=parquet_and_info_config.committer_hf_token, + revision="main", + max_dataset_size=parquet_and_info_config.max_dataset_size, + supported_datasets=[dataset] if in_list else ["another_dataset"], + blocked_datasets=[], + ) + else: + raise_if_not_supported( + dataset=dataset, + config=config, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + committer_hf_token=parquet_and_info_config.committer_hf_token, + revision="main", + max_dataset_size=parquet_and_info_config.max_dataset_size, + supported_datasets=[dataset] if in_list else ["another_dataset"], + blocked_datasets=[], + ) + + +def test_not_supported_if_big( + app_config: AppConfig, + get_job_runner: GetJobRunner, + parquet_and_info_config: ParquetAndInfoConfig, + hub_datasets: HubDatasets, +) -> None: + # Not in the list of supported datasets and bigger than the maximum size + # dataset = hub_public_big + dataset = hub_datasets["big"]["name"] + config = hub_datasets["big"]["config_names_response"]["config_names"][0]["config"] + upsert_response( + kind="/config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets["big"]["config_names_response"], + ) + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + assert not job_runner.process() + cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config) + assert cached_response["http_status"] == HTTPStatus.NOT_IMPLEMENTED + assert cached_response["error_code"] == "DatasetTooBigFromDatasetsError" + + +def test_supported_if_gated( + app_config: AppConfig, + get_job_runner: GetJobRunner, + parquet_and_info_config: ParquetAndInfoConfig, + hub_datasets: HubDatasets, +) -> None: + # Access must be granted + dataset = hub_datasets["gated"]["name"] + config = hub_datasets["gated"]["config_names_response"]["config_names"][0]["config"] + upsert_response( + "/config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets["gated"]["config_names_response"], + ) + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + assert job_runner.process() + cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config) + assert cached_response["http_status"] == HTTPStatus.OK + assert cached_response["error_code"] is None + + +def test_not_supported_if_gated_with_extra_fields( + app_config: AppConfig, + get_job_runner: GetJobRunner, + parquet_and_info_config: ParquetAndInfoConfig, + hub_datasets: HubDatasets, +) -> None: + # Access request should fail because extra fields in gated datasets are not supported + dataset = hub_datasets["gated_extra_fields"]["name"] + config = hub_datasets["gated_extra_fields"]["config_names_response"]["config_names"][0]["config"] + upsert_response( + kind="/config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets["gated_extra_fields"]["config_names_response"], + ) + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + assert not job_runner.process() + cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config) + assert cached_response["http_status"] == HTTPStatus.NOT_FOUND + assert cached_response["error_code"] == "GatedExtraFieldsError" + + +def test_blocked( + app_config: AppConfig, + get_job_runner: GetJobRunner, + parquet_and_info_config: ParquetAndInfoConfig, + hub_datasets: HubDatasets, +) -> None: + # In the list of blocked datasets + dataset = hub_datasets["jsonl"]["name"] + config = hub_datasets["jsonl"]["config_names_response"]["config_names"][0]["config"] + upsert_response( + kind="/config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets["jsonl"]["config_names_response"], + ) + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + assert not job_runner.process() + cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config) + assert cached_response["http_status"] == HTTPStatus.NOT_IMPLEMENTED + assert cached_response["error_code"] == "DatasetInBlockListError" + + [email protected]( + "name", + ["public", "audio", "gated"], +) +def test_compute_splits_response_simple_csv_ok( + hub_datasets: HubDatasets, + get_job_runner: GetJobRunner, + name: str, + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, + data_df: pd.DataFrame, +) -> None: + dataset = hub_datasets[name]["name"] + config = hub_datasets[name]["config_names_response"]["config_names"][0]["config"] + upsert_response( + "/config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets[name]["config_names_response"], + ) + expected_parquet_and_info_response = hub_datasets[name]["parquet_and_info_response"] + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + result = job_runner.compute().content + assert_content_is_equal(result, expected_parquet_and_info_response) + + # download the parquet file and check that it is valid + if name == "audio": + return + + if name == "public": + df = pd.read_parquet(result["parquet_files"][0]["url"], engine="auto") + else: + # in all these cases, the parquet files are not accessible without a token + with pytest.raises(Exception): + pd.read_parquet(result["parquet_files"][0]["url"], engine="auto") + r = requests.get( + result["parquet_files"][0]["url"], headers={"Authorization": f"Bearer {app_config.common.hf_token}"} + ) + assert r.status_code == HTTPStatus.OK, r.text + df = pd.read_parquet(io.BytesIO(r.content), engine="auto") + assert df.equals(data_df), df + + [email protected]( + "name,error_code,cause", + [ + ("does_not_exist", "ParameterMissingError", None), + ("gated_extra_fields", "GatedExtraFieldsError", "HTTPError"), + ("private", "DatasetNotFoundError", None), + ("public", "DatasetNotFoundError", "DoesNotExist"), # no cache for /config-names -> DatasetNotFoundError + ], +) +def test_compute_splits_response_simple_csv_error( + hub_datasets: HubDatasets, + get_job_runner: GetJobRunner, + name: str, + error_code: str, + cause: str, + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, +) -> None: + dataset = hub_datasets[name]["name"] + config_names_response = hub_datasets[name]["config_names_response"] + config = config_names_response["config_names"][0]["config"] if config_names_response else None + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.code == error_code + assert exc_info.value.cause_exception == cause + if exc_info.value.disclose_cause: + response = exc_info.value.as_response() + assert set(response.keys()) == {"error", "cause_exception", "cause_message", "cause_traceback"} + response_dict = dict(response) + # ^ to remove mypy warnings + assert response_dict["cause_exception"] == cause + assert isinstance(response_dict["cause_traceback"], list) + assert response_dict["cause_traceback"][0] == "Traceback (most recent call last):\n" + + [email protected]( + "upstream_status,upstream_content,error_code", + [ + (HTTPStatus.NOT_FOUND, {"error": "error"}, "PreviousStepStatusError"), + (HTTPStatus.OK, {"not_config_names": "wrong_format"}, "PreviousStepFormatError"), + (HTTPStatus.OK, {"config_names": "not a list"}, "PreviousStepFormatError"), + ], +) +def test_previous_step_error( + get_job_runner: GetJobRunner, + upstream_status: HTTPStatus, + upstream_content: Any, + error_code: str, + hub_public_csv: str, + hub_datasets: HubDatasets, + app_config: AppConfig, + parquet_and_info_config: ParquetAndInfoConfig, +) -> None: + dataset = hub_datasets["public"]["name"] + config = hub_datasets["public"]["config_names_response"]["config_names"][0]["config"] + job_runner = get_job_runner(dataset, config, app_config, parquet_and_info_config, False) + upsert_response( + "/config-names", + dataset=dataset, + http_status=upstream_status, + content=upstream_content, + ) + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.code == error_code + + [email protected]( + "filename,split,config,raises", + [ + ("config/builder-split.parquet", "split", "config", False), + ("config/builder-with-dashes-split.parquet", "split", "config", False), + ("config/builder-split-00000-of-00001.parquet", "split", "config", False), + ("config/builder-with-dashes-split-00000-of-00001.parquet", "split", "config", False), + ("config/builder-split.with.dots-00000-of-00001.parquet", "split.with.dots", "config", False), + ( + "config/builder-with-dashes-caveat-asplitwithdashesisnotsupported-00000-of-00001.parquet", + "asplitwithdashesisnotsupported", + "config", + False, + ), + ("builder-split-00000-of-00001.parquet", "split", "config", True), + ("plain_text/openwebtext-10k-train.parquet", "train", "plain_text", False), + ("plain_text/openwebtext-10k-train-00000-of-00001.parquet", "train", "plain_text", False), + ], +) +def test_parse_repo_filename(filename: str, split: str, config: str, raises: bool) -> None: + if raises: + with pytest.raises(Exception): + parse_repo_filename(filename) + else: + assert parse_repo_filename(filename) == (config, split) diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py index 8ecdc637..fa19e076 100644 --- a/services/worker/tests/job_runners/config/test_size.py +++ b/services/worker/tests/job_runners/config/test_size.py @@ -79,3 +78,0 @@ def get_job_runner( - {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 1234}, - {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 6789}, - {"dataset": "dataset_ok", "config": "config_2", "split": "test", "size": 2383903}, @@ -84,7 +81,5 @@ def get_job_runner( - "config_1": { - "features": { - "image": {"_type": "Image"}, - "label": { - "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], - "_type": "ClassLabel", - }, + "features": { + "image": {"_type": "Image"}, + "label": { + "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], + "_type": "ClassLabel", @@ -92,13 +87,7 @@ def get_job_runner( - "splits": { - "train": { - "name": "train", - "num_bytes": 17470800, - "num_examples": 60000, - "dataset_name": "dataset_ok", - }, - "test": { - "name": "test", - "num_bytes": 2916432, - "num_examples": 10000, - "dataset_name": "dataset_ok", - }, + }, + "splits": { + "train": { + "name": "train", + "num_bytes": 17470800, + "num_examples": 60000, + "dataset_name": "dataset_ok", @@ -106,17 +95,5 @@ def get_job_runner( - "download_checksums": { - "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": { - "num_bytes": 9912422, - "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609", - }, - "https://storage.googleapis.com/cvdf-datasets/mnist/train-labels-idx1-ubyte.gz": { - "num_bytes": 28881, - "checksum": "3552534a0a558bbed6aed32b30c495cca23d567ec52cac8be1a0730e8010255c", - }, - "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-images-idx3-ubyte.gz": { - "num_bytes": 1648877, - "checksum": "8d422c7b0a1c1c79245a5bcf07fe86e33eeafee792b84584aec276f5a2dbc4e6", - }, - "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-labels-idx1-ubyte.gz": { - "num_bytes": 4542, - "checksum": "f7ae60f92e00ec6debd23a6088c31dbd2371eca3ffa0defaefb259924204aec6", - }, + "test": { + "name": "test", + "num_bytes": 2916432, + "num_examples": 10000, + "dataset_name": "dataset_ok", @@ -124,3 +100,0 @@ def get_job_runner( - "download_size": 11594722, - "dataset_size": 20387232, - "size_in_bytes": 31981954, @@ -128,8 +102,8 @@ def get_job_runner( - "config_2": { - "features": { - "image": {"_type": "Image"}, - "image2": {"_type": "Image"}, - "label": { - "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], - "_type": "ClassLabel", - }, + "download_checksums": { + "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": { + "num_bytes": 9912422, + "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609", + }, + "https://storage.googleapis.com/cvdf-datasets/mnist/train-labels-idx1-ubyte.gz": { + "num_bytes": 28881, + "checksum": "3552534a0a558bbed6aed32b30c495cca23d567ec52cac8be1a0730e8010255c", @@ -137,13 +111,3 @@ def get_job_runner( - "splits": { - "train": { - "name": "train", - "num_bytes": 5678, - "num_examples": 3000, - "dataset_name": "dataset_ok", - }, - "test": { - "name": "test", - "num_bytes": 1234, - "num_examples": 1000, - "dataset_name": "dataset_ok", - }, + "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-images-idx3-ubyte.gz": { + "num_bytes": 1648877, + "checksum": "8d422c7b0a1c1c79245a5bcf07fe86e33eeafee792b84584aec276f5a2dbc4e6", @@ -151,5 +115,3 @@ def get_job_runner( - "download_checksums": { - "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": { - "num_bytes": 9912422, - "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609", - }, + "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-labels-idx1-ubyte.gz": { + "num_bytes": 4542, + "checksum": "f7ae60f92e00ec6debd23a6088c31dbd2371eca3ffa0defaefb259924204aec6", @@ -157,3 +118,0 @@ def get_job_runner( - "download_size": 9912422, - "dataset_size": 6912, - "size_in_bytes": 9919334, @@ -160,0 +120,3 @@ def get_job_runner( + "download_size": 11594722, + "dataset_size": 20387232, + "size_in_bytes": 31981954, @@ -231 +193,5 @@ def test_compute( - kind="/parquet-and-dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status + kind="config-parquet-and-info", + dataset=dataset, + config=config, + content=upstream_content, + http_status=upstream_status, diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index a2b68d9d..8ded0a66 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -15,0 +16 @@ from worker.job_runners.config.parquet import ConfigParquetResponse +from worker.job_runners.config.parquet_and_info import ParquetFileItem @@ -22 +22,0 @@ from worker.job_runners.dataset.parquet import ( -from worker.job_runners.parquet_and_dataset_info import ParquetFileItem diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index b5b1fb10..9bb30190 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -18 +18 @@ from libcommon.simple_cache import DoesNotExist, get_response -from worker.config import AppConfig, ParquetAndDatasetInfoConfig +from worker.config import AppConfig, ParquetAndInfoConfig @@ -44 +44 @@ def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyP - "PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS", + "PARQUET_AND_INFO_BLOCKED_DATASETS", @@ -48 +48 @@ def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyP - "PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS", + "PARQUET_AND_INFO_SUPPORTED_DATASETS", @@ -58,2 +58,2 @@ def parquet_and_dataset_info_config( -) -> ParquetAndDatasetInfoConfig: - return ParquetAndDatasetInfoConfig.from_env() +) -> ParquetAndInfoConfig: + return ParquetAndInfoConfig.from_env() @@ -62 +62 @@ def parquet_and_dataset_info_config( -GetJobRunner = Callable[[str, AppConfig, ParquetAndDatasetInfoConfig, bool], ParquetAndDatasetInfoJobRunner] +GetJobRunner = Callable[[str, AppConfig, ParquetAndInfoConfig, bool], ParquetAndDatasetInfoJobRunner] @@ -74 +74 @@ def get_job_runner( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -124 +124 @@ def test_compute( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -141 +141 @@ def test_doesnotexist( - app_config: AppConfig, get_job_runner: GetJobRunner, parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig + app_config: AppConfig, get_job_runner: GetJobRunner, parquet_and_dataset_info_config: ParquetAndInfoConfig @@ -176 +176 @@ def test_raise_if_too_big_from_hub( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -205 +205 @@ def test_raise_if_too_big_from_datasets( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -240 +240 @@ def test_raise_if_too_big_external_files( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -274 +274 @@ def test_raise_if_too_many_external_files( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -305 +305 @@ def test_raise_if_not_supported( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -337 +337 @@ def test_not_supported_if_big( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -352 +352 @@ def test_supported_if_gated( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -355 +355 @@ def test_supported_if_gated( - # Access should must be granted + # Access must be granted @@ -367 +367 @@ def test_not_supported_if_gated_with_extra_fields( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -382 +382 @@ def test_blocked( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -403 +403 @@ def test_compute_splits_response_simple_csv_ok( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, @@ -446 +446 @@ def test_compute_splits_response_simple_csv_error( - parquet_and_dataset_info_config: ParquetAndDatasetInfoConfig, + parquet_and_dataset_info_config: ParquetAndInfoConfig, diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py index a6b335a5..7987caa7 100644 --- a/services/worker/tests/test_job_runner_factory.py +++ b/services/worker/tests/test_job_runner_factory.py @@ -26,0 +27 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph: + ("config-parquet-and-info", "ConfigParquetAndInfoJobRunner"), diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml index bab6e8be..d2efc006 100644 --- a/tools/docker-compose-datasets-server.yml +++ b/tools/docker-compose-datasets-server.yml @@ -91,8 +91,8 @@ services: - PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS: ${PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS-} - PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE: ${PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE-Update parquet files} - PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-} - PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE-100_000_000} - PARQUET_AND_DATASET_INFO_SOURCE_REVISION: ${PARQUET_AND_DATASET_INFO_SOURCE_REVISION-main} - PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS-} - PARQUET_AND_DATASET_INFO_TARGET_REVISION: ${PARQUET_AND_DATASET_INFO_TARGET_REVISION-refs/convert/parquet} - PARQUET_AND_DATASET_INFO_URL_TEMPLATE: ${PARQUET_AND_DATASET_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s} + PARQUET_AND_INFO_BLOCKED_DATASETS: ${PARQUET_AND_INFO_BLOCKED_DATASETS-} + PARQUET_AND_INFO_COMMIT_MESSAGE: ${PARQUET_AND_INFO_COMMIT_MESSAGE-Update parquet files} + PARQUET_AND_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_INFO_COMMITTER_HF_TOKEN-} + PARQUET_AND_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_INFO_MAX_DATASET_SIZE-100_000_000} + PARQUET_AND_INFO_SOURCE_REVISION: ${PARQUET_AND_INFO_SOURCE_REVISION-main} + PARQUET_AND_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_INFO_SUPPORTED_DATASETS-} + PARQUET_AND_INFO_TARGET_REVISION: ${PARQUET_AND_INFO_TARGET_REVISION-refs/convert/parquet} + PARQUET_AND_INFO_URL_TEMPLATE: ${PARQUET_AND_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s} diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index a5b22ac7..b4be1f51 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -91,8 +91,8 @@ services: - PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS: ${PARQUET_AND_DATASET_INFO_BLOCKED_DATASETS-} - PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE: ${PARQUET_AND_DATASET_INFO_COMMIT_MESSAGE-Update parquet files} - PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_DATASET_INFO_COMMITTER_HF_TOKEN-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} - PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_DATASET_INFO_MAX_DATASET_SIZE-100_000_000} - PARQUET_AND_DATASET_INFO_SOURCE_REVISION: ${PARQUET_AND_DATASET_INFO_SOURCE_REVISION-main} - PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_DATASET_INFO_SUPPORTED_DATASETS-} - PARQUET_AND_DATASET_INFO_TARGET_REVISION: ${PARQUET_AND_DATASET_INFO_TARGET_REVISION-refs/convert/parquet} - PARQUET_AND_DATASET_INFO_URL_TEMPLATE: ${PARQUET_AND_DATASET_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s} + PARQUET_AND_INFO_BLOCKED_DATASETS: ${PARQUET_AND_INFO_BLOCKED_DATASETS-} + PARQUET_AND_INFO_COMMIT_MESSAGE: ${PARQUET_AND_INFO_COMMIT_MESSAGE-Update parquet files} + PARQUET_AND_INFO_COMMITTER_HF_TOKEN: ${PARQUET_AND_INFO_COMMITTER_HF_TOKEN-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} + PARQUET_AND_INFO_MAX_DATASET_SIZE: ${PARQUET_AND_INFO_MAX_DATASET_SIZE-100_000_000} + PARQUET_AND_INFO_SOURCE_REVISION: ${PARQUET_AND_INFO_SOURCE_REVISION-main} + PARQUET_AND_INFO_SUPPORTED_DATASETS: ${PARQUET_AND_INFO_SUPPORTED_DATASETS-} + PARQUET_AND_INFO_TARGET_REVISION: ${PARQUET_AND_INFO_TARGET_REVISION-refs/convert/parquet} + PARQUET_AND_INFO_URL_TEMPLATE: ${PARQUET_AND_INFO_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
401cb3f3c288269e67b313e0b60bc6d742db46f7
Andrea Francis Soria Jimenez
2023-04-11T11:47:23
Dataset status in admin ui (#1040)
diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace index 724166ba..d73dd79a 100644 --- a/.vscode/monorepo.code-workspace +++ b/.vscode/monorepo.code-workspace @@ -37,0 +38,4 @@ + }, + { + "name": "front/admin_ui", + "path": "../front/admin_ui" diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py index 065ea00a..b7f69f66 100644 --- a/front/admin_ui/app.py +++ b/front/admin_ui/app.py @@ -73,0 +74,8 @@ with gr.Blocks() as demo: + with gr.Tab("Dataset status"): + dataset_name = gr.Textbox(label="dataset", placeholder="c4") + dataset_status_button = gr.Button("Get dataset status") + gr.Markdown("### Cached responses") + cached_responses_table = gr.DataFrame() + gr.Markdown("### Pending jobs") + jobs_table = gr.DataFrame() + @@ -124 +132,47 @@ with gr.Blocks() as demo: - + + def get_dataset_status(token, dataset): + headers = {"Authorization": f"Bearer {token}"} + response = requests.get(f"{DSS_ENDPOINT}/admin/dataset-status?dataset={dataset}", headers=headers, timeout=60) + if response.status_code == 200: + dataset_status = response.json() + cached_responses_df = pd.DataFrame([{ + "type": job_type, + "dataset": cached_response["dataset"], + "config": cached_response["config"], + "split": cached_response["split"], + "http_status": cached_response["http_status"], + "error_code": cached_response["error_code"], + "job_runner_version": cached_response["job_runner_version"], + "dataset_git_revision": cached_response["dataset_git_revision"], + "progress": cached_response["progress"] + } + for job_type, content in dataset_status.items() + for cached_response in content["cached_responses"] + ]) + jobs_df = pd.DataFrame([{ + "type": job_type, + "dataset": job["dataset"], + "config": job["config"], + "split": job["split"], + "namespace": job["namespace"], + "force": job["force"], + "priority": job["priority"], + "status": job["status"], + "created_at": job["created_at"], + "started_at": job["started_at"], + "finished_at": job["finished_at"], + "last_heartbeat": job["last_heartbeat"] + } + for job_type, content in dataset_status.items() + for job in content["jobs"] + ]) + return { + cached_responses_table: gr.update(value=cached_responses_df), + jobs_table: gr.update(value=jobs_df) + } + else: + return { + cached_responses_table: gr.update(value=None), + jobs_table: gr.update(value=None) + } + @@ -171,0 +226 @@ with gr.Blocks() as demo: + dataset_status_button.click(get_dataset_status, inputs=[token_box, dataset_name], outputs=[cached_responses_table, jobs_table])
768b9aa626e762f229b676191eb32483004a7d9b
Andrea Francis Soria Jimenez
2023-04-11T11:44:48
Change response type and adding progress (#1041)
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index 28024da6..bbf451c4 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -190,16 +189,0 @@ def get_response_without_content( -def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> List[CacheEntryWithoutContent]: - responses = CachedResponse.objects(kind=kind, dataset=dataset).only( - "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress" - ) - return [ - { - "http_status": response.http_status, - "error_code": response.error_code, - "job_runner_version": response.job_runner_version, - "dataset_git_revision": response.dataset_git_revision, - "progress": response.progress, - } - for response in responses - ] - - @@ -353,0 +338 @@ class CacheReport(TypedDict): + progress: Optional[float] @@ -415,0 +401 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo + "progress", @@ -429,0 +416 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo + "progress": object.progress, @@ -445,0 +433,28 @@ def get_outdated_split_full_names_for_step(kind: str, current_version: int) -> L +def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> List[CacheReport]: + responses = CachedResponse.objects(kind=kind, dataset=dataset).only( + "kind", + "dataset", + "config", + "split", + "http_status", + "error_code", + "job_runner_version", + "dataset_git_revision", + "progress", + ) + return [ + { + "kind": response.kind, + "dataset": response.dataset, + "config": response.config, + "split": response.split, + "http_status": response.http_status, + "error_code": response.error_code, + "job_runner_version": response.job_runner_version, + "dataset_git_revision": response.dataset_git_revision, + "progress": response.progress, + } + for response in responses + ] + + @@ -506,0 +522 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int) + "progress", @@ -523,0 +540 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int) + "progress": object.progress, diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index 6b921123..04fa6b68 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -490,0 +491 @@ def test_get_cache_reports() -> None: + "progress": None, @@ -500,0 +502 @@ def test_get_cache_reports() -> None: + "progress": None, @@ -516,0 +519 @@ def test_get_cache_reports() -> None: + "progress": None, @@ -539,0 +543 @@ def test_get_cache_reports() -> None: + "progress": None, @@ -552,0 +557 @@ def test_get_cache_reports() -> None: + "progress": None, @@ -573,0 +579 @@ def test_get_cache_reports() -> None: + "progress": None,
2d7a5e63c3c1c6bd1b1e5947a655c0cf0bd33798
Polina Kazakova
2023-04-10T16:04:48
Update `huggingface_hub` version to 0.13 (#1039)
diff --git a/e2e/poetry.lock b/e2e/poetry.lock index 210871b4..8cebc09a 100644 --- a/e2e/poetry.lock +++ b/e2e/poetry.lock @@ -423 +423 @@ name = "huggingface-hub" -version = "0.12.1" +version = "0.13.4" @@ -429,2 +429,2 @@ files = [ - {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"}, - {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"}, + {file = "huggingface_hub-0.13.4-py3-none-any.whl", hash = "sha256:4d3d40593de6673d624a4baaaf249b9bf5165bfcafd1ad58de361931f0b4fda5"}, + {file = "huggingface_hub-0.13.4.tar.gz", hash = "sha256:db83d9c2f76aed8cf49893ffadd6be24e82074da2f64b1d36b8ba40eb255e115"}, @@ -442 +442 @@ typing-extensions = ">=3.7.4.3" -all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -444 +444 @@ cli = ["InquirerPy (==0.3.4)"] -dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -446 +446 @@ fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] -quality = ["black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "mypy (==0.982)"] +quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"] @@ -448 +448 @@ tensorflow = ["graphviz", "pydot", "tensorflow"] -testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "isort (>=5.5.4)", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] @@ -1412 +1412 @@ python-versions = "3.9.15" -content-hash = "9d2cb8cda0199417c4c2d89abb558f93702f57571a0fadfae10f122b3774f827" +content-hash = "e903d839571339372972d5d533fc7386aa70b612240f5ee6f3d0f1673be1d0ba" diff --git a/e2e/pyproject.toml b/e2e/pyproject.toml index 066d81f7..189e1efa 100644 --- a/e2e/pyproject.toml +++ b/e2e/pyproject.toml @@ -16 +16 @@ flake8 = "^3.9.2" -huggingface-hub = "^0.12.0" +huggingface-hub = "^0.13.0" @@ -43,0 +44,7 @@ strict = true + +[[tool.mypy.overrides]] +module = [ + "huggingface_hub.*", +] +# ^ huggingface_hub is not typed since version 0.13.0 +ignore_missing_imports = true diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock index 3aba1994..5840dc3e 100644 --- a/front/admin_ui/poetry.lock +++ b/front/admin_ui/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -802 +802 @@ name = "huggingface-hub" -version = "0.12.1" +version = "0.13.4" @@ -808,2 +808,2 @@ files = [ - {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"}, - {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"}, + {file = "huggingface_hub-0.13.4-py3-none-any.whl", hash = "sha256:4d3d40593de6673d624a4baaaf249b9bf5165bfcafd1ad58de361931f0b4fda5"}, + {file = "huggingface_hub-0.13.4.tar.gz", hash = "sha256:db83d9c2f76aed8cf49893ffadd6be24e82074da2f64b1d36b8ba40eb255e115"}, @@ -821 +821 @@ typing-extensions = ">=3.7.4.3" -all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -823 +823 @@ cli = ["InquirerPy (==0.3.4)"] -dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -825 +825 @@ fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] -quality = ["black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "mypy (==0.982)"] +quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"] @@ -827 +827 @@ tensorflow = ["graphviz", "pydot", "tensorflow"] -testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "isort (>=5.5.4)", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] @@ -1428,0 +1429,7 @@ files = [ + {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"}, + {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"}, + {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"}, + {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"}, + {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"}, + {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"}, + {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"}, @@ -2114 +2121 @@ python-versions = "~3.9.15" -content-hash = "f5c77c5c5e2d965f1fe77cd0ab7f7299fc5cc788c3614b289bae08e8014348c1" +content-hash = "a352b5418336bced98c9f784c0b3ac87ecc738b0fc2ed1149842d8eeb72ef64f" diff --git a/front/admin_ui/pyproject.toml b/front/admin_ui/pyproject.toml index f45088d0..67e5696d 100644 --- a/front/admin_ui/pyproject.toml +++ b/front/admin_ui/pyproject.toml @@ -12 +12 @@ python = "~3.9.15" -huggingface-hub = "^0.12.0" +huggingface-hub = "^0.13.0" diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock index e4c1e6fb..c1bc5c87 100644 --- a/libs/libcommon/poetry.lock +++ b/libs/libcommon/poetry.lock @@ -473 +473 @@ name = "huggingface-hub" -version = "0.12.1" +version = "0.13.4" @@ -479,2 +479,2 @@ files = [ - {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"}, - {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"}, + {file = "huggingface_hub-0.13.4-py3-none-any.whl", hash = "sha256:4d3d40593de6673d624a4baaaf249b9bf5165bfcafd1ad58de361931f0b4fda5"}, + {file = "huggingface_hub-0.13.4.tar.gz", hash = "sha256:db83d9c2f76aed8cf49893ffadd6be24e82074da2f64b1d36b8ba40eb255e115"}, @@ -492 +492 @@ typing-extensions = ">=3.7.4.3" -all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -494 +494 @@ cli = ["InquirerPy (==0.3.4)"] -dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -496 +496 @@ fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] -quality = ["black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "mypy (==0.982)"] +quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"] @@ -498 +498 @@ tensorflow = ["graphviz", "pydot", "tensorflow"] -testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "isort (>=5.5.4)", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] @@ -1608 +1608 @@ python-versions = "3.9.15" -content-hash = "93a0454f579cd3797eb20ee68dac51f5a0cb9fb1d432fa21f6c0d74222884090" +content-hash = "7a54f8bf0afda6421a0b8af6d43de261f0bb88e4e6b2f41552d2bc64ee521f20" diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml index 4345c277..5f64c2a9 100644 --- a/libs/libcommon/pyproject.toml +++ b/libs/libcommon/pyproject.toml @@ -11 +11 @@ environs = "^9.5.0" -huggingface-hub = "^0.12.0" +huggingface-hub = "^0.13.0" @@ -60,0 +61 @@ module = [ + "huggingface_hub.*", @@ -61,0 +63 @@ module = [ +# ^ huggingface_hub is not typed since version 0.13.0 diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py index f5607512..54b39585 100644 --- a/libs/libcommon/src/libcommon/dataset.py +++ b/libs/libcommon/src/libcommon/dataset.py @@ -289 +289 @@ def get_dataset_git_revision( - return get_dataset_info_for_supported_datasets( + return get_dataset_info_for_supported_datasets( # type: ignore diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 22665d7c..06d634c6 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1437 +1437 @@ name = "huggingface-hub" -version = "0.12.1" +version = "0.13.4" @@ -1443,2 +1443,2 @@ files = [ - {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"}, - {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"}, + {file = "huggingface_hub-0.13.4-py3-none-any.whl", hash = "sha256:4d3d40593de6673d624a4baaaf249b9bf5165bfcafd1ad58de361931f0b4fda5"}, + {file = "huggingface_hub-0.13.4.tar.gz", hash = "sha256:db83d9c2f76aed8cf49893ffadd6be24e82074da2f64b1d36b8ba40eb255e115"}, @@ -1456 +1456 @@ typing-extensions = ">=3.7.4.3" -all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -1458 +1458 @@ cli = ["InquirerPy (==0.3.4)"] -dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] @@ -1460 +1460 @@ fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] -quality = ["black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "mypy (==0.982)"] +quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"] @@ -1462 +1462 @@ tensorflow = ["graphviz", "pydot", "tensorflow"] -testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "isort (>=5.5.4)", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] @@ -1701,0 +1702 @@ files = [ + {file = "libclang-15.0.6.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:0bf192c48a8d2992fc5034393ddc99e772ac30e105df84927d62fc88ef8a659f"}, @@ -1724 +1725 @@ environs = "^9.5.0" -huggingface-hub = "^0.12.0" +huggingface-hub = "^0.13.0" @@ -2746,0 +2748 @@ files = [ + {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -2766,0 +2769,7 @@ files = [ + {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"}, + {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"}, + {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"}, + {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"}, + {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"}, + {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"}, + {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"}, @@ -3781,0 +3791,8 @@ files = [ + {file = "pyzstd-0.15.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:df80324759cb74a10818f5ea4dfb0332b6064a31305246f66b043c4f73b37918"}, + {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b4833842d5e6660c1b6cd316f102ffddb95566f12ec1f89c5630d4b511e7e084"}, + {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9069da3bb94cac8af3fd2d7b622c7a78a248d3295a478587d880553287d56e6"}, + {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:9db032965b8fabf0bac13b75706031444ebfa7214794afe80be41799d8a349fb"}, + {file = "pyzstd-0.15.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:16eb2b2b206489bf8a7b690cf3b06b24352a16ee25717cb38c4381aefb3a94ce"}, + {file = "pyzstd-0.15.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:821555ce9bc7f50130bf52dc7b7d342913e00435c68a5cac564c5f2e57f33cbf"}, + {file = "pyzstd-0.15.4-cp36-cp36m-win32.whl", hash = "sha256:4f4a96bdffa773f8820e065ec840c5d3f252b0611951b73567121757388f171c"}, + {file = "pyzstd-0.15.4-cp36-cp36m-win_amd64.whl", hash = "sha256:3b0d384a48592683edf812c1a3d75a0cf5af253b1629596567ce3a2991d6ccbc"}, @@ -4214,0 +4232 @@ files = [ + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -4587,0 +4606,2 @@ files = [ + {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"}, + {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"}, @@ -4592,0 +4613 @@ files = [ + {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"}, @@ -4600,0 +4622 @@ files = [ + {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"}, @@ -5391 +5413 @@ python-versions = "3.9.15" -content-hash = "ad80aa527fe00375e13bb2be4ba2eb8fa4b658fb34756c14db4aa84b9ecef75e" +content-hash = "90df140fdcf886f4fae9406c7d5c0c5c3e18278e31d84674f9ae9444569d056d" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index f8db5f27..79322c36 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -18 +18 @@ hffs = {git = "https://github.com/huggingface/hffs.git", rev="0e187e74d38e943635 -huggingface-hub = "^0.12.0" +huggingface-hub = "^0.13.0" @@ -91,0 +92 @@ module = [ + "huggingface_hub.*",
9f7edded8358034a1034a4aad98f6c4ac2037772
Andrea Francis Soria Jimenez
2023-04-10T12:06:55
Moving cache_refresh to upgrade maintenance action (#1029)
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml index 5c0d5340..d938a7f6 100644 --- a/.github/workflows/cd.yml +++ b/.github/workflows/cd.yml @@ -26,2 +25,0 @@ jobs: - - directory: jobs - project: cache_refresh @@ -106,2 +103,0 @@ jobs: - cacheRefresh: - tag: sha-${{ steps.vars.outputs.sha_short }} diff --git a/.github/workflows/j-cache-refresh.yml b/.github/workflows/j-cache-refresh.yml deleted file mode 100644 index 35ef8434..00000000 --- a/.github/workflows/j-cache-refresh.yml +++ /dev/null @@ -1,33 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -name: jobs/cache_refresh -on: - workflow_dispatch: - push: - branches: - - main - paths: - - 'jobs/cache_refresh/**' - - 'libs/libcommon/**' - - '.github/workflows/j-cache-refresh.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' - pull_request: - paths: - - 'jobs/cache_refresh/**' - - 'libs/libcommon/**' - - '.github/workflows/j-cache-refresh.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' -jobs: - quality: - uses: ./.github/workflows/_quality-python.yml - with: - working-directory: jobs/cache_refresh - unit-tests: - uses: ./.github/workflows/_unit-tests-python.yml - with: - working-directory: jobs/cache_refresh diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace index 847cd8df..724166ba 100644 --- a/.vscode/monorepo.code-workspace +++ b/.vscode/monorepo.code-workspace @@ -15,4 +14,0 @@ - { - "name": "jobs/cache_refresh", - "path": "../jobs/cache_refresh" - }, diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 6a7bfe6a..302923c0 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -35,6 +34,0 @@ images: - # cacheRefresh: - # registry: huggingface - # useGlobalRegistry: false - # repository: datasets-server-jobs-cache_refresh - # tag: sha-fb3399a - # --- temporary disabling this job --- @@ -106,0 +101 @@ cacheMaintenance: + # ^ allowed values are {skip,backfill,upgrade} diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index fe24f4e6..32732890 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -34,6 +33,0 @@ images: - # cacheRefresh: - # registry: huggingface - # useGlobalRegistry: false - # repository: datasets-server-jobs-cache_refresh - # tag: sha-fb3399a - # --- temporary disabling this job --- @@ -112,0 +107 @@ cacheMaintenance: + # ^ allowed values are {skip,backfill,upgrade} diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index b2bad103..98cf42eb 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -37,4 +36,0 @@ Docker image management -{{- define "jobs.cacheRefresh.image" -}} -{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.jobs.cacheRefresh "global" .Values.global.huggingface) }} -{{- end -}} - @@ -78,5 +73,0 @@ app.kubernetes.io/component: "{{ include "name" . }}-mongodb-migration" -{{- end -}} - -{{- define "labels.cacheRefresh" -}} -{{ include "hf.labels.commons" . }} -app.kubernetes.io/component: "{{ include "name" . }}-cache-refresh" diff --git a/chart/templates/jobs/cache-refresh/_container.tpl b/chart/templates/jobs/cache-refresh/_container.tpl deleted file mode 100644 index 9bf1c796..00000000 --- a/chart/templates/jobs/cache-refresh/_container.tpl +++ /dev/null @@ -1,15 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -{{- define "containerCacheRefresh" -}} -- name: "{{ include "name" . }}-cache-refresh" - image: {{ include "jobs.cacheRefresh.image" . }} - imagePullPolicy: {{ .Values.images.pullPolicy }} - env: - {{ include "envCache" . | nindent 2 }} - {{ include "envQueue" . | nindent 2 }} - {{ include "envLog" . | nindent 2 }} - securityContext: - allowPrivilegeEscalation: false - resources: {{ toYaml .Values.cacheRefresh.resources | nindent 4 }} -{{- end -}} diff --git a/chart/values.yaml b/chart/values.yaml index c650dde2..269f31c1 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -40,6 +39,0 @@ images: - # cacheRefresh: - # registry: huggingface - # useGlobalRegistry: false - # repository: datasets-server-jobs-cache_refresh - # tag: sha-fb3399a - # --- temporary disabling this job --- @@ -194,9 +187,0 @@ mongodbMigration: -cacheRefresh: - nodeSelector: {} - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - @@ -204,0 +190 @@ cacheMaintenance: + # ^ allowed values are {skip,backfill,upgrade} diff --git a/jobs/cache_maintenance/src/cache_maintenance/config.py b/jobs/cache_maintenance/src/cache_maintenance/config.py index 7340409d..69ded5c9 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/config.py +++ b/jobs/cache_maintenance/src/cache_maintenance/config.py @@ -8 +8,7 @@ from environs import Env -from libcommon.config import CommonConfig, LogConfig, ProcessingGraphConfig, QueueConfig +from libcommon.config import ( + CacheConfig, + CommonConfig, + LogConfig, + ProcessingGraphConfig, + QueueConfig, +) @@ -15,0 +22 @@ class JobConfig: + cache: CacheConfig = field(default_factory=CacheConfig) @@ -26,0 +34 @@ class JobConfig: + cache=CacheConfig.from_env(), diff --git a/jobs/cache_maintenance/src/cache_maintenance/main.py b/jobs/cache_maintenance/src/cache_maintenance/main.py index f330b7c4..0479017d 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/main.py +++ b/jobs/cache_maintenance/src/cache_maintenance/main.py @@ -10 +10 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.resources import QueueMongoResource +from libcommon.resources import CacheMongoResource, QueueMongoResource @@ -13,0 +14 @@ from cache_maintenance.config import JobConfig +from cache_maintenance.upgrade import upgrade_cache @@ -19 +20 @@ def run_job() -> None: - supported_actions = ["backfill"] + supported_actions = ["backfill", "upgrade"] @@ -30,0 +32,3 @@ def run_job() -> None: + CacheMongoResource( + database=job_config.cache.mongo_database, host=job_config.cache.mongo_url + ) as cache_resource, @@ -34,0 +39,5 @@ def run_job() -> None: + if not cache_resource.is_available(): + logging.warning( + "The connection to the cache database could not be established. The cache refresh job is skipped." + ) + return @@ -43 +52 @@ def run_job() -> None: - + processing_steps = list(processing_graph.steps.values()) @@ -51,0 +61,2 @@ def run_job() -> None: + if action == "upgrade": + upgrade_cache(processing_steps) diff --git a/jobs/cache_refresh/src/cache_refresh/outdated_cache.py b/jobs/cache_maintenance/src/cache_maintenance/upgrade.py similarity index 93% rename from jobs/cache_refresh/src/cache_refresh/outdated_cache.py rename to jobs/cache_maintenance/src/cache_maintenance/upgrade.py index 9fee3cd7..52764614 100644 --- a/jobs/cache_refresh/src/cache_refresh/outdated_cache.py +++ b/jobs/cache_maintenance/src/cache_maintenance/upgrade.py @@ -12 +12,2 @@ from libcommon.simple_cache import get_outdated_split_full_names_for_step -def refresh_cache(processing_steps: List[ProcessingStep]) -> None: +def upgrade_cache(processing_steps: List[ProcessingStep]) -> None: + logging.info("upgrade cache") diff --git a/jobs/cache_maintenance/tests/conftest.py b/jobs/cache_maintenance/tests/conftest.py index 72f04b52..ad491836 100644 --- a/jobs/cache_maintenance/tests/conftest.py +++ b/jobs/cache_maintenance/tests/conftest.py @@ -7 +7,2 @@ from libcommon.queue import _clean_queue_database -from libcommon.resources import QueueMongoResource +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import _clean_cache_database @@ -16,0 +18 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]: + monkeypatch_session.setenv("CACHE_MONGO_DATABASE", "datasets_server_cache_test") @@ -25 +27 @@ def job_config(monkeypatch_session: MonkeyPatch) -> JobConfig: - if "test" not in job_config.queue.mongo_database: + if "test" not in job_config.cache.mongo_database or "test" not in job_config.queue.mongo_database: @@ -29,0 +32,7 @@ def job_config(monkeypatch_session: MonkeyPatch) -> JobConfig: +@fixture(autouse=True) +def cache_mongo_resource(job_config: JobConfig) -> Iterator[CacheMongoResource]: + with CacheMongoResource(database=job_config.cache.mongo_database, host=job_config.cache.mongo_url) as resource: + yield resource + _clean_cache_database() + + diff --git a/jobs/cache_refresh/tests/test_outdated_cache.py b/jobs/cache_maintenance/tests/test_upgrade_cache.py similarity index 89% rename from jobs/cache_refresh/tests/test_outdated_cache.py rename to jobs/cache_maintenance/tests/test_upgrade_cache.py index d2115233..5fb6a24c 100644 --- a/jobs/cache_refresh/tests/test_outdated_cache.py +++ b/jobs/cache_maintenance/tests/test_upgrade_cache.py @@ -11 +11 @@ from libcommon.simple_cache import upsert_response -from cache_refresh.outdated_cache import refresh_cache +from cache_maintenance.upgrade import upgrade_cache @@ -22 +22 @@ from cache_refresh.outdated_cache import refresh_cache -def test_refresh_cache( +def test_upgrade_cache( @@ -46 +46 @@ def test_refresh_cache( - refresh_cache(processing_steps=processing_steps) + upgrade_cache(processing_steps=processing_steps) diff --git a/jobs/cache_refresh/.flake8 b/jobs/cache_refresh/.flake8 deleted file mode 100644 index f7d6157c..00000000 --- a/jobs/cache_refresh/.flake8 +++ /dev/null @@ -1,5 +0,0 @@ -[flake8] -# Recommend matching the black line length (119), -# rather than using the flake8 default of 79: -max-line-length = 119 -extend-ignore = "E203" diff --git a/jobs/cache_refresh/.python-version b/jobs/cache_refresh/.python-version deleted file mode 100644 index b326afbc..00000000 --- a/jobs/cache_refresh/.python-version +++ /dev/null @@ -1 +0,0 @@ -3.9.15 diff --git a/jobs/cache_refresh/Dockerfile b/jobs/cache_refresh/Dockerfile deleted file mode 100644 index 0a9188b2..00000000 --- a/jobs/cache_refresh/Dockerfile +++ /dev/null @@ -1,33 +0,0 @@ -# build with -# docker build -t some_tag_admin -f Dockerfile ../.. -FROM python:3.9.15-slim - -ENV PYTHONFAULTHANDLER=1 \ - PYTHONUNBUFFERED=1 \ - PYTHONHASHSEED=random \ - PIP_NO_CACHE_DIR=1 \ - PIP_DISABLE_PIP_VERSION_CHECK=on \ - PIP_DEFAULT_TIMEOUT=100 \ - POETRY_NO_INTERACTION=1 \ - # Versions: - POETRY_VERSION=1.3.2 \ - POETRY_VIRTUALENVS_IN_PROJECT=true \ - PATH="$PATH:/root/.local/bin" - -# System deps: -RUN apt-get update \ - && apt-get install -y build-essential python3-dev make \ - && rm -rf /var/lib/apt/lists/* -RUN pip install -U pip -RUN pip install "poetry==$POETRY_VERSION" - -WORKDIR /src -COPY jobs/cache_refresh/poetry.lock ./jobs/cache_refresh/poetry.lock -COPY jobs/cache_refresh/pyproject.toml ./jobs/cache_refresh/pyproject.toml -COPY libs/libcommon ./libs/libcommon -WORKDIR /src/jobs/cache_refresh/ -RUN poetry install --no-cache -COPY jobs/cache_refresh/src ./src -RUN poetry install --no-cache - -ENTRYPOINT ["poetry", "run", "python", "src/cache_refresh/main.py"] diff --git a/jobs/cache_refresh/Makefile b/jobs/cache_refresh/Makefile deleted file mode 100644 index f421f1ee..00000000 --- a/jobs/cache_refresh/Makefile +++ /dev/null @@ -1,15 +0,0 @@ -# environment variables for the commands (docker compose, poetry) -export COMPOSE_PROJECT_NAME := cache_refresh -export MONGO_PORT := 27080 -export CACHE_MONGO_URL := mongodb://localhost:${MONGO_PORT} -export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT} -# makefile variables -DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml - -include ../../tools/Python.mk -include ../../tools/PythonTest.mk -include ../../tools/Docker.mk - -.PHONY: run -run: - poetry run python src/cache_refresh/main.py diff --git a/jobs/cache_refresh/poetry.lock b/jobs/cache_refresh/poetry.lock deleted file mode 100644 index 84bcce0c..00000000 --- a/jobs/cache_refresh/poetry.lock +++ /dev/null @@ -1,1571 +0,0 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. - -[[package]] -name = "appdirs" -version = "1.4.4" -description = "A small Python module for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." -category = "main" -optional = false -python-versions = "*" -files = [ - {file = "appdirs-1.4.4-py2.py3-none-any.whl", hash = "sha256:a841dacd6b99318a741b166adb07e19ee71a274450e68237b4650ca1055ab128"}, - {file = "appdirs-1.4.4.tar.gz", hash = "sha256:7d5d0167b2b1ba821647616af46a749d1c653740dd0d2415100fe26e27afdf41"}, -] - -[[package]] -name = "attrs" -version = "22.2.0" -description = "Classes Without Boilerplate" -category = "dev" -optional = false -python-versions = ">=3.6" -files = [ - {file = "attrs-22.2.0-py3-none-any.whl", hash = "sha256:29e95c7f6778868dbd49170f98f8818f78f3dc5e0e37c0b1f474e3561b240836"}, - {file = "attrs-22.2.0.tar.gz", hash = "sha256:c9227bfc2f01993c03f68db37d1d15c9690188323c067c641f1a35ca58185f99"}, -] - -[package.extras] -cov = ["attrs[tests]", "coverage-enable-subprocess", "coverage[toml] (>=5.3)"] -dev = ["attrs[docs,tests]"] -docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope.interface"] -tests = ["attrs[tests-no-zope]", "zope.interface"] -tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy (>=0.971,<0.990)", "mypy (>=0.971,<0.990)", "pympler", "pympler", "pytest (>=4.3.0)", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-mypy-plugins", "pytest-xdist[psutil]", "pytest-xdist[psutil]"] - -[[package]] -name = "bandit" -version = "1.7.4" -description = "Security oriented static analyser for python code." -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "bandit-1.7.4-py3-none-any.whl", hash = "sha256:412d3f259dab4077d0e7f0c11f50f650cc7d10db905d98f6520a95a18049658a"}, - {file = "bandit-1.7.4.tar.gz", hash = "sha256:2d63a8c573417bae338962d4b9b06fbc6080f74ecd955a092849e1e65c717bd2"}, -] - -[package.dependencies] -colorama = {version = ">=0.3.9", markers = "platform_system == \"Windows\""} -GitPython = ">=1.0.1" -PyYAML = ">=5.3.1" -stevedore = ">=1.20.0" - -[package.extras] -test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "toml"] -toml = ["toml"] -yaml = ["PyYAML"] - -[[package]] -name = "black" -version = "22.12.0" -description = "The uncompromising code formatter." -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "black-22.12.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eedd20838bd5d75b80c9f5487dbcb06836a43833a37846cf1d8c1cc01cef59d"}, - {file = "black-22.12.0-cp310-cp310-win_amd64.whl", hash = "sha256:159a46a4947f73387b4d83e87ea006dbb2337eab6c879620a3ba52699b1f4351"}, - {file = "black-22.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d30b212bffeb1e252b31dd269dfae69dd17e06d92b87ad26e23890f3efea366f"}, - {file = "black-22.12.0-cp311-cp311-win_amd64.whl", hash = "sha256:7412e75863aa5c5411886804678b7d083c7c28421210180d67dfd8cf1221e1f4"}, - {file = "black-22.12.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c116eed0efb9ff870ded8b62fe9f28dd61ef6e9ddd28d83d7d264a38417dcee2"}, - {file = "black-22.12.0-cp37-cp37m-win_amd64.whl", hash = "sha256:1f58cbe16dfe8c12b7434e50ff889fa479072096d79f0a7f25e4ab8e94cd8350"}, - {file = "black-22.12.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77d86c9f3db9b1bf6761244bc0b3572a546f5fe37917a044e02f3166d5aafa7d"}, - {file = "black-22.12.0-cp38-cp38-win_amd64.whl", hash = "sha256:82d9fe8fee3401e02e79767016b4907820a7dc28d70d137eb397b92ef3cc5bfc"}, - {file = "black-22.12.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:101c69b23df9b44247bd88e1d7e90154336ac4992502d4197bdac35dd7ee3320"}, - {file = "black-22.12.0-cp39-cp39-win_amd64.whl", hash = "sha256:559c7a1ba9a006226f09e4916060982fd27334ae1998e7a38b3f33a37f7a2148"}, - {file = "black-22.12.0-py3-none-any.whl", hash = "sha256:436cc9167dd28040ad90d3b404aec22cedf24a6e4d7de221bec2730ec0c97bcf"}, - {file = "black-22.12.0.tar.gz", hash = "sha256:229351e5a18ca30f447bf724d007f890f97e13af070bb6ad4c0a441cd7596a2f"}, -] - -[package.dependencies] -click = ">=8.0.0" -mypy-extensions = ">=0.4.3" -pathspec = ">=0.9.0" -platformdirs = ">=2" -tomli = {version = ">=1.1.0", markers = "python_full_version < \"3.11.0a7\""} -typing-extensions = {version = ">=3.10.0.0", markers = "python_version < \"3.10\""} - -[package.extras] -colorama = ["colorama (>=0.4.3)"] -d = ["aiohttp (>=3.7.4)"] -jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"] -uvloop = ["uvloop (>=0.15.2)"] - -[[package]] -name = "cachecontrol" -version = "0.12.11" -description = "httplib2 caching for requests" -category = "dev" -optional = false -python-versions = ">=3.6" -files = [ - {file = "CacheControl-0.12.11-py2.py3-none-any.whl", hash = "sha256:2c75d6a8938cb1933c75c50184549ad42728a27e9f6b92fd677c3151aa72555b"}, - {file = "CacheControl-0.12.11.tar.gz", hash = "sha256:a5b9fcc986b184db101aa280b42ecdcdfc524892596f606858e0b7a8b4d9e144"}, -] - -[package.dependencies] -lockfile = {version = ">=0.9", optional = true, markers = "extra == \"filecache\""} -msgpack = ">=0.5.2" -requests = "*" - -[package.extras] -filecache = ["lockfile (>=0.9)"] -redis = ["redis (>=2.10.5)"] - -[[package]] -name = "certifi" -version = "2022.12.7" -description = "Python package for providing Mozilla's CA Bundle." -category = "main" -optional = false -python-versions = ">=3.6" -files = [ - {file = "certifi-2022.12.7-py3-none-any.whl", hash = "sha256:4ad3232f5e926d6718ec31cfc1fcadfde020920e278684144551c91769c7bc18"}, - {file = "certifi-2022.12.7.tar.gz", hash = "sha256:35824b4c3a97115964b408844d64aa14db1cc518f6562e8d7261699d1350a9e3"}, -] - -[[package]] -name = "charset-normalizer" -version = "3.1.0" -description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." -category = "main" -optional = false -python-versions = ">=3.7.0" -files = [ - {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"}, - {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"}, - {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"}, - {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"}, - {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"}, - {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"}, - {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"}, -] - -[[package]] -name = "click" -version = "8.1.3" -description = "Composable command line interface toolkit" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "click-8.1.3-py3-none-any.whl", hash = "sha256:bb4d8133cb15a609f44e8213d9b391b0809795062913b383c62be0ee95b1db48"}, - {file = "click-8.1.3.tar.gz", hash = "sha256:7682dc8afb30297001674575ea00d1814d808d6a36af415a82bd481d37ba7b8e"}, -] - -[package.dependencies] -colorama = {version = "*", markers = "platform_system == \"Windows\""} - -[[package]] -name = "colorama" -version = "0.4.6" -description = "Cross-platform colored terminal text." -category = "main" -optional = false -python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7" -files = [ - {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"}, - {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"}, -] - -[[package]] -name = "coverage" -version = "7.2.1" -description = "Code coverage measurement for Python" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"}, - {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"}, - {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"}, - {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"}, - {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"}, - {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"}, - {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"}, - {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"}, - {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"}, - {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"}, - {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"}, - {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"}, - {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"}, - {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"}, - {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"}, - {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"}, - {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"}, - {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"}, - {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"}, - {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"}, - {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"}, - {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"}, - {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"}, - {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"}, - {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"}, - {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"}, - {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"}, - {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"}, - {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"}, - {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"}, - {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"}, - {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"}, - {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"}, - {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"}, - {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"}, - {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"}, - {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"}, - {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"}, - {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"}, - {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"}, - {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"}, - {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"}, - {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"}, - {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"}, - {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"}, - {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"}, - {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"}, - {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"}, - {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"}, - {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"}, - {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"}, -] - -[package.extras] -toml = ["tomli"] - -[[package]] -name = "cyclonedx-python-lib" -version = "3.1.5" -description = "A library for producing CycloneDX SBOM (Software Bill of Materials) files." -category = "dev" -optional = false -python-versions = ">=3.6,<4.0" -files = [ - {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, - {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, -] - -[package.dependencies] -packageurl-python = ">=0.9" -setuptools = ">=47.0.0" -sortedcontainers = ">=2.4.0,<3.0.0" -toml = ">=0.10.0,<0.11.0" - -[[package]] -name = "dnspython" -version = "1.16.0" -description = "DNS toolkit" -category = "main" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" -files = [ - {file = "dnspython-1.16.0-py2.py3-none-any.whl", hash = "sha256:f69c21288a962f4da86e56c4905b49d11aba7938d3d740e80d9e366ee4f1632d"}, - {file = "dnspython-1.16.0.zip", hash = "sha256:36c5e8e38d4369a08b6780b7f27d790a292b2b08eea01607865bf0936c558e01"}, -] - -[package.extras] -dnssec = ["ecdsa (>=0.13)", "pycryptodome"] -idna = ["idna (>=2.1)"] - -[[package]] -name = "environs" -version = "9.5.0" -description = "simplified environment variable parsing" -category = "main" -optional = false -python-versions = ">=3.6" -files = [ - {file = "environs-9.5.0-py2.py3-none-any.whl", hash = "sha256:1e549569a3de49c05f856f40bce86979e7d5ffbbc4398e7f338574c220189124"}, - {file = "environs-9.5.0.tar.gz", hash = "sha256:a76307b36fbe856bdca7ee9161e6c466fd7fcffc297109a118c59b54e27e30c9"}, -] - -[package.dependencies] -marshmallow = ">=3.0.0" -python-dotenv = "*" - -[package.extras] -dev = ["dj-database-url", "dj-email-url", "django-cache-url", "flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)", "pytest", "tox"] -django = ["dj-database-url", "dj-email-url", "django-cache-url"] -lint = ["flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)"] -tests = ["dj-database-url", "dj-email-url", "django-cache-url", "pytest"] - -[[package]] -name = "exceptiongroup" -version = "1.1.0" -description = "Backport of PEP 654 (exception groups)" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "exceptiongroup-1.1.0-py3-none-any.whl", hash = "sha256:327cbda3da756e2de031a3107b81ab7b3770a602c4d16ca618298c526f4bec1e"}, - {file = "exceptiongroup-1.1.0.tar.gz", hash = "sha256:bcb67d800a4497e1b404c2dd44fca47d3b7a5e5433dbab67f96c1a685cdfdf23"}, -] - -[package.extras] -test = ["pytest (>=6)"] - -[[package]] -name = "filelock" -version = "3.9.0" -description = "A platform independent file lock." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "filelock-3.9.0-py3-none-any.whl", hash = "sha256:f58d535af89bb9ad5cd4df046f741f8553a418c01a7856bf0d173bbc9f6bd16d"}, - {file = "filelock-3.9.0.tar.gz", hash = "sha256:7b319f24340b51f55a2bf7a12ac0755a9b03e718311dac567a0f4f7fabd2f5de"}, -] - -[package.extras] -docs = ["furo (>=2022.12.7)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"] -testing = ["covdefaults (>=2.2.2)", "coverage (>=7.0.1)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-timeout (>=2.1)"] - -[[package]] -name = "flake8" -version = "3.9.2" -description = "the modular source code checker: pep8 pyflakes and co" -category = "dev" -optional = false -python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7" -files = [ - {file = "flake8-3.9.2-py2.py3-none-any.whl", hash = "sha256:bf8fd333346d844f616e8d47905ef3a3384edae6b4e9beb0c5101e25e3110907"}, - {file = "flake8-3.9.2.tar.gz", hash = "sha256:07528381786f2a6237b061f6e96610a4167b226cb926e2aa2b6b1d78057c576b"}, -] - -[package.dependencies] -mccabe = ">=0.6.0,<0.7.0" -pycodestyle = ">=2.7.0,<2.8.0" -pyflakes = ">=2.3.0,<2.4.0" - -[[package]] -name = "gitdb" -version = "4.0.10" -description = "Git Object Database" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "gitdb-4.0.10-py3-none-any.whl", hash = "sha256:c286cf298426064079ed96a9e4a9d39e7f3e9bf15ba60701e95f5492f28415c7"}, - {file = "gitdb-4.0.10.tar.gz", hash = "sha256:6eb990b69df4e15bad899ea868dc46572c3f75339735663b81de79b06f17eb9a"}, -] - -[package.dependencies] -smmap = ">=3.0.1,<6" - -[[package]] -name = "gitpython" -version = "3.1.31" -description = "GitPython is a Python library used to interact with Git repositories" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"}, - {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"}, -] - -[package.dependencies] -gitdb = ">=4.0.1,<5" - -[[package]] -name = "html5lib" -version = "1.1" -description = "HTML parser based on the WHATWG HTML specification" -category = "dev" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" -files = [ - {file = "html5lib-1.1-py2.py3-none-any.whl", hash = "sha256:0d78f8fde1c230e99fe37986a60526d7049ed4bf8a9fadbad5f00e22e58e041d"}, - {file = "html5lib-1.1.tar.gz", hash = "sha256:b2e5b40261e20f354d198eae92afc10d750afb487ed5e50f9c4eaf07c184146f"}, -] - -[package.dependencies] -six = ">=1.9" -webencodings = "*" - -[package.extras] -all = ["chardet (>=2.2)", "genshi", "lxml"] -chardet = ["chardet (>=2.2)"] -genshi = ["genshi"] -lxml = ["lxml"] - -[[package]] -name = "huggingface-hub" -version = "0.12.1" -description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub" -category = "main" -optional = false -python-versions = ">=3.7.0" -files = [ - {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"}, - {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"}, -] - -[package.dependencies] -filelock = "*" -packaging = ">=20.9" -pyyaml = ">=5.1" -requests = "*" -tqdm = ">=4.42.1" -typing-extensions = ">=3.7.4.3" - -[package.extras] -all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] -cli = ["InquirerPy (==0.3.4)"] -dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] -fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] -quality = ["black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "mypy (==0.982)"] -tensorflow = ["graphviz", "pydot", "tensorflow"] -testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "isort (>=5.5.4)", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] -torch = ["torch"] -typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] - -[[package]] -name = "idna" -version = "3.4" -description = "Internationalized Domain Names in Applications (IDNA)" -category = "main" -optional = false -python-versions = ">=3.5" -files = [ - {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"}, - {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"}, -] - -[[package]] -name = "iniconfig" -version = "2.0.0" -description = "brain-dead simple config-ini parsing" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "iniconfig-2.0.0-py3-none-any.whl", hash = "sha256:b6a85871a79d2e3b22d2d1b94ac2824226a63c6b741c88f7ae975f18b6778374"}, - {file = "iniconfig-2.0.0.tar.gz", hash = "sha256:2d91e135bf72d31a410b17c16da610a82cb55f6b0477d1a902134b24a455b8b3"}, -] - -[[package]] -name = "isort" -version = "5.12.0" -description = "A Python utility / library to sort Python imports." -category = "dev" -optional = false -python-versions = ">=3.8.0" -files = [ - {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"}, - {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"}, -] - -[package.extras] -colors = ["colorama (>=0.4.3)"] -pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"] -plugins = ["setuptools"] -requirements-deprecated-finder = ["pip-api", "pipreqs"] - -[[package]] -name = "libcommon" -version = "0.6.8" -description = "Library for utils common to all the services" -category = "main" -optional = false -python-versions = "3.9.15" -files = [] -develop = true - -[package.dependencies] -appdirs = "^1.4.4" -environs = "^9.5.0" -huggingface-hub = "^0.12.0" -mongo-types = "0.15.1" -mongoengine = "^0.24.2" -networkx = "^3.0" -orjson = "^3.8.6" -psutil = "^5.9.4" -pymongo = {version = "^3.13.0", extras = ["srv"]} -pytz = "^2020.1" -requests = "^2.28.2" - -[package.source] -type = "directory" -url = "../../libs/libcommon" - -[[package]] -name = "lockfile" -version = "0.12.2" -description = "Platform-independent file locking module" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "lockfile-0.12.2-py2.py3-none-any.whl", hash = "sha256:6c3cb24f344923d30b2785d5ad75182c8ea7ac1b6171b08657258ec7429d50fa"}, - {file = "lockfile-0.12.2.tar.gz", hash = "sha256:6aed02de03cba24efabcd600b30540140634fc06cfa603822d508d5361e9f799"}, -] - -[[package]] -name = "markdown-it-py" -version = "2.2.0" -description = "Python port of markdown-it. Markdown parsing, done right!" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"}, - {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"}, -] - -[package.dependencies] -mdurl = ">=0.1,<1.0" - -[package.extras] -benchmarking = ["psutil", "pytest", "pytest-benchmark"] -code-style = ["pre-commit (>=3.0,<4.0)"] -compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"] -linkify = ["linkify-it-py (>=1,<3)"] -plugins = ["mdit-py-plugins"] -profiling = ["gprof2dot"] -rtd = ["attrs", "myst-parser", "pyyaml", "sphinx", "sphinx-copybutton", "sphinx-design", "sphinx_book_theme"] -testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"] - -[[package]] -name = "marshmallow" -version = "3.19.0" -description = "A lightweight library for converting complex datatypes to and from native Python datatypes." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "marshmallow-3.19.0-py3-none-any.whl", hash = "sha256:93f0958568da045b0021ec6aeb7ac37c81bfcccbb9a0e7ed8559885070b3a19b"}, - {file = "marshmallow-3.19.0.tar.gz", hash = "sha256:90032c0fd650ce94b6ec6dc8dfeb0e3ff50c144586462c389b81a07205bedb78"}, -] - -[package.dependencies] -packaging = ">=17.0" - -[package.extras] -dev = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)", "pytest", "pytz", "simplejson", "tox"] -docs = ["alabaster (==0.7.12)", "autodocsumm (==0.2.9)", "sphinx (==5.3.0)", "sphinx-issues (==3.0.1)", "sphinx-version-warning (==1.1.2)"] -lint = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)"] -tests = ["pytest", "pytz", "simplejson"] - -[[package]] -name = "mccabe" -version = "0.6.1" -description = "McCabe checker, plugin for flake8" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "mccabe-0.6.1-py2.py3-none-any.whl", hash = "sha256:ab8a6258860da4b6677da4bd2fe5dc2c659cff31b3ee4f7f5d64e79735b80d42"}, - {file = "mccabe-0.6.1.tar.gz", hash = "sha256:dd8d182285a0fe56bace7f45b5e7d1a6ebcbf524e8f3bd87eb0f125271b8831f"}, -] - -[[package]] -name = "mdurl" -version = "0.1.2" -description = "Markdown URL utilities" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8"}, - {file = "mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba"}, -] - -[[package]] -name = "mongo-types" -version = "0.15.1" -description = "Type stubs for mongoengine w/ basic support for bson and pymongo" -category = "main" -optional = false -python-versions = ">=3.7,<4.0" -files = [ - {file = "mongo-types-0.15.1.tar.gz", hash = "sha256:0a9deeb7733ea7da5db3711d92e22d93556b522f860bbff82e5df44c53bd06a9"}, - {file = "mongo_types-0.15.1-py3-none-any.whl", hash = "sha256:9417ae5b9a759c09630b5ec7d66904cc333c2d2fcfe75e2760a332ed5e267309"}, -] - -[[package]] -name = "mongoengine" -version = "0.24.2" -description = "MongoEngine is a Python Object-Document Mapper for working with MongoDB." -category = "main" -optional = false -python-versions = ">=3.6" -files = [ - {file = "mongoengine-0.24.2-py3-none-any.whl", hash = "sha256:f5c4e1b206b2ccffe4adc7a6283ed26dd799bd115a5fb1d2e885a075132cdb88"}, - {file = "mongoengine-0.24.2.tar.gz", hash = "sha256:c76d49658575bb995682e2e77c8ef7cda63faf939415b32ee923745d120f8b02"}, -] - -[package.dependencies] -pymongo = ">=3.4,<5.0" - -[[package]] -name = "msgpack" -version = "1.0.4" -description = "MessagePack serializer" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:4ab251d229d10498e9a2f3b1e68ef64cb393394ec477e3370c457f9430ce9250"}, - {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:112b0f93202d7c0fef0b7810d465fde23c746a2d482e1e2de2aafd2ce1492c88"}, - {file = "msgpack-1.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:002b5c72b6cd9b4bafd790f364b8480e859b4712e91f43014fe01e4f957b8467"}, - {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:35bc0faa494b0f1d851fd29129b2575b2e26d41d177caacd4206d81502d4c6a6"}, - {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4733359808c56d5d7756628736061c432ded018e7a1dff2d35a02439043321aa"}, - {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:eb514ad14edf07a1dbe63761fd30f89ae79b42625731e1ccf5e1f1092950eaa6"}, - {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c23080fdeec4716aede32b4e0ef7e213c7b1093eede9ee010949f2a418ced6ba"}, - {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:49565b0e3d7896d9ea71d9095df15b7f75a035c49be733051c34762ca95bbf7e"}, - {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:aca0f1644d6b5a73eb3e74d4d64d5d8c6c3d577e753a04c9e9c87d07692c58db"}, - {file = "msgpack-1.0.4-cp310-cp310-win32.whl", hash = "sha256:0dfe3947db5fb9ce52aaea6ca28112a170db9eae75adf9339a1aec434dc954ef"}, - {file = "msgpack-1.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dea20515f660aa6b7e964433b1808d098dcfcabbebeaaad240d11f909298075"}, - {file = "msgpack-1.0.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:e83f80a7fec1a62cf4e6c9a660e39c7f878f603737a0cdac8c13131d11d97f52"}, - {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3c11a48cf5e59026ad7cb0dc29e29a01b5a66a3e333dc11c04f7e991fc5510a9"}, - {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1276e8f34e139aeff1c77a3cefb295598b504ac5314d32c8c3d54d24fadb94c9"}, - {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6c9566f2c39ccced0a38d37c26cc3570983b97833c365a6044edef3574a00c08"}, - {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:fcb8a47f43acc113e24e910399376f7277cf8508b27e5b88499f053de6b115a8"}, - {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:76ee788122de3a68a02ed6f3a16bbcd97bc7c2e39bd4d94be2f1821e7c4a64e6"}, - {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0a68d3ac0104e2d3510de90a1091720157c319ceeb90d74f7b5295a6bee51bae"}, - {file = "msgpack-1.0.4-cp36-cp36m-win32.whl", hash = "sha256:85f279d88d8e833ec015650fd15ae5eddce0791e1e8a59165318f371158efec6"}, - {file = "msgpack-1.0.4-cp36-cp36m-win_amd64.whl", hash = "sha256:c1683841cd4fa45ac427c18854c3ec3cd9b681694caf5bff04edb9387602d661"}, - {file = "msgpack-1.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:a75dfb03f8b06f4ab093dafe3ddcc2d633259e6c3f74bb1b01996f5d8aa5868c"}, - {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9667bdfdf523c40d2511f0e98a6c9d3603be6b371ae9a238b7ef2dc4e7a427b0"}, - {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11184bc7e56fd74c00ead4f9cc9a3091d62ecb96e97653add7a879a14b003227"}, - {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ac5bd7901487c4a1dd51a8c58f2632b15d838d07ceedaa5e4c080f7190925bff"}, - {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1e91d641d2bfe91ba4c52039adc5bccf27c335356055825c7f88742c8bb900dd"}, - {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:2a2df1b55a78eb5f5b7d2a4bb221cd8363913830145fad05374a80bf0877cb1e"}, - {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:545e3cf0cf74f3e48b470f68ed19551ae6f9722814ea969305794645da091236"}, - {file = "msgpack-1.0.4-cp37-cp37m-win32.whl", hash = "sha256:2cc5ca2712ac0003bcb625c96368fd08a0f86bbc1a5578802512d87bc592fe44"}, - {file = "msgpack-1.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:eba96145051ccec0ec86611fe9cf693ce55f2a3ce89c06ed307de0e085730ec1"}, - {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:7760f85956c415578c17edb39eed99f9181a48375b0d4a94076d84148cf67b2d"}, - {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:449e57cc1ff18d3b444eb554e44613cffcccb32805d16726a5494038c3b93dab"}, - {file = "msgpack-1.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d603de2b8d2ea3f3bcb2efe286849aa7a81531abc52d8454da12f46235092bcb"}, - {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:48f5d88c99f64c456413d74a975bd605a9b0526293218a3b77220a2c15458ba9"}, - {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6916c78f33602ecf0509cc40379271ba0f9ab572b066bd4bdafd7434dee4bc6e"}, - {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:81fc7ba725464651190b196f3cd848e8553d4d510114a954681fd0b9c479d7e1"}, - {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:d5b5b962221fa2c5d3a7f8133f9abffc114fe218eb4365e40f17732ade576c8e"}, - {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:77ccd2af37f3db0ea59fb280fa2165bf1b096510ba9fe0cc2bf8fa92a22fdb43"}, - {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b17be2478b622939e39b816e0aa8242611cc8d3583d1cd8ec31b249f04623243"}, - {file = "msgpack-1.0.4-cp38-cp38-win32.whl", hash = "sha256:2bb8cdf50dd623392fa75525cce44a65a12a00c98e1e37bf0fb08ddce2ff60d2"}, - {file = "msgpack-1.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:26b8feaca40a90cbe031b03d82b2898bf560027160d3eae1423f4a67654ec5d6"}, - {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:462497af5fd4e0edbb1559c352ad84f6c577ffbbb708566a0abaaa84acd9f3ae"}, - {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2999623886c5c02deefe156e8f869c3b0aaeba14bfc50aa2486a0415178fce55"}, - {file = "msgpack-1.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f0029245c51fd9473dc1aede1160b0a29f4a912e6b1dd353fa6d317085b219da"}, - {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ed6f7b854a823ea44cf94919ba3f727e230da29feb4a99711433f25800cf747f"}, - {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0df96d6eaf45ceca04b3f3b4b111b86b33785683d682c655063ef8057d61fd92"}, - {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6a4192b1ab40f8dca3f2877b70e63799d95c62c068c84dc028b40a6cb03ccd0f"}, - {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:0e3590f9fb9f7fbc36df366267870e77269c03172d086fa76bb4eba8b2b46624"}, - {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:1576bd97527a93c44fa856770197dec00d223b0b9f36ef03f65bac60197cedf8"}, - {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:63e29d6e8c9ca22b21846234913c3466b7e4ee6e422f205a2988083de3b08cae"}, - {file = "msgpack-1.0.4-cp39-cp39-win32.whl", hash = "sha256:fb62ea4b62bfcb0b380d5680f9a4b3f9a2d166d9394e9bbd9666c0ee09a3645c"}, - {file = "msgpack-1.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:4d5834a2a48965a349da1c5a79760d94a1a0172fbb5ab6b5b33cbf8447e109ce"}, - {file = "msgpack-1.0.4.tar.gz", hash = "sha256:f5d869c18f030202eb412f08b28d2afeea553d6613aee89e200d7aca7ef01f5f"}, -] - -[[package]] -name = "mypy" -version = "1.1.1" -description = "Optional static typing for Python" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "mypy-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39c7119335be05630611ee798cc982623b9e8f0cff04a0b48dfc26100e0b97af"}, - {file = "mypy-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:61bf08362e93b6b12fad3eab68c4ea903a077b87c90ac06c11e3d7a09b56b9c1"}, - {file = "mypy-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dbb19c9f662e41e474e0cff502b7064a7edc6764f5262b6cd91d698163196799"}, - {file = "mypy-1.1.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:315ac73cc1cce4771c27d426b7ea558fb4e2836f89cb0296cbe056894e3a1f78"}, - {file = "mypy-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:5cb14ff9919b7df3538590fc4d4c49a0f84392237cbf5f7a816b4161c061829e"}, - {file = "mypy-1.1.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:26cdd6a22b9b40b2fd71881a8a4f34b4d7914c679f154f43385ca878a8297389"}, - {file = "mypy-1.1.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5b5f81b40d94c785f288948c16e1f2da37203c6006546c5d947aab6f90aefef2"}, - {file = "mypy-1.1.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21b437be1c02712a605591e1ed1d858aba681757a1e55fe678a15c2244cd68a5"}, - {file = "mypy-1.1.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d809f88734f44a0d44959d795b1e6f64b2bbe0ea4d9cc4776aa588bb4229fc1c"}, - {file = "mypy-1.1.1-cp311-cp311-win_amd64.whl", hash = "sha256:a380c041db500e1410bb5b16b3c1c35e61e773a5c3517926b81dfdab7582be54"}, - {file = "mypy-1.1.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:b7c7b708fe9a871a96626d61912e3f4ddd365bf7f39128362bc50cbd74a634d5"}, - {file = "mypy-1.1.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c1c10fa12df1232c936830839e2e935d090fc9ee315744ac33b8a32216b93707"}, - {file = "mypy-1.1.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:0a28a76785bf57655a8ea5eb0540a15b0e781c807b5aa798bd463779988fa1d5"}, - {file = "mypy-1.1.1-cp37-cp37m-win_amd64.whl", hash = "sha256:ef6a01e563ec6a4940784c574d33f6ac1943864634517984471642908b30b6f7"}, - {file = "mypy-1.1.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d64c28e03ce40d5303450f547e07418c64c241669ab20610f273c9e6290b4b0b"}, - {file = "mypy-1.1.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:64cc3afb3e9e71a79d06e3ed24bb508a6d66f782aff7e56f628bf35ba2e0ba51"}, - {file = "mypy-1.1.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce61663faf7a8e5ec6f456857bfbcec2901fbdb3ad958b778403f63b9e606a1b"}, - {file = "mypy-1.1.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:2b0c373d071593deefbcdd87ec8db91ea13bd8f1328d44947e88beae21e8d5e9"}, - {file = "mypy-1.1.1-cp38-cp38-win_amd64.whl", hash = "sha256:2888ce4fe5aae5a673386fa232473014056967f3904f5abfcf6367b5af1f612a"}, - {file = "mypy-1.1.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:19ba15f9627a5723e522d007fe708007bae52b93faab00f95d72f03e1afa9598"}, - {file = "mypy-1.1.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:59bbd71e5c58eed2e992ce6523180e03c221dcd92b52f0e792f291d67b15a71c"}, - {file = "mypy-1.1.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9401e33814cec6aec8c03a9548e9385e0e228fc1b8b0a37b9ea21038e64cdd8a"}, - {file = "mypy-1.1.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4b398d8b1f4fba0e3c6463e02f8ad3346f71956b92287af22c9b12c3ec965a9f"}, - {file = "mypy-1.1.1-cp39-cp39-win_amd64.whl", hash = "sha256:69b35d1dcb5707382810765ed34da9db47e7f95b3528334a3c999b0c90fe523f"}, - {file = "mypy-1.1.1-py3-none-any.whl", hash = "sha256:4e4e8b362cdf99ba00c2b218036002bdcdf1e0de085cdb296a49df03fb31dfc4"}, - {file = "mypy-1.1.1.tar.gz", hash = "sha256:ae9ceae0f5b9059f33dbc62dea087e942c0ccab4b7a003719cb70f9b8abfa32f"}, -] - -[package.dependencies] -mypy-extensions = ">=1.0.0" -tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""} -typing-extensions = ">=3.10" - -[package.extras] -dmypy = ["psutil (>=4.0)"] -install-types = ["pip"] -python2 = ["typed-ast (>=1.4.0,<2)"] -reports = ["lxml"] - -[[package]] -name = "mypy-extensions" -version = "1.0.0" -description = "Type system extensions for programs checked with the mypy type checker." -category = "dev" -optional = false -python-versions = ">=3.5" -files = [ - {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"}, - {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"}, -] - -[[package]] -name = "networkx" -version = "3.1" -description = "Python package for creating and manipulating graphs and networks" -category = "main" -optional = false -python-versions = ">=3.8" -files = [ - {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, - {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, -] - -[package.extras] -default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] -developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] -doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] -extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] -test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] - -[[package]] -name = "orjson" -version = "3.8.7" -description = "Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy" -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "orjson-3.8.7-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:f98c82850b7b4b7e27785ca43706fa86c893cdb88d54576bbb9b0d9c1070e421"}, - {file = "orjson-3.8.7-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:1dee503c6c1a0659c5b46f5f39d9ca9d3657b11ca8bb4af8506086df416887d9"}, - {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cc4fa83831f42ce5c938f8cefc2e175fa1df6f661fdeaba3badf26d2b8cfcf73"}, - {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9e432c6c9c8b97ad825276d5795286f7cc9689f377a97e3b7ecf14918413303f"}, - {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee519964a5a0efb9633f38b1129fd242807c5c57162844efeeaab1c8de080051"}, - {file = "orjson-3.8.7-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:109b539ce5bf60a121454d008fa67c3b67e5a3249e47d277012645922cf74bd0"}, - {file = "orjson-3.8.7-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ad4d441fbde4133af6fee37f67dbf23181b9c537ecc317346ec8c3b4c8ec7705"}, - {file = "orjson-3.8.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:89dc786419e1ce2588345f58dd6a434e6728bce66b94989644234bcdbe39b603"}, - {file = "orjson-3.8.7-cp310-none-win_amd64.whl", hash = "sha256:697abde7350fb8076d44bcb6b4ab3ce415ae2b5a9bb91efc460e5ab0d96bb5d3"}, - {file = "orjson-3.8.7-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:1c19f47b35b9966a3abadf341b18ee4a860431bf2b00fd8d58906d51cf78aa70"}, - {file = "orjson-3.8.7-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:3ffaabb380cd0ee187b4fc362516df6bf739808130b1339445c7d8878fca36e7"}, - {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d88837002c5a8af970745b8e0ca1b0fdb06aafbe7f1279e110d338ea19f3d23"}, - {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ff60187d1b7e0bfab376b6002b08c560b7de06c87cf3a8ac639ecf58f84c5f3b"}, - {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0110970aed35dec293f30ed1e09f8604afd5d15c5ef83de7f6c427619b3ba47b"}, - {file = "orjson-3.8.7-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:51b275475d4e36118b65ad56f9764056a09d985c5d72e64579bf8816f1356a5e"}, - {file = "orjson-3.8.7-cp311-none-win_amd64.whl", hash = "sha256:63144d27735f3b60f079f247ac9a289d80dfe49a7f03880dfa0c0ba64d6491d5"}, - {file = "orjson-3.8.7-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:a16273d77db746bb1789a2bbfded81148a60743fd6f9d5185e02d92e3732fa18"}, - {file = "orjson-3.8.7-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:5bb32259ea22cc9dd47a6fdc4b8f9f1e2f798fcf56c7c1122a7df0f4c5d33bf3"}, - {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ad02e9102d4ba67db30a136e631e32aeebd1dce26c9f5942a457b02df131c5d0"}, - {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:dbcfcec2b7ac52deb7be3685b551addc28ee8fa454ef41f8b714df6ba0e32a27"}, - {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e1a0e5504a5fc86083cc210c6946e8d61e13fe9f1d7a7bf81b42f7050a49d4fb"}, - {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:7bd4fd37adb03b1f2a1012d43c9f95973a02164e131dfe3ff804d7e180af5653"}, - {file = "orjson-3.8.7-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:188ed9f9a781333ad802af54c55d5a48991e292239aef41bd663b6e314377eb8"}, - {file = "orjson-3.8.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:cc52f58c688cb10afd810280e450f56fbcb27f52c053463e625c8335c95db0dc"}, - {file = "orjson-3.8.7-cp37-none-win_amd64.whl", hash = "sha256:403c8c84ac8a02c40613b0493b74d5256379e65196d39399edbf2ed3169cbeb5"}, - {file = "orjson-3.8.7-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:7d6ac5f8a2a17095cd927c4d52abbb38af45918e0d3abd60fb50cfd49d71ae24"}, - {file = "orjson-3.8.7-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:0295a7bfd713fa89231fd0822c995c31fc2343c59a1d13aa1b8b6651335654f5"}, - {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:feb32aaaa34cf2f891eb793ad320d4bb6731328496ae59b6c9eb1b620c42b529"}, - {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7a3ab1a473894e609b6f1d763838c6689ba2b97620c256a32c4d9f10595ac179"}, - {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e8c430d82b532c5ab95634e034bbf6ca7432ffe175a3e63eadd493e00b3a555"}, - {file = "orjson-3.8.7-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:366cc75f7e09106f9dac95a675aef413367b284f25507d21e55bd7f45f445e80"}, - {file = "orjson-3.8.7-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:84d154d07e8b17d97e990d5d710b719a031738eb1687d8a05b9089f0564ff3e0"}, - {file = "orjson-3.8.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:06180014afcfdc167ca984b312218aa62ce20093965c437c5f9166764cb65ef7"}, - {file = "orjson-3.8.7-cp38-none-win_amd64.whl", hash = "sha256:41244431ba13f2e6ef22b52c5cf0202d17954489f4a3c0505bd28d0e805c3546"}, - {file = "orjson-3.8.7-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:b20f29fa8371b8023f1791df035a2c3ccbd98baa429ac3114fc104768f7db6f8"}, - {file = "orjson-3.8.7-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:226bfc1da2f21ee74918cee2873ea9a0fec1a8830e533cb287d192d593e99d02"}, - {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e75c11023ac29e29fd3e75038d0e8dd93f9ea24d7b9a5e871967a8921a88df24"}, - {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:78604d3acfd7cd502f6381eea0c42281fe2b74755b334074ab3ebc0224100be1"}, - {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e7129a6847f0494aa1427167486ef6aea2e835ba05f6c627df522692ee228f65"}, - {file = "orjson-3.8.7-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:1a1a8f4980059f48483782c608145b0f74538c266e01c183d9bcd9f8b71dbada"}, - {file = "orjson-3.8.7-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:d60304172a33705ce4bd25a6261ab84bed2dab0b3d3b79672ea16c7648af4832"}, - {file = "orjson-3.8.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4f733062d84389c32c0492e5a4929056fac217034a94523debe0430bcc602cda"}, - {file = "orjson-3.8.7-cp39-none-win_amd64.whl", hash = "sha256:010e2970ec9e826c332819e0da4b14b29b19641da0f1a6af4cec91629ef9b988"}, - {file = "orjson-3.8.7.tar.gz", hash = "sha256:8460c8810652dba59c38c80d27c325b5092d189308d8d4f3e688dbd8d4f3b2dc"}, -] - -[[package]] -name = "packageurl-python" -version = "0.10.4" -description = "A purl aka. Package URL parser and builder" -category = "dev" -optional = false -python-versions = ">=3.6" -files = [ - {file = "packageurl-python-0.10.4.tar.gz", hash = "sha256:5c91334f942cd55d45eb0c67dd339a535ef90e25f05b9ec016ad188ed0ef9048"}, - {file = "packageurl_python-0.10.4-py3-none-any.whl", hash = "sha256:bf8a1ffe755634776f6563904d792fb0aa13b377fc86115c36fe17f69b6e59db"}, -] - -[package.extras] -build = ["wheel"] -test = ["black", "isort", "pytest"] - -[[package]] -name = "packaging" -version = "23.0" -description = "Core utilities for Python packages" -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "packaging-23.0-py3-none-any.whl", hash = "sha256:714ac14496c3e68c99c29b00845f7a2b85f3bb6f1078fd9f72fd20f0570002b2"}, - {file = "packaging-23.0.tar.gz", hash = "sha256:b6ad297f8907de0fa2fe1ccbd26fdaf387f5f47c7275fedf8cce89f99446cf97"}, -] - -[[package]] -name = "pathspec" -version = "0.11.0" -description = "Utility library for gitignore style pattern matching of file paths." -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "pathspec-0.11.0-py3-none-any.whl", hash = "sha256:3a66eb970cbac598f9e5ccb5b2cf58930cd8e3ed86d393d541eaf2d8b1705229"}, - {file = "pathspec-0.11.0.tar.gz", hash = "sha256:64d338d4e0914e91c1792321e6907b5a593f1ab1851de7fc269557a21b30ebbc"}, -] - -[[package]] -name = "pbr" -version = "5.11.1" -description = "Python Build Reasonableness" -category = "dev" -optional = false -python-versions = ">=2.6" -files = [ - {file = "pbr-5.11.1-py2.py3-none-any.whl", hash = "sha256:567f09558bae2b3ab53cb3c1e2e33e726ff3338e7bae3db5dc954b3a44eef12b"}, - {file = "pbr-5.11.1.tar.gz", hash = "sha256:aefc51675b0b533d56bb5fd1c8c6c0522fe31896679882e1c4c63d5e4a0fccb3"}, -] - -[[package]] -name = "pip" -version = "23.0.1" -description = "The PyPA recommended tool for installing Python packages." -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"}, - {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"}, -] - -[[package]] -name = "pip-api" -version = "0.0.30" -description = "An unofficial, importable pip API" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "pip-api-0.0.30.tar.gz", hash = "sha256:a05df2c7aa9b7157374bcf4273544201a0c7bae60a9c65bcf84f3959ef3896f3"}, - {file = "pip_api-0.0.30-py3-none-any.whl", hash = "sha256:2a0314bd31522eb9ffe8a99668b0d07fee34ebc537931e7b6483001dbedcbdc9"}, -] - -[package.dependencies] -pip = "*" - -[[package]] -name = "pip-audit" -version = "2.4.14" -description = "A tool for scanning Python environments for known vulnerabilities" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, - {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, -] - -[package.dependencies] -CacheControl = {version = ">=0.12.10", extras = ["filecache"]} -cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" -html5lib = ">=1.1" -packaging = ">=23.0.0" -pip-api = ">=0.0.28" -pip-requirements-parser = ">=32.0.0" -resolvelib = ">=0.8.0" -rich = ">=12.4" -toml = ">=0.10" - -[package.extras] -dev = ["build", "bump (>=1.3.2)", "pip-audit[doc,lint,test]"] -doc = ["pdoc"] -lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] -test = ["coverage[toml]", "pretend", "pytest", "pytest-cov"] - -[[package]] -name = "pip-requirements-parser" -version = "32.0.1" -description = "pip requirements parser - a mostly correct pip requirements parsing library because it uses pip's own code." -category = "dev" -optional = false -python-versions = ">=3.6.0" -files = [ - {file = "pip-requirements-parser-32.0.1.tar.gz", hash = "sha256:b4fa3a7a0be38243123cf9d1f3518da10c51bdb165a2b2985566247f9155a7d3"}, - {file = "pip_requirements_parser-32.0.1-py3-none-any.whl", hash = "sha256:4659bc2a667783e7a15d190f6fccf8b2486685b6dba4c19c3876314769c57526"}, -] - -[package.dependencies] -packaging = "*" -pyparsing = "*" - -[package.extras] -docs = ["Sphinx (>=3.3.1)", "doc8 (>=0.8.1)", "sphinx-rtd-theme (>=0.5.0)"] -testing = ["aboutcode-toolkit (>=6.0.0)", "black", "pytest (>=6,!=7.0.0)", "pytest-xdist (>=2)"] - -[[package]] -name = "platformdirs" -version = "3.1.0" -description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "platformdirs-3.1.0-py3-none-any.whl", hash = "sha256:13b08a53ed71021350c9e300d4ea8668438fb0046ab3937ac9a29913a1a1350a"}, - {file = "platformdirs-3.1.0.tar.gz", hash = "sha256:accc3665857288317f32c7bebb5a8e482ba717b474f3fc1d18ca7f9214be0cef"}, -] - -[package.extras] -docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"] -test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"] - -[[package]] -name = "pluggy" -version = "1.0.0" -description = "plugin and hook calling mechanisms for python" -category = "dev" -optional = false -python-versions = ">=3.6" -files = [ - {file = "pluggy-1.0.0-py2.py3-none-any.whl", hash = "sha256:74134bbf457f031a36d68416e1509f34bd5ccc019f0bcc952c7b909d06b37bd3"}, - {file = "pluggy-1.0.0.tar.gz", hash = "sha256:4224373bacce55f955a878bf9cfa763c1e360858e330072059e10bad68531159"}, -] - -[package.extras] -dev = ["pre-commit", "tox"] -testing = ["pytest", "pytest-benchmark"] - -[[package]] -name = "psutil" -version = "5.9.4" -description = "Cross-platform lib for process and system monitoring in Python." -category = "main" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" -files = [ - {file = "psutil-5.9.4-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:c1ca331af862803a42677c120aff8a814a804e09832f166f226bfd22b56feee8"}, - {file = "psutil-5.9.4-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:68908971daf802203f3d37e78d3f8831b6d1014864d7a85937941bb35f09aefe"}, - {file = "psutil-5.9.4-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:3ff89f9b835100a825b14c2808a106b6fdcc4b15483141482a12c725e7f78549"}, - {file = "psutil-5.9.4-cp27-cp27m-win32.whl", hash = "sha256:852dd5d9f8a47169fe62fd4a971aa07859476c2ba22c2254d4a1baa4e10b95ad"}, - {file = "psutil-5.9.4-cp27-cp27m-win_amd64.whl", hash = "sha256:9120cd39dca5c5e1c54b59a41d205023d436799b1c8c4d3ff71af18535728e94"}, - {file = "psutil-5.9.4-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:6b92c532979bafc2df23ddc785ed116fced1f492ad90a6830cf24f4d1ea27d24"}, - {file = "psutil-5.9.4-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:efeae04f9516907be44904cc7ce08defb6b665128992a56957abc9b61dca94b7"}, - {file = "psutil-5.9.4-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:54d5b184728298f2ca8567bf83c422b706200bcbbfafdc06718264f9393cfeb7"}, - {file = "psutil-5.9.4-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16653106f3b59386ffe10e0bad3bb6299e169d5327d3f187614b1cb8f24cf2e1"}, - {file = "psutil-5.9.4-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:54c0d3d8e0078b7666984e11b12b88af2db11d11249a8ac8920dd5ef68a66e08"}, - {file = "psutil-5.9.4-cp36-abi3-win32.whl", hash = "sha256:149555f59a69b33f056ba1c4eb22bb7bf24332ce631c44a319cec09f876aaeff"}, - {file = "psutil-5.9.4-cp36-abi3-win_amd64.whl", hash = "sha256:fd8522436a6ada7b4aad6638662966de0d61d241cb821239b2ae7013d41a43d4"}, - {file = "psutil-5.9.4-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:6001c809253a29599bc0dfd5179d9f8a5779f9dffea1da0f13c53ee568115e1e"}, - {file = "psutil-5.9.4.tar.gz", hash = "sha256:3d7f9739eb435d4b1338944abe23f49584bde5395f27487d2ee25ad9a8774a62"}, -] - -[package.extras] -test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] - -[[package]] -name = "pycodestyle" -version = "2.7.0" -description = "Python style guide checker" -category = "dev" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" -files = [ - {file = "pycodestyle-2.7.0-py2.py3-none-any.whl", hash = "sha256:514f76d918fcc0b55c6680472f0a37970994e07bbb80725808c17089be302068"}, - {file = "pycodestyle-2.7.0.tar.gz", hash = "sha256:c389c1d06bf7904078ca03399a4816f974a1d590090fecea0c63ec26ebaf1cef"}, -] - -[[package]] -name = "pyflakes" -version = "2.3.1" -description = "passive checker of Python programs" -category = "dev" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" -files = [ - {file = "pyflakes-2.3.1-py2.py3-none-any.whl", hash = "sha256:7893783d01b8a89811dd72d7dfd4d84ff098e5eed95cfa8905b22bbffe52efc3"}, - {file = "pyflakes-2.3.1.tar.gz", hash = "sha256:f5bc8ecabc05bb9d291eb5203d6810b49040f6ff446a756326104746cc00c1db"}, -] - -[[package]] -name = "pygments" -version = "2.14.0" -description = "Pygments is a syntax highlighting package written in Python." -category = "dev" -optional = false -python-versions = ">=3.6" -files = [ - {file = "Pygments-2.14.0-py3-none-any.whl", hash = "sha256:fa7bd7bd2771287c0de303af8bfdfc731f51bd2c6a47ab69d117138893b82717"}, - {file = "Pygments-2.14.0.tar.gz", hash = "sha256:b3ed06a9e8ac9a9aae5a6f5dbe78a8a58655d17b43b93c078f094ddc476ae297"}, -] - -[package.extras] -plugins = ["importlib-metadata"] - -[[package]] -name = "pymongo" -version = "3.13.0" -description = "Python driver for MongoDB <http://www.mongodb.org>" -category = "main" -optional = false -python-versions = "*" -files = [ - {file = "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl", hash = "sha256:3ad3a3df830f7df7e0856c2bdb54d19f5bf188bd7420985e18643b8e4d2a075f"}, - {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:b96e0e9d2d48948240b510bac81614458fc10adcd3a93240c2fd96448b4efd35"}, - {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9f592b202d77923498b32ddc5b376e5fa9ba280d3e16ed56cb8c932fe6d6a478"}, - {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:851f2bb52b5cb2f4711171ca925e0e05344a8452972a748a8a8ffdda1e1d72a7"}, - {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:1c9d23f62a3fa7523d849c4942acc0d9ff7081ebc00c808ee7cfdc070df0687f"}, - {file = "pymongo-3.13.0-cp27-cp27m-win32.whl", hash = "sha256:a17b81f22398e3e0f72bdf938e98c810286994b2bcc0a125cd5ad8fd4ea54ad7"}, - {file = "pymongo-3.13.0-cp27-cp27m-win_amd64.whl", hash = "sha256:4f6dd55dab77adf60b445c11f426ee5cdfa1b86f6d54cb937bfcbf09572333ab"}, - {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:776f90bf2252f90a4ae838e7917638894c6356bef7265f424592e2fd1f577d05"}, - {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:50b99f4d3eee6f03778fe841d6f470e6c18e744dc665156da6da3bc6e65b398d"}, - {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50a81b2d9f188c7909e0a1084fa969bb92a788076809c437ac1ae80393f46df9"}, - {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c7c45a8a1a752002b0a7c81ab3a4c5e3b6f67f9826b16fbe3943f5329f565f24"}, - {file = "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1037097708498bdc85f23c8798a5c46c7bce432d77d23608ff14e0d831f1a971"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl", hash = "sha256:b5b733694e7df22d5c049581acfc487695a6ff813322318bed8dd66f79978636"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:d7c91747ec8dde51440dd594603158cc98abb3f7df84b2ed8a836f138285e4fb"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl", hash = "sha256:f4175fcdddf764d371ee52ec4505a40facee2533e84abf2953cda86d050cfa1f"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl", hash = "sha256:93d4e9a02c17813b34e4bd9f6fbf07310c140c8f74341537c24d07c1cdeb24d1"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl", hash = "sha256:3b261d593f2563299062733ae003a925420a86ff4ddda68a69097d67204e43f3"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl", hash = "sha256:172db03182a22e9002157b262c1ea3b0045c73d4ff465adc152ce5b4b0e7b8d4"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09de3bfc995ae8cb955abb0c9ae963c134dba1b5622be3bcc527b89b0fd4091c"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c0379447587ee4b8f983ba183202496e86c0358f47c45612619d634d1fcd82bd"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30245a8747dc90019a3c9ad9df987e0280a3ea632ad36227cde7d1d8dcba0830"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65b6fddf6a7b91da044f202771a38e71bbb9bf42720a406b26b25fe2256e7102"}, - {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5831a377d15a626fbec10890ffebc4c6abcd37e4126737932cd780a171eabdc1"}, - {file = "pymongo-3.13.0-cp310-cp310-win32.whl", hash = "sha256:944249aa83dee314420c37d0f40c30a8f6dc4a3877566017b87062e53af449f4"}, - {file = "pymongo-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea8824ebc9a1a5c8269e8f1e3989b5a6bec876726e2f3c33ebd036cb488277f0"}, - {file = "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:bdd34c57b4da51a7961beb33645646d197e41f8517801dc76b37c1441e7a4e10"}, - {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26f9cc42a162faa241c82e117ac85734ae9f14343dc2df1c90c6b2181f791b22"}, - {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a82a1c10f5608e6494913faa169e213d703194bfca0aa710901f303be212414"}, - {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8927f22ef6a16229da7f18944deac8605bdc2c0858be5184259f2f7ce7fd4459"}, - {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e6f8191a282ef77e526f8f8f63753a437e4aa4bc78f5edd8b6b6ed0eaebd5363"}, - {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4d9ed67c987bf9ac2ac684590ba3d2599cdfb0f331ee3db607f9684469b3b59d"}, - {file = "pymongo-3.13.0-cp311-cp311-win32.whl", hash = "sha256:e8f6979664ff477cd61b06bf8aba206df7b2334209815ab3b1019931dab643d6"}, - {file = "pymongo-3.13.0-cp311-cp311-win_amd64.whl", hash = "sha256:174fd1000e896d0dfbc7f6d7e6a1992a4868796c7dec31679e38218c78d6a942"}, - {file = "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl", hash = "sha256:d1ee773fb72ba024e7e3bb6ea8907fe52bccafcb5184aaced6bad995bd30ea20"}, - {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:28565e3dbd69fe5fe35a210067064dbb6ed5abe997079f653c19c873c3896fe6"}, - {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:5c1db7d366004d6c699eb08c716a63ae0a3e946d061cbebea65d7ce361950265"}, - {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e1956f3338c10308e2f99c2c9ff46ae412035cbcd7aaa76c39ccdb806854a247"}, - {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:10f0fddc1d63ba3d4a4bffcc7720184c1b7efd570726ad5e2f55818da320239f"}, - {file = "pymongo-3.13.0-cp35-cp35m-win32.whl", hash = "sha256:570ae3365b23d4fd8c669cb57613b1a90b2757e993588d3370ef90945dbeec4b"}, - {file = "pymongo-3.13.0-cp35-cp35m-win_amd64.whl", hash = "sha256:79f777eaf3f5b2c6d81f9ef00d87837001d7063302503bbcbfdbf3e9bc27c96f"}, - {file = "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl", hash = "sha256:d42eb29ba314adfd9c11234b4b646f61b0448bf9b00f14db4b317e6e4b947e77"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:e5e87c0eb774561c546f979342a8ff36ebee153c60a0b6c6b03ba989ceb9538c"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:0f2c5a5984599a88d087a15859860579b825098b473d8c843f1979a83d159f2e"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:59c98e86c5e861032b71e6e5b65f23e6afaacea6e82483b66f1191a5021a7b4f"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:70b67390e27e58876853efbb87e43c85252de2515e2887f7dd901b4fa3d21973"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:42ba8606492d76e6f9e4c7a458ed4bc712603be393259a52450345f0945da2cf"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:0e5536994cf2d8488c6fd9dea71df3c4dbb3e0d2ba5e695da06d9142a29a0969"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:fe8194f107f0fa3cabd14e9e809f174eca335993c1db72d1e74e0f496e7afe1f"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d593d50815771f517d3ac4367ff716e3f3c78edae51d98e1e25791459f8848ff"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5136ebe8da6a1604998a8eb96be55935aa5f7129c41cc7bddc400d48e8df43be"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a424bdedfd84454d2905a861e0d4bb947cc5bd024fdeb3600c1a97d2be0f4255"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5161167b3840e9c84c80f2534ea6a099f51749d5673b662a3dd248be17c3208"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:644470442beaf969df99c4e00367a817eee05f0bba5d888f1ba6fe97b5e1c102"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2406df90b2335371706c59b7d79e9633b81ed2a7ecd48c1faf8584552bdf2d90"}, - {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:222591b828de10ac90064047b5d4916953f38c38b155009c4b8b5e0d33117c2b"}, - {file = "pymongo-3.13.0-cp36-cp36m-win32.whl", hash = "sha256:7cb987b199fa223ad78eebaa9fbc183d5a5944bfe568a9d6f617316ca1c1f32f"}, - {file = "pymongo-3.13.0-cp36-cp36m-win_amd64.whl", hash = "sha256:a6cbb73d9fc2282677e2b7a137d13da987bd0b13abd88ed27bba5534c226db06"}, - {file = "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl", hash = "sha256:b1223b826acbef07a7f5eb9bf37247b0b580119916dca9eae19d92b1290f5855"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:398fb86d374dc351a4abc2e24cd15e5e14b2127f6d90ce0df3fdf2adcc55ac1b"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:9c3d07ea19cd2856d9943dce37e75d69ecbb5baf93c3e4c82f73b6075c481292"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:2943d739715f265a2983ac43747595b6af3312d0a370614040959fd293763adf"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:c3b70ed82f20d18d22eafc9bda0ea656605071762f7d31f3c5afc35c59d3393b"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:7ec2bb598847569ae34292f580842d37619eea3e546005042f485e15710180d5"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:8cc37b437cba909bef06499dadd91a39c15c14225e8d8c7870020049f8a549fe"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:65a063970e15a4f338f14b820561cf6cdaf2839691ac0adb2474ddff9d0b8b0b"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02f0e1a75d3bc0e16c7e15daf9c56185642be055e425f3b34888fc6eb1b22401"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16e74b9c2aca2734c7f49f00fe68d6830a30d26df60e2ace7fe40ccb92087b94"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:24e954be35ad4537840f20bbc8d75320ae647d3cb4fab12cb8fcd2d55f408e76"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a149377d1ff766fd618500798d0d94637f66d0ae222bb6d28f41f3e15c626297"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61660710b054ae52c8fc10368e91d74719eb05554b631d7f8ca93d21d2bff2e6"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4bbc0d27dfef7689285e54f2e0a224f0c7cd9d5c46d2638fabad5500b951c92f"}, - {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:9b2ed9c3b30f11cd4a3fbfc22167af7987b01b444215c2463265153fe7cf66d6"}, - {file = "pymongo-3.13.0-cp37-cp37m-win32.whl", hash = "sha256:1c2c5e2b00e2fadcd590c0b2e293d71215e98ed1cb635cfca2be4998d197e534"}, - {file = "pymongo-3.13.0-cp37-cp37m-win_amd64.whl", hash = "sha256:32eac95bbb030b2376ffd897376c6f870222a3457f01a9ce466b9057876132f8"}, - {file = "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a796ef39dadf9d73af05d24937644d386495e43a7d13617aa3651d836da542c8"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:b6793baf4639c72a500698a49e9250b293e17ae1faf11ac1699d8141194786fe"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:80d8576b04d0824f63bf803190359c0d3bcb6e7fa63fefbd4bc0ceaa7faae38c"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:db2e11507fe9cc2a722be21ccc62c1b1295398fe9724c1f14900cdc7166fc0d7"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:b01ce58eec5edeededf1992d2dce63fb8565e437be12d6f139d75b15614c4d08"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:d1a19d6c5098f1f4e11430cd74621699453cbc534dd7ade9167e582f50814b19"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:7219b1a726ced3bacecabef9bd114529bbb69477901373e800d7d0140baadc95"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:2dae3b353a10c3767e0aa1c1492f2af388f1012b08117695ab3fd1f219e5814e"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:12721d926d43d33dd3318e58dce9b0250e8a9c6e1093fa8e09f4805193ff4b43"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6af0a4b17faf26779d5caee8542a4f2cba040cea27d3bffc476cbc6ccbd4c8ee"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:09b9d0f5a445c7e0ddcc021b09835aa6556f0166afc498f57dfdd72cdf6f02ad"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db5b4f8ad8607a3d612da1d4c89a84e4cf5c88f98b46365820d9babe5884ba45"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:34dbf5fecf653c152edb75a35a8b15dfdc4549473484ee768aeb12c97983cead"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:34cd48df7e1fc69222f296d8f69e3957eb7c6b5aa0709d3467184880ed7538c0"}, - {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c8f755ff1f4ab4ca790d1d6d3229006100b301475948021b6b2757822e0d6c97"}, - {file = "pymongo-3.13.0-cp38-cp38-win32.whl", hash = "sha256:b0746d0d4535f56bbaa63a8f6da362f330804d578e66e126b226eebe76c2bf00"}, - {file = "pymongo-3.13.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ad0515abb132f52ce9d8abd1a29681a1e65dba7b7fe13ea01e1a8db5715bf80"}, - {file = "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:3c5cb6c93c94df76a879bad4b89db0104b01806d17c2b803c1316ba50962b6d6"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2e0854170813238f0c3131050c67cb1fb1ade75c93bf6cd156c1bd9a16095528"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:1410faa51ce835cc1234c99ec42e98ab4f3c6f50d92d86a2d4f6e11c97ee7a4e"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:d7910135f5de1c5c3578e61d6f4b087715b15e365f11d4fa51a9cee92988b2bd"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:028175dd8d2979a889153a2308e8e500b3df7d9e3fd1c33ca7fdeadf61cc87a2"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:2bfc39276c0e6d07c95bd1088b5003f049e986e089509f7dbd68bb7a4b1e65ac"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:4092b660ec720d44d3ca81074280dc25c7a3718df1b6c0fe9fe36ac6ed2833e4"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:5bdeb71a610a7b801416268e500e716d0fe693fb10d809e17f0fb3dac5be5a34"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aa3bca8e76f5c00ed2bb4325e0e383a547d71595926d5275d7c88175aaf7435e"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7c7cab8155f430ca460a6fc7ae8a705b34f3e279a57adb5f900eb81943ec777c"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4a32f3dfcca4a4816373bdb6256c18c78974ebb3430e7da988516cd95b2bd6e4"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30ed2788a6ec68743e2040ab1d16573d7d9f6e7333e45070ce9268cbc93d148c"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21e61a536ffed84d10376c21c13a6ed1ebefb61989a844952547c229d6aeedf3"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0665412dce26b2318092a33bd2d2327d487c4490cfcde158d6946d39b1e28d78"}, - {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:64ed1a5ce5e5926727eb0f87c698c4d9a7a9f7b0953683a65e9ce2b7cc5f8e91"}, - {file = "pymongo-3.13.0-cp39-cp39-win32.whl", hash = "sha256:7593cb1214185a0c5b43b96effc51ce82ddc933298ee36db7dc2bd45d61b4adc"}, - {file = "pymongo-3.13.0-cp39-cp39-win_amd64.whl", hash = "sha256:3cfc9bc1e8b5667bc1f3dbe46d2f85b3f24ff7533893bdc1203058012db2c046"}, - {file = "pymongo-3.13.0.tar.gz", hash = "sha256:e22d6cf5802cd09b674c307cc9e03870b8c37c503ebec3d25b86f2ce8c535dc7"}, -] - -[package.dependencies] -dnspython = {version = ">=1.16.0,<1.17.0", optional = true, markers = "extra == \"srv\""} - -[package.extras] -aws = ["pymongo-auth-aws (<2.0.0)"] -encryption = ["pymongocrypt (>=1.1.0,<2.0.0)"] -gssapi = ["pykerberos"] -ocsp = ["certifi", "pyopenssl (>=17.2.0)", "requests (<3.0.0)", "service-identity (>=18.1.0)"] -snappy = ["python-snappy"] -srv = ["dnspython (>=1.16.0,<1.17.0)"] -tls = ["ipaddress"] -zstd = ["zstandard"] - -[[package]] -name = "pyparsing" -version = "3.0.9" -description = "pyparsing module - Classes and methods to define and execute parsing grammars" -category = "dev" -optional = false -python-versions = ">=3.6.8" -files = [ - {file = "pyparsing-3.0.9-py3-none-any.whl", hash = "sha256:5026bae9a10eeaefb61dab2f09052b9f4307d44aee4eda64b309723d8d206bbc"}, - {file = "pyparsing-3.0.9.tar.gz", hash = "sha256:2b020ecf7d21b687f219b71ecad3631f644a47f01403fa1d1036b0c6416d70fb"}, -] - -[package.extras] -diagrams = ["jinja2", "railroad-diagrams"] - -[[package]] -name = "pytest" -version = "7.2.2" -description = "pytest: simple powerful testing with Python" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "pytest-7.2.2-py3-none-any.whl", hash = "sha256:130328f552dcfac0b1cec75c12e3f005619dc5f874f0a06e8ff7263f0ee6225e"}, - {file = "pytest-7.2.2.tar.gz", hash = "sha256:c99ab0c73aceb050f68929bc93af19ab6db0558791c6a0715723abe9d0ade9d4"}, -] - -[package.dependencies] -attrs = ">=19.2.0" -colorama = {version = "*", markers = "sys_platform == \"win32\""} -exceptiongroup = {version = ">=1.0.0rc8", markers = "python_version < \"3.11\""} -iniconfig = "*" -packaging = "*" -pluggy = ">=0.12,<2.0" -tomli = {version = ">=1.0.0", markers = "python_version < \"3.11\""} - -[package.extras] -testing = ["argcomplete", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.7.2)", "requests", "xmlschema"] - -[[package]] -name = "pytest-cov" -version = "2.12.1" -description = "Pytest plugin for measuring coverage." -category = "dev" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" -files = [ - {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"}, - {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"}, -] - -[package.dependencies] -coverage = ">=5.2.1" -pytest = ">=4.6" -toml = "*" - -[package.extras] -testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"] - -[[package]] -name = "python-dotenv" -version = "1.0.0" -description = "Read key-value pairs from a .env file and set them as environment variables" -category = "main" -optional = false -python-versions = ">=3.8" -files = [ - {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"}, - {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"}, -] - -[package.extras] -cli = ["click (>=5.0)"] - -[[package]] -name = "pytz" -version = "2020.5" -description = "World timezone definitions, modern and historical" -category = "main" -optional = false -python-versions = "*" -files = [ - {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"}, - {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"}, -] - -[[package]] -name = "pyyaml" -version = "6.0" -description = "YAML parser and emitter for Python" -category = "main" -optional = false -python-versions = ">=3.6" -files = [ - {file = "PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d4db7c7aef085872ef65a8fd7d6d09a14ae91f691dec3e87ee5ee0539d516f53"}, - {file = "PyYAML-6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9df7ed3b3d2e0ecfe09e14741b857df43adb5a3ddadc919a2d94fbdf78fea53c"}, - {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77f396e6ef4c73fdc33a9157446466f1cff553d979bd00ecb64385760c6babdc"}, - {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a80a78046a72361de73f8f395f1f1e49f956c6be882eed58505a15f3e430962b"}, - {file = "PyYAML-6.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f84fbc98b019fef2ee9a1cb3ce93e3187a6df0b2538a651bfb890254ba9f90b5"}, - {file = "PyYAML-6.0-cp310-cp310-win32.whl", hash = "sha256:2cd5df3de48857ed0544b34e2d40e9fac445930039f3cfe4bcc592a1f836d513"}, - {file = "PyYAML-6.0-cp310-cp310-win_amd64.whl", hash = "sha256:daf496c58a8c52083df09b80c860005194014c3698698d1a57cbcfa182142a3a"}, - {file = "PyYAML-6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d4b0ba9512519522b118090257be113b9468d804b19d63c71dbcf4a48fa32358"}, - {file = "PyYAML-6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:81957921f441d50af23654aa6c5e5eaf9b06aba7f0a19c18a538dc7ef291c5a1"}, - {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:afa17f5bc4d1b10afd4466fd3a44dc0e245382deca5b3c353d8b757f9e3ecb8d"}, - {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dbad0e9d368bb989f4515da330b88a057617d16b6a8245084f1b05400f24609f"}, - {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:432557aa2c09802be39460360ddffd48156e30721f5e8d917f01d31694216782"}, - {file = "PyYAML-6.0-cp311-cp311-win32.whl", hash = "sha256:bfaef573a63ba8923503d27530362590ff4f576c626d86a9fed95822a8255fd7"}, - {file = "PyYAML-6.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b45c0191e6d66c470b6cf1b9531a771a83c1c4208272ead47a3ae4f2f603bf"}, - {file = "PyYAML-6.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:897b80890765f037df3403d22bab41627ca8811ae55e9a722fd0392850ec4d86"}, - {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50602afada6d6cbfad699b0c7bb50d5ccffa7e46a3d738092afddc1f9758427f"}, - {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:48c346915c114f5fdb3ead70312bd042a953a8ce5c7106d5bfb1a5254e47da92"}, - {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:98c4d36e99714e55cfbaaee6dd5badbc9a1ec339ebfc3b1f52e293aee6bb71a4"}, - {file = "PyYAML-6.0-cp36-cp36m-win32.whl", hash = "sha256:0283c35a6a9fbf047493e3a0ce8d79ef5030852c51e9d911a27badfde0605293"}, - {file = "PyYAML-6.0-cp36-cp36m-win_amd64.whl", hash = "sha256:07751360502caac1c067a8132d150cf3d61339af5691fe9e87803040dbc5db57"}, - {file = "PyYAML-6.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:819b3830a1543db06c4d4b865e70ded25be52a2e0631ccd2f6a47a2822f2fd7c"}, - {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:473f9edb243cb1935ab5a084eb238d842fb8f404ed2193a915d1784b5a6b5fc0"}, - {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0ce82d761c532fe4ec3f87fc45688bdd3a4c1dc5e0b4a19814b9009a29baefd4"}, - {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:231710d57adfd809ef5d34183b8ed1eeae3f76459c18fb4a0b373ad56bedcdd9"}, - {file = "PyYAML-6.0-cp37-cp37m-win32.whl", hash = "sha256:c5687b8d43cf58545ade1fe3e055f70eac7a5a1a0bf42824308d868289a95737"}, - {file = "PyYAML-6.0-cp37-cp37m-win_amd64.whl", hash = "sha256:d15a181d1ecd0d4270dc32edb46f7cb7733c7c508857278d3d378d14d606db2d"}, - {file = "PyYAML-6.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b4624f379dab24d3725ffde76559cff63d9ec94e1736b556dacdfebe5ab6d4b"}, - {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:213c60cd50106436cc818accf5baa1aba61c0189ff610f64f4a3e8c6726218ba"}, - {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9fa600030013c4de8165339db93d182b9431076eb98eb40ee068700c9c813e34"}, - {file = "PyYAML-6.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:277a0ef2981ca40581a47093e9e2d13b3f1fbbeffae064c1d21bfceba2030287"}, - {file = "PyYAML-6.0-cp38-cp38-win32.whl", hash = "sha256:d4eccecf9adf6fbcc6861a38015c2a64f38b9d94838ac1810a9023a0609e1b78"}, - {file = "PyYAML-6.0-cp38-cp38-win_amd64.whl", hash = "sha256:1e4747bc279b4f613a09eb64bba2ba602d8a6664c6ce6396a4d0cd413a50ce07"}, - {file = "PyYAML-6.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:055d937d65826939cb044fc8c9b08889e8c743fdc6a32b33e2390f66013e449b"}, - {file = "PyYAML-6.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e61ceaab6f49fb8bdfaa0f92c4b57bcfbea54c09277b1b4f7ac376bfb7a7c174"}, - {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d67d839ede4ed1b28a4e8909735fc992a923cdb84e618544973d7dfc71540803"}, - {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba8c411ef271aa037d7357a2bc8f9ee8b58b9965831d9e51baf703280dc73d3"}, - {file = "PyYAML-6.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:40527857252b61eacd1d9af500c3337ba8deb8fc298940291486c465c8b46ec0"}, - {file = "PyYAML-6.0-cp39-cp39-win32.whl", hash = "sha256:b5b9eccad747aabaaffbc6064800670f0c297e52c12754eb1d976c57e4f74dcb"}, - {file = "PyYAML-6.0-cp39-cp39-win_amd64.whl", hash = "sha256:b3d267842bf12586ba6c734f89d1f5b871df0273157918b0ccefa29deb05c21c"}, - {file = "PyYAML-6.0.tar.gz", hash = "sha256:68fb519c14306fec9720a2a5b45bc9f0c8d1b9c72adf45c37baedfcd949c35a2"}, -] - -[[package]] -name = "requests" -version = "2.28.2" -description = "Python HTTP for Humans." -category = "main" -optional = false -python-versions = ">=3.7, <4" -files = [ - {file = "requests-2.28.2-py3-none-any.whl", hash = "sha256:64299f4909223da747622c030b781c0d7811e359c37124b4bd368fb8c6518baa"}, - {file = "requests-2.28.2.tar.gz", hash = "sha256:98b1b2782e3c6c4904938b84c0eb932721069dfdb9134313beff7c83c2df24bf"}, -] - -[package.dependencies] -certifi = ">=2017.4.17" -charset-normalizer = ">=2,<4" -idna = ">=2.5,<4" -urllib3 = ">=1.21.1,<1.27" - -[package.extras] -socks = ["PySocks (>=1.5.6,!=1.5.7)"] -use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] - -[[package]] -name = "resolvelib" -version = "0.9.0" -description = "Resolve abstract dependencies into concrete ones" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"}, - {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"}, -] - -[package.extras] -examples = ["html5lib", "packaging", "pygraphviz", "requests"] -lint = ["black", "flake8", "isort", "mypy", "types-requests"] -release = ["build", "towncrier", "twine"] -test = ["commentjson", "packaging", "pytest"] - -[[package]] -name = "rich" -version = "13.3.2" -description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal" -category = "dev" -optional = false -python-versions = ">=3.7.0" -files = [ - {file = "rich-13.3.2-py3-none-any.whl", hash = "sha256:a104f37270bf677148d8acb07d33be1569eeee87e2d1beb286a4e9113caf6f2f"}, - {file = "rich-13.3.2.tar.gz", hash = "sha256:91954fe80cfb7985727a467ca98a7618e5dd15178cc2da10f553b36a93859001"}, -] - -[package.dependencies] -markdown-it-py = ">=2.2.0,<3.0.0" -pygments = ">=2.13.0,<3.0.0" - -[package.extras] -jupyter = ["ipywidgets (>=7.5.1,<9)"] - -[[package]] -name = "setuptools" -version = "67.5.1" -description = "Easily download, build, install, upgrade, and uninstall Python packages" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "setuptools-67.5.1-py3-none-any.whl", hash = "sha256:1c39d42bda4cb89f7fdcad52b6762e3c309ec8f8715b27c684176b7d71283242"}, - {file = "setuptools-67.5.1.tar.gz", hash = "sha256:15136a251127da2d2e77ac7a1bc231eb504654f7e3346d93613a13f2e2787535"}, -] - -[package.extras] -docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-hoverxref (<2)", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (==0.8.3)", "sphinx-reredirects", "sphinxcontrib-towncrier"] -testing = ["build[virtualenv]", "filelock (>=3.4.0)", "flake8 (<5)", "flake8-2020", "ini2toml[lite] (>=0.9)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pip (>=19.1)", "pip-run (>=8.8)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)", "pytest-perf", "pytest-timeout", "pytest-xdist", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel"] -testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pytest", "pytest-enabler", "pytest-xdist", "tomli", "virtualenv (>=13.0.0)", "wheel"] - -[[package]] -name = "six" -version = "1.16.0" -description = "Python 2 and 3 compatibility utilities" -category = "dev" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*" -files = [ - {file = "six-1.16.0-py2.py3-none-any.whl", hash = "sha256:8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254"}, - {file = "six-1.16.0.tar.gz", hash = "sha256:1e61c37477a1626458e36f7b1d82aa5c9b094fa4802892072e49de9c60c4c926"}, -] - -[[package]] -name = "smmap" -version = "5.0.0" -description = "A pure Python implementation of a sliding window memory map manager" -category = "dev" -optional = false -python-versions = ">=3.6" -files = [ - {file = "smmap-5.0.0-py3-none-any.whl", hash = "sha256:2aba19d6a040e78d8b09de5c57e96207b09ed71d8e55ce0959eeee6c8e190d94"}, - {file = "smmap-5.0.0.tar.gz", hash = "sha256:c840e62059cd3be204b0c9c9f74be2c09d5648eddd4580d9314c3ecde0b30936"}, -] - -[[package]] -name = "sortedcontainers" -version = "2.4.0" -description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"}, - {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"}, -] - -[[package]] -name = "stevedore" -version = "5.0.0" -description = "Manage dynamic plugins for Python applications" -category = "dev" -optional = false -python-versions = ">=3.8" -files = [ - {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"}, - {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"}, -] - -[package.dependencies] -pbr = ">=2.0.0,<2.1.0 || >2.1.0" - -[[package]] -name = "toml" -version = "0.10.2" -description = "Python Library for Tom's Obvious, Minimal Language" -category = "dev" -optional = false -python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*" -files = [ - {file = "toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b"}, - {file = "toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f"}, -] - -[[package]] -name = "tomli" -version = "2.0.1" -description = "A lil' TOML parser" -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "tomli-2.0.1-py3-none-any.whl", hash = "sha256:939de3e7a6161af0c887ef91b7d41a53e7c5a1ca976325f429cb46ea9bc30ecc"}, - {file = "tomli-2.0.1.tar.gz", hash = "sha256:de526c12914f0c550d15924c62d72abc48d6fe7364aa87328337a31007fe8a4f"}, -] - -[[package]] -name = "tqdm" -version = "4.65.0" -description = "Fast, Extensible Progress Meter" -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"}, - {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"}, -] - -[package.dependencies] -colorama = {version = "*", markers = "platform_system == \"Windows\""} - -[package.extras] -dev = ["py-make (>=0.1.0)", "twine", "wheel"] -notebook = ["ipywidgets (>=6)"] -slack = ["slack-sdk"] -telegram = ["requests"] - -[[package]] -name = "typing-extensions" -version = "4.5.0" -description = "Backported and Experimental Type Hints for Python 3.7+" -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"}, - {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"}, -] - -[[package]] -name = "urllib3" -version = "1.26.14" -description = "HTTP library with thread-safe connection pooling, file post, and more." -category = "main" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*, !=3.5.*" -files = [ - {file = "urllib3-1.26.14-py2.py3-none-any.whl", hash = "sha256:75edcdc2f7d85b137124a6c3c9fc3933cdeaa12ecb9a6a959f22797a0feca7e1"}, - {file = "urllib3-1.26.14.tar.gz", hash = "sha256:076907bf8fd355cde77728471316625a4d2f7e713c125f51953bb5b3eecf4f72"}, -] - -[package.extras] -brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)", "brotlipy (>=0.6.0)"] -secure = ["certifi", "cryptography (>=1.3.4)", "idna (>=2.0.0)", "ipaddress", "pyOpenSSL (>=0.14)", "urllib3-secure-extra"] -socks = ["PySocks (>=1.5.6,!=1.5.7,<2.0)"] - -[[package]] -name = "webencodings" -version = "0.5.1" -description = "Character encoding aliases for legacy web content" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78"}, - {file = "webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923"}, -] - -[metadata] -lock-version = "2.0" -python-versions = "3.9.15" -content-hash = "67ae1eb17c3921f944b323ddcaa58919c1deff1e7cc25e63cf3a82ec23340c62" diff --git a/jobs/cache_refresh/poetry.toml b/jobs/cache_refresh/poetry.toml deleted file mode 100644 index 5fcef8cd..00000000 --- a/jobs/cache_refresh/poetry.toml +++ /dev/null @@ -1,3 +0,0 @@ -[virtualenvs] -in-project = true -prefer-active-python = true diff --git a/jobs/cache_refresh/pyproject.toml b/jobs/cache_refresh/pyproject.toml deleted file mode 100644 index 17fcdbbf..00000000 --- a/jobs/cache_refresh/pyproject.toml +++ /dev/null @@ -1,44 +0,0 @@ -[tool.poetry] -authors = ["Andrea Soria <[email protected]>"] -description = "Refresh outdated cache entries based on job runner version" -name = "cache_refresh" -version = "0.0.1" -license = "Apache-2.0" - -[tool.poetry.dependencies] -environs = "^9.5.0" -libcommon = {path = "../../libs/libcommon", develop = true} -python = "3.9.15" - -[tool.poetry.group.dev.dependencies] -bandit = "^1.7.4" -black = "^22.12.0" -flake8 = "^3.9.2" -isort = "^5.12.0" -mypy = "^1.0.0" -pip-audit = "^2.4.14" -pytest = "^7.2.1" -pytest-cov = "^2.12.1" - -[build-system] -build-backend = "poetry.core.masonry.api" -requires = ["poetry-core>=1.0.0"] - -[tool.pytest.ini_options] -filterwarnings = ["ignore::DeprecationWarning"] -markers = [ - "wip: tests being developed" -] - -[tool.coverage.run] -source = ["cache_refresh"] - -[tool.isort] -profile = "black" - -[tool.black] -line-length = 119 -preview = true - -[tool.mypy] -strict = true diff --git a/jobs/cache_refresh/src/cache_refresh/__init__.py b/jobs/cache_refresh/src/cache_refresh/__init__.py deleted file mode 100644 index 1e9d0c5a..00000000 --- a/jobs/cache_refresh/src/cache_refresh/__init__.py +++ /dev/null @@ -1,2 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. diff --git a/jobs/cache_refresh/src/cache_refresh/config.py b/jobs/cache_refresh/src/cache_refresh/config.py deleted file mode 100644 index 27dc2944..00000000 --- a/jobs/cache_refresh/src/cache_refresh/config.py +++ /dev/null @@ -1,21 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from dataclasses import dataclass, field - -from libcommon.config import CacheConfig, LogConfig, QueueConfig - - -@dataclass(frozen=True) -class JobConfig: - cache: CacheConfig = field(default_factory=CacheConfig) - log: LogConfig = field(default_factory=LogConfig) - queue: QueueConfig = field(default_factory=QueueConfig) - - @classmethod - def from_env(cls) -> "JobConfig": - return cls( - cache=CacheConfig.from_env(), - log=LogConfig.from_env(), - queue=QueueConfig.from_env(), - ) diff --git a/jobs/cache_refresh/src/cache_refresh/main.py b/jobs/cache_refresh/src/cache_refresh/main.py deleted file mode 100644 index 1a181227..00000000 --- a/jobs/cache_refresh/src/cache_refresh/main.py +++ /dev/null @@ -1,52 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import logging -import sys - -from libcommon.config import ProcessingGraphConfig -from libcommon.log import init_logging -from libcommon.processing_graph import ProcessingGraph -from libcommon.resources import CacheMongoResource, QueueMongoResource - -from cache_refresh.config import JobConfig -from cache_refresh.outdated_cache import refresh_cache - - -def run_job() -> None: - job_config = JobConfig.from_env() - - init_logging(level=job_config.log.level) - # ^ set first to have logs as soon as possible - - with ( - CacheMongoResource( - database=job_config.cache.mongo_database, host=job_config.cache.mongo_url - ) as cache_resource, - QueueMongoResource( - database=job_config.queue.mongo_database, host=job_config.queue.mongo_url - ) as queue_resource, - ): - if not cache_resource.is_available(): - logging.warning( - "The connection to the cache database could not be established. The cache refresh job is skipped." - ) - return - if not queue_resource.is_available(): - logging.warning( - "The connection to the queue database could not be established. The cache refresh job is skipped." - ) - return - - processing_graph_config = ProcessingGraphConfig.from_env() - processing_graph = ProcessingGraph(processing_graph_config.specification) - processing_steps = list(processing_graph.steps.values()) - refresh_cache(processing_steps) - - -if __name__ == "__main__": - try: - run_job() - sys.exit(0) - except Exception: - sys.exit(1) diff --git a/jobs/cache_refresh/tests/__init__.py b/jobs/cache_refresh/tests/__init__.py deleted file mode 100644 index 1e9d0c5a..00000000 --- a/jobs/cache_refresh/tests/__init__.py +++ /dev/null @@ -1,2 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. diff --git a/jobs/cache_refresh/tests/conftest.py b/jobs/cache_refresh/tests/conftest.py deleted file mode 100644 index 793b4e18..00000000 --- a/jobs/cache_refresh/tests/conftest.py +++ /dev/null @@ -1,43 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from typing import Iterator - -from libcommon.queue import _clean_queue_database -from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import _clean_cache_database -from pytest import MonkeyPatch, fixture - -from cache_refresh.config import JobConfig - - -# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200 -@fixture(scope="session") -def monkeypatch_session() -> Iterator[MonkeyPatch]: - monkeypatch_session = MonkeyPatch() - monkeypatch_session.setenv("CACHE_MONGO_DATABASE", "datasets_server_cache_test") - monkeypatch_session.setenv("QUEUE_MONGO_DATABASE", "datasets_server_queue_test") - yield monkeypatch_session - monkeypatch_session.undo() - - -@fixture(scope="session") -def job_config(monkeypatch_session: MonkeyPatch) -> JobConfig: - job_config = JobConfig.from_env() - if "test" not in job_config.cache.mongo_database or "test" not in job_config.queue.mongo_database: - raise ValueError("Test must be launched on a test mongo database") - return job_config - - -@fixture(autouse=True) -def cache_mongo_resource(job_config: JobConfig) -> Iterator[CacheMongoResource]: - with CacheMongoResource(database=job_config.cache.mongo_database, host=job_config.cache.mongo_url) as resource: - yield resource - _clean_cache_database() - - -@fixture(autouse=True) -def queue_mongo_resource(job_config: JobConfig) -> Iterator[QueueMongoResource]: - with QueueMongoResource(database=job_config.queue.mongo_database, host=job_config.queue.mongo_url) as resource: - yield resource - _clean_queue_database()
63e69ea939f5a329ba0d84daaa729a305e29edb6
Sylvain Lesage
2023-04-07T12:27:33
fix: 🐛 reduce the k8s job TTL to 5 minutes (#1036)
diff --git a/chart/templates/jobs/cache-maintenance/job.yaml b/chart/templates/jobs/cache-maintenance/job.yaml index eb4b793d..6cd51f57 100644 --- a/chart/templates/jobs/cache-maintenance/job.yaml +++ b/chart/templates/jobs/cache-maintenance/job.yaml @@ -14 +14 @@ spec: - ttlSecondsAfterFinished: 86400 + ttlSecondsAfterFinished: 300
c5b3b0570d7ea0876660659533fb04a4d876babd
Sylvain Lesage
2023-04-07T12:27:02
Create step dataset-is-valid (#1032)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index ac01cd8a..86bbf204 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -16,0 +17 @@ from libcommon.constants import ( + PROCESSING_STEP_DATASET_IS_VALID_VERSION, @@ -199,0 +201,9 @@ class ProcessingGraphConfig: + "dataset-is-valid": { + "input_type": "dataset", + "requires": [ + "dataset-split-names", + "split-first-rows-from-parquet", + "split-first-rows-from-streaming", + ], + "job_runner_version": PROCESSING_STEP_DATASET_IS_VALID_VERSION, + }, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index bd3c1ca3..db3c31cb 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -13,0 +14 @@ PROCESSING_STEP_DATASET_INFO_VERSION = 1 +PROCESSING_STEP_DATASET_IS_VALID_VERSION = 1 diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index 9b592ce3..28024da6 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -300 +300 @@ def get_valid_datasets(kind: str) -> Set[str]: -def get_validity_by_kind(dataset: str) -> Mapping[str, bool]: +def get_validity_by_kind(dataset: str, kinds: Optional[List[str]] = None) -> Mapping[str, bool]: @@ -302 +302,5 @@ def get_validity_by_kind(dataset: str) -> Mapping[str, bool]: - entries = CachedResponse.objects(dataset=dataset).only("kind", "http_status") + entries = ( + CachedResponse.objects(dataset=dataset) + if kinds is None + else CachedResponse.objects(dataset=dataset, kind__in=kinds) + ).only("kind", "http_status") diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 43890d48..59c7cb84 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -87 +87 @@ def graph() -> ProcessingGraph: - [], + ["dataset-is-valid"], @@ -96 +96 @@ def graph() -> ProcessingGraph: - ("split-first-rows-from-parquet", [], ["config-parquet", "/parquet-and-dataset-info"]), + ("split-first-rows-from-parquet", ["dataset-is-valid"], ["config-parquet", "/parquet-and-dataset-info"]), @@ -99 +99 @@ def graph() -> ProcessingGraph: - [], + ["dataset-is-valid"], @@ -114,0 +115,15 @@ def graph() -> ProcessingGraph: + ( + "dataset-is-valid", + [], + [ + "/config-names", + "/parquet-and-dataset-info", + "dataset-split-names", + "config-info", + "config-parquet", + "/split-names-from-dataset-info", + "/split-names-from-streaming", + "split-first-rows-from-parquet", + "split-first-rows-from-streaming", + ], + ), diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index 3445d65d..6b921123 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -318,0 +319 @@ def test_get_validity_by_kind_two_valid_datasets() -> None: + other_kind = "other_kind" @@ -324,0 +326,3 @@ def test_get_validity_by_kind_two_valid_datasets() -> None: + assert get_validity_by_kind(dataset=dataset_b, kinds=[kind]) == {kind: True} + assert not get_validity_by_kind(dataset=dataset_b, kinds=[other_kind]) + assert get_validity_by_kind(dataset=dataset_b, kinds=[kind, other_kind]) == {kind: True} diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index e85ae818..b035dba8 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -24,0 +25 @@ from worker.job_runners.dataset.info import DatasetInfoJobRunner +from worker.job_runners.dataset.is_valid import DatasetIsValidJobRunner @@ -187,0 +189,7 @@ class JobRunnerFactory(BaseJobRunnerFactory): + if job_type == DatasetIsValidJobRunner.get_job_type(): + return DatasetIsValidJobRunner( + job_info=job_info, + processing_step=processing_step, + common_config=self.app_config.common, + worker_config=self.app_config.worker, + ) @@ -202,0 +211 @@ class JobRunnerFactory(BaseJobRunnerFactory): + DatasetIsValidJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/dataset/is_valid.py b/services/worker/src/worker/job_runners/dataset/is_valid.py new file mode 100644 index 00000000..049e3844 --- /dev/null +++ b/services/worker/src/worker/job_runners/dataset/is_valid.py @@ -0,0 +1,73 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from typing import Any, Literal, Mapping, Tuple, TypedDict + +from libcommon.constants import PROCESSING_STEP_DATASET_IS_VALID_VERSION +from libcommon.simple_cache import SplitFullName, get_validity_by_kind + +from worker.job_runner import JobResult, JobRunner + +IsValidJobRunnerErrorCode = Literal[ + "PreviousStepStatusError", + "PreviousStepFormatError", +] + + +class DatasetIsValidResponse(TypedDict): + valid: bool + + +SPLIT_KINDS = [ + "/splits", + "/split-names-from-streaming", + "/split-names-from-dataset-info", +] +FIRST_ROWS_KINDS = ["split-first-rows-from-streaming", "split-first-rows-from-parquet"] + + +def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, float]: + """ + Get the response of /is-valid for one specific dataset on huggingface.co. + + A dataset is valid if: + - /splits is valid for at least one of the configs + - /first-rows is valid for at least one of the splits + + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + Returns: + `DatasetIsValidResponse`: An object with the is_valid_response. + """ + logging.info(f"get is-valid response for dataset={dataset}") + + validity_by_kind = get_validity_by_kind(dataset=dataset, kinds=SPLIT_KINDS + FIRST_ROWS_KINDS) + is_valid = any(validity_by_kind[kind] for kind in SPLIT_KINDS if kind in validity_by_kind) and any( + validity_by_kind[kind] for kind in FIRST_ROWS_KINDS if kind in validity_by_kind + ) + + return (DatasetIsValidResponse({"valid": is_valid}), 1.0) + + +class DatasetIsValidJobRunner(JobRunner): + @staticmethod + def get_job_type() -> str: + return "dataset-is-valid" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_DATASET_IS_VALID_VERSION + + def compute(self) -> JobResult: + if self.dataset is None: + raise ValueError("dataset is required") + response_content, progress = compute_is_valid_response(dataset=self.dataset) + return JobResult(response_content, progress=progress) + + def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by the compute.""" + return set() + # ^ it does not make sense to depend on this step. Not sure what should be returned here. diff --git a/services/worker/tests/job_runners/dataset/test_is_valid.py b/services/worker/tests/job_runners/dataset/test_is_valid.py new file mode 100644 index 00000000..b1abc270 --- /dev/null +++ b/services/worker/tests/job_runners/dataset/test_is_valid.py @@ -0,0 +1,186 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Callable, List + +import pytest +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response + +from worker.config import AppConfig +from worker.job_runners.dataset.is_valid import DatasetIsValidJobRunner + +from ..utils import UpstreamResponse + + [email protected](autouse=True) +def prepare_and_clean_mongo(app_config: AppConfig) -> None: + # prepare the database before each test, and clean it afterwards + pass + + +GetJobRunner = Callable[[str, AppConfig, bool], DatasetIsValidJobRunner] + + +UPSTREAM_RESPONSE_SPLITS: UpstreamResponse = UpstreamResponse( + kind="/splits", dataset="dataset_ok", config=None, http_status=HTTPStatus.OK, content={} +) +UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING: UpstreamResponse = UpstreamResponse( + kind="/split-names-from-streaming", dataset="dataset_ok", config=None, http_status=HTTPStatus.OK, content={} +) +UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO: UpstreamResponse = UpstreamResponse( + kind="/split-names-from-dataset-info", dataset="dataset_ok", config=None, http_status=HTTPStatus.OK, content={} +) +UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING: UpstreamResponse = UpstreamResponse( + kind="split-first-rows-from-streaming", + dataset="dataset_ok", + config="config", + http_status=HTTPStatus.OK, + content={}, +) +UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET: UpstreamResponse = UpstreamResponse( + kind="split-first-rows-from-parquet", dataset="dataset_ok", config="config", http_status=HTTPStatus.OK, content={} +) +UPSTREAM_RESPONSE_SPLITS_ERROR: UpstreamResponse = UpstreamResponse( + kind="/splits", dataset="dataset_ok", config=None, http_status=HTTPStatus.INTERNAL_SERVER_ERROR, content={} +) +UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR: UpstreamResponse = UpstreamResponse( + kind="split-first-rows-from-streaming", + dataset="dataset_ok", + config="config", + http_status=HTTPStatus.INTERNAL_SERVER_ERROR, + content={}, +) +EXPECTED_OK = ( + {"valid": True}, + 1.0, +) +EXPECTED_ERROR = ( + {"valid": False}, + 1.0, +) + + [email protected] +def get_job_runner( + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + app_config: AppConfig, + force: bool = False, + ) -> DatasetIsValidJobRunner: + return DatasetIsValidJobRunner( + job_info={ + "type": DatasetIsValidJobRunner.get_job_type(), + "dataset": dataset, + "config": None, + "split": None, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + common_config=app_config.common, + worker_config=app_config.worker, + processing_step=ProcessingStep( + name=DatasetIsValidJobRunner.get_job_type(), + input_type="dataset", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + job_runner_version=DatasetIsValidJobRunner.get_job_runner_version(), + ), + ) + + return _get_job_runner + + [email protected]( + "dataset,upstream_responses,expected_error_code,expected,should_raise", + [ + ( + "dataset_ok", + [ + UPSTREAM_RESPONSE_SPLITS, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING, + ], + None, + EXPECTED_OK, + False, + ), + ( + "dataset_ok", + [ + UPSTREAM_RESPONSE_SPLITS, + UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING, + UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET, + ], + None, + EXPECTED_OK, + False, + ), + ("dataset_ok", [], None, EXPECTED_ERROR, False), + ("dataset_ok", [UPSTREAM_RESPONSE_SPLITS], None, EXPECTED_ERROR, False), + ("dataset_ok", [UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING], None, EXPECTED_ERROR, False), + ( + "dataset_ok", + [UPSTREAM_RESPONSE_SPLITS_ERROR, UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING], + None, + EXPECTED_ERROR, + False, + ), + ( + "dataset_ok", + [UPSTREAM_RESPONSE_SPLITS, UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR], + None, + EXPECTED_ERROR, + False, + ), + ( + "dataset_ok", + [ + UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING, + UPSTREAM_RESPONSE_SPLITS_ERROR, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR, + ], + None, + EXPECTED_OK, + False, + ), + ], +) +def test_compute( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset: str, + upstream_responses: List[UpstreamResponse], + expected_error_code: str, + expected: Any, + should_raise: bool, +) -> None: + for upstream_response in upstream_responses: + upsert_response(**upstream_response) + job_runner = get_job_runner(dataset, app_config, False) + if should_raise: + with pytest.raises(Exception) as e: + job_runner.compute() + assert e.type.__name__ == expected_error_code + else: + compute_result = job_runner.compute() + assert compute_result.content == expected[0] + assert compute_result.progress == expected[1] + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = "doesnotexist" + job_runner = get_job_runner(dataset, app_config, False) + compute_result = job_runner.compute() + assert compute_result.content == {"valid": False}
d1ab91e178a3d782ea4c650f0614d90e2d3a23c6
Sylvain Lesage
2023-04-07T11:28:58
fix: 🐛 enable the two last migrations (#1034)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 053469d8..02cf9a88 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -43,0 +44,6 @@ from mongodb_migration.migrations._20230323160000_queue_dataset_info import ( +from mongodb_migration.migrations._20230407091400_queue_delete_splits import ( + MigrationQueueDeleteSplits, +) +from mongodb_migration.migrations._20230407091500_cache_delete_splits import ( + MigrationCacheDeleteSplits, +) @@ -99,0 +106,8 @@ class MigrationsCollector: + MigrationQueueDeleteSplits( + version="20230407091400", + description="delete the jobs of type '/splits'", + ), + MigrationCacheDeleteSplits( + version="20230407091500", + description="delete the cache entries of kind '/splits'", + ),
fc169202811648e4c53bfca558a064e63f58d6aa
Sylvain Lesage
2023-04-07T09:26:47
Remove /splits (#1033)
diff --git a/e2e/tests/test_31_admin_metrics.py b/e2e/tests/test_31_admin_metrics.py index c3ef9de7..65a3cd6c 100644 --- a/e2e/tests/test_31_admin_metrics.py +++ b/e2e/tests/test_31_admin_metrics.py @@ -33 +33 @@ def test_metrics() -> None: - for queue in ["/splits", "split-first-rows-from-streaming", "dataset-parquet"]: + for queue in ["/config-names", "split-first-rows-from-streaming", "dataset-parquet"]: @@ -40 +40 @@ def test_metrics() -> None: - for cache_kind in ["/splits", "split-first-rows-from-streaming", "dataset-parquet"]: + for cache_kind in ["/config-names", "split-first-rows-from-streaming", "dataset-parquet"]: @@ -42 +42 @@ def test_metrics() -> None: - # eg. 'responses_in_cache_total{error_code="None",http_status="200",path="/splits",pid="10"}' + # eg. 'responses_in_cache_total{error_code="None",http_status="200",path="/config-names",pid="10"}' diff --git a/jobs/mongodb_migration/src/mongodb_migration/migration.py b/jobs/mongodb_migration/src/mongodb_migration/migration.py index 5ac444f5..61b95ad4 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migration.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migration.py @@ -7 +7 @@ from abc import ABC, abstractmethod -class IrreversibleMigration(Exception): +class IrreversibleMigrationError(Exception): @@ -32 +32 @@ class Migration(ABC): - raise IrreversibleMigration() + raise IrreversibleMigrationError() diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py index bf25e6b2..2f261957 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py @@ -10 +10 @@ from mongodb_migration.check import check_documents -from mongodb_migration.migration import IrreversibleMigration, Migration +from mongodb_migration.migration import IrreversibleMigrationError, Migration @@ -21 +21 @@ class MigrationRemoveWorkerVersionFromCachedResponse(Migration): - raise IrreversibleMigration("This migration does not support rollback") + raise IrreversibleMigrationError("This migration does not support rollback") diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091400_queue_delete_splits.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091400_queue_delete_splits.py new file mode 100644 index 00000000..e56ac03c --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091400_queue_delete_splits.py @@ -0,0 +1,27 @@ +import logging + +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +job_type = "/splits" +db_name = "queue" + + +# connection already occurred in the main.py (caveat: we use globals) +class MigrationQueueDeleteSplits(Migration): + def up(self) -> None: + logging.info(f"Delete jobs of type {job_type}") + + db = get_db(db_name) + db["jobsBlue"].delete_many({"type": job_type}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {job_type} type") + + db = get_db(db_name) + if db[db_name].count_documents({"type": job_type}): + raise ValueError(f"Found documents with type {job_type}") diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091500_cache_delete_splits.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091500_cache_delete_splits.py new file mode 100644 index 00000000..19aaa468 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230407091500_cache_delete_splits.py @@ -0,0 +1,30 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from mongoengine.connection import get_db + +from mongodb_migration.migration import IrreversibleMigrationError, Migration + +cache_kind = "/splits" +db_name = "cache" + + +class MigrationCacheDeleteSplits(Migration): + def up(self) -> None: + logging.info(f"Delete cahe entries of kind {cache_kind}") + db = get_db(db_name) + + # update existing documents with the old kind + db["cachedResponsesBlue"].delete_many({"kind": cache_kind}) + + def down(self) -> None: + raise IrreversibleMigrationError("This migration does not support rollback") + + def validate(self) -> None: + logging.info(f"Check that none of the documents has the {cache_kind} kind") + + db = get_db(db_name) + if db[db_name].count_documents({"kind": cache_kind}): + raise ValueError(f"Found documents with kind {cache_kind}") diff --git a/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py index d8078f00..8a94eec7 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230313164200_cache_remove_worker_version.py @@ -8 +8 @@ from pytest import raises -from mongodb_migration.migration import IrreversibleMigration +from mongodb_migration.migration import IrreversibleMigrationError @@ -36 +36 @@ def test_cache_remove_worker_version(mongo_host: str) -> None: - with raises(IrreversibleMigration): + with raises(IrreversibleMigrationError): diff --git a/jobs/mongodb_migration/tests/migrations/test_20230407091400_queue_delete_splits.py b/jobs/mongodb_migration/tests/migrations/test_20230407091400_queue_delete_splits.py new file mode 100644 index 00000000..96b722a6 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230407091400_queue_delete_splits.py @@ -0,0 +1,36 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230407091400_queue_delete_splits import ( + MigrationQueueDeleteSplits, +) + + +def test_queue_remove_splits(mongo_host: str) -> None: + job_type = "/splits" + with MongoResource(database="test_queue_remove_splits", host=mongo_host, mongoengine_alias="queue"): + db = get_db("queue") + db["jobsBlue"].insert_many( + [ + { + "type": job_type, + "unicity_id": f"Job[{job_type}][dataset][config][split]", + "dataset": "dataset", + "http_status": 200, + } + ] + ) + assert db["jobsBlue"].find_one({"type": job_type}) # Ensure there is at least one record to delete + + migration = MigrationQueueDeleteSplits( + version="20230407091400", + description=f"remove jobs of type '{job_type}'", + ) + migration.up() + + assert not db["jobsBlue"].find_one({"type": job_type}) # Ensure 0 records with old type + + db["jobsBlue"].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230407091500_cache_remove_splits.py b/jobs/mongodb_migration/tests/migrations/test_20230407091500_cache_remove_splits.py new file mode 100644 index 00000000..5be77a69 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230407091500_cache_remove_splits.py @@ -0,0 +1,27 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230407091500_cache_delete_splits import ( + MigrationCacheDeleteSplits, +) + + +def test_cache_remove_splits(mongo_host: str) -> None: + kind = "/splits" + with MongoResource(database="test_cache_remove_splits", host=mongo_host, mongoengine_alias="cache"): + db = get_db("cache") + db["cachedResponsesBlue"].insert_many([{"kind": kind, "dataset": "dataset", "http_status": 200}]) + assert db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure there is at least one record to update + + migration = MigrationCacheDeleteSplits( + version="20230407091500", + description=f"remove cache for kind {kind}", + ) + migration.up() + + assert not db["cachedResponsesBlue"].find_one({"kind": kind}) # Ensure 0 records with old kind + + db["cachedResponsesBlue"].drop() diff --git a/jobs/mongodb_migration/tests/test_plan.py b/jobs/mongodb_migration/tests/test_plan.py index 3a0fbc6c..b268d822 100644 --- a/jobs/mongodb_migration/tests/test_plan.py +++ b/jobs/mongodb_migration/tests/test_plan.py @@ -12 +12 @@ from mongodb_migration.database_migrations import ( -from mongodb_migration.migration import IrreversibleMigration, Migration +from mongodb_migration.migration import IrreversibleMigrationError, Migration @@ -76 +76 @@ class MigrationErrorIrreversible(Migration): - raise IrreversibleMigration("Error in down") + raise IrreversibleMigrationError("Error in down") @@ -125 +125 @@ def test_collected_migrations_order_dont_matter(collected_migrations: List[Migra - ([migration_error_irreversible], [migration_error_irreversible], IrreversibleMigration), + ([migration_error_irreversible], [migration_error_irreversible], IrreversibleMigrationError), diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index d52e7277..ac01cd8a 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -27 +26,0 @@ from libcommon.constants import ( - PROCESSING_STEP_SPLITS_VERSION, @@ -136,5 +134,0 @@ class ProcessingGraphConfig: - "/splits": { - "input_type": "dataset", - "required_by_dataset_viewer": True, - "job_runner_version": PROCESSING_STEP_SPLITS_VERSION, - }, # to be deprecated @@ -195 +189 @@ class ProcessingGraphConfig: - }, + }, # to be deprecated @@ -200 +194 @@ class ProcessingGraphConfig: - }, + }, # to be deprecated diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 1ede9ada..bd3c1ca3 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -24 +23,0 @@ PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 1 -PROCESSING_STEP_SPLITS_VERSION = 2 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 8e4f0ee2..43890d48 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -69 +68,0 @@ def graph() -> ProcessingGraph: - ("/splits", [], []), @@ -131 +130 @@ def test_default_graph_first_steps(graph: ProcessingGraph) -> None: - [graph.get_step(step_name) for step_name in {"/config-names", "/splits", "/parquet-and-dataset-info"}], + [graph.get_step(step_name) for step_name in {"/config-names", "/parquet-and-dataset-info"}], @@ -138 +137 @@ def test_default_graph_required_by_dataset_viewer(graph: ProcessingGraph) -> Non - [graph.get_step(step_name) for step_name in {"/splits", "split-first-rows-from-streaming"}], + [graph.get_step(step_name) for step_name in {"split-first-rows-from-streaming"}], diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index ff95be34..96eb6363 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -120 +119,0 @@ class EndpointConfig: - "/splits", diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index d43f0e00..c2964873 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -39 +39 @@ def test_endpoints_definition() -> None: - assert len(splits["dataset"]) == 4 # Has four processing steps + assert len(splits["dataset"]) == 3 # Has four processing steps @@ -143,2 +143,2 @@ def test_get_cache_entry_from_steps() -> None: - queue.upsert_job(job_type="/splits", dataset=dataset, config=config, force=True) - non_existent_step = graph.get_step("/splits") + queue.upsert_job(job_type="dataset-split-names", dataset=dataset, config=config, force=True) + non_existent_step = graph.get_step("dataset-split-names") diff --git a/services/worker/README.md b/services/worker/README.md index dd27e6b2..8ce3ce69 100644 --- a/services/worker/README.md +++ b/services/worker/README.md @@ -20 +20 @@ Set environment variables to configure the worker. -- `WORKER_ONLY_JOB_TYPES`: comma-separated list of the job types to process, e.g. "/splits,/first-rows". If empty, the worker processes all the jobs. Defaults to empty. +- `WORKER_ONLY_JOB_TYPES`: comma-separated list of the job types to process, e.g. "/config-names,dataset-split-names". If empty, the worker processes all the jobs. Defaults to empty. diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 3eb21bba..e85ae818 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -41 +40,0 @@ from worker.job_runners.split.first_rows_from_streaming import ( -from worker.job_runners.splits import SplitsJobRunner @@ -91,7 +89,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - if job_type == SplitsJobRunner.get_job_type(): - return SplitsJobRunner( - job_info=job_info, - app_config=self.app_config, - processing_step=processing_step, - hf_datasets_cache=self.hf_datasets_cache, - ) @@ -199 +190,0 @@ class JobRunnerFactory(BaseJobRunnerFactory): - SplitsJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/splits.py b/services/worker/src/worker/job_runners/splits.py deleted file mode 100644 index cc9d611b..00000000 --- a/services/worker/src/worker/job_runners/splits.py +++ /dev/null @@ -1,142 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import logging -from http import HTTPStatus -from typing import Any, List, Literal, Mapping, Optional, TypedDict, Union - -from datasets import get_dataset_config_names, get_dataset_split_names -from datasets.data_files import EmptyDatasetError as _EmptyDatasetError -from libcommon.constants import PROCESSING_STEP_SPLITS_VERSION -from libcommon.simple_cache import SplitFullName - -from worker.job_runner import CompleteJobResult, JobRunnerError -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner - -SplitsJobRunnerErrorCode = Literal[ - "EmptyDatasetError", - "SplitsNamesError", -] - - -class SplitsJobRunnerError(JobRunnerError): - """Base class for splits job runner exceptions.""" - - def __init__( - self, - message: str, - status_code: HTTPStatus, - code: SplitsJobRunnerErrorCode, - cause: Optional[BaseException] = None, - disclose_cause: bool = False, - ): - super().__init__( - message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause - ) - - -class SplitsNamesError(SplitsJobRunnerError): - """Raised when the split names could not be fetched.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "SplitsNamesError", cause, True) - - -class EmptyDatasetError(SplitsJobRunnerError): - """Raised when the dataset has no data.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "EmptyDatasetError", cause, True) - - -class SplitItem(TypedDict): - dataset: str - config: str - split: str - - -class SplitsResponseContent(TypedDict): - splits: List[SplitItem] - - -def get_dataset_split_full_names(dataset: str, use_auth_token: Union[bool, str, None] = False) -> List[SplitItem]: - """Get the list of splits full names (split and config) for a dataset. - - Args: - dataset (str): A dataset name. If the repository is namespaced (a user or an organization), the namespace and - the dataset name are separated with a slash (`/`), for example: `user/dataset`. - use_auth_token (Union[bool, str, None], optional): user token. It allows to retrieve the splits for gated - datasets. Defaults to False (no authentication). - - Returns: - List[SplitItem]: a list of splits full names: objects with the keys `dataset`, `config` and `split`. They - are sorted alphabetically by configuration (config), but the splits order for a given configuration is - preserved. - """ - logging.info(f"get dataset '{dataset}' split full names") - return [ - {"dataset": dataset, "config": str(config), "split": str(split)} - for config in sorted(get_dataset_config_names(path=dataset, use_auth_token=use_auth_token)) - for split in get_dataset_split_names(path=dataset, config_name=config, use_auth_token=use_auth_token) - ] - - -def compute_splits_response( - dataset: str, - hf_token: Optional[str] = None, -) -> SplitsResponseContent: - """ - Get the response of /splits for one specific dataset on huggingface.co. - Dataset can be private or gated if you pass an acceptable token. - - It is assumed that the dataset exists and can be accessed using the token. - - The list of splits might require the dataset to support the streaming mode. See - https://github.dev/huggingface/datasets/blob/e183a269067575db8765ee979bd8523d14a1adae/src/datasets/inspect.py#L389-L390 - - The /splits response generated by this function does not include stats about the split, like the size or number - of samples. See dataset-info or dataset-size for that. - - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - hf_token (`str`, *optional*): - An authentication token (See https://huggingface.co/settings/token) - Returns: - `SplitsResponseContent`: An object with the list of split names. - <Tip> - Raises the following errors: - - [`~job_runners.splits.EmptyDatasetError`] - The dataset is empty. - - [`~job_runners.splits.SplitsNamesError`] - If the list of splits could not be obtained using the datasets library. - </Tip> - """ - logging.info(f"get splits for dataset={dataset}") - use_auth_token: Union[bool, str, None] = hf_token if hf_token is not None else False - # get the list of splits in streaming mode - try: - split_items = get_dataset_split_full_names(dataset=dataset, use_auth_token=use_auth_token) - except _EmptyDatasetError as err: - raise EmptyDatasetError("The dataset is empty.", cause=err) from err - except Exception as err: - raise SplitsNamesError("Cannot get the split names for the dataset.", cause=err) from err - return {"splits": split_items} - - -class SplitsJobRunner(DatasetsBasedJobRunner): - @staticmethod - def get_job_type() -> str: - return "/splits" - - @staticmethod - def get_job_runner_version() -> int: - return PROCESSING_STEP_SPLITS_VERSION - - def compute(self) -> CompleteJobResult: - return CompleteJobResult(compute_splits_response(dataset=self.dataset, hf_token=self.common_config.hf_token)) - - def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]: - """Get the set of new splits, from the content created by the compute.""" - return {SplitFullName(dataset=s["dataset"], config=s["config"], split=s["split"]) for s in content["splits"]} diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py index 6a56ad8a..40554d94 100644 --- a/services/worker/tests/job_runners/test__datasets_based_worker.py +++ b/services/worker/tests/job_runners/test__datasets_based_worker.py @@ -28 +28 @@ class DummyJobRunner(DatasetsBasedJobRunner): - return "/splits" + return "/config-names" @@ -88 +88 @@ def get_job_runner( - ("user/dataset", "config", "split", True, "2022-11-07-12-34-56--splits-user-dataset-775e7212"), + ("user/dataset", "config", "split", True, "2022-11-07-12-34-56--config-names-user-dataset-cdf8effa"), @@ -90,5 +90,5 @@ def get_job_runner( - ("user/dataset", None, "split", True, "2022-11-07-12-34-56--splits-user-dataset-73c4b810"), - ("user/dataset", "config2", "split", True, "2022-11-07-12-34-56--splits-user-dataset-b6920bfb"), - ("user/dataset", "config", None, True, "2022-11-07-12-34-56--splits-user-dataset-36d21623"), - ("user/dataset", "config", "split2", True, "2022-11-07-12-34-56--splits-user-dataset-f60adde1"), - ("user/dataset", "config", "split", False, "2022-11-07-12-34-56--splits-user-dataset-f7985698"), + ("user/dataset", None, "split", True, "2022-11-07-12-34-56--config-names-user-dataset-54ba8b96"), + ("user/dataset", "config2", "split", True, "2022-11-07-12-34-56--config-names-user-dataset-1ad0bdcb"), + ("user/dataset", "config", None, True, "2022-11-07-12-34-56--config-names-user-dataset-49c90a57"), + ("user/dataset", "config", "split2", True, "2022-11-07-12-34-56--config-names-user-dataset-9a5cd356"), + ("user/dataset", "config", "split", False, "2022-11-07-12-34-56--config-names-user-dataset-abec311a"), @@ -101 +101 @@ def get_job_runner( - "2022-11-07-12-34-56--splits-very_long_dataset_name_0123456789012-1457d125", + "2022-11-07-12-34-56--config-names-very_long_dataset_name_0123456-30acf104", @@ -136 +136 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner - assert "-splits-user-dataset" in str(datasets.config.HF_DATASETS_CACHE) + assert "-config-names-user-dataset" in str(datasets.config.HF_DATASETS_CACHE) diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py deleted file mode 100644 index d2f7a155..00000000 --- a/services/worker/tests/job_runners/test_splits.py +++ /dev/null @@ -1,142 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from dataclasses import replace -from http import HTTPStatus -from typing import Callable - -import pytest -from libcommon.exceptions import CustomError -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Priority -from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import DoesNotExist, get_response - -from worker.config import AppConfig -from worker.job_runners.splits import SplitsJobRunner -from worker.resources import LibrariesResource - -from ..fixtures.hub import HubDatasets - -GetJobRunner = Callable[[str, AppConfig, bool], SplitsJobRunner] - - [email protected] -def get_job_runner( - libraries_resource: LibrariesResource, - cache_mongo_resource: CacheMongoResource, - queue_mongo_resource: QueueMongoResource, -) -> GetJobRunner: - def _get_job_runner( - dataset: str, - app_config: AppConfig, - force: bool = False, - ) -> SplitsJobRunner: - return SplitsJobRunner( - job_info={ - "type": SplitsJobRunner.get_job_type(), - "dataset": dataset, - "config": None, - "split": None, - "job_id": "job_id", - "force": force, - "priority": Priority.NORMAL, - }, - app_config=app_config, - processing_step=ProcessingStep( - name=SplitsJobRunner.get_job_type(), - input_type="dataset", - requires=[], - required_by_dataset_viewer=True, - ancestors=[], - children=[], - job_runner_version=SplitsJobRunner.get_job_runner_version(), - ), - hf_datasets_cache=libraries_resource.hf_datasets_cache, - ) - - return _get_job_runner - - -def should_skip_job(app_config: AppConfig, hub_public_csv: str, get_job_runner: GetJobRunner) -> None: - dataset = hub_public_csv - job_runner = get_job_runner(dataset, app_config, False) - assert not job_runner.should_skip_job() - # we add an entry to the cache - job_runner.process() - assert job_runner.should_skip_job() - job_runner = get_job_runner(dataset, app_config, True) - assert not job_runner.should_skip_job() - - -def test_process(app_config: AppConfig, hub_public_csv: str, get_job_runner: GetJobRunner) -> None: - dataset = hub_public_csv - job_runner = get_job_runner(dataset, app_config, False) - assert job_runner.process() - cached_response = get_response(kind=job_runner.processing_step.cache_kind, dataset=hub_public_csv) - assert cached_response["http_status"] == HTTPStatus.OK - assert cached_response["error_code"] is None - assert cached_response["job_runner_version"] == job_runner.get_job_runner_version() - assert cached_response["dataset_git_revision"] is not None - assert cached_response["error_code"] is None - content = cached_response["content"] - assert len(content["splits"]) == 1 - assert "stats" not in content["splits"][0] - - -def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset = "doesnotexist" - job_runner = get_job_runner(dataset, app_config, False) - assert not job_runner.process() - with pytest.raises(DoesNotExist): - get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset) - - [email protected]( - "name,use_token,error_code,cause", - [ - ("public", False, None, None), - ("audio", False, None, None), - ("gated", True, None, None), - ("private", True, None, None), - ("empty", False, "EmptyDatasetError", "EmptyDatasetError"), - # should we really test the following cases? - # The assumption is that the dataset exists and is accessible with the token - ("does_not_exist", False, "SplitsNamesError", "FileNotFoundError"), - ("gated", False, "SplitsNamesError", "FileNotFoundError"), - ("private", False, "SplitsNamesError", "FileNotFoundError"), - ], -) -def test_compute_splits_response_simple_csv( - hub_datasets: HubDatasets, - get_job_runner: GetJobRunner, - name: str, - use_token: bool, - error_code: str, - cause: str, - app_config: AppConfig, -) -> None: - dataset = hub_datasets[name]["name"] - expected_splits_response = hub_datasets[name]["splits_response"] - job_runner = get_job_runner( - dataset, - app_config if use_token else replace(app_config, common=replace(app_config.common, hf_token=None)), - False, - ) - if error_code is None: - result = job_runner.compute().content - assert result == expected_splits_response - return - - with pytest.raises(CustomError) as exc_info: - job_runner.compute() - assert exc_info.value.code == error_code - assert exc_info.value.cause_exception == cause - if exc_info.value.disclose_cause: - response = exc_info.value.as_response() - assert set(response.keys()) == {"error", "cause_exception", "cause_message", "cause_traceback"} - response_dict = dict(response) - # ^ to remove mypy warnings - assert response_dict["cause_exception"] == cause - assert isinstance(response_dict["cause_traceback"], list) - assert response_dict["cause_traceback"][0] == "Traceback (most recent call last):\n" diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py index f09eadc9..ee8db769 100644 --- a/services/worker/tests/test_executor.py +++ b/services/worker/tests/test_executor.py @@ -44 +44 @@ def get_job_info(prefix: str = "base") -> JobInfo: - type="/splits", + type="/config-names", diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py index 3e07c047..a6b335a5 100644 --- a/services/worker/tests/test_job_runner_factory.py +++ b/services/worker/tests/test_job_runner_factory.py @@ -25 +24,0 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph: - ("/splits", "SplitsJobRunner"),
020d0eec506aaf0dd1650b6885c07f8b71884b77
Andrea Francis Soria Jimenez
2023-04-07T06:55:25
Disable full backfill (#1028)
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 3adb6f9c..6a7bfe6a 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -106 +106 @@ cacheMaintenance: - action: "backfill" + action: "skip" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 20a5ba11..fe24f4e6 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -112 +112 @@ cacheMaintenance: - action: "backfill" + action: "skip"
34d49c9a8383bf3521844ed507a28a141ef193fc
Andrea Francis Soria Jimenez
2023-04-06T18:40:10
Increase backgill resources and adding logs (#1031)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 63c6cf9c..20a5ba11 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -115,2 +114,0 @@ cacheMaintenance: - cpu: 100m - limits: @@ -117,0 +116,4 @@ cacheMaintenance: + memory: "2Gi" + limits: + cpu: 2 + memory: "8Gi" diff --git a/chart/templates/jobs/cache-maintenance/job.yaml b/chart/templates/jobs/cache-maintenance/job.yaml index 1178ffc6..eb4b793d 100644 --- a/chart/templates/jobs/cache-maintenance/job.yaml +++ b/chart/templates/jobs/cache-maintenance/job.yaml @@ -25 +25 @@ spec: - backoffLimit: 0 + backoffLimit: 1 diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index cdb7ac1f..a304a7df 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -20,0 +21,2 @@ def backfill_cache( + backfilled_datasets = 0 + log_batch = 100 @@ -30,0 +33,3 @@ def backfill_cache( + backfilled_datasets += 1 + if backfilled_datasets % log_batch == 0: + logging.info(f"{backfilled_datasets} datasets have been backfilled")
2e6ace6c8ce33322d8c9bb8839f89d7f8a96f922
Andrea Francis Soria Jimenez
2023-04-06T15:39:46
Adding more replicas (#1027)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 68954414..63c6cf9c 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -222 +222 @@ workers: - replicas: 40 + replicas: 80 @@ -237 +237 @@ workers: - replicas: 2 + replicas: 4
ff889862f54cd05141c77a16008b35c659afb4ea
Andrea Francis Soria Jimenez
2023-04-06T15:14:55
Moving /backfill to a k8s job (#1017)
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml index 42f13df5..5c0d5340 100644 --- a/.github/workflows/cd.yml +++ b/.github/workflows/cd.yml @@ -27,0 +28,2 @@ jobs: + - directory: jobs + project: cache_maintenance @@ -105,0 +108,2 @@ jobs: + cacheMaintenance: + tag: sha-${{ steps.vars.outputs.sha_short }} diff --git a/.github/workflows/j-cache-maintenance.yml b/.github/workflows/j-cache-maintenance.yml new file mode 100644 index 00000000..74052dcf --- /dev/null +++ b/.github/workflows/j-cache-maintenance.yml @@ -0,0 +1,33 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +name: jobs/cache_maintenance +on: + workflow_dispatch: + push: + branches: + - main + paths: + - 'jobs/cache_maintenance/**' + - 'libs/libcommon/**' + - '.github/workflows/j-cache-maintenance.yml' + - '.github/workflows/_quality-python.yml' + - '.github/workflows/_unit-tests-python.yml' + - 'tools/docker-compose-mongo.yml' + pull_request: + paths: + - 'jobs/cache_maintenance/**' + - 'libs/libcommon/**' + - '.github/workflows/j-cache-maintenance.yml' + - '.github/workflows/_quality-python.yml' + - '.github/workflows/_unit-tests-python.yml' + - 'tools/docker-compose-mongo.yml' +jobs: + quality: + uses: ./.github/workflows/_quality-python.yml + with: + working-directory: jobs/cache_maintenance + unit-tests: + uses: ./.github/workflows/_unit-tests-python.yml + with: + working-directory: jobs/cache_maintenance diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace index 604b61a1..847cd8df 100644 --- a/.vscode/monorepo.code-workspace +++ b/.vscode/monorepo.code-workspace @@ -18,0 +19,4 @@ + { + "name": "jobs/cache_maintenance", + "path": "../jobs/cache_maintenance" + }, diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 84411992..2f782c84 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.6.0 +version: 1.7.0 diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 6509195f..3adb6f9c 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -29,0 +30,5 @@ images: + cacheMaintenance: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-jobs-cache_maintenance + tag: sha-fb3399a @@ -97,0 +103,10 @@ mongodbMigration: +# --- jobs (post-upgrade hooks) --- + +cacheMaintenance: + action: "backfill" + resources: + requests: + cpu: 100m + limits: + cpu: 1 + diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index a0465b2c..68954414 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -28,0 +29,5 @@ images: + cacheMaintenance: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-jobs-cache_maintenance + tag: sha-fb3399a @@ -103,0 +109,10 @@ mongodbMigration: +# --- jobs (post-upgrade hooks) --- + +cacheMaintenance: + action: "backfill" + resources: + requests: + cpu: 100m + limits: + cpu: 1 + diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index 4a601cf6..b2bad103 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -40,0 +41,4 @@ Docker image management +{{- define "jobs.cacheMaintenance.image" -}} +{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.jobs.cacheMaintenance "global" .Values.global.huggingface) }} +{{- end -}} + @@ -80,0 +85,5 @@ app.kubernetes.io/component: "{{ include "name" . }}-cache-refresh" +{{- define "labels.cacheMaintenance" -}} +{{ include "hf.labels.commons" . }} +app.kubernetes.io/component: "{{ include "name" . }}-cache-maintenance" +{{- end -}} + diff --git a/chart/templates/jobs/cache-maintenance/_container.tpl b/chart/templates/jobs/cache-maintenance/_container.tpl new file mode 100644 index 00000000..648d0b06 --- /dev/null +++ b/chart/templates/jobs/cache-maintenance/_container.tpl @@ -0,0 +1,17 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "containerCacheMaintenance" -}} +- name: "{{ include "name" . }}-cache-maintenance" + image: {{ include "jobs.cacheMaintenance.image" . }} + imagePullPolicy: {{ .Values.images.pullPolicy }} + env: + {{ include "envLog" . | nindent 2 }} + {{ include "envQueue" . | nindent 2 }} + {{ include "envCommon" . | nindent 2 }} + - name: CACHE_MAINTENANCE_ACTION + value: {{ .Values.cacheMaintenance.action | quote }} + securityContext: + allowPrivilegeEscalation: false + resources: {{ toYaml .Values.cacheMaintenance.resources | nindent 4 }} +{{- end -}} diff --git a/chart/templates/jobs/cache-maintenance/job.yaml b/chart/templates/jobs/cache-maintenance/job.yaml new file mode 100644 index 00000000..1178ffc6 --- /dev/null +++ b/chart/templates/jobs/cache-maintenance/job.yaml @@ -0,0 +1,26 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- if .Values.images.jobs.cacheMaintenance }} +apiVersion: batch/v1 +kind: Job +metadata: + labels: {{ include "labels.cacheMaintenance" . | nindent 4 }} + name: "{{ include "name" . }}-job-cache-maintenance" + namespace: {{ .Release.Namespace }} + annotations: + "helm.sh/hook": post-install,post-upgrade +spec: + ttlSecondsAfterFinished: 86400 + template: + metadata: + labels: {{ include "labels.cacheMaintenance" . | nindent 8 }} + spec: + restartPolicy: Never + {{- include "image.imagePullSecrets" . | nindent 6 }} + containers: {{ include "containerCacheMaintenance" . | nindent 8 }} + nodeSelector: {{ toYaml .Values.cacheMaintenance.nodeSelector | nindent 8 }} + tolerations: {{ toYaml .Values.cacheMaintenance.tolerations | nindent 8 }} + securityContext: {{ include "securityContext" . | nindent 8 }} + backoffLimit: 0 +{{- end}} diff --git a/chart/values.yaml b/chart/values.yaml index 7fac12fc..c650dde2 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -34,0 +35,5 @@ images: + cacheMaintenance: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-jobs-cache_maintenance + tag: sha-fb3399a @@ -197,0 +203,10 @@ cacheRefresh: +cacheMaintenance: + action: "skip" + nodeSelector: {} + resources: + requests: + cpu: 0 + limits: + cpu: 0 + tolerations: [] + diff --git a/jobs/cache_maintenance/.flake8 b/jobs/cache_maintenance/.flake8 new file mode 100644 index 00000000..f7d6157c --- /dev/null +++ b/jobs/cache_maintenance/.flake8 @@ -0,0 +1,5 @@ +[flake8] +# Recommend matching the black line length (119), +# rather than using the flake8 default of 79: +max-line-length = 119 +extend-ignore = "E203" diff --git a/jobs/cache_maintenance/.python-version b/jobs/cache_maintenance/.python-version new file mode 100644 index 00000000..b326afbc --- /dev/null +++ b/jobs/cache_maintenance/.python-version @@ -0,0 +1 @@ +3.9.15 diff --git a/jobs/cache_maintenance/Dockerfile b/jobs/cache_maintenance/Dockerfile new file mode 100644 index 00000000..1cf31830 --- /dev/null +++ b/jobs/cache_maintenance/Dockerfile @@ -0,0 +1,33 @@ +# build with +# docker build -t some_tag_admin -f Dockerfile ../.. +FROM python:3.9.15-slim + +ENV PYTHONFAULTHANDLER=1 \ + PYTHONUNBUFFERED=1 \ + PYTHONHASHSEED=random \ + PIP_NO_CACHE_DIR=1 \ + PIP_DISABLE_PIP_VERSION_CHECK=on \ + PIP_DEFAULT_TIMEOUT=100 \ + POETRY_NO_INTERACTION=1 \ + # Versions: + POETRY_VERSION=1.3.2 \ + POETRY_VIRTUALENVS_IN_PROJECT=true \ + PATH="$PATH:/root/.local/bin" + +# System deps: +RUN apt-get update \ + && apt-get install -y build-essential python3-dev make \ + && rm -rf /var/lib/apt/lists/* +RUN pip install -U pip +RUN pip install "poetry==$POETRY_VERSION" + +WORKDIR /src +COPY jobs/cache_maintenance/poetry.lock ./jobs/cache_maintenance/poetry.lock +COPY jobs/cache_maintenance/pyproject.toml ./jobs/cache_maintenance/pyproject.toml +COPY libs/libcommon ./libs/libcommon +WORKDIR /src/jobs/cache_maintenance/ +RUN poetry install --no-cache +COPY jobs/cache_maintenance/src ./src +RUN poetry install --no-cache + +ENTRYPOINT ["poetry", "run", "python", "src/cache_maintenance/main.py"] diff --git a/jobs/cache_maintenance/Makefile b/jobs/cache_maintenance/Makefile new file mode 100644 index 00000000..a56a0338 --- /dev/null +++ b/jobs/cache_maintenance/Makefile @@ -0,0 +1,15 @@ +# environment variables for the commands (docker compose, poetry) +export COMPOSE_PROJECT_NAME := cache_maintenance +export MONGO_PORT := 27080 +export CACHE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +# makefile variables +DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml + +include ../../tools/Python.mk +include ../../tools/PythonTest.mk +include ../../tools/Docker.mk + +.PHONY: run +run: + poetry run python src/cache_maintenance/main.py diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock new file mode 100644 index 00000000..ee4e2a3a --- /dev/null +++ b/jobs/cache_maintenance/poetry.lock @@ -0,0 +1,1589 @@ +# This file is automatically @generated by Poetry and should not be changed by hand. + +[[package]] +name = "appdirs" +version = "1.4.4" +description = "A small Python module for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "appdirs-1.4.4-py2.py3-none-any.whl", hash = "sha256:a841dacd6b99318a741b166adb07e19ee71a274450e68237b4650ca1055ab128"}, + {file = "appdirs-1.4.4.tar.gz", hash = "sha256:7d5d0167b2b1ba821647616af46a749d1c653740dd0d2415100fe26e27afdf41"}, +] + +[[package]] +name = "attrs" +version = "22.2.0" +description = "Classes Without Boilerplate" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "attrs-22.2.0-py3-none-any.whl", hash = "sha256:29e95c7f6778868dbd49170f98f8818f78f3dc5e0e37c0b1f474e3561b240836"}, + {file = "attrs-22.2.0.tar.gz", hash = "sha256:c9227bfc2f01993c03f68db37d1d15c9690188323c067c641f1a35ca58185f99"}, +] + +[package.extras] +cov = ["attrs[tests]", "coverage-enable-subprocess", "coverage[toml] (>=5.3)"] +dev = ["attrs[docs,tests]"] +docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope.interface"] +tests = ["attrs[tests-no-zope]", "zope.interface"] +tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy (>=0.971,<0.990)", "mypy (>=0.971,<0.990)", "pympler", "pympler", "pytest (>=4.3.0)", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-mypy-plugins", "pytest-xdist[psutil]", "pytest-xdist[psutil]"] + +[[package]] +name = "bandit" +version = "1.7.5" +description = "Security oriented static analyser for python code." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "bandit-1.7.5-py3-none-any.whl", hash = "sha256:75665181dc1e0096369112541a056c59d1c5f66f9bb74a8d686c3c362b83f549"}, + {file = "bandit-1.7.5.tar.gz", hash = "sha256:bdfc739baa03b880c2d15d0431b31c658ffc348e907fe197e54e0389dd59e11e"}, +] + +[package.dependencies] +colorama = {version = ">=0.3.9", markers = "platform_system == \"Windows\""} +GitPython = ">=1.0.1" +PyYAML = ">=5.3.1" +rich = "*" +stevedore = ">=1.20.0" + +[package.extras] +test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "tomli (>=1.1.0)"] +toml = ["tomli (>=1.1.0)"] +yaml = ["PyYAML"] + +[[package]] +name = "black" +version = "22.12.0" +description = "The uncompromising code formatter." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "black-22.12.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eedd20838bd5d75b80c9f5487dbcb06836a43833a37846cf1d8c1cc01cef59d"}, + {file = "black-22.12.0-cp310-cp310-win_amd64.whl", hash = "sha256:159a46a4947f73387b4d83e87ea006dbb2337eab6c879620a3ba52699b1f4351"}, + {file = "black-22.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d30b212bffeb1e252b31dd269dfae69dd17e06d92b87ad26e23890f3efea366f"}, + {file = "black-22.12.0-cp311-cp311-win_amd64.whl", hash = "sha256:7412e75863aa5c5411886804678b7d083c7c28421210180d67dfd8cf1221e1f4"}, + {file = "black-22.12.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c116eed0efb9ff870ded8b62fe9f28dd61ef6e9ddd28d83d7d264a38417dcee2"}, + {file = "black-22.12.0-cp37-cp37m-win_amd64.whl", hash = "sha256:1f58cbe16dfe8c12b7434e50ff889fa479072096d79f0a7f25e4ab8e94cd8350"}, + {file = "black-22.12.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77d86c9f3db9b1bf6761244bc0b3572a546f5fe37917a044e02f3166d5aafa7d"}, + {file = "black-22.12.0-cp38-cp38-win_amd64.whl", hash = "sha256:82d9fe8fee3401e02e79767016b4907820a7dc28d70d137eb397b92ef3cc5bfc"}, + {file = "black-22.12.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:101c69b23df9b44247bd88e1d7e90154336ac4992502d4197bdac35dd7ee3320"}, + {file = "black-22.12.0-cp39-cp39-win_amd64.whl", hash = "sha256:559c7a1ba9a006226f09e4916060982fd27334ae1998e7a38b3f33a37f7a2148"}, + {file = "black-22.12.0-py3-none-any.whl", hash = "sha256:436cc9167dd28040ad90d3b404aec22cedf24a6e4d7de221bec2730ec0c97bcf"}, + {file = "black-22.12.0.tar.gz", hash = "sha256:229351e5a18ca30f447bf724d007f890f97e13af070bb6ad4c0a441cd7596a2f"}, +] + +[package.dependencies] +click = ">=8.0.0" +mypy-extensions = ">=0.4.3" +pathspec = ">=0.9.0" +platformdirs = ">=2" +tomli = {version = ">=1.1.0", markers = "python_full_version < \"3.11.0a7\""} +typing-extensions = {version = ">=3.10.0.0", markers = "python_version < \"3.10\""} + +[package.extras] +colorama = ["colorama (>=0.4.3)"] +d = ["aiohttp (>=3.7.4)"] +jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"] +uvloop = ["uvloop (>=0.15.2)"] + +[[package]] +name = "cachecontrol" +version = "0.12.11" +description = "httplib2 caching for requests" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "CacheControl-0.12.11-py2.py3-none-any.whl", hash = "sha256:2c75d6a8938cb1933c75c50184549ad42728a27e9f6b92fd677c3151aa72555b"}, + {file = "CacheControl-0.12.11.tar.gz", hash = "sha256:a5b9fcc986b184db101aa280b42ecdcdfc524892596f606858e0b7a8b4d9e144"}, +] + +[package.dependencies] +lockfile = {version = ">=0.9", optional = true, markers = "extra == \"filecache\""} +msgpack = ">=0.5.2" +requests = "*" + +[package.extras] +filecache = ["lockfile (>=0.9)"] +redis = ["redis (>=2.10.5)"] + +[[package]] +name = "certifi" +version = "2022.12.7" +description = "Python package for providing Mozilla's CA Bundle." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "certifi-2022.12.7-py3-none-any.whl", hash = "sha256:4ad3232f5e926d6718ec31cfc1fcadfde020920e278684144551c91769c7bc18"}, + {file = "certifi-2022.12.7.tar.gz", hash = "sha256:35824b4c3a97115964b408844d64aa14db1cc518f6562e8d7261699d1350a9e3"}, +] + +[[package]] +name = "charset-normalizer" +version = "3.1.0" +description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"}, + {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"}, +] + +[[package]] +name = "click" +version = "8.1.3" +description = "Composable command line interface toolkit" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "click-8.1.3-py3-none-any.whl", hash = "sha256:bb4d8133cb15a609f44e8213d9b391b0809795062913b383c62be0ee95b1db48"}, + {file = "click-8.1.3.tar.gz", hash = "sha256:7682dc8afb30297001674575ea00d1814d808d6a36af415a82bd481d37ba7b8e"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[[package]] +name = "colorama" +version = "0.4.6" +description = "Cross-platform colored terminal text." +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7" +files = [ + {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"}, + {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"}, +] + +[[package]] +name = "coverage" +version = "7.2.2" +description = "Code coverage measurement for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "coverage-7.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c90e73bdecb7b0d1cea65a08cb41e9d672ac6d7995603d6465ed4914b98b9ad7"}, + {file = "coverage-7.2.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e2926b8abedf750c2ecf5035c07515770944acf02e1c46ab08f6348d24c5f94d"}, + {file = "coverage-7.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57b77b9099f172804e695a40ebaa374f79e4fb8b92f3e167f66facbf92e8e7f5"}, + {file = "coverage-7.2.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:efe1c0adad110bf0ad7fb59f833880e489a61e39d699d37249bdf42f80590169"}, + {file = "coverage-7.2.2-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2199988e0bc8325d941b209f4fd1c6fa007024b1442c5576f1a32ca2e48941e6"}, + {file = "coverage-7.2.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:81f63e0fb74effd5be736cfe07d710307cc0a3ccb8f4741f7f053c057615a137"}, + {file = "coverage-7.2.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:186e0fc9cf497365036d51d4d2ab76113fb74f729bd25da0975daab2e107fd90"}, + {file = "coverage-7.2.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:420f94a35e3e00a2b43ad5740f935358e24478354ce41c99407cddd283be00d2"}, + {file = "coverage-7.2.2-cp310-cp310-win32.whl", hash = "sha256:38004671848b5745bb05d4d621526fca30cee164db42a1f185615f39dc997292"}, + {file = "coverage-7.2.2-cp310-cp310-win_amd64.whl", hash = "sha256:0ce383d5f56d0729d2dd40e53fe3afeb8f2237244b0975e1427bfb2cf0d32bab"}, + {file = "coverage-7.2.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3eb55b7b26389dd4f8ae911ba9bc8c027411163839dea4c8b8be54c4ee9ae10b"}, + {file = "coverage-7.2.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d2b96123a453a2d7f3995ddb9f28d01fd112319a7a4d5ca99796a7ff43f02af5"}, + {file = "coverage-7.2.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:299bc75cb2a41e6741b5e470b8c9fb78d931edbd0cd009c58e5c84de57c06731"}, + {file = "coverage-7.2.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5e1df45c23d4230e3d56d04414f9057eba501f78db60d4eeecfcb940501b08fd"}, + {file = "coverage-7.2.2-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:006ed5582e9cbc8115d2e22d6d2144a0725db542f654d9d4fda86793832f873d"}, + {file = "coverage-7.2.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d683d230b5774816e7d784d7ed8444f2a40e7a450e5720d58af593cb0b94a212"}, + {file = "coverage-7.2.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:8efb48fa743d1c1a65ee8787b5b552681610f06c40a40b7ef94a5b517d885c54"}, + {file = "coverage-7.2.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:4c752d5264053a7cf2fe81c9e14f8a4fb261370a7bb344c2a011836a96fb3f57"}, + {file = "coverage-7.2.2-cp311-cp311-win32.whl", hash = "sha256:55272f33da9a5d7cccd3774aeca7a01e500a614eaea2a77091e9be000ecd401d"}, + {file = "coverage-7.2.2-cp311-cp311-win_amd64.whl", hash = "sha256:92ebc1619650409da324d001b3a36f14f63644c7f0a588e331f3b0f67491f512"}, + {file = "coverage-7.2.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:5afdad4cc4cc199fdf3e18088812edcf8f4c5a3c8e6cb69127513ad4cb7471a9"}, + {file = "coverage-7.2.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0484d9dd1e6f481b24070c87561c8d7151bdd8b044c93ac99faafd01f695c78e"}, + {file = "coverage-7.2.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d530191aa9c66ab4f190be8ac8cc7cfd8f4f3217da379606f3dd4e3d83feba69"}, + {file = "coverage-7.2.2-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ac0f522c3b6109c4b764ffec71bf04ebc0523e926ca7cbe6c5ac88f84faced0"}, + {file = "coverage-7.2.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:ba279aae162b20444881fc3ed4e4f934c1cf8620f3dab3b531480cf602c76b7f"}, + {file = "coverage-7.2.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:53d0fd4c17175aded9c633e319360d41a1f3c6e352ba94edcb0fa5167e2bad67"}, + {file = "coverage-7.2.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8c99cb7c26a3039a8a4ee3ca1efdde471e61b4837108847fb7d5be7789ed8fd9"}, + {file = "coverage-7.2.2-cp37-cp37m-win32.whl", hash = "sha256:5cc0783844c84af2522e3a99b9b761a979a3ef10fb87fc4048d1ee174e18a7d8"}, + {file = "coverage-7.2.2-cp37-cp37m-win_amd64.whl", hash = "sha256:817295f06eacdc8623dc4df7d8b49cea65925030d4e1e2a7c7218380c0072c25"}, + {file = "coverage-7.2.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:6146910231ece63facfc5984234ad1b06a36cecc9fd0c028e59ac7c9b18c38c6"}, + {file = "coverage-7.2.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:387fb46cb8e53ba7304d80aadca5dca84a2fbf6fe3faf6951d8cf2d46485d1e5"}, + {file = "coverage-7.2.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:046936ab032a2810dcaafd39cc4ef6dd295df1a7cbead08fe996d4765fca9fe4"}, + {file = "coverage-7.2.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e627dee428a176ffb13697a2c4318d3f60b2ccdde3acdc9b3f304206ec130ccd"}, + {file = "coverage-7.2.2-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4fa54fb483decc45f94011898727802309a109d89446a3c76387d016057d2c84"}, + {file = "coverage-7.2.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:3668291b50b69a0c1ef9f462c7df2c235da3c4073f49543b01e7eb1dee7dd540"}, + {file = "coverage-7.2.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:7c20b731211261dc9739bbe080c579a1835b0c2d9b274e5fcd903c3a7821cf88"}, + {file = "coverage-7.2.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:5764e1f7471cb8f64b8cda0554f3d4c4085ae4b417bfeab236799863703e5de2"}, + {file = "coverage-7.2.2-cp38-cp38-win32.whl", hash = "sha256:4f01911c010122f49a3e9bdc730eccc66f9b72bd410a3a9d3cb8448bb50d65d3"}, + {file = "coverage-7.2.2-cp38-cp38-win_amd64.whl", hash = "sha256:c448b5c9e3df5448a362208b8d4b9ed85305528313fca1b479f14f9fe0d873b8"}, + {file = "coverage-7.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bfe7085783cda55e53510482fa7b5efc761fad1abe4d653b32710eb548ebdd2d"}, + {file = "coverage-7.2.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:9d22e94e6dc86de981b1b684b342bec5e331401599ce652900ec59db52940005"}, + {file = "coverage-7.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:507e4720791977934bba016101579b8c500fb21c5fa3cd4cf256477331ddd988"}, + {file = "coverage-7.2.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bc4803779f0e4b06a2361f666e76f5c2e3715e8e379889d02251ec911befd149"}, + {file = "coverage-7.2.2-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db8c2c5ace167fd25ab5dd732714c51d4633f58bac21fb0ff63b0349f62755a8"}, + {file = "coverage-7.2.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:4f68ee32d7c4164f1e2c8797535a6d0a3733355f5861e0f667e37df2d4b07140"}, + {file = "coverage-7.2.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:d52f0a114b6a58305b11a5cdecd42b2e7f1ec77eb20e2b33969d702feafdd016"}, + {file = "coverage-7.2.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:797aad79e7b6182cb49c08cc5d2f7aa7b2128133b0926060d0a8889ac43843be"}, + {file = "coverage-7.2.2-cp39-cp39-win32.whl", hash = "sha256:db45eec1dfccdadb179b0f9ca616872c6f700d23945ecc8f21bb105d74b1c5fc"}, + {file = "coverage-7.2.2-cp39-cp39-win_amd64.whl", hash = "sha256:8dbe2647bf58d2c5a6c5bcc685f23b5f371909a5624e9f5cd51436d6a9f6c6ef"}, + {file = "coverage-7.2.2-pp37.pp38.pp39-none-any.whl", hash = "sha256:872d6ce1f5be73f05bea4df498c140b9e7ee5418bfa2cc8204e7f9b817caa968"}, + {file = "coverage-7.2.2.tar.gz", hash = "sha256:36dd42da34fe94ed98c39887b86db9d06777b1c8f860520e21126a75507024f2"}, +] + +[package.extras] +toml = ["tomli"] + +[[package]] +name = "cyclonedx-python-lib" +version = "2.7.1" +description = "A library for producing CycloneDX SBOM (Software Bill of Materials) files." +category = "dev" +optional = false +python-versions = ">=3.6,<4.0" +files = [ + {file = "cyclonedx-python-lib-2.7.1.tar.gz", hash = "sha256:493bf2f30e26c48f305f745ed8580ce10d05a8d68d62a598fe95f05a0d9007dc"}, + {file = "cyclonedx_python_lib-2.7.1-py3-none-any.whl", hash = "sha256:fabc4c8baf722faeea01c3bbca83730e3489dfb37d85c6036baa67a9a7519d40"}, +] + +[package.dependencies] +packageurl-python = ">=0.9" +setuptools = ">=47.0.0" +sortedcontainers = ">=2.4.0,<3.0.0" +toml = ">=0.10.0,<0.11.0" + +[[package]] +name = "dnspython" +version = "1.16.0" +description = "DNS toolkit" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "dnspython-1.16.0-py2.py3-none-any.whl", hash = "sha256:f69c21288a962f4da86e56c4905b49d11aba7938d3d740e80d9e366ee4f1632d"}, + {file = "dnspython-1.16.0.zip", hash = "sha256:36c5e8e38d4369a08b6780b7f27d790a292b2b08eea01607865bf0936c558e01"}, +] + +[package.extras] +dnssec = ["ecdsa (>=0.13)", "pycryptodome"] +idna = ["idna (>=2.1)"] + +[[package]] +name = "environs" +version = "9.5.0" +description = "simplified environment variable parsing" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "environs-9.5.0-py2.py3-none-any.whl", hash = "sha256:1e549569a3de49c05f856f40bce86979e7d5ffbbc4398e7f338574c220189124"}, + {file = "environs-9.5.0.tar.gz", hash = "sha256:a76307b36fbe856bdca7ee9161e6c466fd7fcffc297109a118c59b54e27e30c9"}, +] + +[package.dependencies] +marshmallow = ">=3.0.0" +python-dotenv = "*" + +[package.extras] +dev = ["dj-database-url", "dj-email-url", "django-cache-url", "flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)", "pytest", "tox"] +django = ["dj-database-url", "dj-email-url", "django-cache-url"] +lint = ["flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)"] +tests = ["dj-database-url", "dj-email-url", "django-cache-url", "pytest"] + +[[package]] +name = "exceptiongroup" +version = "1.1.1" +description = "Backport of PEP 654 (exception groups)" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "exceptiongroup-1.1.1-py3-none-any.whl", hash = "sha256:232c37c63e4f682982c8b6459f33a8981039e5fb8756b2074364e5055c498c9e"}, + {file = "exceptiongroup-1.1.1.tar.gz", hash = "sha256:d484c3090ba2889ae2928419117447a14daf3c1231d5e30d0aae34f354f01785"}, +] + +[package.extras] +test = ["pytest (>=6)"] + +[[package]] +name = "filelock" +version = "3.10.7" +description = "A platform independent file lock." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "filelock-3.10.7-py3-none-any.whl", hash = "sha256:bde48477b15fde2c7e5a0713cbe72721cb5a5ad32ee0b8f419907960b9d75536"}, + {file = "filelock-3.10.7.tar.gz", hash = "sha256:892be14aa8efc01673b5ed6589dbccb95f9a8596f0507e232626155495c18105"}, +] + +[package.extras] +docs = ["furo (>=2022.12.7)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"] +testing = ["covdefaults (>=2.3)", "coverage (>=7.2.2)", "diff-cover (>=7.5)", "pytest (>=7.2.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)", "pytest-timeout (>=2.1)"] + +[[package]] +name = "flake8" +version = "3.9.2" +description = "the modular source code checker: pep8 pyflakes and co" +category = "dev" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7" +files = [ + {file = "flake8-3.9.2-py2.py3-none-any.whl", hash = "sha256:bf8fd333346d844f616e8d47905ef3a3384edae6b4e9beb0c5101e25e3110907"}, + {file = "flake8-3.9.2.tar.gz", hash = "sha256:07528381786f2a6237b061f6e96610a4167b226cb926e2aa2b6b1d78057c576b"}, +] + +[package.dependencies] +mccabe = ">=0.6.0,<0.7.0" +pycodestyle = ">=2.7.0,<2.8.0" +pyflakes = ">=2.3.0,<2.4.0" + +[[package]] +name = "gitdb" +version = "4.0.10" +description = "Git Object Database" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "gitdb-4.0.10-py3-none-any.whl", hash = "sha256:c286cf298426064079ed96a9e4a9d39e7f3e9bf15ba60701e95f5492f28415c7"}, + {file = "gitdb-4.0.10.tar.gz", hash = "sha256:6eb990b69df4e15bad899ea868dc46572c3f75339735663b81de79b06f17eb9a"}, +] + +[package.dependencies] +smmap = ">=3.0.1,<6" + +[[package]] +name = "gitpython" +version = "3.1.31" +description = "GitPython is a Python library used to interact with Git repositories" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"}, + {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"}, +] + +[package.dependencies] +gitdb = ">=4.0.1,<5" + +[[package]] +name = "html5lib" +version = "1.1" +description = "HTML parser based on the WHATWG HTML specification" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "html5lib-1.1-py2.py3-none-any.whl", hash = "sha256:0d78f8fde1c230e99fe37986a60526d7049ed4bf8a9fadbad5f00e22e58e041d"}, + {file = "html5lib-1.1.tar.gz", hash = "sha256:b2e5b40261e20f354d198eae92afc10d750afb487ed5e50f9c4eaf07c184146f"}, +] + +[package.dependencies] +six = ">=1.9" +webencodings = "*" + +[package.extras] +all = ["chardet (>=2.2)", "genshi", "lxml"] +chardet = ["chardet (>=2.2)"] +genshi = ["genshi"] +lxml = ["lxml"] + +[[package]] +name = "huggingface-hub" +version = "0.12.1" +description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"}, + {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"}, +] + +[package.dependencies] +filelock = "*" +packaging = ">=20.9" +pyyaml = ">=5.1" +requests = "*" +tqdm = ">=4.42.1" +typing-extensions = ">=3.7.4.3" + +[package.extras] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +cli = ["InquirerPy (==0.3.4)"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] +quality = ["black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "mypy (==0.982)"] +tensorflow = ["graphviz", "pydot", "tensorflow"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "isort (>=5.5.4)", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] +torch = ["torch"] +typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] + +[[package]] +name = "idna" +version = "3.4" +description = "Internationalized Domain Names in Applications (IDNA)" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"}, + {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"}, +] + +[[package]] +name = "iniconfig" +version = "2.0.0" +description = "brain-dead simple config-ini parsing" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "iniconfig-2.0.0-py3-none-any.whl", hash = "sha256:b6a85871a79d2e3b22d2d1b94ac2824226a63c6b741c88f7ae975f18b6778374"}, + {file = "iniconfig-2.0.0.tar.gz", hash = "sha256:2d91e135bf72d31a410b17c16da610a82cb55f6b0477d1a902134b24a455b8b3"}, +] + +[[package]] +name = "isort" +version = "5.12.0" +description = "A Python utility / library to sort Python imports." +category = "dev" +optional = false +python-versions = ">=3.8.0" +files = [ + {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"}, + {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"}, +] + +[package.extras] +colors = ["colorama (>=0.4.3)"] +pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"] +plugins = ["setuptools"] +requirements-deprecated-finder = ["pip-api", "pipreqs"] + +[[package]] +name = "libcommon" +version = "0.6.8" +description = "Library for utils common to all the services" +category = "main" +optional = false +python-versions = "3.9.15" +files = [] +develop = true + +[package.dependencies] +appdirs = "^1.4.4" +environs = "^9.5.0" +huggingface-hub = "^0.12.0" +mongo-types = "0.15.1" +mongoengine = "^0.24.2" +networkx = "^3.0" +orjson = "^3.8.6" +psutil = "^5.9.4" +pymongo = {version = "^3.13.0", extras = ["srv"]} +pytz = "^2020.1" +requests = "^2.28.2" + +[package.source] +type = "directory" +url = "../../libs/libcommon" + +[[package]] +name = "lockfile" +version = "0.12.2" +description = "Platform-independent file locking module" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "lockfile-0.12.2-py2.py3-none-any.whl", hash = "sha256:6c3cb24f344923d30b2785d5ad75182c8ea7ac1b6171b08657258ec7429d50fa"}, + {file = "lockfile-0.12.2.tar.gz", hash = "sha256:6aed02de03cba24efabcd600b30540140634fc06cfa603822d508d5361e9f799"}, +] + +[[package]] +name = "markdown-it-py" +version = "2.2.0" +description = "Python port of markdown-it. Markdown parsing, done right!" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"}, + {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"}, +] + +[package.dependencies] +mdurl = ">=0.1,<1.0" + +[package.extras] +benchmarking = ["psutil", "pytest", "pytest-benchmark"] +code-style = ["pre-commit (>=3.0,<4.0)"] +compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"] +linkify = ["linkify-it-py (>=1,<3)"] +plugins = ["mdit-py-plugins"] +profiling = ["gprof2dot"] +rtd = ["attrs", "myst-parser", "pyyaml", "sphinx", "sphinx-copybutton", "sphinx-design", "sphinx_book_theme"] +testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"] + +[[package]] +name = "marshmallow" +version = "3.19.0" +description = "A lightweight library for converting complex datatypes to and from native Python datatypes." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "marshmallow-3.19.0-py3-none-any.whl", hash = "sha256:93f0958568da045b0021ec6aeb7ac37c81bfcccbb9a0e7ed8559885070b3a19b"}, + {file = "marshmallow-3.19.0.tar.gz", hash = "sha256:90032c0fd650ce94b6ec6dc8dfeb0e3ff50c144586462c389b81a07205bedb78"}, +] + +[package.dependencies] +packaging = ">=17.0" + +[package.extras] +dev = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)", "pytest", "pytz", "simplejson", "tox"] +docs = ["alabaster (==0.7.12)", "autodocsumm (==0.2.9)", "sphinx (==5.3.0)", "sphinx-issues (==3.0.1)", "sphinx-version-warning (==1.1.2)"] +lint = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)"] +tests = ["pytest", "pytz", "simplejson"] + +[[package]] +name = "mccabe" +version = "0.6.1" +description = "McCabe checker, plugin for flake8" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "mccabe-0.6.1-py2.py3-none-any.whl", hash = "sha256:ab8a6258860da4b6677da4bd2fe5dc2c659cff31b3ee4f7f5d64e79735b80d42"}, + {file = "mccabe-0.6.1.tar.gz", hash = "sha256:dd8d182285a0fe56bace7f45b5e7d1a6ebcbf524e8f3bd87eb0f125271b8831f"}, +] + +[[package]] +name = "mdurl" +version = "0.1.2" +description = "Markdown URL utilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8"}, + {file = "mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba"}, +] + +[[package]] +name = "mongo-types" +version = "0.15.1" +description = "Type stubs for mongoengine w/ basic support for bson and pymongo" +category = "main" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "mongo-types-0.15.1.tar.gz", hash = "sha256:0a9deeb7733ea7da5db3711d92e22d93556b522f860bbff82e5df44c53bd06a9"}, + {file = "mongo_types-0.15.1-py3-none-any.whl", hash = "sha256:9417ae5b9a759c09630b5ec7d66904cc333c2d2fcfe75e2760a332ed5e267309"}, +] + +[[package]] +name = "mongoengine" +version = "0.24.2" +description = "MongoEngine is a Python Object-Document Mapper for working with MongoDB." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "mongoengine-0.24.2-py3-none-any.whl", hash = "sha256:f5c4e1b206b2ccffe4adc7a6283ed26dd799bd115a5fb1d2e885a075132cdb88"}, + {file = "mongoengine-0.24.2.tar.gz", hash = "sha256:c76d49658575bb995682e2e77c8ef7cda63faf939415b32ee923745d120f8b02"}, +] + +[package.dependencies] +pymongo = ">=3.4,<5.0" + +[[package]] +name = "msgpack" +version = "1.0.5" +description = "MessagePack serializer" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:525228efd79bb831cf6830a732e2e80bc1b05436b086d4264814b4b2955b2fa9"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4f8d8b3bf1ff2672567d6b5c725a1b347fe838b912772aa8ae2bf70338d5a198"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cdc793c50be3f01106245a61b739328f7dccc2c648b501e237f0699fe1395b81"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb47c21a8a65b165ce29f2bec852790cbc04936f502966768e4aae9fa763cb7"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e42b9594cc3bf4d838d67d6ed62b9e59e201862a25e9a157019e171fbe672dd3"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:55b56a24893105dc52c1253649b60f475f36b3aa0fc66115bffafb624d7cb30b"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:1967f6129fc50a43bfe0951c35acbb729be89a55d849fab7686004da85103f1c"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:20a97bf595a232c3ee6d57ddaadd5453d174a52594bf9c21d10407e2a2d9b3bd"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d25dd59bbbbb996eacf7be6b4ad082ed7eacc4e8f3d2df1ba43822da9bfa122a"}, + {file = "msgpack-1.0.5-cp310-cp310-win32.whl", hash = "sha256:382b2c77589331f2cb80b67cc058c00f225e19827dbc818d700f61513ab47bea"}, + {file = "msgpack-1.0.5-cp310-cp310-win_amd64.whl", hash = "sha256:4867aa2df9e2a5fa5f76d7d5565d25ec76e84c106b55509e78c1ede0f152659a"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9f5ae84c5c8a857ec44dc180a8b0cc08238e021f57abdf51a8182e915e6299f0"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:9e6ca5d5699bcd89ae605c150aee83b5321f2115695e741b99618f4856c50898"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5494ea30d517a3576749cad32fa27f7585c65f5f38309c88c6d137877fa28a5a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ab2f3331cb1b54165976a9d976cb251a83183631c88076613c6c780f0d6e45a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:28592e20bbb1620848256ebc105fc420436af59515793ed27d5c77a217477705"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe5c63197c55bce6385d9aee16c4d0641684628f63ace85f73571e65ad1c1e8d"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:ed40e926fa2f297e8a653c954b732f125ef97bdd4c889f243182299de27e2aa9"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b2de4c1c0538dcb7010902a2b97f4e00fc4ddf2c8cda9749af0e594d3b7fa3d7"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf22a83f973b50f9d38e55c6aade04c41ddda19b00c4ebc558930d78eecc64ed"}, + {file = "msgpack-1.0.5-cp311-cp311-win32.whl", hash = "sha256:c396e2cc213d12ce017b686e0f53497f94f8ba2b24799c25d913d46c08ec422c"}, + {file = "msgpack-1.0.5-cp311-cp311-win_amd64.whl", hash = "sha256:6c4c68d87497f66f96d50142a2b73b97972130d93677ce930718f68828b382e2"}, + {file = "msgpack-1.0.5-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:a2b031c2e9b9af485d5e3c4520f4220d74f4d222a5b8dc8c1a3ab9448ca79c57"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f837b93669ce4336e24d08286c38761132bc7ab29782727f8557e1eb21b2080"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1d46dfe3832660f53b13b925d4e0fa1432b00f5f7210eb3ad3bb9a13c6204a6"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:366c9a7b9057e1547f4ad51d8facad8b406bab69c7d72c0eb6f529cf76d4b85f"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4c075728a1095efd0634a7dccb06204919a2f67d1893b6aa8e00497258bf926c"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:f933bbda5a3ee63b8834179096923b094b76f0c7a73c1cfe8f07ad608c58844b"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:36961b0568c36027c76e2ae3ca1132e35123dcec0706c4b7992683cc26c1320c"}, + {file = "msgpack-1.0.5-cp36-cp36m-win32.whl", hash = "sha256:b5ef2f015b95f912c2fcab19c36814963b5463f1fb9049846994b007962743e9"}, + {file = "msgpack-1.0.5-cp36-cp36m-win_amd64.whl", hash = "sha256:288e32b47e67f7b171f86b030e527e302c91bd3f40fd9033483f2cacc37f327a"}, + {file = "msgpack-1.0.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:137850656634abddfb88236008339fdaba3178f4751b28f270d2ebe77a563b6c"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c05a4a96585525916b109bb85f8cb6511db1c6f5b9d9cbcbc940dc6b4be944b"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:56a62ec00b636583e5cb6ad313bbed36bb7ead5fa3a3e38938503142c72cba4f"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef8108f8dedf204bb7b42994abf93882da1159728a2d4c5e82012edd92c9da9f"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1835c84d65f46900920b3708f5ba829fb19b1096c1800ad60bae8418652a951d"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:e57916ef1bd0fee4f21c4600e9d1da352d8816b52a599c46460e93a6e9f17086"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:17358523b85973e5f242ad74aa4712b7ee560715562554aa2134d96e7aa4cbbf"}, + {file = "msgpack-1.0.5-cp37-cp37m-win32.whl", hash = "sha256:cb5aaa8c17760909ec6cb15e744c3ebc2ca8918e727216e79607b7bbce9c8f77"}, + {file = "msgpack-1.0.5-cp37-cp37m-win_amd64.whl", hash = "sha256:ab31e908d8424d55601ad7075e471b7d0140d4d3dd3272daf39c5c19d936bd82"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:b72d0698f86e8d9ddf9442bdedec15b71df3598199ba33322d9711a19f08145c"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:379026812e49258016dd84ad79ac8446922234d498058ae1d415f04b522d5b2d"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:332360ff25469c346a1c5e47cbe2a725517919892eda5cfaffe6046656f0b7bb"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:476a8fe8fae289fdf273d6d2a6cb6e35b5a58541693e8f9f019bfe990a51e4ba"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a9985b214f33311df47e274eb788a5893a761d025e2b92c723ba4c63936b69b1"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:48296af57cdb1d885843afd73c4656be5c76c0c6328db3440c9601a98f303d87"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:addab7e2e1fcc04bd08e4eb631c2a90960c340e40dfc4a5e24d2ff0d5a3b3edb"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:916723458c25dfb77ff07f4c66aed34e47503b2eb3188b3adbec8d8aa6e00f48"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:821c7e677cc6acf0fd3f7ac664c98803827ae6de594a9f99563e48c5a2f27eb0"}, + {file = "msgpack-1.0.5-cp38-cp38-win32.whl", hash = "sha256:1c0f7c47f0087ffda62961d425e4407961a7ffd2aa004c81b9c07d9269512f6e"}, + {file = "msgpack-1.0.5-cp38-cp38-win_amd64.whl", hash = "sha256:bae7de2026cbfe3782c8b78b0db9cbfc5455e079f1937cb0ab8d133496ac55e1"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:20c784e66b613c7f16f632e7b5e8a1651aa5702463d61394671ba07b2fc9e025"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:266fa4202c0eb94d26822d9bfd7af25d1e2c088927fe8de9033d929dd5ba24c5"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:18334484eafc2b1aa47a6d42427da7fa8f2ab3d60b674120bce7a895a0a85bdd"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57e1f3528bd95cc44684beda696f74d3aaa8a5e58c816214b9046512240ef437"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:586d0d636f9a628ddc6a17bfd45aa5b5efaf1606d2b60fa5d87b8986326e933f"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a740fa0e4087a734455f0fc3abf5e746004c9da72fbd541e9b113013c8dc3282"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3055b0455e45810820db1f29d900bf39466df96ddca11dfa6d074fa47054376d"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:a61215eac016f391129a013c9e46f3ab308db5f5ec9f25811e811f96962599a8"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:362d9655cd369b08fda06b6657a303eb7172d5279997abe094512e919cf74b11"}, + {file = "msgpack-1.0.5-cp39-cp39-win32.whl", hash = "sha256:ac9dd47af78cae935901a9a500104e2dea2e253207c924cc95de149606dc43cc"}, + {file = "msgpack-1.0.5-cp39-cp39-win_amd64.whl", hash = "sha256:06f5174b5f8ed0ed919da0e62cbd4ffde676a374aba4020034da05fab67b9164"}, + {file = "msgpack-1.0.5.tar.gz", hash = "sha256:c075544284eadc5cddc70f4757331d99dcbc16b2bbd4849d15f8aae4cf36d31c"}, +] + +[[package]] +name = "mypy" +version = "1.1.1" +description = "Optional static typing for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mypy-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39c7119335be05630611ee798cc982623b9e8f0cff04a0b48dfc26100e0b97af"}, + {file = "mypy-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:61bf08362e93b6b12fad3eab68c4ea903a077b87c90ac06c11e3d7a09b56b9c1"}, + {file = "mypy-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dbb19c9f662e41e474e0cff502b7064a7edc6764f5262b6cd91d698163196799"}, + {file = "mypy-1.1.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:315ac73cc1cce4771c27d426b7ea558fb4e2836f89cb0296cbe056894e3a1f78"}, + {file = "mypy-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:5cb14ff9919b7df3538590fc4d4c49a0f84392237cbf5f7a816b4161c061829e"}, + {file = "mypy-1.1.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:26cdd6a22b9b40b2fd71881a8a4f34b4d7914c679f154f43385ca878a8297389"}, + {file = "mypy-1.1.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5b5f81b40d94c785f288948c16e1f2da37203c6006546c5d947aab6f90aefef2"}, + {file = "mypy-1.1.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21b437be1c02712a605591e1ed1d858aba681757a1e55fe678a15c2244cd68a5"}, + {file = "mypy-1.1.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d809f88734f44a0d44959d795b1e6f64b2bbe0ea4d9cc4776aa588bb4229fc1c"}, + {file = "mypy-1.1.1-cp311-cp311-win_amd64.whl", hash = "sha256:a380c041db500e1410bb5b16b3c1c35e61e773a5c3517926b81dfdab7582be54"}, + {file = "mypy-1.1.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:b7c7b708fe9a871a96626d61912e3f4ddd365bf7f39128362bc50cbd74a634d5"}, + {file = "mypy-1.1.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c1c10fa12df1232c936830839e2e935d090fc9ee315744ac33b8a32216b93707"}, + {file = "mypy-1.1.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:0a28a76785bf57655a8ea5eb0540a15b0e781c807b5aa798bd463779988fa1d5"}, + {file = "mypy-1.1.1-cp37-cp37m-win_amd64.whl", hash = "sha256:ef6a01e563ec6a4940784c574d33f6ac1943864634517984471642908b30b6f7"}, + {file = "mypy-1.1.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d64c28e03ce40d5303450f547e07418c64c241669ab20610f273c9e6290b4b0b"}, + {file = "mypy-1.1.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:64cc3afb3e9e71a79d06e3ed24bb508a6d66f782aff7e56f628bf35ba2e0ba51"}, + {file = "mypy-1.1.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce61663faf7a8e5ec6f456857bfbcec2901fbdb3ad958b778403f63b9e606a1b"}, + {file = "mypy-1.1.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:2b0c373d071593deefbcdd87ec8db91ea13bd8f1328d44947e88beae21e8d5e9"}, + {file = "mypy-1.1.1-cp38-cp38-win_amd64.whl", hash = "sha256:2888ce4fe5aae5a673386fa232473014056967f3904f5abfcf6367b5af1f612a"}, + {file = "mypy-1.1.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:19ba15f9627a5723e522d007fe708007bae52b93faab00f95d72f03e1afa9598"}, + {file = "mypy-1.1.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:59bbd71e5c58eed2e992ce6523180e03c221dcd92b52f0e792f291d67b15a71c"}, + {file = "mypy-1.1.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9401e33814cec6aec8c03a9548e9385e0e228fc1b8b0a37b9ea21038e64cdd8a"}, + {file = "mypy-1.1.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4b398d8b1f4fba0e3c6463e02f8ad3346f71956b92287af22c9b12c3ec965a9f"}, + {file = "mypy-1.1.1-cp39-cp39-win_amd64.whl", hash = "sha256:69b35d1dcb5707382810765ed34da9db47e7f95b3528334a3c999b0c90fe523f"}, + {file = "mypy-1.1.1-py3-none-any.whl", hash = "sha256:4e4e8b362cdf99ba00c2b218036002bdcdf1e0de085cdb296a49df03fb31dfc4"}, + {file = "mypy-1.1.1.tar.gz", hash = "sha256:ae9ceae0f5b9059f33dbc62dea087e942c0ccab4b7a003719cb70f9b8abfa32f"}, +] + +[package.dependencies] +mypy-extensions = ">=1.0.0" +tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""} +typing-extensions = ">=3.10" + +[package.extras] +dmypy = ["psutil (>=4.0)"] +install-types = ["pip"] +python2 = ["typed-ast (>=1.4.0,<2)"] +reports = ["lxml"] + +[[package]] +name = "mypy-extensions" +version = "1.0.0" +description = "Type system extensions for programs checked with the mypy type checker." +category = "dev" +optional = false +python-versions = ">=3.5" +files = [ + {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"}, + {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"}, +] + +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + +[[package]] +name = "orjson" +version = "3.8.9" +description = "Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "orjson-3.8.9-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:5d029843eae7b6cbd6468b63517b8b61471afed6572162171d8b6471b6dbf41f"}, + {file = "orjson-3.8.9-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:405933c05490efb209d0f940d8ef1403d2932a97e47010a26d2694e9dd49f84d"}, + {file = "orjson-3.8.9-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:183de66eff4d41c330a3006f210ab0bce7affe398da6f6eda9579b67245a34ff"}, + {file = "orjson-3.8.9-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bb4081fe340ed1df42dddfd055e1d50479cb0ccb976d13e6b5e8667a07fec6f4"}, + {file = "orjson-3.8.9-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d11593a2e736055dd7b9587dbf89cd1cbe4a42a70e70f186e51aee7e1b38902e"}, + {file = "orjson-3.8.9-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e20649359e28f34d01b2570e4650a076f439a959bae3a8bbe7f5923ad80f54e8"}, + {file = "orjson-3.8.9-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:c02ece4f36a160c83efe74adfba5f189c7c7702361f02b809ab73744923ee139"}, + {file = "orjson-3.8.9-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:f0e19801836cf1b30f333d475b05d79051b8ae8639a8e2422fb5f64e82676ae7"}, + {file = "orjson-3.8.9-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d4850fe5650cead3c0f8822192e381cee9d4c3b8162eb082c86c927124572dc6"}, + {file = "orjson-3.8.9-cp310-none-win_amd64.whl", hash = "sha256:5fd4193f260d9d30112b5e379d0870b54dc88040807c93cbe8d67bfea148ba5a"}, + {file = "orjson-3.8.9-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:70eae063ad8d7405dc63873760567b600fc10728ba0da24a69d49c1a5d318d6d"}, + {file = "orjson-3.8.9-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:251653437632583d02203e6b118b72b99c04425175853f35340f4bac7034a36e"}, + {file = "orjson-3.8.9-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6ea833751f017ba321c277e7425b51c0b1a18a2c60f8c9c0f4c6c4d7e16cbd6c"}, + {file = "orjson-3.8.9-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:8563c2cdeb923b82a5cc5bfc76c28c786777428263ee39292d928e9687165fb4"}, + {file = "orjson-3.8.9-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6f33e9ea45b4c9457eedca0c40f38cf5732c91b0fb68f091ac59e6ea68e03eb2"}, + {file = "orjson-3.8.9-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:855dee152daecb7de7b4cd7069d7854e11aa291687bffe8433156af0a224417e"}, + {file = "orjson-3.8.9-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:74fa9e02589339defc9d3662de9e7eef51d8f9f3a7f6304b43b18b39d7bbf10f"}, + {file = "orjson-3.8.9-cp311-none-win_amd64.whl", hash = "sha256:6c5b10ba1e62df8f96cbc37f6d5ae9acb3f6475926dea8b1b6a1a60f201a64f7"}, + {file = "orjson-3.8.9-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:a651123d01bc399fcd866e56acc2d76512e62aae3673652b13b470ea69faf1f4"}, + {file = "orjson-3.8.9-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:73019b6d2cc998c99556020c6bd8f8bc28420c69583186ca290c66a27916a3b7"}, + {file = "orjson-3.8.9-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8f5c3daa8b02786ad5f0e14ae16a59bbb4e02cbae3a41989a25188e5a6c962ff"}, + {file = "orjson-3.8.9-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:598598b7f81f8fda7c3e09c88165f844152b7be223bc4ea929ec8ad59b00ea17"}, + {file = "orjson-3.8.9-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:090b10bdb06baae6d5cd3550d772ecbabd833bfceed7592ff167c0a82f5b4c20"}, + {file = "orjson-3.8.9-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bd46f688ddf9c2ea10367446fe9bf3ceba0f7490c15b4f96420491c7f00bb283"}, + {file = "orjson-3.8.9-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:b8ed8d780e9fab01bc404a70d755a8b2b34ea6c0b6604b65de135daaaadaf9a9"}, + {file = "orjson-3.8.9-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:8a32c9fb742868a34346f3c52e12d893a9d27f8e0c0bf3c480db7e6903d8be28"}, + {file = "orjson-3.8.9-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:2ba366009b98ac8899e935eff6fef7672d3ea43d3ce9deb3ee33452134b6cc3a"}, + {file = "orjson-3.8.9-cp37-none-win_amd64.whl", hash = "sha256:236b9313425cb2570626c64dd5cb6caff13882d1717d491da542cff228b96e97"}, + {file = "orjson-3.8.9-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:e8efc7e9ec35336f7cc98b6692536b1262046ff1d2a545295a4d89b8a2495903"}, + {file = "orjson-3.8.9-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:8c7eba3610ae69f4aba4032ecb61b0a6fbd1e4537283d1553eb8c1cb136e9118"}, + {file = "orjson-3.8.9-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7742649e4c357d4e7ad483a35ff5f55d519e895de56772cc486913614ee7d23b"}, + {file = "orjson-3.8.9-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:b6566fb8daa538c7848fd6822e2409a7e1c41dae8e65e6536598d505f641a318"}, + {file = "orjson-3.8.9-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:0ce8a2a667221e2e5160021e26b09e9c13eeedafb5cda1981340c8c0c0bc8f9d"}, + {file = "orjson-3.8.9-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4c0399631b88fa4868956badef2561fba07dffcaf050bf53959ee50d26edf6f6"}, + {file = "orjson-3.8.9-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:189ccb16ed140a824d133fa1c55175cf0d2207edaade54f1db0456a526cb5fd8"}, + {file = "orjson-3.8.9-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:b707fa4481e1af19b3052ec9352c688bad3f539d7bdd8aa4a451f6dd7e4bae73"}, + {file = "orjson-3.8.9-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:c3d988eb562da1dda7d49e9abd8a64b3cabc632b4299d177fb9e0c0ca9f06b8c"}, + {file = "orjson-3.8.9-cp38-none-win_amd64.whl", hash = "sha256:b30240eb6b22daab604f1595f6aacf92bcdac0d29e2d7ad507dfac68d2b39182"}, + {file = "orjson-3.8.9-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:81869a6de00bc676d10056fa8bb28cbe805b1cf498a45c14cb7b1765eee33fcb"}, + {file = "orjson-3.8.9-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:a25a5a215b19d414de8d416a3c5414f29165843a06f704cc0345ded9eac34ac1"}, + {file = "orjson-3.8.9-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:dec0f2bea52e30ea98ce095f1f42da04535791f9a31b2aab2499caa88307bc49"}, + {file = "orjson-3.8.9-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7b91d88fe96b698b28bb1b95b1fce226f72757ab3ab7d8d97551e23bc629c84f"}, + {file = "orjson-3.8.9-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7629841ccdcccd3c43ebc6a4165abe9844909fcedb2041994c0153470f610801"}, + {file = "orjson-3.8.9-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d875b304e19f4b2758d233bbf2b9d627c66fac50b3150b8d31a35ba6cda3db67"}, + {file = "orjson-3.8.9-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:723ec880c5290fe4de330febb8030e57c1978fbd624fc5b9399969e7d7d74984"}, + {file = "orjson-3.8.9-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b11f8a71c82d19fce11ce487efeec2ca0dc3bcf5b4564445fecfc68d9c268744"}, + {file = "orjson-3.8.9-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:b2079bf86dec62731c1b90fdfea3211f993f0c894d9261e0ce9b68ed9c9dfbec"}, + {file = "orjson-3.8.9-cp39-none-win_amd64.whl", hash = "sha256:97d94322a2eaab767ba8d52f6bf9d0ec0f35313fe36287be6e6085dd65d55d37"}, + {file = "orjson-3.8.9.tar.gz", hash = "sha256:c40bece58c11cb09aff17424d21b41f6f767d2b1252b2f745ec3ff29cce6a240"}, +] + +[[package]] +name = "packageurl-python" +version = "0.11.1" +description = "A purl aka. Package URL parser and builder" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "packageurl-python-0.11.1.tar.gz", hash = "sha256:bbcc53d2cb5920c815c1626c75992f319bfc450b73893fa7bd8aac5869aa49fe"}, + {file = "packageurl_python-0.11.1-py3-none-any.whl", hash = "sha256:4bad1d3ea4feb5e7a1db5ca8fb690ac9c82ab18e08d500755947b853df68817d"}, +] + +[package.extras] +build = ["wheel"] +lint = ["black", "isort", "mypy"] +test = ["pytest"] + +[[package]] +name = "packaging" +version = "23.0" +description = "Core utilities for Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "packaging-23.0-py3-none-any.whl", hash = "sha256:714ac14496c3e68c99c29b00845f7a2b85f3bb6f1078fd9f72fd20f0570002b2"}, + {file = "packaging-23.0.tar.gz", hash = "sha256:b6ad297f8907de0fa2fe1ccbd26fdaf387f5f47c7275fedf8cce89f99446cf97"}, +] + +[[package]] +name = "pathspec" +version = "0.11.1" +description = "Utility library for gitignore style pattern matching of file paths." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pathspec-0.11.1-py3-none-any.whl", hash = "sha256:d8af70af76652554bd134c22b3e8a1cc46ed7d91edcdd721ef1a0c51a84a5293"}, + {file = "pathspec-0.11.1.tar.gz", hash = "sha256:2798de800fa92780e33acca925945e9a19a133b715067cf165b8866c15a31687"}, +] + +[[package]] +name = "pbr" +version = "5.11.1" +description = "Python Build Reasonableness" +category = "dev" +optional = false +python-versions = ">=2.6" +files = [ + {file = "pbr-5.11.1-py2.py3-none-any.whl", hash = "sha256:567f09558bae2b3ab53cb3c1e2e33e726ff3338e7bae3db5dc954b3a44eef12b"}, + {file = "pbr-5.11.1.tar.gz", hash = "sha256:aefc51675b0b533d56bb5fd1c8c6c0522fe31896679882e1c4c63d5e4a0fccb3"}, +] + +[[package]] +name = "pip" +version = "23.0.1" +description = "The PyPA recommended tool for installing Python packages." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"}, + {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"}, +] + +[[package]] +name = "pip-api" +version = "0.0.30" +description = "An unofficial, importable pip API" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-api-0.0.30.tar.gz", hash = "sha256:a05df2c7aa9b7157374bcf4273544201a0c7bae60a9c65bcf84f3959ef3896f3"}, + {file = "pip_api-0.0.30-py3-none-any.whl", hash = "sha256:2a0314bd31522eb9ffe8a99668b0d07fee34ebc537931e7b6483001dbedcbdc9"}, +] + +[package.dependencies] +pip = "*" + +[[package]] +name = "pip-audit" +version = "2.4.14" +description = "A tool for scanning Python environments for known vulnerabilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, + {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, +] + +[package.dependencies] +CacheControl = {version = ">=0.12.10", extras = ["filecache"]} +cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +html5lib = ">=1.1" +packaging = ">=23.0.0" +pip-api = ">=0.0.28" +pip-requirements-parser = ">=32.0.0" +resolvelib = ">=0.8.0" +rich = ">=12.4" +toml = ">=0.10" + +[package.extras] +dev = ["build", "bump (>=1.3.2)", "pip-audit[doc,lint,test]"] +doc = ["pdoc"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +test = ["coverage[toml]", "pretend", "pytest", "pytest-cov"] + +[[package]] +name = "pip-requirements-parser" +version = "32.0.1" +description = "pip requirements parser - a mostly correct pip requirements parsing library because it uses pip's own code." +category = "dev" +optional = false +python-versions = ">=3.6.0" +files = [ + {file = "pip-requirements-parser-32.0.1.tar.gz", hash = "sha256:b4fa3a7a0be38243123cf9d1f3518da10c51bdb165a2b2985566247f9155a7d3"}, + {file = "pip_requirements_parser-32.0.1-py3-none-any.whl", hash = "sha256:4659bc2a667783e7a15d190f6fccf8b2486685b6dba4c19c3876314769c57526"}, +] + +[package.dependencies] +packaging = "*" +pyparsing = "*" + +[package.extras] +docs = ["Sphinx (>=3.3.1)", "doc8 (>=0.8.1)", "sphinx-rtd-theme (>=0.5.0)"] +testing = ["aboutcode-toolkit (>=6.0.0)", "black", "pytest (>=6,!=7.0.0)", "pytest-xdist (>=2)"] + +[[package]] +name = "platformdirs" +version = "3.2.0" +description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "platformdirs-3.2.0-py3-none-any.whl", hash = "sha256:ebe11c0d7a805086e99506aa331612429a72ca7cd52a1f0d277dc4adc20cb10e"}, + {file = "platformdirs-3.2.0.tar.gz", hash = "sha256:d5b638ca397f25f979350ff789db335903d7ea010ab28903f57b27e1b16c2b08"}, +] + +[package.extras] +docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"] +test = ["appdirs (==1.4.4)", "covdefaults (>=2.3)", "pytest (>=7.2.2)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"] + +[[package]] +name = "pluggy" +version = "1.0.0" +description = "plugin and hook calling mechanisms for python" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pluggy-1.0.0-py2.py3-none-any.whl", hash = "sha256:74134bbf457f031a36d68416e1509f34bd5ccc019f0bcc952c7b909d06b37bd3"}, + {file = "pluggy-1.0.0.tar.gz", hash = "sha256:4224373bacce55f955a878bf9cfa763c1e360858e330072059e10bad68531159"}, +] + +[package.extras] +dev = ["pre-commit", "tox"] +testing = ["pytest", "pytest-benchmark"] + +[[package]] +name = "psutil" +version = "5.9.4" +description = "Cross-platform lib for process and system monitoring in Python." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "psutil-5.9.4-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:c1ca331af862803a42677c120aff8a814a804e09832f166f226bfd22b56feee8"}, + {file = "psutil-5.9.4-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:68908971daf802203f3d37e78d3f8831b6d1014864d7a85937941bb35f09aefe"}, + {file = "psutil-5.9.4-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:3ff89f9b835100a825b14c2808a106b6fdcc4b15483141482a12c725e7f78549"}, + {file = "psutil-5.9.4-cp27-cp27m-win32.whl", hash = "sha256:852dd5d9f8a47169fe62fd4a971aa07859476c2ba22c2254d4a1baa4e10b95ad"}, + {file = "psutil-5.9.4-cp27-cp27m-win_amd64.whl", hash = "sha256:9120cd39dca5c5e1c54b59a41d205023d436799b1c8c4d3ff71af18535728e94"}, + {file = "psutil-5.9.4-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:6b92c532979bafc2df23ddc785ed116fced1f492ad90a6830cf24f4d1ea27d24"}, + {file = "psutil-5.9.4-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:efeae04f9516907be44904cc7ce08defb6b665128992a56957abc9b61dca94b7"}, + {file = "psutil-5.9.4-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:54d5b184728298f2ca8567bf83c422b706200bcbbfafdc06718264f9393cfeb7"}, + {file = "psutil-5.9.4-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16653106f3b59386ffe10e0bad3bb6299e169d5327d3f187614b1cb8f24cf2e1"}, + {file = "psutil-5.9.4-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:54c0d3d8e0078b7666984e11b12b88af2db11d11249a8ac8920dd5ef68a66e08"}, + {file = "psutil-5.9.4-cp36-abi3-win32.whl", hash = "sha256:149555f59a69b33f056ba1c4eb22bb7bf24332ce631c44a319cec09f876aaeff"}, + {file = "psutil-5.9.4-cp36-abi3-win_amd64.whl", hash = "sha256:fd8522436a6ada7b4aad6638662966de0d61d241cb821239b2ae7013d41a43d4"}, + {file = "psutil-5.9.4-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:6001c809253a29599bc0dfd5179d9f8a5779f9dffea1da0f13c53ee568115e1e"}, + {file = "psutil-5.9.4.tar.gz", hash = "sha256:3d7f9739eb435d4b1338944abe23f49584bde5395f27487d2ee25ad9a8774a62"}, +] + +[package.extras] +test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] + +[[package]] +name = "pycodestyle" +version = "2.7.0" +description = "Python style guide checker" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycodestyle-2.7.0-py2.py3-none-any.whl", hash = "sha256:514f76d918fcc0b55c6680472f0a37970994e07bbb80725808c17089be302068"}, + {file = "pycodestyle-2.7.0.tar.gz", hash = "sha256:c389c1d06bf7904078ca03399a4816f974a1d590090fecea0c63ec26ebaf1cef"}, +] + +[[package]] +name = "pyflakes" +version = "2.3.1" +description = "passive checker of Python programs" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pyflakes-2.3.1-py2.py3-none-any.whl", hash = "sha256:7893783d01b8a89811dd72d7dfd4d84ff098e5eed95cfa8905b22bbffe52efc3"}, + {file = "pyflakes-2.3.1.tar.gz", hash = "sha256:f5bc8ecabc05bb9d291eb5203d6810b49040f6ff446a756326104746cc00c1db"}, +] + +[[package]] +name = "pygments" +version = "2.14.0" +description = "Pygments is a syntax highlighting package written in Python." +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "Pygments-2.14.0-py3-none-any.whl", hash = "sha256:fa7bd7bd2771287c0de303af8bfdfc731f51bd2c6a47ab69d117138893b82717"}, + {file = "Pygments-2.14.0.tar.gz", hash = "sha256:b3ed06a9e8ac9a9aae5a6f5dbe78a8a58655d17b43b93c078f094ddc476ae297"}, +] + +[package.extras] +plugins = ["importlib-metadata"] + +[[package]] +name = "pymongo" +version = "3.13.0" +description = "Python driver for MongoDB <http://www.mongodb.org>" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl", hash = "sha256:3ad3a3df830f7df7e0856c2bdb54d19f5bf188bd7420985e18643b8e4d2a075f"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:b96e0e9d2d48948240b510bac81614458fc10adcd3a93240c2fd96448b4efd35"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9f592b202d77923498b32ddc5b376e5fa9ba280d3e16ed56cb8c932fe6d6a478"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:851f2bb52b5cb2f4711171ca925e0e05344a8452972a748a8a8ffdda1e1d72a7"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:1c9d23f62a3fa7523d849c4942acc0d9ff7081ebc00c808ee7cfdc070df0687f"}, + {file = "pymongo-3.13.0-cp27-cp27m-win32.whl", hash = "sha256:a17b81f22398e3e0f72bdf938e98c810286994b2bcc0a125cd5ad8fd4ea54ad7"}, + {file = "pymongo-3.13.0-cp27-cp27m-win_amd64.whl", hash = "sha256:4f6dd55dab77adf60b445c11f426ee5cdfa1b86f6d54cb937bfcbf09572333ab"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:776f90bf2252f90a4ae838e7917638894c6356bef7265f424592e2fd1f577d05"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:50b99f4d3eee6f03778fe841d6f470e6c18e744dc665156da6da3bc6e65b398d"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50a81b2d9f188c7909e0a1084fa969bb92a788076809c437ac1ae80393f46df9"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c7c45a8a1a752002b0a7c81ab3a4c5e3b6f67f9826b16fbe3943f5329f565f24"}, + {file = "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1037097708498bdc85f23c8798a5c46c7bce432d77d23608ff14e0d831f1a971"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl", hash = "sha256:b5b733694e7df22d5c049581acfc487695a6ff813322318bed8dd66f79978636"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:d7c91747ec8dde51440dd594603158cc98abb3f7df84b2ed8a836f138285e4fb"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl", hash = "sha256:f4175fcdddf764d371ee52ec4505a40facee2533e84abf2953cda86d050cfa1f"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl", hash = "sha256:93d4e9a02c17813b34e4bd9f6fbf07310c140c8f74341537c24d07c1cdeb24d1"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl", hash = "sha256:3b261d593f2563299062733ae003a925420a86ff4ddda68a69097d67204e43f3"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl", hash = "sha256:172db03182a22e9002157b262c1ea3b0045c73d4ff465adc152ce5b4b0e7b8d4"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09de3bfc995ae8cb955abb0c9ae963c134dba1b5622be3bcc527b89b0fd4091c"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c0379447587ee4b8f983ba183202496e86c0358f47c45612619d634d1fcd82bd"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30245a8747dc90019a3c9ad9df987e0280a3ea632ad36227cde7d1d8dcba0830"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65b6fddf6a7b91da044f202771a38e71bbb9bf42720a406b26b25fe2256e7102"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5831a377d15a626fbec10890ffebc4c6abcd37e4126737932cd780a171eabdc1"}, + {file = "pymongo-3.13.0-cp310-cp310-win32.whl", hash = "sha256:944249aa83dee314420c37d0f40c30a8f6dc4a3877566017b87062e53af449f4"}, + {file = "pymongo-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea8824ebc9a1a5c8269e8f1e3989b5a6bec876726e2f3c33ebd036cb488277f0"}, + {file = "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:bdd34c57b4da51a7961beb33645646d197e41f8517801dc76b37c1441e7a4e10"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26f9cc42a162faa241c82e117ac85734ae9f14343dc2df1c90c6b2181f791b22"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a82a1c10f5608e6494913faa169e213d703194bfca0aa710901f303be212414"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8927f22ef6a16229da7f18944deac8605bdc2c0858be5184259f2f7ce7fd4459"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e6f8191a282ef77e526f8f8f63753a437e4aa4bc78f5edd8b6b6ed0eaebd5363"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4d9ed67c987bf9ac2ac684590ba3d2599cdfb0f331ee3db607f9684469b3b59d"}, + {file = "pymongo-3.13.0-cp311-cp311-win32.whl", hash = "sha256:e8f6979664ff477cd61b06bf8aba206df7b2334209815ab3b1019931dab643d6"}, + {file = "pymongo-3.13.0-cp311-cp311-win_amd64.whl", hash = "sha256:174fd1000e896d0dfbc7f6d7e6a1992a4868796c7dec31679e38218c78d6a942"}, + {file = "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl", hash = "sha256:d1ee773fb72ba024e7e3bb6ea8907fe52bccafcb5184aaced6bad995bd30ea20"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:28565e3dbd69fe5fe35a210067064dbb6ed5abe997079f653c19c873c3896fe6"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:5c1db7d366004d6c699eb08c716a63ae0a3e946d061cbebea65d7ce361950265"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e1956f3338c10308e2f99c2c9ff46ae412035cbcd7aaa76c39ccdb806854a247"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:10f0fddc1d63ba3d4a4bffcc7720184c1b7efd570726ad5e2f55818da320239f"}, + {file = "pymongo-3.13.0-cp35-cp35m-win32.whl", hash = "sha256:570ae3365b23d4fd8c669cb57613b1a90b2757e993588d3370ef90945dbeec4b"}, + {file = "pymongo-3.13.0-cp35-cp35m-win_amd64.whl", hash = "sha256:79f777eaf3f5b2c6d81f9ef00d87837001d7063302503bbcbfdbf3e9bc27c96f"}, + {file = "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl", hash = "sha256:d42eb29ba314adfd9c11234b4b646f61b0448bf9b00f14db4b317e6e4b947e77"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:e5e87c0eb774561c546f979342a8ff36ebee153c60a0b6c6b03ba989ceb9538c"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:0f2c5a5984599a88d087a15859860579b825098b473d8c843f1979a83d159f2e"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:59c98e86c5e861032b71e6e5b65f23e6afaacea6e82483b66f1191a5021a7b4f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:70b67390e27e58876853efbb87e43c85252de2515e2887f7dd901b4fa3d21973"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:42ba8606492d76e6f9e4c7a458ed4bc712603be393259a52450345f0945da2cf"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:0e5536994cf2d8488c6fd9dea71df3c4dbb3e0d2ba5e695da06d9142a29a0969"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:fe8194f107f0fa3cabd14e9e809f174eca335993c1db72d1e74e0f496e7afe1f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d593d50815771f517d3ac4367ff716e3f3c78edae51d98e1e25791459f8848ff"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5136ebe8da6a1604998a8eb96be55935aa5f7129c41cc7bddc400d48e8df43be"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a424bdedfd84454d2905a861e0d4bb947cc5bd024fdeb3600c1a97d2be0f4255"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5161167b3840e9c84c80f2534ea6a099f51749d5673b662a3dd248be17c3208"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:644470442beaf969df99c4e00367a817eee05f0bba5d888f1ba6fe97b5e1c102"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2406df90b2335371706c59b7d79e9633b81ed2a7ecd48c1faf8584552bdf2d90"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:222591b828de10ac90064047b5d4916953f38c38b155009c4b8b5e0d33117c2b"}, + {file = "pymongo-3.13.0-cp36-cp36m-win32.whl", hash = "sha256:7cb987b199fa223ad78eebaa9fbc183d5a5944bfe568a9d6f617316ca1c1f32f"}, + {file = "pymongo-3.13.0-cp36-cp36m-win_amd64.whl", hash = "sha256:a6cbb73d9fc2282677e2b7a137d13da987bd0b13abd88ed27bba5534c226db06"}, + {file = "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl", hash = "sha256:b1223b826acbef07a7f5eb9bf37247b0b580119916dca9eae19d92b1290f5855"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:398fb86d374dc351a4abc2e24cd15e5e14b2127f6d90ce0df3fdf2adcc55ac1b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:9c3d07ea19cd2856d9943dce37e75d69ecbb5baf93c3e4c82f73b6075c481292"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:2943d739715f265a2983ac43747595b6af3312d0a370614040959fd293763adf"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:c3b70ed82f20d18d22eafc9bda0ea656605071762f7d31f3c5afc35c59d3393b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:7ec2bb598847569ae34292f580842d37619eea3e546005042f485e15710180d5"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:8cc37b437cba909bef06499dadd91a39c15c14225e8d8c7870020049f8a549fe"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:65a063970e15a4f338f14b820561cf6cdaf2839691ac0adb2474ddff9d0b8b0b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02f0e1a75d3bc0e16c7e15daf9c56185642be055e425f3b34888fc6eb1b22401"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16e74b9c2aca2734c7f49f00fe68d6830a30d26df60e2ace7fe40ccb92087b94"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:24e954be35ad4537840f20bbc8d75320ae647d3cb4fab12cb8fcd2d55f408e76"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a149377d1ff766fd618500798d0d94637f66d0ae222bb6d28f41f3e15c626297"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61660710b054ae52c8fc10368e91d74719eb05554b631d7f8ca93d21d2bff2e6"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4bbc0d27dfef7689285e54f2e0a224f0c7cd9d5c46d2638fabad5500b951c92f"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:9b2ed9c3b30f11cd4a3fbfc22167af7987b01b444215c2463265153fe7cf66d6"}, + {file = "pymongo-3.13.0-cp37-cp37m-win32.whl", hash = "sha256:1c2c5e2b00e2fadcd590c0b2e293d71215e98ed1cb635cfca2be4998d197e534"}, + {file = "pymongo-3.13.0-cp37-cp37m-win_amd64.whl", hash = "sha256:32eac95bbb030b2376ffd897376c6f870222a3457f01a9ce466b9057876132f8"}, + {file = "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a796ef39dadf9d73af05d24937644d386495e43a7d13617aa3651d836da542c8"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:b6793baf4639c72a500698a49e9250b293e17ae1faf11ac1699d8141194786fe"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:80d8576b04d0824f63bf803190359c0d3bcb6e7fa63fefbd4bc0ceaa7faae38c"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:db2e11507fe9cc2a722be21ccc62c1b1295398fe9724c1f14900cdc7166fc0d7"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:b01ce58eec5edeededf1992d2dce63fb8565e437be12d6f139d75b15614c4d08"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:d1a19d6c5098f1f4e11430cd74621699453cbc534dd7ade9167e582f50814b19"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:7219b1a726ced3bacecabef9bd114529bbb69477901373e800d7d0140baadc95"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:2dae3b353a10c3767e0aa1c1492f2af388f1012b08117695ab3fd1f219e5814e"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:12721d926d43d33dd3318e58dce9b0250e8a9c6e1093fa8e09f4805193ff4b43"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6af0a4b17faf26779d5caee8542a4f2cba040cea27d3bffc476cbc6ccbd4c8ee"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:09b9d0f5a445c7e0ddcc021b09835aa6556f0166afc498f57dfdd72cdf6f02ad"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db5b4f8ad8607a3d612da1d4c89a84e4cf5c88f98b46365820d9babe5884ba45"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:34dbf5fecf653c152edb75a35a8b15dfdc4549473484ee768aeb12c97983cead"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:34cd48df7e1fc69222f296d8f69e3957eb7c6b5aa0709d3467184880ed7538c0"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c8f755ff1f4ab4ca790d1d6d3229006100b301475948021b6b2757822e0d6c97"}, + {file = "pymongo-3.13.0-cp38-cp38-win32.whl", hash = "sha256:b0746d0d4535f56bbaa63a8f6da362f330804d578e66e126b226eebe76c2bf00"}, + {file = "pymongo-3.13.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ad0515abb132f52ce9d8abd1a29681a1e65dba7b7fe13ea01e1a8db5715bf80"}, + {file = "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:3c5cb6c93c94df76a879bad4b89db0104b01806d17c2b803c1316ba50962b6d6"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2e0854170813238f0c3131050c67cb1fb1ade75c93bf6cd156c1bd9a16095528"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:1410faa51ce835cc1234c99ec42e98ab4f3c6f50d92d86a2d4f6e11c97ee7a4e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:d7910135f5de1c5c3578e61d6f4b087715b15e365f11d4fa51a9cee92988b2bd"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:028175dd8d2979a889153a2308e8e500b3df7d9e3fd1c33ca7fdeadf61cc87a2"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:2bfc39276c0e6d07c95bd1088b5003f049e986e089509f7dbd68bb7a4b1e65ac"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:4092b660ec720d44d3ca81074280dc25c7a3718df1b6c0fe9fe36ac6ed2833e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:5bdeb71a610a7b801416268e500e716d0fe693fb10d809e17f0fb3dac5be5a34"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aa3bca8e76f5c00ed2bb4325e0e383a547d71595926d5275d7c88175aaf7435e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7c7cab8155f430ca460a6fc7ae8a705b34f3e279a57adb5f900eb81943ec777c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4a32f3dfcca4a4816373bdb6256c18c78974ebb3430e7da988516cd95b2bd6e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30ed2788a6ec68743e2040ab1d16573d7d9f6e7333e45070ce9268cbc93d148c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21e61a536ffed84d10376c21c13a6ed1ebefb61989a844952547c229d6aeedf3"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0665412dce26b2318092a33bd2d2327d487c4490cfcde158d6946d39b1e28d78"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:64ed1a5ce5e5926727eb0f87c698c4d9a7a9f7b0953683a65e9ce2b7cc5f8e91"}, + {file = "pymongo-3.13.0-cp39-cp39-win32.whl", hash = "sha256:7593cb1214185a0c5b43b96effc51ce82ddc933298ee36db7dc2bd45d61b4adc"}, + {file = "pymongo-3.13.0-cp39-cp39-win_amd64.whl", hash = "sha256:3cfc9bc1e8b5667bc1f3dbe46d2f85b3f24ff7533893bdc1203058012db2c046"}, + {file = "pymongo-3.13.0.tar.gz", hash = "sha256:e22d6cf5802cd09b674c307cc9e03870b8c37c503ebec3d25b86f2ce8c535dc7"}, +] + +[package.dependencies] +dnspython = {version = ">=1.16.0,<1.17.0", optional = true, markers = "extra == \"srv\""} + +[package.extras] +aws = ["pymongo-auth-aws (<2.0.0)"] +encryption = ["pymongocrypt (>=1.1.0,<2.0.0)"] +gssapi = ["pykerberos"] +ocsp = ["certifi", "pyopenssl (>=17.2.0)", "requests (<3.0.0)", "service-identity (>=18.1.0)"] +snappy = ["python-snappy"] +srv = ["dnspython (>=1.16.0,<1.17.0)"] +tls = ["ipaddress"] +zstd = ["zstandard"] + +[[package]] +name = "pyparsing" +version = "3.0.9" +description = "pyparsing module - Classes and methods to define and execute parsing grammars" +category = "dev" +optional = false +python-versions = ">=3.6.8" +files = [ + {file = "pyparsing-3.0.9-py3-none-any.whl", hash = "sha256:5026bae9a10eeaefb61dab2f09052b9f4307d44aee4eda64b309723d8d206bbc"}, + {file = "pyparsing-3.0.9.tar.gz", hash = "sha256:2b020ecf7d21b687f219b71ecad3631f644a47f01403fa1d1036b0c6416d70fb"}, +] + +[package.extras] +diagrams = ["jinja2", "railroad-diagrams"] + +[[package]] +name = "pytest" +version = "7.2.2" +description = "pytest: simple powerful testing with Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pytest-7.2.2-py3-none-any.whl", hash = "sha256:130328f552dcfac0b1cec75c12e3f005619dc5f874f0a06e8ff7263f0ee6225e"}, + {file = "pytest-7.2.2.tar.gz", hash = "sha256:c99ab0c73aceb050f68929bc93af19ab6db0558791c6a0715723abe9d0ade9d4"}, +] + +[package.dependencies] +attrs = ">=19.2.0" +colorama = {version = "*", markers = "sys_platform == \"win32\""} +exceptiongroup = {version = ">=1.0.0rc8", markers = "python_version < \"3.11\""} +iniconfig = "*" +packaging = "*" +pluggy = ">=0.12,<2.0" +tomli = {version = ">=1.0.0", markers = "python_version < \"3.11\""} + +[package.extras] +testing = ["argcomplete", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.7.2)", "requests", "xmlschema"] + +[[package]] +name = "pytest-cov" +version = "2.12.1" +description = "Pytest plugin for measuring coverage." +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"}, + {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"}, +] + +[package.dependencies] +coverage = ">=5.2.1" +pytest = ">=4.6" +toml = "*" + +[package.extras] +testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"] + +[[package]] +name = "python-dotenv" +version = "1.0.0" +description = "Read key-value pairs from a .env file and set them as environment variables" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"}, + {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"}, +] + +[package.extras] +cli = ["click (>=5.0)"] + +[[package]] +name = "pytz" +version = "2020.5" +description = "World timezone definitions, modern and historical" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"}, + {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"}, +] + +[[package]] +name = "pyyaml" +version = "6.0" +description = "YAML parser and emitter for Python" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d4db7c7aef085872ef65a8fd7d6d09a14ae91f691dec3e87ee5ee0539d516f53"}, + {file = "PyYAML-6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9df7ed3b3d2e0ecfe09e14741b857df43adb5a3ddadc919a2d94fbdf78fea53c"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77f396e6ef4c73fdc33a9157446466f1cff553d979bd00ecb64385760c6babdc"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a80a78046a72361de73f8f395f1f1e49f956c6be882eed58505a15f3e430962b"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f84fbc98b019fef2ee9a1cb3ce93e3187a6df0b2538a651bfb890254ba9f90b5"}, + {file = "PyYAML-6.0-cp310-cp310-win32.whl", hash = "sha256:2cd5df3de48857ed0544b34e2d40e9fac445930039f3cfe4bcc592a1f836d513"}, + {file = "PyYAML-6.0-cp310-cp310-win_amd64.whl", hash = "sha256:daf496c58a8c52083df09b80c860005194014c3698698d1a57cbcfa182142a3a"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d4b0ba9512519522b118090257be113b9468d804b19d63c71dbcf4a48fa32358"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:81957921f441d50af23654aa6c5e5eaf9b06aba7f0a19c18a538dc7ef291c5a1"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:afa17f5bc4d1b10afd4466fd3a44dc0e245382deca5b3c353d8b757f9e3ecb8d"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dbad0e9d368bb989f4515da330b88a057617d16b6a8245084f1b05400f24609f"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:432557aa2c09802be39460360ddffd48156e30721f5e8d917f01d31694216782"}, + {file = "PyYAML-6.0-cp311-cp311-win32.whl", hash = "sha256:bfaef573a63ba8923503d27530362590ff4f576c626d86a9fed95822a8255fd7"}, + {file = "PyYAML-6.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b45c0191e6d66c470b6cf1b9531a771a83c1c4208272ead47a3ae4f2f603bf"}, + {file = "PyYAML-6.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:897b80890765f037df3403d22bab41627ca8811ae55e9a722fd0392850ec4d86"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50602afada6d6cbfad699b0c7bb50d5ccffa7e46a3d738092afddc1f9758427f"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:48c346915c114f5fdb3ead70312bd042a953a8ce5c7106d5bfb1a5254e47da92"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:98c4d36e99714e55cfbaaee6dd5badbc9a1ec339ebfc3b1f52e293aee6bb71a4"}, + {file = "PyYAML-6.0-cp36-cp36m-win32.whl", hash = "sha256:0283c35a6a9fbf047493e3a0ce8d79ef5030852c51e9d911a27badfde0605293"}, + {file = "PyYAML-6.0-cp36-cp36m-win_amd64.whl", hash = "sha256:07751360502caac1c067a8132d150cf3d61339af5691fe9e87803040dbc5db57"}, + {file = "PyYAML-6.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:819b3830a1543db06c4d4b865e70ded25be52a2e0631ccd2f6a47a2822f2fd7c"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:473f9edb243cb1935ab5a084eb238d842fb8f404ed2193a915d1784b5a6b5fc0"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0ce82d761c532fe4ec3f87fc45688bdd3a4c1dc5e0b4a19814b9009a29baefd4"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:231710d57adfd809ef5d34183b8ed1eeae3f76459c18fb4a0b373ad56bedcdd9"}, + {file = "PyYAML-6.0-cp37-cp37m-win32.whl", hash = "sha256:c5687b8d43cf58545ade1fe3e055f70eac7a5a1a0bf42824308d868289a95737"}, + {file = "PyYAML-6.0-cp37-cp37m-win_amd64.whl", hash = "sha256:d15a181d1ecd0d4270dc32edb46f7cb7733c7c508857278d3d378d14d606db2d"}, + {file = "PyYAML-6.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b4624f379dab24d3725ffde76559cff63d9ec94e1736b556dacdfebe5ab6d4b"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:213c60cd50106436cc818accf5baa1aba61c0189ff610f64f4a3e8c6726218ba"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9fa600030013c4de8165339db93d182b9431076eb98eb40ee068700c9c813e34"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:277a0ef2981ca40581a47093e9e2d13b3f1fbbeffae064c1d21bfceba2030287"}, + {file = "PyYAML-6.0-cp38-cp38-win32.whl", hash = "sha256:d4eccecf9adf6fbcc6861a38015c2a64f38b9d94838ac1810a9023a0609e1b78"}, + {file = "PyYAML-6.0-cp38-cp38-win_amd64.whl", hash = "sha256:1e4747bc279b4f613a09eb64bba2ba602d8a6664c6ce6396a4d0cd413a50ce07"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:055d937d65826939cb044fc8c9b08889e8c743fdc6a32b33e2390f66013e449b"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e61ceaab6f49fb8bdfaa0f92c4b57bcfbea54c09277b1b4f7ac376bfb7a7c174"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d67d839ede4ed1b28a4e8909735fc992a923cdb84e618544973d7dfc71540803"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba8c411ef271aa037d7357a2bc8f9ee8b58b9965831d9e51baf703280dc73d3"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:40527857252b61eacd1d9af500c3337ba8deb8fc298940291486c465c8b46ec0"}, + {file = "PyYAML-6.0-cp39-cp39-win32.whl", hash = "sha256:b5b9eccad747aabaaffbc6064800670f0c297e52c12754eb1d976c57e4f74dcb"}, + {file = "PyYAML-6.0-cp39-cp39-win_amd64.whl", hash = "sha256:b3d267842bf12586ba6c734f89d1f5b871df0273157918b0ccefa29deb05c21c"}, + {file = "PyYAML-6.0.tar.gz", hash = "sha256:68fb519c14306fec9720a2a5b45bc9f0c8d1b9c72adf45c37baedfcd949c35a2"}, +] + +[[package]] +name = "requests" +version = "2.28.2" +description = "Python HTTP for Humans." +category = "main" +optional = false +python-versions = ">=3.7, <4" +files = [ + {file = "requests-2.28.2-py3-none-any.whl", hash = "sha256:64299f4909223da747622c030b781c0d7811e359c37124b4bd368fb8c6518baa"}, + {file = "requests-2.28.2.tar.gz", hash = "sha256:98b1b2782e3c6c4904938b84c0eb932721069dfdb9134313beff7c83c2df24bf"}, +] + +[package.dependencies] +certifi = ">=2017.4.17" +charset-normalizer = ">=2,<4" +idna = ">=2.5,<4" +urllib3 = ">=1.21.1,<1.27" + +[package.extras] +socks = ["PySocks (>=1.5.6,!=1.5.7)"] +use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] + +[[package]] +name = "resolvelib" +version = "1.0.1" +description = "Resolve abstract dependencies into concrete ones" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "resolvelib-1.0.1-py2.py3-none-any.whl", hash = "sha256:d2da45d1a8dfee81bdd591647783e340ef3bcb104b54c383f70d422ef5cc7dbf"}, + {file = "resolvelib-1.0.1.tar.gz", hash = "sha256:04ce76cbd63fded2078ce224785da6ecd42b9564b1390793f64ddecbe997b309"}, +] + +[package.extras] +examples = ["html5lib", "packaging", "pygraphviz", "requests"] +lint = ["black", "flake8", "isort", "mypy", "types-requests"] +release = ["build", "towncrier", "twine"] +test = ["commentjson", "packaging", "pytest"] + +[[package]] +name = "rich" +version = "13.3.3" +description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal" +category = "dev" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "rich-13.3.3-py3-none-any.whl", hash = "sha256:540c7d6d26a1178e8e8b37e9ba44573a3cd1464ff6348b99ee7061b95d1c6333"}, + {file = "rich-13.3.3.tar.gz", hash = "sha256:dc84400a9d842b3a9c5ff74addd8eb798d155f36c1c91303888e0a66850d2a15"}, +] + +[package.dependencies] +markdown-it-py = ">=2.2.0,<3.0.0" +pygments = ">=2.13.0,<3.0.0" + +[package.extras] +jupyter = ["ipywidgets (>=7.5.1,<9)"] + +[[package]] +name = "setuptools" +version = "67.6.1" +description = "Easily download, build, install, upgrade, and uninstall Python packages" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "setuptools-67.6.1-py3-none-any.whl", hash = "sha256:e728ca814a823bf7bf60162daf9db95b93d532948c4c0bea762ce62f60189078"}, + {file = "setuptools-67.6.1.tar.gz", hash = "sha256:257de92a9d50a60b8e22abfcbb771571fde0dbf3ec234463212027a4eeecbe9a"}, +] + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-hoverxref (<2)", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (==0.8.3)", "sphinx-reredirects", "sphinxcontrib-towncrier"] +testing = ["build[virtualenv]", "filelock (>=3.4.0)", "flake8 (<5)", "flake8-2020", "ini2toml[lite] (>=0.9)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pip (>=19.1)", "pip-run (>=8.8)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)", "pytest-perf", "pytest-timeout", "pytest-xdist", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel"] +testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pytest", "pytest-enabler", "pytest-xdist", "tomli", "virtualenv (>=13.0.0)", "wheel"] + +[[package]] +name = "six" +version = "1.16.0" +description = "Python 2 and 3 compatibility utilities" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "six-1.16.0-py2.py3-none-any.whl", hash = "sha256:8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254"}, + {file = "six-1.16.0.tar.gz", hash = "sha256:1e61c37477a1626458e36f7b1d82aa5c9b094fa4802892072e49de9c60c4c926"}, +] + +[[package]] +name = "smmap" +version = "5.0.0" +description = "A pure Python implementation of a sliding window memory map manager" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "smmap-5.0.0-py3-none-any.whl", hash = "sha256:2aba19d6a040e78d8b09de5c57e96207b09ed71d8e55ce0959eeee6c8e190d94"}, + {file = "smmap-5.0.0.tar.gz", hash = "sha256:c840e62059cd3be204b0c9c9f74be2c09d5648eddd4580d9314c3ecde0b30936"}, +] + +[[package]] +name = "sortedcontainers" +version = "2.4.0" +description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"}, + {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"}, +] + +[[package]] +name = "stevedore" +version = "5.0.0" +description = "Manage dynamic plugins for Python applications" +category = "dev" +optional = false +python-versions = ">=3.8" +files = [ + {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"}, + {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"}, +] + +[package.dependencies] +pbr = ">=2.0.0,<2.1.0 || >2.1.0" + +[[package]] +name = "toml" +version = "0.10.2" +description = "Python Library for Tom's Obvious, Minimal Language" +category = "dev" +optional = false +python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b"}, + {file = "toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f"}, +] + +[[package]] +name = "tomli" +version = "2.0.1" +description = "A lil' TOML parser" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tomli-2.0.1-py3-none-any.whl", hash = "sha256:939de3e7a6161af0c887ef91b7d41a53e7c5a1ca976325f429cb46ea9bc30ecc"}, + {file = "tomli-2.0.1.tar.gz", hash = "sha256:de526c12914f0c550d15924c62d72abc48d6fe7364aa87328337a31007fe8a4f"}, +] + +[[package]] +name = "tqdm" +version = "4.65.0" +description = "Fast, Extensible Progress Meter" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"}, + {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[package.extras] +dev = ["py-make (>=0.1.0)", "twine", "wheel"] +notebook = ["ipywidgets (>=6)"] +slack = ["slack-sdk"] +telegram = ["requests"] + +[[package]] +name = "typing-extensions" +version = "4.5.0" +description = "Backported and Experimental Type Hints for Python 3.7+" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"}, + {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"}, +] + +[[package]] +name = "urllib3" +version = "1.26.15" +description = "HTTP library with thread-safe connection pooling, file post, and more." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*, !=3.5.*" +files = [ + {file = "urllib3-1.26.15-py2.py3-none-any.whl", hash = "sha256:aa751d169e23c7479ce47a0cb0da579e3ede798f994f5816a74e4f4500dcea42"}, + {file = "urllib3-1.26.15.tar.gz", hash = "sha256:8a388717b9476f934a21484e8c8e61875ab60644d29b9b39e11e4b9dc1c6b305"}, +] + +[package.extras] +brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)", "brotlipy (>=0.6.0)"] +secure = ["certifi", "cryptography (>=1.3.4)", "idna (>=2.0.0)", "ipaddress", "pyOpenSSL (>=0.14)", "urllib3-secure-extra"] +socks = ["PySocks (>=1.5.6,!=1.5.7,<2.0)"] + +[[package]] +name = "webencodings" +version = "0.5.1" +description = "Character encoding aliases for legacy web content" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78"}, + {file = "webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923"}, +] + +[metadata] +lock-version = "2.0" +python-versions = "3.9.15" +content-hash = "67ae1eb17c3921f944b323ddcaa58919c1deff1e7cc25e63cf3a82ec23340c62" diff --git a/jobs/cache_maintenance/poetry.toml b/jobs/cache_maintenance/poetry.toml new file mode 100644 index 00000000..5fcef8cd --- /dev/null +++ b/jobs/cache_maintenance/poetry.toml @@ -0,0 +1,3 @@ +[virtualenvs] +in-project = true +prefer-active-python = true diff --git a/jobs/cache_maintenance/pyproject.toml b/jobs/cache_maintenance/pyproject.toml new file mode 100644 index 00000000..d7481fff --- /dev/null +++ b/jobs/cache_maintenance/pyproject.toml @@ -0,0 +1,44 @@ +[tool.poetry] +authors = ["Andrea Soria <[email protected]>"] +description = "Monitor cache status entries and apply corrections if needed." +name = "cache_maintenance" +version = "0.0.1" +license = "Apache-2.0" + +[tool.poetry.dependencies] +environs = "^9.5.0" +libcommon = {path = "../../libs/libcommon", develop = true} +python = "3.9.15" + +[tool.poetry.group.dev.dependencies] +bandit = "^1.7.4" +black = "^22.12.0" +flake8 = "^3.9.2" +isort = "^5.12.0" +mypy = "^1.0.0" +pip-audit = "^2.4.14" +pytest = "^7.2.1" +pytest-cov = "^2.12.1" + +[build-system] +build-backend = "poetry.core.masonry.api" +requires = ["poetry-core>=1.0.0"] + +[tool.pytest.ini_options] +filterwarnings = ["ignore::DeprecationWarning"] +markers = [ + "wip: tests being developed" +] + +[tool.coverage.run] +source = ["cache_maintenance"] + +[tool.isort] +profile = "black" + +[tool.black] +line-length = 119 +preview = true + +[tool.mypy] +strict = true diff --git a/jobs/cache_maintenance/src/cache_maintenance/__init__.py b/jobs/cache_maintenance/src/cache_maintenance/__init__.py new file mode 100644 index 00000000..1e9d0c5a --- /dev/null +++ b/jobs/cache_maintenance/src/cache_maintenance/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py new file mode 100644 index 00000000..cdb7ac1f --- /dev/null +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -0,0 +1,31 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from typing import Optional + +import libcommon +from libcommon.operations import update_dataset +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority + + +def backfill_cache( + init_processing_steps: list[ProcessingStep], + hf_endpoint: str, + hf_token: Optional[str] = None, +) -> None: + logging.info("backfill init processing steps for supported datasets") + supported_datasets = libcommon.dataset.get_supported_datasets(hf_endpoint=hf_endpoint, hf_token=hf_token) + logging.info(f"about to backfill {len(supported_datasets)} datasets") + for dataset in libcommon.dataset.get_supported_datasets(hf_endpoint=hf_endpoint, hf_token=hf_token): + update_dataset( + dataset=dataset, + init_processing_steps=init_processing_steps, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + force=False, + priority=Priority.LOW, + do_check_support=False, + ) + logging.info("backfill completed") diff --git a/jobs/cache_maintenance/src/cache_maintenance/config.py b/jobs/cache_maintenance/src/cache_maintenance/config.py new file mode 100644 index 00000000..7340409d --- /dev/null +++ b/jobs/cache_maintenance/src/cache_maintenance/config.py @@ -0,0 +1,31 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from dataclasses import dataclass, field +from typing import Optional + +from environs import Env +from libcommon.config import CommonConfig, LogConfig, ProcessingGraphConfig, QueueConfig + +CACHE_MAINTENANCE_ACTION = None + + +@dataclass(frozen=True) +class JobConfig: + log: LogConfig = field(default_factory=LogConfig) + queue: QueueConfig = field(default_factory=QueueConfig) + common: CommonConfig = field(default_factory=CommonConfig) + graph: ProcessingGraphConfig = field(default_factory=ProcessingGraphConfig) + action: Optional[str] = CACHE_MAINTENANCE_ACTION + + @classmethod + def from_env(cls) -> "JobConfig": + env = Env(expand_vars=True) + + return cls( + log=LogConfig.from_env(), + queue=QueueConfig.from_env(), + common=CommonConfig.from_env(), + graph=ProcessingGraphConfig.from_env(), + action=env.str(name="CACHE_MAINTENANCE_ACTION", default=CACHE_MAINTENANCE_ACTION), + ) diff --git a/jobs/cache_maintenance/src/cache_maintenance/main.py b/jobs/cache_maintenance/src/cache_maintenance/main.py new file mode 100644 index 00000000..f330b7c4 --- /dev/null +++ b/jobs/cache_maintenance/src/cache_maintenance/main.py @@ -0,0 +1,62 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +import sys +from datetime import datetime + +from libcommon.log import init_logging +from libcommon.processing_graph import ProcessingGraph +from libcommon.resources import QueueMongoResource + +from cache_maintenance.backfill import backfill_cache +from cache_maintenance.config import JobConfig + + +def run_job() -> None: + job_config = JobConfig.from_env() + action = job_config.action + supported_actions = ["backfill"] + # In the future we will support other kind of actions + if not action: + logging.warning("No action mode was selected, skipping tasks.") + return + if action not in supported_actions: + logging.warning(f"Wrong action mode selected, supported actions are {supported_actions}.") + return + + init_logging(level=job_config.log.level) + + with ( + QueueMongoResource( + database=job_config.queue.mongo_database, host=job_config.queue.mongo_url + ) as queue_resource, + ): + if not queue_resource.is_available(): + logging.warning( + "The connection to the queue database could not be established. The cache refresh job is skipped." + ) + return + + processing_graph = ProcessingGraph(job_config.graph.specification) + init_processing_steps = processing_graph.get_first_steps() + + start_time = datetime.now() + + if action == "backfill": + backfill_cache( + init_processing_steps=init_processing_steps, + hf_endpoint=job_config.common.hf_endpoint, + hf_token=job_config.common.hf_token, + ) + + end_time = datetime.now() + logging.info(f"Duration: {end_time - start_time}") + + +if __name__ == "__main__": + try: + run_job() + sys.exit(0) + except Exception: + sys.exit(1) diff --git a/jobs/cache_maintenance/tests/__init__.py b/jobs/cache_maintenance/tests/__init__.py new file mode 100644 index 00000000..1e9d0c5a --- /dev/null +++ b/jobs/cache_maintenance/tests/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. diff --git a/jobs/cache_maintenance/tests/conftest.py b/jobs/cache_maintenance/tests/conftest.py new file mode 100644 index 00000000..72f04b52 --- /dev/null +++ b/jobs/cache_maintenance/tests/conftest.py @@ -0,0 +1,34 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from typing import Iterator + +from libcommon.queue import _clean_queue_database +from libcommon.resources import QueueMongoResource +from pytest import MonkeyPatch, fixture + +from cache_maintenance.config import JobConfig + + +# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200 +@fixture(scope="session") +def monkeypatch_session() -> Iterator[MonkeyPatch]: + monkeypatch_session = MonkeyPatch() + monkeypatch_session.setenv("QUEUE_MONGO_DATABASE", "datasets_server_queue_test") + yield monkeypatch_session + monkeypatch_session.undo() + + +@fixture(scope="session") +def job_config(monkeypatch_session: MonkeyPatch) -> JobConfig: + job_config = JobConfig.from_env() + if "test" not in job_config.queue.mongo_database: + raise ValueError("Test must be launched on a test mongo database") + return job_config + + +@fixture(autouse=True) +def queue_mongo_resource(job_config: JobConfig) -> Iterator[QueueMongoResource]: + with QueueMongoResource(database=job_config.queue.mongo_database, host=job_config.queue.mongo_url) as resource: + yield resource + _clean_queue_database() diff --git a/jobs/cache_maintenance/tests/test_backfill_cache.py b/jobs/cache_maintenance/tests/test_backfill_cache.py new file mode 100644 index 00000000..c89d8870 --- /dev/null +++ b/jobs/cache_maintenance/tests/test_backfill_cache.py @@ -0,0 +1,31 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from unittest.mock import patch + +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Queue, Status + +from cache_maintenance.backfill import backfill_cache + + +def test_backfill_cache() -> None: + some_datasets = ["dummy", "dummy_2", "dummy_3"] + datasets_number = len(some_datasets) + step_name = "dummy_step" + with patch("libcommon.dataset.get_supported_datasets", return_value=some_datasets): + init_processing_steps = [ + ProcessingStep( + name=step_name, + input_type="dataset", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + job_runner_version=1, + ) + ] + backfill_cache(init_processing_steps=init_processing_steps, hf_endpoint="hf_endpoint", hf_token="hf_token") + queue = Queue() + assert queue.count_jobs(Status.WAITING, step_name) == datasets_number + assert queue.count_jobs(Status.STARTED, step_name) == 0 diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py index 83dab1a3..0e29841f 100644 --- a/services/admin/src/admin/app.py +++ b/services/admin/src/admin/app.py @@ -18 +17,0 @@ from admin.prometheus import Prometheus -from admin.routes.backfill import create_backfill_endpoint @@ -42 +40,0 @@ def create_app() -> Starlette: - init_processing_steps = processing_graph.get_first_steps() @@ -99,13 +96,0 @@ def create_app() -> Starlette: - + [ - Route( - "/backfill", - endpoint=create_backfill_endpoint( - init_processing_steps=init_processing_steps, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - external_auth_url=app_config.admin.external_auth_url, - organization=app_config.admin.hf_organization, - ), - methods=["POST"], - ) - ] diff --git a/services/admin/src/admin/routes/backfill.py b/services/admin/src/admin/routes/backfill.py deleted file mode 100644 index 95ca237d..00000000 --- a/services/admin/src/admin/routes/backfill.py +++ /dev/null @@ -1,60 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import logging -from typing import Optional - -from libcommon.dataset import get_supported_datasets -from libcommon.operations import update_dataset -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Priority -from starlette.requests import Request -from starlette.responses import Response - -from admin.authentication import auth_check -from admin.utils import ( - AdminCustomError, - Endpoint, - UnexpectedError, - get_json_admin_error_response, - get_json_ok_response, -) - - -def create_backfill_endpoint( - init_processing_steps: list[ProcessingStep], - hf_endpoint: str, - hf_token: Optional[str] = None, - external_auth_url: Optional[str] = None, - organization: Optional[str] = None, -) -> Endpoint: - async def backfill_endpoint(request: Request) -> Response: - try: - logging.info("/backfill") - - # if auth_check fails, it will raise an exception that will be caught below - auth_check(external_auth_url=external_auth_url, request=request, organization=organization) - for dataset in get_supported_datasets(hf_endpoint=hf_endpoint, hf_token=hf_token): - update_dataset( - dataset=dataset, - init_processing_steps=init_processing_steps, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - force=False, - priority=Priority.LOW, - do_check_support=False, - ) - # ^ we simply ask an update for all the datasets on the Hub, supported by the datasets-server - # we could be more precise and only ask for updates for the datasets that have some missing - # cache entries, but it's not easy to check. - # Also: we could try to do a batch update of the database, instead of one query per dataset - return get_json_ok_response( - {"status": "ok"}, - max_age=0, - ) - except AdminCustomError as e: - return get_json_admin_error_response(e, max_age=0) - except Exception as e: - return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=0) - - return backfill_endpoint
3e100ad5ed29d3842f41cc01956ed679181df3ec
Sylvain Lesage
2023-04-06T15:05:59
Create step `dataset-split-names` (#1025)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index d3765678..d52e7277 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -20,0 +21 @@ from libcommon.constants import ( + PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION, @@ -199,0 +201,5 @@ class ProcessingGraphConfig: + "dataset-split-names": { + "input_type": "dataset", + "requires": ["/split-names-from-dataset-info", "/split-names-from-streaming"], + "job_runner_version": PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION, + }, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 1d5683b2..1ede9ada 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -22,0 +23 @@ PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION = 2 +PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 1 diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index 37287402..9b592ce3 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -3,0 +4 @@ +import contextlib @@ -4,0 +6 @@ import types +from dataclasses import dataclass @@ -224,0 +227,62 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O +@dataclass +class BestResponse: + kind: str + response: Any + + +def _get_progress(best_response: BestResponse) -> float: + return ( + best_response.response["progress"] + if ("progress" in best_response.response and isinstance(best_response.response["progress"], float)) + else 0.0 + ) + + +def get_best_response( + kinds: List[str], dataset: str, config: Optional[str] = None, split: Optional[str] = None +) -> BestResponse: + """ + Get the best response from a list of cache kinds. + + Best means: + - the first success response with progress=1.0 is returned + - if no success response with progress=1.0 is found, the success response with the highest progress is returned + - if no success response is found, the first error response is returned + - if no response is found, a `~libcommon.simple_cache.DoesNotExist` error is raised + + Args: + kinds (`List[str]`): + A list of cache kinds to look responses for. + dataset (`str`): + A namespace (user or an organization) and a repo name separated by a `/`. + config (`str`, optional): + A config name. + split (`str`, optional): + A split name. + Returns: + BestResponse: The best response (object with fields: kind and response). The response can be an error. + <Tip> + Raises the following errors: + - [`~libcommon.simple_cache.DoesNotExist`] + If no response was found + </Tip> + """ + in_progress_success_responses = [] + error_responses = [] + for kind in kinds: + with contextlib.suppress(DoesNotExist): + response = get_response(kind=kind, dataset=dataset, config=config, split=split) + best_response = BestResponse(kind=kind, response=response) + if best_response.response["http_status"] == HTTPStatus.OK: + if best_response.response["progress"] == 1.0: + return best_response + in_progress_success_responses.append(best_response) + else: + error_responses.append(best_response) + if in_progress_success_responses: + return sorted(in_progress_success_responses, key=_get_progress, reverse=True)[0] + if error_responses: + return error_responses[0] + raise DoesNotExist("No cached response found.") + + diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 40bf26fd..8e4f0ee2 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -5,0 +6,2 @@ from typing import List +import pytest + @@ -57 +59,2 @@ def test_graph() -> None: -def test_default_graph() -> None: [email protected](scope="module") +def graph() -> ProcessingGraph: @@ -59,61 +62,59 @@ def test_default_graph() -> None: - graph = ProcessingGraph(config.specification) - - config_names = graph.get_step("/config-names") - split_names_from_streaming = graph.get_step("/split-names-from-streaming") - splits = graph.get_step("/splits") - split_first_rows_from_streaming = graph.get_step("split-first-rows-from-streaming") - parquet_and_dataset_info = graph.get_step("/parquet-and-dataset-info") - config_parquet = graph.get_step("config-parquet") - dataset_parquet = graph.get_step("dataset-parquet") - config_info = graph.get_step("config-info") - dataset_info = graph.get_step("dataset-info") - config_size = graph.get_step("config-size") - dataset_size = graph.get_step("dataset-size") - split_names_from_dataset_info = graph.get_step("/split-names-from-dataset-info") - dataset_split_names_from_streaming = graph.get_step("dataset-split-names-from-streaming") - dataset_split_names_from_dataset_info = graph.get_step("dataset-split-names-from-dataset-info") - split_first_rows_from_parquet = graph.get_step("split-first-rows-from-parquet") - - assert_step(config_names, children=[split_names_from_streaming], ancestors=[]) - assert_step( - split_names_from_streaming, - children=[split_first_rows_from_streaming, dataset_split_names_from_streaming], - ancestors=[config_names], - ) - assert_step(splits, children=[], ancestors=[]) - assert_step( - split_first_rows_from_streaming, - children=[], - ancestors=[ - config_names, - split_names_from_streaming, - split_names_from_dataset_info, - parquet_and_dataset_info, - config_info, - ], - ) - assert_step(parquet_and_dataset_info, children=[config_parquet, config_info, config_size], ancestors=[]) - assert_step( - config_parquet, - children=[split_first_rows_from_parquet, dataset_parquet], - ancestors=[parquet_and_dataset_info], - ) - assert_step(dataset_parquet, children=[], ancestors=[parquet_and_dataset_info, config_parquet]) - assert_step( - config_info, - children=[dataset_info, split_names_from_dataset_info], - ancestors=[parquet_and_dataset_info], - ) - assert_step(dataset_info, children=[], ancestors=[parquet_and_dataset_info, config_info]) - assert_step(config_size, children=[dataset_size], ancestors=[parquet_and_dataset_info]) - assert_step(dataset_size, children=[], ancestors=[parquet_and_dataset_info, config_size]) - assert_step( - split_names_from_dataset_info, - children=[split_first_rows_from_streaming, dataset_split_names_from_dataset_info], - ancestors=[parquet_and_dataset_info, config_info], - ) - assert_step( - dataset_split_names_from_streaming, - children=[], - ancestors=[config_names, split_names_from_streaming], - ) + return ProcessingGraph(config.specification) + + [email protected]( + "step_name,children,ancestors", + [ + ("/config-names", ["/split-names-from-streaming"], []), + ("/splits", [], []), + ( + "/split-names-from-dataset-info", + ["dataset-split-names-from-dataset-info", "split-first-rows-from-streaming", "dataset-split-names"], + ["/parquet-and-dataset-info", "config-info"], + ), + ( + "/split-names-from-streaming", + ["split-first-rows-from-streaming", "dataset-split-names-from-streaming", "dataset-split-names"], + ["/config-names"], + ), + ( + "dataset-split-names-from-dataset-info", + [], + ["/parquet-and-dataset-info", "config-info", "/split-names-from-dataset-info"], + ), + ("dataset-split-names-from-streaming", [], ["/config-names", "/split-names-from-streaming"]), + ( + "dataset-split-names", + [], + [ + "/parquet-and-dataset-info", + "config-info", + "/split-names-from-dataset-info", + "/config-names", + "/split-names-from-streaming", + ], + ), + ("split-first-rows-from-parquet", [], ["config-parquet", "/parquet-and-dataset-info"]), + ( + "split-first-rows-from-streaming", + [], + [ + "/config-names", + "/split-names-from-streaming", + "/split-names-from-dataset-info", + "/parquet-and-dataset-info", + "config-info", + ], + ), + ("/parquet-and-dataset-info", ["config-parquet", "config-info", "config-size"], []), + ("config-parquet", ["split-first-rows-from-parquet", "dataset-parquet"], ["/parquet-and-dataset-info"]), + ("dataset-parquet", [], ["/parquet-and-dataset-info", "config-parquet"]), + ("config-info", ["dataset-info", "/split-names-from-dataset-info"], ["/parquet-and-dataset-info"]), + ("dataset-info", [], ["/parquet-and-dataset-info", "config-info"]), + ("config-size", ["dataset-size"], ["/parquet-and-dataset-info"]), + ("dataset-size", [], ["/parquet-and-dataset-info", "config-size"]), + ], +) +def test_default_graph_steps( + graph: ProcessingGraph, step_name: str, children: List[str], ancestors: List[str] +) -> None: @@ -121,3 +122,3 @@ def test_default_graph() -> None: - dataset_split_names_from_dataset_info, - children=[], - ancestors=[parquet_and_dataset_info, config_info, split_names_from_dataset_info], + graph.get_step(step_name), + children=[graph.get_step(child) for child in children], + ancestors=[graph.get_step(ancestor) for ancestor in ancestors], @@ -125,4 +126,6 @@ def test_default_graph() -> None: - assert_step( - split_first_rows_from_parquet, - children=[], - ancestors=[parquet_and_dataset_info, config_parquet], + + +def test_default_graph_first_steps(graph: ProcessingGraph) -> None: + assert_lists_are_equal( + graph.get_first_steps(), + [graph.get_step(step_name) for step_name in {"/config-names", "/splits", "/parquet-and-dataset-info"}], @@ -131,2 +134,6 @@ def test_default_graph() -> None: - assert graph.get_first_steps() == [config_names, splits, parquet_and_dataset_info] - assert graph.get_steps_required_by_dataset_viewer() == [splits, split_first_rows_from_streaming] + +def test_default_graph_required_by_dataset_viewer(graph: ProcessingGraph) -> None: + assert_lists_are_equal( + graph.get_steps_required_by_dataset_viewer(), + [graph.get_step(step_name) for step_name in {"/splits", "split-first-rows-from-streaming"}], + ) diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index b939fa39..3445d65d 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -7 +7 @@ from time import process_time -from typing import Iterator, Optional +from typing import Dict, Iterator, List, Optional, TypedDict @@ -23,0 +24 @@ from libcommon.simple_cache import ( + get_best_response, @@ -646,0 +648,112 @@ def test_get_outdated_split_full_names_for_step() -> None: + + +class EntrySpec(TypedDict): + kind: str + dataset: str + config: Optional[str] + http_status: HTTPStatus + progress: float + + [email protected]( + "selected_entries,kinds,dataset,config,best_entry", + [ + # Best means: + # - the first success response with progress=1.0 is returned + (["ok1"], ["kind1"], "dataset", None, "ok1"), + (["ok_config1"], ["kind1"], "dataset", "config", "ok_config1"), + (["ok1", "ok2"], ["kind1", "kind2"], "dataset", None, "ok1"), + (["ok1", "ok2"], ["kind2", "kind1"], "dataset", None, "ok2"), + (["partial1", "ok2"], ["kind1", "kind2"], "dataset", None, "ok2"), + (["error1", "ok2"], ["kind1", "kind2"], "dataset", None, "ok2"), + # - if no success response with progress=1.0 is found, the success response with the highest progress is + # returned + (["partial1", "partial2"], ["kind1", "kind2"], "dataset", None, "partial2"), + (["partial1", "error2"], ["kind1", "kind2"], "dataset", None, "partial1"), + # - if no success response is found, the first error response is returned + (["error1", "error2"], ["kind1", "kind2"], "dataset", None, "error1"), + (["error1", "error2"], ["kind2", "kind1"], "dataset", None, "error2"), + # - if no response is found, a `~libcommon.simple_cache.DoesNotExist` error is raised + ([], ["kind1"], "dataset", None, None), + (["ok_config1"], ["kind1"], "dataset", None, None), + (["ok1"], ["kind1"], "dataset", "config", None), + ], +) +def test_get_best_response( + selected_entries: List[str], kinds: List[str], dataset: str, config: Optional[str], best_entry: Optional[str] +) -> None: + # arrange + entries: Dict[str, EntrySpec] = { + "ok1": { + "kind": "kind1", + "dataset": "dataset", + "config": None, + "http_status": HTTPStatus.OK, + "progress": 1.0, + }, + "ok2": { + "kind": "kind2", + "dataset": "dataset", + "config": None, + "http_status": HTTPStatus.OK, + "progress": 1.0, + }, + "partial1": { + "kind": "kind1", + "dataset": "dataset", + "config": None, + "http_status": HTTPStatus.OK, + "progress": 0, + }, + "partial2": { + "kind": "kind2", + "dataset": "dataset", + "config": None, + "http_status": HTTPStatus.OK, + "progress": 0.5, + }, + "ok_config1": { + "kind": "kind1", + "dataset": "dataset", + "config": "config", + "http_status": HTTPStatus.OK, + "progress": 1.0, + }, + "error1": { + "kind": "kind1", + "dataset": "dataset", + "config": None, + "http_status": HTTPStatus.INTERNAL_SERVER_ERROR, + "progress": 1.0, + }, + "error2": { + "kind": "kind2", + "dataset": "dataset", + "config": None, + "http_status": HTTPStatus.NOT_FOUND, + "progress": 1.0, + }, + } + + for entry in selected_entries: + upsert_response( + kind=entries[entry]["kind"], + dataset=entries[entry]["dataset"], + config=entries[entry]["config"], + http_status=entries[entry]["http_status"], + progress=entries[entry]["progress"], + content={"entry": entry}, + ) + + # act + if best_entry is None: + with pytest.raises(DoesNotExist): + get_best_response(kinds, dataset, config) + return + best_response = get_best_response(kinds, dataset, config) + + # assert + assert best_response.kind == entries[best_entry]["kind"] + assert best_response.response["content"]["entry"] == best_entry + assert best_response.response["http_status"] == entries[best_entry]["http_status"].value + assert best_response.response["progress"] == entries[best_entry]["progress"] diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index cc24a640..ff95be34 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -119 +119,6 @@ class EndpointConfig: - "dataset": ["/splits", "dataset-split-names-from-streaming", "dataset-split-names-from-dataset-info"], + "dataset": [ + "/splits", + "dataset-split-names", + "dataset-split-names-from-streaming", + "dataset-split-names-from-dataset-info", + ], diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index c8f1f7c7..d43f0e00 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -39 +39 @@ def test_endpoints_definition() -> None: - assert len(splits["dataset"]) == 3 # Has three processing steps + assert len(splits["dataset"]) == 4 # Has four processing steps diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 370bdc73..3eb21bba 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -26,0 +27 @@ from worker.job_runners.dataset.size import DatasetSizeJobRunner +from worker.job_runners.dataset.split_names import DatasetSplitNamesJobRunner @@ -163,0 +165,7 @@ class JobRunnerFactory(BaseJobRunnerFactory): + if job_type == DatasetSplitNamesJobRunner.get_job_type(): + return DatasetSplitNamesJobRunner( + job_info=job_info, + processing_step=processing_step, + common_config=self.app_config.common, + worker_config=self.app_config.worker, + ) diff --git a/services/worker/src/worker/job_runners/dataset/split_names.py b/services/worker/src/worker/job_runners/dataset/split_names.py new file mode 100644 index 00000000..c2087b02 --- /dev/null +++ b/services/worker/src/worker/job_runners/dataset/split_names.py @@ -0,0 +1,176 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from http import HTTPStatus +from typing import Any, List, Literal, Mapping, Optional, Tuple + +from libcommon.constants import PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION +from libcommon.dataset import DatasetNotFoundError +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_best_response + +from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.utils import ( + ConfigItem, + DatasetSplitNamesResponse, + FailedConfigItem, + SplitItem, +) + +DatasetSplitNamesErrorCode = Literal[ + "PreviousStepStatusError", + "PreviousStepFormatError", +] + + +class DatasetSplitNamesJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: DatasetSplitNamesErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class PreviousStepStatusError(DatasetSplitNamesJobRunnerError): + """Raised when the previous step gave an error. The job should not have been created.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) + + +class PreviousStepFormatError(DatasetSplitNamesJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +def compute_dataset_split_names_response(dataset: str) -> Tuple[DatasetSplitNamesResponse, float]: + """ + Get the response of /splits for one specific dataset on huggingface.co + computed from responses cached in /split-names-from-dataset-info or /split-names-from-streaming steps. + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated by a `/`. + Returns: + `DatasetSplitNamesResponse`: An object with a list of split names for the dataset [splits], + a list of pending configs to be processed [pending] and the list of errors [failed] by config. + <Tip> + Raises the following errors: + - [`~job_runners.dataset_split_names.PreviousStepStatusError`] + If the the previous step gave an error. + - [`~job_runners.dataset_split_names.PreviousStepFormatError`] + If the content of the previous step has not the expected format + - [`~libcommon.dataset.DatasetNotFoundError`] + If previous step content was not found for the dataset + </Tip> + """ + logging.info(f"get dataset split names for dataset={dataset}") + + # Get the config names from the previous steps + try: + best_response = get_best_response(["/config-names", "dataset-info"], dataset) + if best_response.kind == "/config-names": + config_names = [ + config_name_item["config"] for config_name_item in best_response.response["content"]["config_names"] + ] + elif best_response.kind == "dataset-info": + config_names = list(best_response.response["content"]["dataset-info"].keys()) + else: + raise PreviousStepFormatError( + "Previous step '/config-names' or 'dataset-info' did not return the expected content." + ) + except DoesNotExist as e: + raise DatasetNotFoundError( + "No response found in previous step for this dataset: '/config-names' or 'dataset-info'.", e + ) from e + except KeyError as e: + raise PreviousStepFormatError( + "Previous steps '/config-names' or 'dataset-info' did not return the expected content.", e + ) from e + if any(not isinstance(config_name, str) for config_name in config_names): + raise PreviousStepFormatError( + "Previous steps '/config-names' or 'dataset-info' did not return a list of config names." + ) + + split_names_cache_kinds = ["/split-names-from-dataset-info", "/split-names-from-streaming"] + try: + splits: List[SplitItem] = [] + pending: List[ConfigItem] = [] + failed: List[FailedConfigItem] = [] + total = 0 + for config in config_names: + total += 1 + try: + best_response = get_best_response(split_names_cache_kinds, dataset=dataset, config=config) + except DoesNotExist: + logging.debug( + "No response (successful or erroneous) found in cache for the previous steps" + f" '{split_names_cache_kinds}' for this dataset." + ) + pending.append(ConfigItem({"dataset": dataset, "config": config})) + continue + if best_response.response["http_status"] != HTTPStatus.OK: + logging.debug(f"No successful response found in the previous steps {split_names_cache_kinds}.") + failed.append( + FailedConfigItem( + { + "dataset": dataset, + "config": config, + "error": best_response.response["content"], + } + ) + ) + continue + splits.extend( + [ + SplitItem({"dataset": dataset, "config": config, "split": split_content["split"]}) + for split_content in best_response.response["content"]["splits"] + ] + ) + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + progress = (total - len(pending)) / total if total else 1.0 + + return ( + DatasetSplitNamesResponse( + { + "splits": splits, + "pending": pending, + "failed": failed, + } + ), + progress, + ) + + +class DatasetSplitNamesJobRunner(JobRunner): + @staticmethod + def get_job_type() -> str: + return "dataset-split-names" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION + + def compute(self) -> JobResult: + if self.dataset is None: + raise ValueError("dataset is required") + response_content, progress = compute_dataset_split_names_response(dataset=self.dataset) + return JobResult(response_content, progress=progress) + + def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by the compute.""" + return { + SplitFullName(dataset=split_item["dataset"], config=split_item["config"], split=split_item["split"]) + for split_item in content["splits"] + } diff --git a/services/worker/tests/job_runners/dataset/test_split_names.py b/services/worker/tests/job_runners/dataset/test_split_names.py new file mode 100644 index 00000000..e473ab4e --- /dev/null +++ b/services/worker/tests/job_runners/dataset/test_split_names.py @@ -0,0 +1,300 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Callable + +import pytest +from libcommon.dataset import DatasetNotFoundError +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import SplitFullName, upsert_response + +from worker.config import AppConfig +from worker.job_runners.dataset.split_names import ( + DatasetSplitNamesJobRunner, + PreviousStepFormatError, +) + +GetJobRunner = Callable[[str, AppConfig, bool], DatasetSplitNamesJobRunner] + + [email protected] +def get_job_runner( + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + app_config: AppConfig, + force: bool = False, + ) -> DatasetSplitNamesJobRunner: + return DatasetSplitNamesJobRunner( + job_info={ + "type": DatasetSplitNamesJobRunner.get_job_type(), + "dataset": dataset, + "config": None, + "split": None, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + common_config=app_config.common, + worker_config=app_config.worker, + processing_step=ProcessingStep( + name=DatasetSplitNamesJobRunner.get_job_type(), + input_type="dataset", + requires=[], + required_by_dataset_viewer=False, + ancestors=[], + children=[], + job_runner_version=DatasetSplitNamesJobRunner.get_job_runner_version(), + ), + ) + + return _get_job_runner + + [email protected]( + "dataset,split_names,expected_content,progress", + [ + ( + "pending_response", + [ + { + "config": "config_a", + "response": { + "splits": [ + { + "dataset": "pending_response", + "config": "config_a", + "split": "split_a", + } + ] + }, + } + ], + { + "splits": [ + { + "dataset": "pending_response", + "config": "config_a", + "split": "split_a", + }, + ], + "pending": [{"dataset": "pending_response", "config": "config_b"}], + "failed": [], + }, + 0.5, + ), + ( + "complete", + [ + { + "config": "config_a", + "response": { + "splits": [ + { + "dataset": "complete", + "config": "config_a", + "split": "split_a", + } + ] + }, + }, + { + "config": "config_b", + "response": { + "splits": [ + { + "dataset": "complete", + "config": "config_b", + "split": "split_b", + } + ] + }, + }, + ], + { + "splits": [ + { + "dataset": "complete", + "config": "config_a", + "split": "split_a", + }, + { + "dataset": "complete", + "config": "config_b", + "split": "split_b", + }, + ], + "pending": [], + "failed": [], + }, + 1, + ), + ], +) +def test_compute_progress( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset: str, + split_names: Any, + expected_content: Any, + progress: float, +) -> None: + # we could also have tested if dataset-info has a response (it's one case among many, see + # libcommon.simple_cache.get_best_response) + upsert_response( + kind="/config-names", + dataset=dataset, + content={ + "config_names": [ + { + "dataset": dataset, + "config": "config_a", + }, + {"dataset": dataset, "config": "config_b"}, + ] + }, + http_status=HTTPStatus.OK, + ) + for config in split_names: + # we don't really need both parent responses here, but why not (it's one case among many, see + # libcommon.simple_cache.get_best_response) + upsert_response( + kind="/split-names-from-dataset-info", + dataset=dataset, + config=config["config"], + content=config["response"], + http_status=HTTPStatus.OK, + ) + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config["config"], + content=config["response"], + http_status=HTTPStatus.OK, + ) + job_runner = get_job_runner(dataset, app_config, False) + response = job_runner.compute() + assert response.content == expected_content + assert response.progress == progress + + +def test_compute_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = "error" + config = "error" + # we could also have tested if dataset-info has a response (it's one case among many, see + # libcommon.simple_cache.get_best_response) + upsert_response( + kind="/config-names", + dataset=dataset, + content={ + "config_names": [ + { + "dataset": dataset, + "config": config, + } + ] + }, + http_status=HTTPStatus.OK, + ) + # we don't really need both parent responses here, but why not (it's one case among many, see + # libcommon.simple_cache.get_best_response) + upsert_response( + kind="/split-names-from-dataset-info", + dataset=dataset, + config=config, + content={}, + http_status=HTTPStatus.INTERNAL_SERVER_ERROR, + ) + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={}, + http_status=HTTPStatus.INTERNAL_SERVER_ERROR, + ) + job_runner = get_job_runner(dataset, app_config, False) + response = job_runner.compute() + assert response.content == { + "splits": [], + "failed": [{"dataset": dataset, "config": config, "error": {}}], + "pending": [], + } + assert response.progress == 1.0 + + +def test_compute_format_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = "error" + config = "error" + upsert_response( + kind="/config-names", + dataset=dataset, + content={ + "config_names": [ + { + "dataset": dataset, + "config": config, + } + ] + }, + http_status=HTTPStatus.OK, + ) + # here, /split-names-from-dataset-info will be picked because it's the first success response + # with progress==1.0 (see libcommon.simple_cache.get_best_response), but it's format is wrong + # while the other one (/split-names-from-streaming) is correct + upsert_response( + kind="/split-names-from-dataset-info", + dataset=dataset, + config=config, + content={"wrong_format": []}, + http_status=HTTPStatus.OK, + ) + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={"splits": [{"dataset": "dataset", "config": "config", "split": "split"}]}, + http_status=HTTPStatus.OK, + ) + job_runner = get_job_runner(dataset, app_config, False) + with pytest.raises(PreviousStepFormatError): + job_runner.compute() + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = "doesnotexist" + job_runner = get_job_runner(dataset, app_config, False) + with pytest.raises(DatasetNotFoundError): + job_runner.compute() + + +def test_get_new_splits(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = "dataset" + job_runner = get_job_runner(dataset, app_config, False) + content = { + "splits": [ + { + "dataset": dataset, + "config": "config_a", + "split": "split_a", + }, + { + "dataset": dataset, + "config": "config_b", + "split": "split_b", + }, + ], + "pending": [], + "failed": [], + } + expected = { + SplitFullName(dataset=dataset, config="config_a", split="split_a"), + SplitFullName(dataset=dataset, config="config_b", split="split_b"), + } + new_splits = job_runner.get_new_splits(content=content) + assert new_splits + assert new_splits == expected
e89a249ac3dc7548517f8c516089b9f961258565
Quentin Lhoest
2023-04-06T09:33:09
Fix RowsIndex.query (#1020)
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 0b837ed1..9eb26fd7 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -186 +186 @@ class RowsIndex: - last_row = min(offset, offset + length - 1, last_row_in_parquet) + last_row = min(offset + length - 1, last_row_in_parquet)
ea8df845160f0ca1954f6d78f37ba3f323b1372e
Sylvain Lesage
2023-04-06T08:09:12
feat: 🎸 update first-rows from streaming on 2 split-names jobs (#1021)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 578c49b3..d3765678 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -142 +142 @@ class ProcessingGraphConfig: - "requires": "/split-names-from-streaming", + "requires": ["/split-names-from-streaming", "/split-names-from-dataset-info"], diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index c98f0ade..40bf26fd 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -87 +87,7 @@ def test_default_graph() -> None: - ancestors=[config_names, split_names_from_streaming], + ancestors=[ + config_names, + split_names_from_streaming, + split_names_from_dataset_info, + parquet_and_dataset_info, + config_info, + ], @@ -106 +112 @@ def test_default_graph() -> None: - children=[dataset_split_names_from_dataset_info], + children=[split_first_rows_from_streaming, dataset_split_names_from_dataset_info], diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 2ede73b0..c045229a 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -323,4 +323,8 @@ def compute_first_rows_response( - except DoesNotExist: - raise ConfigNotFoundError(f"The config '{config}' does not exist for the dataset.'") - except Exception as e: - raise PreviousStepFormatError("Previous step did not return the expected content.") from e + except Exception: + try: + upstream_response = get_response(kind="/split-names-from-dataset-info", dataset=dataset, config=config) + splits_content = upstream_response["content"]["splits"] + except DoesNotExist as e: + raise ConfigNotFoundError(f"The config '{config}' does not exist for the dataset.'", e) from e + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index ccd53a32..6719cb68 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -142 +142 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first - ("does_not_exist_config", False, "ConfigNotFoundError", None), + ("does_not_exist_config", False, "ConfigNotFoundError", "DoesNotExist"), @@ -198 +198 @@ def test_number_rows( - elif error_code in ("InfoError", "SplitsNamesError"): + elif error_code in {"InfoError", "SplitsNamesError"}:
4fef16ffe5a8c4cf18172edf1418e6bb09d152eb
Sylvain Lesage
2023-04-05T18:43:10
feat: 🎸 let a step have multiple required steps (#1019)
diff --git a/jobs/cache_refresh/tests/test_outdated_cache.py b/jobs/cache_refresh/tests/test_outdated_cache.py index fe0c1f52..d2115233 100644 --- a/jobs/cache_refresh/tests/test_outdated_cache.py +++ b/jobs/cache_refresh/tests/test_outdated_cache.py @@ -30 +30 @@ def test_refresh_cache( - requires=None, + requires=[], @@ -32 +31,0 @@ def test_refresh_cache( - parent=None, diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index b6a610f9..65f33c78 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -8 +8 @@ from dataclasses import dataclass -from typing import List, Literal, Mapping, Optional, TypedDict +from typing import List, Literal, Mapping, TypedDict, Union @@ -20 +20 @@ class ProcessingStepSpecification(_ProcessingStepSpecification, total=False): - requires: Optional[str] + requires: Union[List[str], str, None] @@ -33,4 +33,6 @@ class ProcessingStep: - - the job parameters (mainly: ['dataset'] or ['dataset', 'config', 'split']) - - the immediately previous step required to compute the response - - the list of all the previous steps required to compute the response (in no particular order) - - the next steps (the steps which previous step is the current one, in no particular order) + - the input type ('dataset', 'config' or 'split') + - the ancestors: all the chain of previous steps, even those that are not required, in no particular order + - the children: steps that will be triggered at the end of the step, in no particular order. + + Beware: the children are computed from "requires", but with a subtlety: if c requires a and b, and if b requires a, + only b will trigger c, i.e. c will be a child of a, but not of a. @@ -41 +43 @@ class ProcessingStep: - requires: Optional[str] + requires: List[str] @@ -43 +44,0 @@ class ProcessingStep: - parent: Optional[ProcessingStep] @@ -70,0 +72,6 @@ ProcessingGraphSpecification = Mapping[str, ProcessingStepSpecification] +def get_required_steps(requires: Union[List[str], str, None]) -> List[str]: + if requires is None: + return [] + return [requires] if isinstance(requires, str) else requires + + @@ -74,2 +81,2 @@ class ProcessingGraph: - For now, the steps can have only one parent (immediate previous step), but can have multiple children - (next steps, found automatically by traversing the graph). + The steps can have multiple parents, and multiple children (next steps, found automatically by traversing the + graph). @@ -93 +100 @@ class ProcessingGraph: - requires=specification.get("requires"), + requires=get_required_steps(specification.get("requires")), @@ -95 +101,0 @@ class ProcessingGraph: - parent=None, @@ -109,2 +115,2 @@ class ProcessingGraph: - if step.requires: - graph.add_edge(step.requires, name) + for step_name in step.requires: + graph.add_edge(step_name, name) @@ -115,5 +120,0 @@ class ProcessingGraph: - if parents := set(graph.predecessors(step.name)): - if len(parents) > 1: - raise ValueError(f"Step {step.name} has multiple parents: {parents}") - step.parent = self.get_step(parents.pop()) - step.children = [self.get_step(name) for name in graph.successors(step.name)] @@ -120,0 +122,9 @@ class ProcessingGraph: + for step in self.steps.values(): + required_steps = [self.get_step(name) for name in graph.predecessors(step.name)] + parents = list(required_steps) + for parent_candidate in required_steps: + for other_parent_candidate in required_steps: + if other_parent_candidate in parent_candidate.ancestors: + parents.remove(other_parent_candidate) + for parent in parents: + parent.children.append(step) diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 495ef1c0..c98f0ade 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -4 +4 @@ -from typing import List, Optional +from typing import List @@ -7 +7,5 @@ from libcommon.config import ProcessingGraphConfig -from libcommon.processing_graph import ProcessingGraph, ProcessingStep +from libcommon.processing_graph import ( + ProcessingGraph, + ProcessingGraphSpecification, + ProcessingStep, +) @@ -20 +23,0 @@ def assert_step( - parent: Optional[ProcessingStep], @@ -25 +27,0 @@ def assert_step( - assert step.parent is parent @@ -29,0 +32,25 @@ def assert_step( +def test_graph() -> None: + specification: ProcessingGraphSpecification = { + "a": {"input_type": "dataset", "job_runner_version": 1}, + "b": {"input_type": "dataset", "job_runner_version": 1}, + "c": {"input_type": "dataset", "requires": "a", "job_runner_version": 1}, + "d": {"input_type": "dataset", "requires": ["a", "c"], "job_runner_version": 1}, + "e": {"input_type": "dataset", "requires": ["c"], "job_runner_version": 1}, + "f": {"input_type": "dataset", "requires": ["a", "b"], "job_runner_version": 1}, + } + graph = ProcessingGraph(ProcessingGraphConfig(specification).specification) + a = graph.get_step("a") + b = graph.get_step("b") + c = graph.get_step("c") + d = graph.get_step("d") + e = graph.get_step("e") + f = graph.get_step("f") + + assert_step(a, children=[c, f], ancestors=[]) + assert_step(b, children=[f], ancestors=[]) + assert_step(c, children=[d, e], ancestors=[a]) + assert_step(d, children=[], ancestors=[a, c]) + assert_step(e, children=[], ancestors=[a, c]) + assert_step(f, children=[], ancestors=[a, b]) + + @@ -50 +77 @@ def test_default_graph() -> None: - assert_step(config_names, parent=None, children=[split_names_from_streaming], ancestors=[]) + assert_step(config_names, children=[split_names_from_streaming], ancestors=[]) @@ -53 +79,0 @@ def test_default_graph() -> None: - parent=config_names, @@ -57 +83 @@ def test_default_graph() -> None: - assert_step(splits, parent=None, children=[], ancestors=[]) + assert_step(splits, children=[], ancestors=[]) @@ -60 +85,0 @@ def test_default_graph() -> None: - parent=split_names_from_streaming, @@ -64,3 +89 @@ def test_default_graph() -> None: - assert_step( - parquet_and_dataset_info, parent=None, children=[config_parquet, config_info, config_size], ancestors=[] - ) + assert_step(parquet_and_dataset_info, children=[config_parquet, config_info, config_size], ancestors=[]) @@ -69 +91,0 @@ def test_default_graph() -> None: - parent=parquet_and_dataset_info, @@ -73,3 +95 @@ def test_default_graph() -> None: - assert_step( - dataset_parquet, parent=config_parquet, children=[], ancestors=[parquet_and_dataset_info, config_parquet] - ) + assert_step(dataset_parquet, children=[], ancestors=[parquet_and_dataset_info, config_parquet]) @@ -78 +97,0 @@ def test_default_graph() -> None: - parent=parquet_and_dataset_info, @@ -82,5 +101,3 @@ def test_default_graph() -> None: - assert_step(dataset_info, parent=config_info, children=[], ancestors=[parquet_and_dataset_info, config_info]) - assert_step( - config_size, parent=parquet_and_dataset_info, children=[dataset_size], ancestors=[parquet_and_dataset_info] - ) - assert_step(dataset_size, parent=config_size, children=[], ancestors=[parquet_and_dataset_info, config_size]) + assert_step(dataset_info, children=[], ancestors=[parquet_and_dataset_info, config_info]) + assert_step(config_size, children=[dataset_size], ancestors=[parquet_and_dataset_info]) + assert_step(dataset_size, children=[], ancestors=[parquet_and_dataset_info, config_size]) @@ -89 +105,0 @@ def test_default_graph() -> None: - parent=config_info, @@ -95 +110,0 @@ def test_default_graph() -> None: - parent=split_names_from_streaming, @@ -101 +115,0 @@ def test_default_graph() -> None: - parent=split_names_from_dataset_info, @@ -107 +120,0 @@ def test_default_graph() -> None: - parent=config_parquet, diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py index 816151ce..f5396b97 100644 --- a/services/api/tests/routes/test_valid.py +++ b/services/api/tests/routes/test_valid.py @@ -14 +14 @@ dataset_step = ProcessingStep( - requires=None, + requires=[], @@ -16 +15,0 @@ dataset_step = ProcessingStep( - parent=None, @@ -24 +23 @@ config_step = ProcessingStep( - requires=None, + requires=[], @@ -26 +24,0 @@ config_step = ProcessingStep( - parent=None, @@ -34 +32 @@ split_step = ProcessingStep( - requires=None, + requires=[], @@ -36 +33,0 @@ split_step = ProcessingStep( - parent=None, diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 38156f40..088065f0 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -121 +121 @@ def test_processing_step() -> ProcessingStep: - requires=None, + requires=[], @@ -123 +122,0 @@ def test_processing_step() -> ProcessingStep: - parent=None, diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py index 7da2a824..b07e5702 100644 --- a/services/worker/tests/job_runners/config/test_info.py +++ b/services/worker/tests/job_runners/config/test_info.py @@ -164 +164 @@ def get_job_runner( - requires=None, + requires=[], @@ -166 +165,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index c48c44af..ae29300c 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -62 +62 @@ def get_job_runner( - requires=None, + requires=[], @@ -64 +63,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py index a257b3b5..8ecdc637 100644 --- a/services/worker/tests/job_runners/config/test_size.py +++ b/services/worker/tests/job_runners/config/test_size.py @@ -57 +57 @@ def get_job_runner( - requires=None, + requires=[], @@ -59 +58,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index ef2f31f8..4ab3c6a2 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -54 +54 @@ def get_job_runner( - requires=None, + requires=[], @@ -56 +55,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index 12e6fa23..150b1036 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -54 +54 @@ def get_job_runner( - requires=None, + requires=[], @@ -56 +55,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py index 0b72cf4b..880ba02a 100644 --- a/services/worker/tests/job_runners/dataset/test_info.py +++ b/services/worker/tests/job_runners/dataset/test_info.py @@ -135 +135 @@ def get_job_runner( - requires=None, + requires=[], @@ -137 +136,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index dc0f643d..a2b68d9d 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -61 +61 @@ def get_job_runner( - requires=None, + requires=[], @@ -63 +62,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/dataset/test_size.py b/services/worker/tests/job_runners/dataset/test_size.py index 668d2576..c8f9760f 100644 --- a/services/worker/tests/job_runners/dataset/test_size.py +++ b/services/worker/tests/job_runners/dataset/test_size.py @@ -58 +58 @@ def get_job_runner( - requires=None, + requires=[], @@ -60 +59,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py index 0c0fccea..8e56cfb5 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py @@ -48 +48 @@ def get_job_runner( - requires=None, + requires=[], @@ -50 +49,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py index 19ca85ad..3f1919f9 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py @@ -48 +48 @@ def get_job_runner( - requires=None, + requires=[], @@ -50 +49,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index 2848c0f3..0bad2976 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -61 +61 @@ def get_job_runner( - requires=None, + requires=[], @@ -63 +62,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index a0c9eb45..ccd53a32 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -60 +60 @@ def get_job_runner( - requires=None, + requires=[], @@ -62 +61,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py index 9af31da0..6a56ad8a 100644 --- a/services/worker/tests/job_runners/test__datasets_based_worker.py +++ b/services/worker/tests/job_runners/test__datasets_based_worker.py @@ -73 +73 @@ def get_job_runner( - requires=None, + requires=[], @@ -75 +74,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py index 28dd2955..bcf755bc 100644 --- a/services/worker/tests/job_runners/test_config_names.py +++ b/services/worker/tests/job_runners/test_config_names.py @@ -49 +49 @@ def get_job_runner( - requires=None, + requires=[], @@ -51 +50,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index b0c8c3f9..b5b1fb10 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -91 +91 @@ def get_job_runner( - requires=None, + requires=[], @@ -93 +92,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py index 25d4370b..d2f7a155 100644 --- a/services/worker/tests/job_runners/test_splits.py +++ b/services/worker/tests/job_runners/test_splits.py @@ -49 +49 @@ def get_job_runner( - requires=None, + requires=[], @@ -51 +50,0 @@ def get_job_runner( - parent=None, diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py index 097ae609..3c50f2d9 100644 --- a/services/worker/tests/test_job_runner.py +++ b/services/worker/tests/test_job_runner.py @@ -228 +228 @@ def test_check_type( - requires=None, + requires=[], @@ -230 +229,0 @@ def test_check_type( - parent=None,
819b57c975c965a987ffef9d9ef8453b5e3e6e3a
Sylvain Lesage
2023-04-05T15:00:19
Use networkx to compute the processing graph (#1016)
diff --git a/jobs/cache_refresh/poetry.lock b/jobs/cache_refresh/poetry.lock index 6c387d05..84bcce0c 100644 --- a/jobs/cache_refresh/poetry.lock +++ b/jobs/cache_refresh/poetry.lock @@ -546,0 +547 @@ mongoengine = "^0.24.2" +networkx = "^3.0" @@ -666,16 +666,0 @@ pymongo = ">=3.4,<5.0" -[[package]] -name = "mongomock" -version = "4.1.2" -description = "Fake pymongo stub for testing simple MongoDB-dependent code" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "mongomock-4.1.2-py2.py3-none-any.whl", hash = "sha256:08a24938a05c80c69b6b8b19a09888d38d8c6e7328547f94d46cadb7f47209f2"}, - {file = "mongomock-4.1.2.tar.gz", hash = "sha256:f06cd62afb8ae3ef63ba31349abd220a657ef0dd4f0243a29587c5213f931b7d"}, -] - -[package.dependencies] -packaging = "*" -sentinels = "*" - @@ -802,0 +788,19 @@ files = [ +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + @@ -1410,11 +1413,0 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] -[[package]] -name = "sentinels" -version = "1.0.0" -description = "Various objects to denote special meanings in python" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "sentinels-1.0.0.tar.gz", hash = "sha256:7be0704d7fe1925e397e92d18669ace2f619c92b5d4eb21a89f31e026f9ff4b1"}, -] - @@ -1578 +1571 @@ python-versions = "3.9.15" -content-hash = "32e1ab78f1abed143644c7bb574041748bb74ff3e8763051f48b5e253a29db56" +content-hash = "67ae1eb17c3921f944b323ddcaa58919c1deff1e7cc25e63cf3a82ec23340c62" diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock index 7989d2a3..d99b420d 100644 --- a/jobs/mongodb_migration/poetry.lock +++ b/jobs/mongodb_migration/poetry.lock @@ -559,0 +560 @@ mongoengine = "^0.24.2" +networkx = "^3.0" @@ -799,0 +801,19 @@ files = [ +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock index ae883283..e4c1e6fb 100644 --- a/libs/libcommon/poetry.lock +++ b/libs/libcommon/poetry.lock @@ -773,0 +774,19 @@ files = [ +[[package]] +name = "networkx" +version = "3.0" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.0-py3-none-any.whl", hash = "sha256:58058d66b1818043527244fab9d41a51fcd7dcc271748015f3c181b8a90c8e2e"}, + {file = "networkx-3.0.tar.gz", hash = "sha256:9a9992345353618ae98339c2b63d8201c381c2944f38a2ab49cb45a4c667e412"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=0.991)", "pre-commit (>=2.20)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.2)", "pydata-sphinx-theme (>=0.11)", "sphinx (==5.2.3)", "sphinx-gallery (>=0.11)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + @@ -1518,12 +1536,0 @@ files = [ -[[package]] -name = "types-pytz" -version = "2022.7.1.2" -description = "Typing stubs for pytz" -category = "dev" -optional = false -python-versions = "*" -files = [ - {file = "types-pytz-2022.7.1.2.tar.gz", hash = "sha256:487d3e8e9f4071eec8081746d53fa982bbc05812e719dcbf2ebf3d55a1a4cd28"}, - {file = "types_pytz-2022.7.1.2-py3-none-any.whl", hash = "sha256:40ca448a928d566f7d44ddfde0066e384f7ffbd4da2778e42a4570eaca572446"}, -] - @@ -1601 +1608 @@ python-versions = "3.9.15" -content-hash = "cfcb2f1c0c312d601ff81a829435ff17e651b1be8959b8302f336b6d6ef9d2aa" +content-hash = "93a0454f579cd3797eb20ee68dac51f5a0cb9fb1d432fa21f6c0d74222884090" diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml index b238db41..4345c277 100644 --- a/libs/libcommon/pyproject.toml +++ b/libs/libcommon/pyproject.toml @@ -13,0 +14 @@ mongoengine = "^0.24.2" +networkx = "^3.0" @@ -55,0 +57,6 @@ strict = true + +[[tool.mypy.overrides]] +module = [ + "networkx.*", +] +ignore_missing_imports = true diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index b6ba8308..b6a610f9 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -9,0 +10,2 @@ from typing import List, Literal, Mapping, Optional, TypedDict +import networkx as nx + @@ -33,2 +35,2 @@ class ProcessingStep: - - the list of all the previous steps required to compute the response - - the next steps (the steps which previous step is the current one) + - the list of all the previous steps required to compute the response (in no particular order) + - the next steps (the steps which previous step is the current one, in no particular order) @@ -63,9 +64,0 @@ class ProcessingStep: - if len(self.ancestors) > 0: - return self.ancestors - if self.parent is None: - self.ancestors = [] - else: - parent_ancestors = self.parent.get_ancestors() - if self in parent_ancestors: - raise ValueError(f"Cycle detected between {self.job_type} and {self.parent.job_type}") - self.ancestors = parent_ancestors + [self.parent] @@ -113,2 +106,3 @@ class ProcessingGraph: - for step in self.steps.values(): - # Set the parent and the children + graph = nx.DiGraph() + for name, step in self.steps.items(): + graph.add_node(name) @@ -116,5 +110,12 @@ class ProcessingGraph: - step.parent = self.get_step(step.requires) - step.parent.children.append(step) - # Set the ancestors (allows to check for cycles) - step.get_ancestors() - self.roots = [step for step in self.steps.values() if step.parent is None] + graph.add_edge(step.requires, name) + if not nx.is_directed_acyclic_graph(graph): + raise ValueError("The graph is not a directed acyclic graph.") + + for step in self.steps.values(): + if parents := set(graph.predecessors(step.name)): + if len(parents) > 1: + raise ValueError(f"Step {step.name} has multiple parents: {parents}") + step.parent = self.get_step(parents.pop()) + step.children = [self.get_step(name) for name in graph.successors(step.name)] + step.ancestors = [self.get_step(name) for name in nx.ancestors(graph, step.name)] + self.roots = [self.get_step(name) for name, degree in graph.in_degree() if degree == 0] diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 605f0b58..495ef1c0 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -3,0 +4,2 @@ +from typing import List, Optional + @@ -5 +7,21 @@ from libcommon.config import ProcessingGraphConfig -from libcommon.processing_graph import ProcessingGraph +from libcommon.processing_graph import ProcessingGraph, ProcessingStep + + +def get_step_name(step: ProcessingStep) -> str: + return step.name + + +def assert_lists_are_equal(a: List[ProcessingStep], b: List[ProcessingStep]) -> None: + assert sorted(a, key=get_step_name) == sorted(b, key=get_step_name) + + +def assert_step( + step: ProcessingStep, + parent: Optional[ProcessingStep], + children: List[ProcessingStep], + ancestors: List[ProcessingStep], +) -> None: + assert step is not None + assert step.parent is parent + assert_lists_are_equal(step.children, children) + assert_lists_are_equal(step.get_ancestors(), ancestors) @@ -28,70 +50,27 @@ def test_default_graph() -> None: - assert config_names is not None - assert config_names.parent is None - assert config_names.children == [split_names_from_streaming] - assert config_names.get_ancestors() == [] - - assert split_names_from_streaming is not None - assert split_names_from_streaming.parent is config_names - assert split_names_from_streaming.children == [split_first_rows_from_streaming, dataset_split_names_from_streaming] - assert split_names_from_streaming.get_ancestors() == [config_names] - - assert splits is not None - assert splits.parent is None - assert splits.children == [] - assert splits.get_ancestors() == [] - - assert split_first_rows_from_streaming is not None - assert split_first_rows_from_streaming.parent is split_names_from_streaming - assert split_first_rows_from_streaming.children == [] - assert split_first_rows_from_streaming.get_ancestors() == [config_names, split_names_from_streaming] - - assert parquet_and_dataset_info is not None - assert parquet_and_dataset_info.parent is None - assert parquet_and_dataset_info.children == [config_parquet, config_info, config_size] - assert parquet_and_dataset_info.get_ancestors() == [] - - assert config_parquet is not None - assert config_parquet.parent is parquet_and_dataset_info - assert config_parquet.children == [split_first_rows_from_parquet, dataset_parquet] - assert config_parquet.get_ancestors() == [parquet_and_dataset_info] - - assert dataset_parquet is not None - assert dataset_parquet.parent is config_parquet - assert dataset_parquet.children == [] - assert dataset_parquet.get_ancestors() == [parquet_and_dataset_info, config_parquet] - - assert config_info is not None - assert config_info.parent is parquet_and_dataset_info - assert config_info.children == [dataset_info, split_names_from_dataset_info] - assert config_info.get_ancestors() == [parquet_and_dataset_info] - - assert dataset_info is not None - assert dataset_info.parent is config_info - assert dataset_info.children == [] - assert dataset_info.get_ancestors() == [parquet_and_dataset_info, config_info] - - assert split_names_from_dataset_info is not None - assert split_names_from_dataset_info.parent is config_info - assert split_names_from_dataset_info.children == [dataset_split_names_from_dataset_info] - assert split_names_from_dataset_info.get_ancestors() == [parquet_and_dataset_info, config_info] - - assert config_size is not None - assert config_size.parent is parquet_and_dataset_info - assert config_size.children == [dataset_size] - assert config_size.get_ancestors() == [parquet_and_dataset_info] - - assert dataset_size is not None - assert dataset_size.parent is config_size - assert dataset_size.children == [] - assert dataset_size.get_ancestors() == [parquet_and_dataset_info, config_size] - - assert dataset_split_names_from_streaming is not None - assert dataset_split_names_from_streaming.parent is split_names_from_streaming - assert dataset_split_names_from_streaming.children == [] - assert dataset_split_names_from_streaming.get_ancestors() == [config_names, split_names_from_streaming] - - assert dataset_split_names_from_dataset_info is not None - assert dataset_split_names_from_dataset_info.parent is split_names_from_dataset_info - assert dataset_split_names_from_dataset_info.children == [] - assert dataset_split_names_from_dataset_info.get_ancestors() == [ - parquet_and_dataset_info, + assert_step(config_names, parent=None, children=[split_names_from_streaming], ancestors=[]) + assert_step( + split_names_from_streaming, + parent=config_names, + children=[split_first_rows_from_streaming, dataset_split_names_from_streaming], + ancestors=[config_names], + ) + assert_step(splits, parent=None, children=[], ancestors=[]) + assert_step( + split_first_rows_from_streaming, + parent=split_names_from_streaming, + children=[], + ancestors=[config_names, split_names_from_streaming], + ) + assert_step( + parquet_and_dataset_info, parent=None, children=[config_parquet, config_info, config_size], ancestors=[] + ) + assert_step( + config_parquet, + parent=parquet_and_dataset_info, + children=[split_first_rows_from_parquet, dataset_parquet], + ancestors=[parquet_and_dataset_info], + ) + assert_step( + dataset_parquet, parent=config_parquet, children=[], ancestors=[parquet_and_dataset_info, config_parquet] + ) + assert_step( @@ -98,0 +78,10 @@ def test_default_graph() -> None: + parent=parquet_and_dataset_info, + children=[dataset_info, split_names_from_dataset_info], + ancestors=[parquet_and_dataset_info], + ) + assert_step(dataset_info, parent=config_info, children=[], ancestors=[parquet_and_dataset_info, config_info]) + assert_step( + config_size, parent=parquet_and_dataset_info, children=[dataset_size], ancestors=[parquet_and_dataset_info] + ) + assert_step(dataset_size, parent=config_size, children=[], ancestors=[parquet_and_dataset_info, config_size]) + assert_step( @@ -100,6 +89,22 @@ def test_default_graph() -> None: - ] - - assert split_first_rows_from_parquet is not None - assert split_first_rows_from_parquet.parent is config_parquet - assert split_first_rows_from_parquet.children == [] - assert split_first_rows_from_parquet.get_ancestors() == [parquet_and_dataset_info, config_parquet] + parent=config_info, + children=[dataset_split_names_from_dataset_info], + ancestors=[parquet_and_dataset_info, config_info], + ) + assert_step( + dataset_split_names_from_streaming, + parent=split_names_from_streaming, + children=[], + ancestors=[config_names, split_names_from_streaming], + ) + assert_step( + dataset_split_names_from_dataset_info, + parent=split_names_from_dataset_info, + children=[], + ancestors=[parquet_and_dataset_info, config_info, split_names_from_dataset_info], + ) + assert_step( + split_first_rows_from_parquet, + parent=config_parquet, + children=[], + ancestors=[parquet_and_dataset_info, config_parquet], + ) diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock index ffb5ed5d..af90c2db 100644 --- a/services/admin/poetry.lock +++ b/services/admin/poetry.lock @@ -638,0 +639 @@ mongoengine = "^0.24.2" +networkx = "^3.0" @@ -878,0 +880,19 @@ files = [ +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + diff --git a/services/api/poetry.lock b/services/api/poetry.lock index e5409739..d5d7da5d 100644 --- a/services/api/poetry.lock +++ b/services/api/poetry.lock @@ -1126,0 +1127 @@ mongoengine = "^0.24.2" +networkx = "^3.0" @@ -1537,0 +1539,19 @@ files = [ +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index c00e6483..22665d7c 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1726,0 +1727 @@ mongoengine = "^0.24.2" +networkx = "^3.0" @@ -2367,0 +2369,19 @@ files = [ +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + @@ -2727 +2746,0 @@ files = [ - {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -2748,7 +2766,0 @@ files = [ - {file = "Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:9d9a62576b68cd90f7075876f4e8444487db5eeea0e4df3ba298ee38a8d067b0"}, - {file = "Pillow-9.4.0-2-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:87708d78a14d56a990fbf4f9cb350b7d89ee8988705e58e39bdf4d82c149210f"}, - {file = "Pillow-9.4.0-2-cp37-cp37m-macosx_10_10_x86_64.whl", hash = "sha256:8a2b5874d17e72dfb80d917213abd55d7e1ed2479f38f001f264f7ce7bae757c"}, - {file = "Pillow-9.4.0-2-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:83125753a60cfc8c412de5896d10a0a405e0bd88d0470ad82e0869ddf0cb3848"}, - {file = "Pillow-9.4.0-2-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9e5f94742033898bfe84c93c831a6f552bb629448d4072dd312306bab3bd96f1"}, - {file = "Pillow-9.4.0-2-pp38-pypy38_pp73-macosx_10_10_x86_64.whl", hash = "sha256:013016af6b3a12a2f40b704677f8b51f72cb007dac785a9933d5c86a72a7fe33"}, - {file = "Pillow-9.4.0-2-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:99d92d148dd03fd19d16175b6d355cc1b01faf80dae93c6c3eb4163709edc0a9"}, @@ -3770,8 +3781,0 @@ files = [ - {file = "pyzstd-0.15.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:df80324759cb74a10818f5ea4dfb0332b6064a31305246f66b043c4f73b37918"}, - {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b4833842d5e6660c1b6cd316f102ffddb95566f12ec1f89c5630d4b511e7e084"}, - {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d9069da3bb94cac8af3fd2d7b622c7a78a248d3295a478587d880553287d56e6"}, - {file = "pyzstd-0.15.4-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:9db032965b8fabf0bac13b75706031444ebfa7214794afe80be41799d8a349fb"}, - {file = "pyzstd-0.15.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:16eb2b2b206489bf8a7b690cf3b06b24352a16ee25717cb38c4381aefb3a94ce"}, - {file = "pyzstd-0.15.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:821555ce9bc7f50130bf52dc7b7d342913e00435c68a5cac564c5f2e57f33cbf"}, - {file = "pyzstd-0.15.4-cp36-cp36m-win32.whl", hash = "sha256:4f4a96bdffa773f8820e065ec840c5d3f252b0611951b73567121757388f171c"}, - {file = "pyzstd-0.15.4-cp36-cp36m-win_amd64.whl", hash = "sha256:3b0d384a48592683edf812c1a3d75a0cf5af253b1629596567ce3a2991d6ccbc"}, @@ -4211 +4214,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -4585,2 +4587,0 @@ files = [ - {file = "tokenizers-0.13.2-cp311-cp311-macosx_10_11_universal2.whl", hash = "sha256:9eee037bb5aa14daeb56b4c39956164b2bebbe6ab4ca7779d88aa16b79bd4e17"}, - {file = "tokenizers-0.13.2-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:d1b079c4c9332048fec4cb9c2055c2373c74fbb336716a5524c9a720206d787e"}, @@ -4592 +4592,0 @@ files = [ - {file = "tokenizers-0.13.2-cp311-cp311-win_amd64.whl", hash = "sha256:fa7ef7ee380b1f49211bbcfac8a006b1a3fa2fa4c7f4ee134ae384eb4ea5e453"}, @@ -4601 +4600,0 @@ files = [ - {file = "tokenizers-0.13.2-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:f44d59bafe3d61e8a56b9e0a963075187c0f0091023120b13fbe37a87936f171"},
56fe67c3ff9bee7c9f8af7406075c0b206794e61
Albert Villanova del Moral
2023-04-04T07:58:10
Raise error when the viewer is disabled (#1005)
diff --git a/docs/source/valid.mdx b/docs/source/valid.mdx index 116db98c..af818e49 100644 --- a/docs/source/valid.mdx +++ b/docs/source/valid.mdx @@ -130,0 +131,6 @@ If a dataset is not valid, then the response looks like: +Some cases where a dataset is not valid are: +- the dataset viewer is disabled +- the dataset is gated but the access is not granted: no token is passed or the passed token is not authorized +- the dataset is private +- the dataset contains no data or the data format is not supported + diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py index 83d3e642..f5607512 100644 --- a/libs/libcommon/src/libcommon/dataset.py +++ b/libs/libcommon/src/libcommon/dataset.py @@ -17,0 +18 @@ DatasetErrorCode = Literal[ + "DisabledViewerError", @@ -77,0 +79,13 @@ class DatasetNotFoundError(DatasetError): +class DisabledViewerError(DatasetError): + """Raised when the dataset viewer is disabled.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__( + message=message, + status_code=HTTPStatus.NOT_FOUND, + code="DisabledViewerError", + cause=cause, + disclose_cause=False, + ) + + @@ -193,0 +208 @@ def get_dataset_info_for_supported_datasets( + - [`~libcommon.dataset.DisabledViewerError`]: if the dataset viewer is disabled. @@ -231,0 +247,2 @@ def get_dataset_info_for_supported_datasets( + if dataset_info.cardData and not dataset_info.cardData.get("viewer", True): + raise DisabledViewerError("The dataset viewer has been disabled on this dataset.") @@ -263,0 +281 @@ def get_dataset_git_revision( + - [`~libcommon.dataset.DisabledViewerError`]: if the dataset viewer is disabled. @@ -303,0 +322 @@ def check_support( + - [`~libcommon.dataset.DisabledViewerError`]: if the dataset viewer is disabled.
2cd18afdb4b48989e090163dd7b5d1900d983d8b
Andrea Francis Soria Jimenez
2023-04-03T14:45:18
First rows skip compute if a response already exists (#1007)
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index 4ffcdfed..6ced35a3 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -35,0 +36 @@ GeneralJobRunnerErrorCode = Literal[ + "ResponseAlreadyComputedError", @@ -181,0 +183,13 @@ class JobRunnerExceededMaximumDurationError(GeneralJobRunnerError): +class ResponseAlreadyComputedError(GeneralJobRunnerError): + """Raised when reponse has been already computed by another job runner.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__( + message=message, + status_code=HTTPStatus.INTERNAL_SERVER_ERROR, + code="ResponseAlreadyComputedError", + cause=cause, + disclose_cause=True, + ) + + @@ -532,0 +547,19 @@ class JobRunner(ABC): + + def raise_if_parallel_response_exists(self, parallel_job_type: str, parallel_job_version: int) -> None: + try: + existing_response = get_response_without_content( + kind=parallel_job_type, dataset=self.dataset, config=self.config, split=self.split + ) + dataset_git_revision = self.get_dataset_git_revision() + if ( + existing_response["http_status"] == HTTPStatus.OK + and existing_response["job_runner_version"] == parallel_job_version + and existing_response["progress"] == 1.0 # completed response + and dataset_git_revision is not None + and existing_response["dataset_git_revision"] == dataset_git_revision + ): + raise ResponseAlreadyComputedError( + f"Response has already been computed by {parallel_job_type}. Compute will be skipped." + ) + except DoesNotExist: + logging.debug(f"no cache found for {parallel_job_type}.") diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index 7d0a6493..3ff1fc3f 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -13,6 +13 @@ from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import ( - DoesNotExist, - SplitFullName, - get_response, - get_response_without_content, -) +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response @@ -68,3 +63 @@ class ResponseAlreadyComputedError(SplitNamesFromDatasetInfoJobRunnerError): -def compute_split_names_from_dataset_info_response( - dataset: str, config: str, dataset_git_revision: Optional[str] -) -> SplitsList: +def compute_split_names_from_dataset_info_response(dataset: str, config: str) -> SplitsList: @@ -99,16 +91,0 @@ def compute_split_names_from_dataset_info_response( - try: - streaming_response = get_response_without_content( - kind="/split-names-from-streaming", dataset=dataset, config=config - ) - if ( - streaming_response["http_status"] == HTTPStatus.OK - and streaming_response["job_runner_version"] == PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION - and streaming_response["progress"] == 1.0 # completed response - and dataset_git_revision is not None - and streaming_response["dataset_git_revision"] == dataset_git_revision - ): - raise ResponseAlreadyComputedError( - "Response has already been computed by /split-names-from-streaming. Compute will be skipped." - ) - except DoesNotExist: - logging.debug("no cache found for /split-names-from-streaming, will proceed to compute from config-info") @@ -150 +127,4 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): - dataset_git_revision = self.get_dataset_git_revision() + self.raise_if_parallel_response_exists( + parallel_job_type="/split-names-from-streaming", + parallel_job_version=PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, + ) @@ -152,3 +132 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): - compute_split_names_from_dataset_info_response( - dataset=self.dataset, config=self.config, dataset_git_revision=dataset_git_revision - ) + compute_split_names_from_dataset_info_response(dataset=self.dataset, config=self.config) diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index f9283ca5..e3964bf9 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -14,5 +14 @@ from libcommon.constants import ( -from libcommon.simple_cache import ( - DoesNotExist, - SplitFullName, - get_response_without_content, -) +from libcommon.simple_cache import SplitFullName @@ -71 +66,0 @@ def compute_split_names_from_streaming_response( - dataset_git_revision: Optional[str], @@ -108,17 +102,0 @@ def compute_split_names_from_streaming_response( - try: - dataset_info_response = get_response_without_content( - kind="/split-names-from-dataset-info", dataset=dataset, config=config - ) - if ( - dataset_info_response["http_status"] == HTTPStatus.OK - and dataset_info_response["job_runner_version"] == PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION - and dataset_info_response["progress"] == 1.0 # completed response - and dataset_git_revision is not None - and dataset_info_response["dataset_git_revision"] == dataset_git_revision - ): - raise ResponseAlreadyComputedError( - "Response has already been computed by /split-names-from-dataset-info. Compute will be skipped." - ) - except DoesNotExist: - logging.debug("no cache found for /split-names-from-dataset-info, will proceed to compute from streaming") - @@ -156 +134,4 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): - dataset_git_revision = self.get_dataset_git_revision() + self.raise_if_parallel_response_exists( + parallel_job_type="/split-names-from-dataset-info", + parallel_job_version=PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + ) @@ -161 +141,0 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): - dataset_git_revision=dataset_git_revision, diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index ce6e4e67..cf3c21df 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -16,0 +17 @@ from libcommon.constants import ( + PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION, @@ -326,0 +328,4 @@ class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): + self.raise_if_parallel_response_exists( + parallel_job_type="split-first-rows-from-streaming", + parallel_job_version=PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION, + ) diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 71723067..2ede73b0 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -21 +21,4 @@ from datasets import ( -from libcommon.constants import PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION +from libcommon.constants import ( + PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION, + PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION, +) @@ -54,0 +58 @@ SplitFirstRowsFromStreamingJobRunnerErrorCode = Literal[ + "ResponseAlreadyComputedError", @@ -498,0 +503,4 @@ class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): + self.raise_if_parallel_response_exists( + parallel_job_type="split-first-rows-from-parquet", + parallel_job_version=PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION, + ) diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index dc9f28af..12e6fa23 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -70 +69,0 @@ def test_process(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public - job_runner.get_dataset_git_revision = Mock(return_value="1.0.0") # type: ignore diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index 2e7cd272..2848c0f3 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -8 +8 @@ from typing import Callable -from unittest.mock import patch +from unittest.mock import Mock, patch @@ -10,0 +11 @@ import pytest +from libcommon.constants import PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION @@ -141,0 +143 @@ def test_compute( + job_runner.get_dataset_git_revision = Mock(return_value="1.0.0") # type: ignore @@ -171,0 +174,47 @@ def test_compute( + + [email protected]( + "streaming_response_status,dataset_git_revision,error_code,status_code", + [ + (HTTPStatus.OK, "CURRENT_GIT_REVISION", "ResponseAlreadyComputedError", HTTPStatus.INTERNAL_SERVER_ERROR), + (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), + (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), + ], +) +def test_response_already_computed( + app_config: AppConfig, + first_rows_config: FirstRowsConfig, + get_job_runner: GetJobRunner, + streaming_response_status: HTTPStatus, + dataset_git_revision: str, + error_code: str, + status_code: HTTPStatus, +) -> None: + dataset = "dataset" + config = "config" + split = "split" + current_dataset_git_revision = "CURRENT_GIT_REVISION" + upsert_response( + kind="split-first-rows-from-streaming", + dataset=dataset, + config=config, + split=split, + content={}, + dataset_git_revision=dataset_git_revision, + job_runner_version=PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION, + progress=1.0, + http_status=streaming_response_status, + ) + job_runner = get_job_runner( + dataset, + config, + split, + app_config, + first_rows_config, + False, + ) + job_runner.get_dataset_git_revision = Mock(return_value=current_dataset_git_revision) # type: ignore + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.status_code == status_code + assert exc_info.value.code == error_code diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index f734fcc9..a0c9eb45 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -6,0 +7 @@ from typing import Callable +from unittest.mock import Mock @@ -9,0 +11 @@ from datasets.packaged_modules import csv +from libcommon.constants import PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION @@ -175,0 +178,2 @@ def test_number_rows( + job_runner.get_dataset_git_revision = Mock(return_value="1.0.0") # type: ignore + @@ -265,0 +270,2 @@ def test_truncation( + job_runner.get_dataset_git_revision = Mock(return_value="1.0.0") # type: ignore + @@ -272,0 +279,47 @@ def test_truncation( + + [email protected]( + "streaming_response_status,dataset_git_revision,error_code,status_code", + [ + (HTTPStatus.OK, "CURRENT_GIT_REVISION", "ResponseAlreadyComputedError", HTTPStatus.INTERNAL_SERVER_ERROR), + (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), + (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "ConfigNotFoundError", HTTPStatus.NOT_FOUND), + ], +) +def test_response_already_computed( + app_config: AppConfig, + first_rows_config: FirstRowsConfig, + get_job_runner: GetJobRunner, + streaming_response_status: HTTPStatus, + dataset_git_revision: str, + error_code: str, + status_code: HTTPStatus, +) -> None: + dataset = "dataset" + config = "config" + split = "split" + current_dataset_git_revision = "CURRENT_GIT_REVISION" + upsert_response( + kind="split-first-rows-from-parquet", + dataset=dataset, + config=config, + split=split, + content={}, + dataset_git_revision=dataset_git_revision, + job_runner_version=PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION, + progress=1.0, + http_status=streaming_response_status, + ) + job_runner = get_job_runner( + dataset, + config, + split, + app_config, + first_rows_config, + False, + ) + job_runner.get_dataset_git_revision = Mock(return_value=current_dataset_git_revision) # type: ignore + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.status_code == status_code + assert exc_info.value.code == error_code diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py index 2ac22706..097ae609 100644 --- a/services/worker/tests/test_job_runner.py +++ b/services/worker/tests/test_job_runner.py @@ -3,0 +4 @@ from typing import Any, Mapping, Optional +from unittest.mock import Mock @@ -6,0 +8 @@ from libcommon.config import CommonConfig +from libcommon.exceptions import CustomError @@ -335,0 +338,39 @@ def test_job_runner_set_crashed( + + +def test_raise_if_parallel_response_exists( + test_processing_step: ProcessingStep, +) -> None: + dataset = "dataset" + config = "config" + split = "split" + current_dataset_git_revision = "CURRENT_GIT_REVISION" + upsert_response( + kind="dummy-parallel", + dataset=dataset, + config=config, + split=split, + content={}, + dataset_git_revision=current_dataset_git_revision, + job_runner_version=1, + progress=1.0, + http_status=HTTPStatus.OK, + ) + job_runner = DummyJobRunner( + job_info={ + "job_id": "job_id", + "type": "/dummy", + "dataset": dataset, + "config": config, + "split": split, + "force": False, + "priority": Priority.NORMAL, + }, + processing_step=test_processing_step, + common_config=CommonConfig(), + worker_config=WorkerConfig(), + ) + job_runner.get_dataset_git_revision = Mock(return_value=current_dataset_git_revision) # type: ignore + with pytest.raises(CustomError) as exc_info: + job_runner.raise_if_parallel_response_exists(parallel_job_type="dummy-parallel", parallel_job_version=1) + assert exc_info.value.status_code == HTTPStatus.INTERNAL_SERVER_ERROR + assert exc_info.value.code == "ResponseAlreadyComputedError"
ecb861b5e8d728b80391f580e63c8d2cad63a1fc
Albert Villanova del Moral
2023-04-03T08:41:02
Update datasets dependency to 2.11.0 version (#1010)
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index d769fb04..c00e6483 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -866 +866 @@ name = "datasets" -version = "2.10.1" +version = "2.11.0" @@ -872,2 +872,2 @@ files = [ - {file = "datasets-2.10.1-py3-none-any.whl", hash = "sha256:bfde7253b31abfd075f9ad55961b13c135a5a9295cea04ae86c47b9c0e0466fd"}, - {file = "datasets-2.10.1.tar.gz", hash = "sha256:e2764c90aa3af96450a9747a934b8893b121f79f58d89e123cb1a7046bb8e81e"}, + {file = "datasets-2.11.0-py3-none-any.whl", hash = "sha256:d946cdb8c4885d3016a2ab3129c9403dd3358fe9107e8ab5e549ceab672774af"}, + {file = "datasets-2.11.0.tar.gz", hash = "sha256:1ca53b9cd6ece7a3fdb81176dadd5b9e646420e52e68e85307b27db3a36ca18c"}, @@ -880 +880 @@ fsspec = {version = ">=2021.11.1", extras = ["http"]} -huggingface-hub = ">=0.2.0,<1.0.0" +huggingface-hub = ">=0.11.0,<1.0.0" @@ -887 +887 @@ Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} -pyarrow = ">=6.0.0" +pyarrow = ">=8.0.0" @@ -890,0 +891 @@ responses = "<0.19" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} @@ -896 +897 @@ apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] -audio = ["librosa"] +audio = ["librosa", "soundfile (>=0.12.1)"] @@ -898 +899 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -901 +902 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "sqlalchemy (<2.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -906 +907 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -4209,0 +4211 @@ files = [ + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -5390 +5392 @@ python-versions = "3.9.15" -content-hash = "22687b53309b318053980c87bcec9741ce1be71bcbb792cf510a721d92919e5e" +content-hash = "ad80aa527fe00375e13bb2be4ba2eb8fa4b658fb34756c14db4aa84b9ecef75e" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index 6a3d4378..f8db5f27 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -14 +14 @@ conllu = "^4.5.2" -datasets = { extras = ["audio", "vision"], version = "^2.10.1" } +datasets = { extras = ["audio", "vision"], version = "^2.11.0" }
2021c40d96a9402a2b864a16d18d18a1a3c4379d
Sylvain Lesage
2023-03-31T15:37:02
feat: 🎸 move COMMON_LOG_LEVEL to LOG_LEVEL (#1006)
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml index 90852c97..e4febb35 100644 --- a/.github/workflows/_e2e_tests.yml +++ b/.github/workflows/_e2e_tests.yml @@ -39 +39 @@ jobs: - COMMON_LOG_LEVEL: "DEBUG" + LOG_LEVEL: "DEBUG" @@ -78 +78 @@ jobs: - COMMON_LOG_LEVEL: "DEBUG" + LOG_LEVEL: "DEBUG" diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 39b97b92..6509195f 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -80,0 +81,2 @@ common: + +log: @@ -82 +84 @@ common: - logLevel: "DEBUG" + level: "DEBUG" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index d03ecce9..a0465b2c 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -78,0 +79,2 @@ common: + +log: @@ -80 +82 @@ common: - logLevel: "INFO" + level: "INFO" diff --git a/chart/templates/_envCommon.tpl b/chart/templates/_envCommon.tpl index 50263629..feae619d 100644 --- a/chart/templates/_envCommon.tpl +++ b/chart/templates/_envCommon.tpl @@ -23,2 +22,0 @@ -- name: COMMON_LOG_LEVEL - value: {{ .Values.common.logLevel | quote }} @@ -41 +39 @@ value: {{ .Values.secrets.mongoUrl.value }} -{{- end -}} \ No newline at end of file +{{- end -}} diff --git a/chart/templates/_envLog.tpl b/chart/templates/_envLog.tpl new file mode 100644 index 00000000..01077c7a --- /dev/null +++ b/chart/templates/_envLog.tpl @@ -0,0 +1,7 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "envLog" -}} +- name: LOG_LEVEL + value: {{ .Values.log.level | quote }} +{{- end -}} diff --git a/chart/templates/jobs/cache-refresh/_container.tpl b/chart/templates/jobs/cache-refresh/_container.tpl index 758baf57..9bf1c796 100644 --- a/chart/templates/jobs/cache-refresh/_container.tpl +++ b/chart/templates/jobs/cache-refresh/_container.tpl @@ -11 +11 @@ - {{ include "envCommon" . | nindent 2 }} + {{ include "envLog" . | nindent 2 }} diff --git a/chart/templates/jobs/mongodb-migration/_container.tpl b/chart/templates/jobs/mongodb-migration/_container.tpl index 0ce5725d..4a78df16 100644 --- a/chart/templates/jobs/mongodb-migration/_container.tpl +++ b/chart/templates/jobs/mongodb-migration/_container.tpl @@ -11 +11 @@ - {{ include "envCommon" . | nindent 2 }} + {{ include "envLog" . | nindent 2 }} diff --git a/chart/templates/services/admin/_container.tpl b/chart/templates/services/admin/_container.tpl index 8ca3220e..dfd0d6df 100644 --- a/chart/templates/services/admin/_container.tpl +++ b/chart/templates/services/admin/_container.tpl @@ -12,0 +13 @@ + {{ include "envLog" . | nindent 2 }} diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl index 35f357c0..c28f92fa 100644 --- a/chart/templates/services/api/_container.tpl +++ b/chart/templates/services/api/_container.tpl @@ -11,0 +12 @@ + {{ include "envLog" . | nindent 2 }} diff --git a/chart/templates/worker/_container.tpl b/chart/templates/worker/_container.tpl index 0df0ade7..bcd65a1e 100644 --- a/chart/templates/worker/_container.tpl +++ b/chart/templates/worker/_container.tpl @@ -12,0 +13 @@ + {{ include "envLog" . | nindent 2 }} diff --git a/chart/values.yaml b/chart/values.yaml index 57d59865..7fac12fc 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -61,0 +62,2 @@ common: + +log: @@ -63 +65 @@ common: - logLevel: "INFO" + level: "INFO" diff --git a/e2e/Makefile b/e2e/Makefile index d26328ae..eca2c0f8 100644 --- a/e2e/Makefile +++ b/e2e/Makefile @@ -14 +14 @@ export COMMON_HF_TOKEN := hf_app_datasets-server_token -export COMMON_LOG_LEVEL := DEBUG +export LOG_LEVEL := DEBUG diff --git a/jobs/cache_refresh/src/cache_refresh/config.py b/jobs/cache_refresh/src/cache_refresh/config.py index 4daeb631..27dc2944 100644 --- a/jobs/cache_refresh/src/cache_refresh/config.py +++ b/jobs/cache_refresh/src/cache_refresh/config.py @@ -6 +6 @@ from dataclasses import dataclass, field -from libcommon.config import CacheConfig, CommonConfig, QueueConfig +from libcommon.config import CacheConfig, LogConfig, QueueConfig @@ -12 +12 @@ class JobConfig: - common: CommonConfig = field(default_factory=CommonConfig) + log: LogConfig = field(default_factory=LogConfig) @@ -19 +19 @@ class JobConfig: - common=CommonConfig.from_env(), + log=LogConfig.from_env(), diff --git a/jobs/cache_refresh/src/cache_refresh/main.py b/jobs/cache_refresh/src/cache_refresh/main.py index e9e7e7f8..1a181227 100644 --- a/jobs/cache_refresh/src/cache_refresh/main.py +++ b/jobs/cache_refresh/src/cache_refresh/main.py @@ -19 +19 @@ def run_job() -> None: - init_logging(log_level=job_config.common.log_level) + init_logging(level=job_config.log.level) diff --git a/jobs/mongodb_migration/src/mongodb_migration/config.py b/jobs/mongodb_migration/src/mongodb_migration/config.py index dcecb7a6..26d60a22 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/config.py +++ b/jobs/mongodb_migration/src/mongodb_migration/config.py @@ -7 +7 @@ from environs import Env -from libcommon.config import CacheConfig, CommonConfig, QueueConfig +from libcommon.config import CacheConfig, LogConfig, QueueConfig @@ -31 +31 @@ class JobConfig: - common: CommonConfig = field(default_factory=CommonConfig) + log: LogConfig = field(default_factory=LogConfig) @@ -38 +38 @@ class JobConfig: - common=CommonConfig.from_env(), + log=LogConfig.from_env(), diff --git a/jobs/mongodb_migration/src/mongodb_migration/main.py b/jobs/mongodb_migration/src/mongodb_migration/main.py index a5f41e6a..6a747aab 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/main.py +++ b/jobs/mongodb_migration/src/mongodb_migration/main.py @@ -19 +19 @@ def run_job() -> None: - init_logging(log_level=job_config.common.log_level) + init_logging(level=job_config.log.level) diff --git a/libs/libcommon/README.md b/libs/libcommon/README.md index 53027901..04b6fba1 100644 --- a/libs/libcommon/README.md +++ b/libs/libcommon/README.md @@ -18 +18,6 @@ Set the common environment variables to configure the following aspects: -- `COMMON_LOG_LEVEL`: log level, among `DEBUG`, `INFO`, `WARNING`, `ERROR`, and `CRITICAL`. Defaults to `INFO`. + +## Logs configuration + +Set the common environment variables to configure the logs: + +- `LOG_LEVEL`: log level, among `DEBUG`, `INFO`, `WARNING`, `ERROR`, and `CRITICAL`. Defaults to `INFO`. diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 084da8ef..578c49b3 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -51 +50,0 @@ COMMON_HF_TOKEN = None -COMMON_LOG_LEVEL = logging.INFO @@ -58 +56,0 @@ class CommonConfig: - log_level: int = COMMON_LOG_LEVEL @@ -67 +65,16 @@ class CommonConfig: - log_level=env.log_level(name="LOG_LEVEL", default=COMMON_LOG_LEVEL), + ) + + +LOG_LEVEL = logging.INFO + + +@dataclass(frozen=True) +class LogConfig: + level: int = LOG_LEVEL + + @classmethod + def from_env(cls) -> "LogConfig": + env = Env(expand_vars=True) + with env.prefixed("LOG_"): + return cls( + level=env.log_level(name="LEVEL", default=LOG_LEVEL), diff --git a/libs/libcommon/src/libcommon/log.py b/libs/libcommon/src/libcommon/log.py index c531c5dd..4f5e15f3 100644 --- a/libs/libcommon/src/libcommon/log.py +++ b/libs/libcommon/src/libcommon/log.py @@ -7,2 +7,2 @@ import logging -def init_logging(log_level: int = logging.INFO) -> None: - logging.basicConfig(level=log_level, format="%(levelname)s: %(asctime)s - %(name)s - %(message)s") +def init_logging(level: int = logging.INFO) -> None: + logging.basicConfig(level=level, format="%(levelname)s: %(asctime)s - %(name)s - %(message)s") diff --git a/libs/libcommon/tests/test_config.py b/libs/libcommon/tests/test_config.py index 83b06a16..71f68f73 100644 --- a/libs/libcommon/tests/test_config.py +++ b/libs/libcommon/tests/test_config.py @@ -4 +4 @@ -from libcommon.config import CommonConfig +from libcommon.config import LogConfig @@ -7,3 +7,3 @@ from libcommon.config import CommonConfig -def test_common_config() -> None: - common_config = CommonConfig() - assert common_config.log_level == 20 +def test_log_config() -> None: + log_config = LogConfig() + assert log_config.level == 20 diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py index 8007127a..83dab1a3 100644 --- a/services/admin/src/admin/app.py +++ b/services/admin/src/admin/app.py @@ -34 +34 @@ def create_app() -> Starlette: - init_logging(log_level=app_config.common.log_level) + init_logging(level=app_config.log.level) diff --git a/services/admin/src/admin/config.py b/services/admin/src/admin/config.py index 64313c20..4434dcb0 100644 --- a/services/admin/src/admin/config.py +++ b/services/admin/src/admin/config.py @@ -11,0 +12 @@ from libcommon.config import ( + LogConfig, @@ -80,0 +82 @@ class AppConfig: + log: LogConfig = field(default_factory=LogConfig) @@ -90,0 +93 @@ class AppConfig: + log=LogConfig.from_env(), diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index b4b43f8b..10ab008a 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -32 +32 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi - init_logging(log_level=app_config.common.log_level) + init_logging(level=app_config.log.level) diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index fff15da2..cc24a640 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -10,0 +11 @@ from libcommon.config import ( + LogConfig, @@ -81,0 +83 @@ class AppConfig: + log: LogConfig = field(default_factory=LogConfig) @@ -90,0 +93 @@ class AppConfig: + log=LogConfig.from_env(), diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index d388b37a..586b4064 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -11,0 +12 @@ from libcommon.config import ( + LogConfig, @@ -183,0 +185 @@ class AppConfig: + log: LogConfig = field(default_factory=LogConfig) @@ -195,0 +198 @@ class AppConfig: + log=LogConfig.from_env(), diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py index 7237718d..1016263a 100644 --- a/services/worker/src/worker/main.py +++ b/services/worker/src/worker/main.py @@ -26 +26 @@ if __name__ == "__main__": - init_logging(log_level=app_config.common.log_level) + init_logging(level=app_config.log.level) diff --git a/services/worker/src/worker/start_worker_loop.py b/services/worker/src/worker/start_worker_loop.py index e5ea2a91..865e292f 100644 --- a/services/worker/src/worker/start_worker_loop.py +++ b/services/worker/src/worker/start_worker_loop.py @@ -25 +25 @@ if __name__ == "__main__": - init_logging(log_level=app_config.common.log_level) + init_logging(level=app_config.log.level) diff --git a/tools/docker-compose-base.yml b/tools/docker-compose-base.yml index 9b7b6284..a8a383e4 100644 --- a/tools/docker-compose-base.yml +++ b/tools/docker-compose-base.yml @@ -8 +8,2 @@ services: - COMMON_LOG_LEVEL: ${COMMON_LOG_LEVEL-INFO} + # log + LOG_LEVEL: ${LOG_LEVEL-INFO} diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index 6f4e91d8..1c7e753e 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -8 +8,2 @@ services: - COMMON_LOG_LEVEL: ${COMMON_LOG_LEVEL-INFO} + # log + LOG_LEVEL: ${LOG_LEVEL-INFO}
3f78bd33bfbbf531037c67f5841d76c7adbd1953
Andrea Francis Soria Jimenez
2023-03-31T13:57:25
Split first rows from parquet new Job Runner (#988)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 05025322..084da8ef 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -21,0 +22 @@ from libcommon.constants import ( + PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION, @@ -140,0 +142,5 @@ class ProcessingGraphConfig: + "split-first-rows-from-parquet": { + "input_type": "split", + "requires": "config-parquet", + "job_runner_version": PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION, + }, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index f1a67f3e..1d5683b2 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -18,0 +19 @@ PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION = 1 +PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION = 1 @@ -22,0 +24,2 @@ PROCESSING_STEP_SPLITS_VERSION = 2 + +PARQUET_REVISION = "refs/convert/parquet" diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index e9cb544c..605f0b58 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -25,0 +26 @@ def test_default_graph() -> None: + split_first_rows_from_parquet = graph.get_step("split-first-rows-from-parquet") @@ -54 +55 @@ def test_default_graph() -> None: - assert config_parquet.children == [dataset_parquet] + assert config_parquet.children == [split_first_rows_from_parquet, dataset_parquet] @@ -100,0 +102,5 @@ def test_default_graph() -> None: + assert split_first_rows_from_parquet is not None + assert split_first_rows_from_parquet.parent is config_parquet + assert split_first_rows_from_parquet.children == [] + assert split_first_rows_from_parquet.get_ancestors() == [parquet_and_dataset_info, config_parquet] + diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 1628c9e2..fff15da2 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -119 +119 @@ class EndpointConfig: - "/first-rows": {"split": ["split-first-rows-from-streaming"]}, + "/first-rows": {"split": ["split-first-rows-from-streaming", "split-first-rows-from-parquet"]}, diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index 051fffe6..c8f1f7c7 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -46 +46 @@ def test_endpoints_definition() -> None: - assert len(first_rows["split"]) == 1 # Only has one processing step + assert len(first_rows["split"]) == 2 # Has two processing steps diff --git a/services/worker/src/worker/features.py b/services/worker/src/worker/features.py index d74931c7..aacd6690 100644 --- a/services/worker/src/worker/features.py +++ b/services/worker/src/worker/features.py @@ -3,0 +4 @@ +import io @@ -7,0 +9,2 @@ from zlib import adler32 +import numpy +import soundfile # type:ignore @@ -22 +24,0 @@ from libcommon.storage import StrPath -from numpy import ndarray @@ -59 +61,5 @@ def image( - raise TypeError("image cell must be a PIL image") + try: + image_bytes = value["bytes"] + value = PILImage.open(io.BytesIO(image_bytes)) + except Exception: + raise TypeError("image cell must be a PIL image") @@ -99,2 +105,6 @@ def audio( - raise TypeError("audio cell must contain 'array' and 'sampling_rate' fields") from e - if type(array) != ndarray: + if "bytes" in value: + bytes_array, sampling_rate = soundfile.read(io.BytesIO(value["bytes"])) + array = numpy.array(bytes_array) + else: + raise TypeError("audio cell must contain 'array' and 'sampling_rate' fields") from e + if type(array) != numpy.ndarray: diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 384fc03d..370bdc73 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -33,0 +34,3 @@ from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoJob +from worker.job_runners.split.first_rows_from_parquet import ( + SplitFirstRowsFromParquetJobRunner, +) @@ -174,0 +178,10 @@ class JobRunnerFactory(BaseJobRunnerFactory): + if job_type == SplitFirstRowsFromParquetJobRunner.get_job_type(): + first_rows_config = FirstRowsConfig.from_env() + return SplitFirstRowsFromParquetJobRunner( + job_info=job_info, + app_config=self.app_config, + processing_step=processing_step, + hf_datasets_cache=self.hf_datasets_cache, + first_rows_config=first_rows_config, + assets_directory=self.assets_directory, + ) @@ -189,0 +203 @@ class JobRunnerFactory(BaseJobRunnerFactory): + SplitFirstRowsFromParquetJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py new file mode 100644 index 00000000..ce6e4e67 --- /dev/null +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -0,0 +1,347 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from functools import partial +from http import HTTPStatus +from pathlib import Path +from typing import Any, List, Literal, Mapping, Optional + +import pyarrow as pa +from datasets import Features +from fsspec import AbstractFileSystem # type: ignore +from hffs.fs import HfFileSystem +from libcommon.constants import ( + PARQUET_REVISION, + PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION, +) +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import JobInfo +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.storage import StrPath +from pyarrow.parquet import ParquetFile +from tqdm.contrib.concurrent import thread_map + +from worker.config import AppConfig, FirstRowsConfig +from worker.features import get_cell_value +from worker.job_runner import CompleteJobResult, ConfigNotFoundError, JobRunnerError +from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.utils import ( + Row, + RowItem, + SplitFirstRowsResponse, + create_truncated_row_items, + get_json_size, + to_features_list, +) + +SplitFirstRowsFromParquetJobRunnerErrorCode = Literal[ + "RowsPostProcessingError", + "TooManyColumnsError", + "TooBigContentError", + "PreviousStepStatusError", + "PreviousStepFormatError", + "ParquetResponseEmptyError", + "FileSystemError", +] + + +class SplitFirstRowsFromParquetJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: SplitFirstRowsFromParquetJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class RowsPostProcessingError(SplitFirstRowsFromParquetJobRunnerError): + """Raised when the rows could not be post-processed successfully.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "RowsPostProcessingError", cause, False) + + +class TooManyColumnsError(SplitFirstRowsFromParquetJobRunnerError): + """Raised when the dataset exceeded the max number of columns.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "TooManyColumnsError", cause, True) + + +class TooBigContentError(SplitFirstRowsFromParquetJobRunnerError): + """Raised when the first rows content exceeded the max size of bytes.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "TooBigContentError", cause, False) + + +class PreviousStepStatusError(SplitFirstRowsFromParquetJobRunnerError): + """Raised when the previous step gave an error. The job should not have been created.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) + + +class PreviousStepFormatError(SplitFirstRowsFromParquetJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +class ParquetResponseEmptyError(SplitFirstRowsFromParquetJobRunnerError): + """Raised when no parquet files were found for split.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ParquetResponseEmptyError", cause, False) + + +class FileSystemError(SplitFirstRowsFromParquetJobRunnerError): + """Raised when an error happen reading from File System.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "FileSystemError", cause, False) + + +def transform_rows( + dataset: str, + config: str, + split: str, + rows: List[RowItem], + features: Features, + assets_base_url: str, + assets_directory: StrPath, +) -> List[Row]: + return [ + { + featureName: get_cell_value( + dataset=dataset, + config=config, + split=split, + row_idx=row_idx, + cell=row["row"][featureName] if featureName in row["row"] else None, + featureName=featureName, + fieldType=fieldType, + assets_base_url=assets_base_url, + assets_directory=assets_directory, + ) + for (featureName, fieldType) in features.items() + } + for row_idx, row in enumerate(rows) + ] + + +def get_parquet_fs(dataset: str, hf_token: Optional[str]) -> AbstractFileSystem: + """Get the parquet filesystem for a dataset. + The parquet files are stored in a separate branch of the dataset repository (see PARQUET_REVISION) + Args: + dataset (str): The dataset name. + hf_token (Optional[str]): The token to access the filesystem. + Returns: + HfFileSystem: The parquet filesystem. + """ + return HfFileSystem(dataset, repo_type="dataset", revision=PARQUET_REVISION, token=hf_token) + + +def compute_first_rows_response( + dataset: str, + config: str, + split: str, + assets_base_url: str, + hf_token: Optional[str], + min_cell_bytes: int, + rows_max_bytes: int, + rows_max_number: int, + rows_min_number: int, + columns_max_number: int, + assets_directory: StrPath, +) -> SplitFirstRowsResponse: + logging.info(f"get first-rows for dataset={dataset} config={config} split={split}") + + # first ensure the tuple (dataset, config, split) exists on the Hub + try: + upstream_response = get_response(kind="config-parquet", dataset=dataset, config=config) + if upstream_response["http_status"] != HTTPStatus.OK: + raise PreviousStepStatusError( + f"Previous step gave an error: {upstream_response['http_status']}. This job should not have been" + " created." + ) + + parquet_files_content = upstream_response["content"]["parquet_files"] + sources = sorted( + f"{config}/{parquet_file['filename']}" + for parquet_file in parquet_files_content + if parquet_file["split"] == split and parquet_file["config"] == config + ) + if not sources: + raise ParquetResponseEmptyError("No parquet files found.") + except DoesNotExist: + raise ConfigNotFoundError(f"The config '{config}' does not exist for the dataset.'") + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.") from e + + logging.debug(f"Found {len(sources)} parquet files for {dataset=}, {config=}, {split=}: {sources}") + + fs = get_parquet_fs(dataset=dataset, hf_token=hf_token) + desc = f"{dataset}/{config}/{split}" + try: + parquet_files: List[ParquetFile] = thread_map( + partial(ParquetFile, filesystem=fs), sources, desc=desc, unit="pq", disable=True + ) + except Exception as e: + raise FileSystemError(f"Could not read the parquet files: {e}") from e + + # get the features + features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema()) + + if features and len(features) > columns_max_number: + raise TooManyColumnsError( + f"The number of columns ({len(features)}) exceeds the maximum supported number of columns" + f" ({columns_max_number}). This is a current limitation of the datasets viewer. You can reduce the number" + " of columns if you want the viewer to work." + ) + + # validate size of response without the rows + features_list = to_features_list(features=features) + response_features_only: SplitFirstRowsResponse = { + "dataset": dataset, + "config": config, + "split": split, + "features": features_list, + "rows": [], + } + + surrounding_json_size = get_json_size(response_features_only) + if surrounding_json_size > rows_max_bytes: + raise TooBigContentError( + f"The size of the content of the first rows ({surrounding_json_size}) exceeds the maximum" + f" supported size ({rows_max_bytes} B) even after truncation. Please report the issue." + ) + + # get the rows + num_rows = 0 + last_row_group_id = 0 + row_group_readers = [] + for parquet_file in parquet_files: + for group_id in range(parquet_file.metadata.num_row_groups): + last_row_group_id = group_id + row_group_readers.append(partial(parquet_file.read_row_group, i=group_id)) + if num_rows + parquet_file.metadata.row_group(group_id).num_rows >= rows_max_number: + num_rows = rows_max_number + break + else: + num_rows += parquet_file.metadata.row_group(group_id).num_rows + else: + continue + break + + if len(row_group_readers) == 0: + raise ParquetResponseEmptyError("No parquet files found.") + + pa_table = pa.concat_tables([row_group_readers[i]() for i in range(0, last_row_group_id + 1)]) + result = pa_table.slice(0, num_rows) + + rows = [ + RowItem( + { + "row_idx": idx, + "row": row, + "truncated_cells": [], + } + ) + for idx, row in enumerate(result.to_pylist()) + ] + + # transform the rows, if needed (e.g. save the images or audio to the assets, and return their URL) + try: + transformed_rows = transform_rows( + dataset=dataset, + config=config, + split=split, + rows=rows, + features=features, + assets_base_url=assets_base_url, + assets_directory=assets_directory, + ) + except Exception as err: + raise RowsPostProcessingError( + "Server error while post-processing the split rows. Please report the issue.", + cause=err, + ) from err + + # truncate the rows to fit within the restrictions, and prepare them as RowItems + row_items = create_truncated_row_items( + rows=transformed_rows, + min_cell_bytes=min_cell_bytes, + rows_max_bytes=rows_max_bytes - surrounding_json_size, + rows_min_number=rows_min_number, + ) + + response = response_features_only + response["rows"] = row_items + return response + + +class SplitFirstRowsFromParquetJobRunner(DatasetsBasedJobRunner): + assets_directory: StrPath + first_rows_config: FirstRowsConfig + + @staticmethod + def get_job_type() -> str: + return "split-first-rows-from-parquet" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION + + def __init__( + self, + job_info: JobInfo, + app_config: AppConfig, + processing_step: ProcessingStep, + first_rows_config: FirstRowsConfig, + hf_datasets_cache: Path, + assets_directory: StrPath, + ) -> None: + super().__init__( + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + hf_datasets_cache=hf_datasets_cache, + ) + self.first_rows_config = first_rows_config + self.assets_directory = assets_directory + self.assets_base_url = app_config.assets.base_url + + def compute(self) -> CompleteJobResult: + if self.config is None or self.split is None: + raise ValueError("config and split are required") + return CompleteJobResult( + compute_first_rows_response( + dataset=self.dataset, + config=self.config, + split=self.split, + assets_base_url=self.assets_base_url, + assets_directory=self.assets_directory, + hf_token=self.common_config.hf_token, + min_cell_bytes=self.first_rows_config.min_cell_bytes, + rows_max_bytes=self.first_rows_config.max_bytes, + rows_max_number=self.first_rows_config.max_number, + rows_min_number=self.first_rows_config.min_number, + columns_max_number=self.first_rows_config.columns_max_number, + ) + ) + + def get_new_splits(self, _: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by compute.""" + if self.config is None or self.split is None: + raise ValueError("config and split are required") + return {SplitFullName(dataset=self.dataset, config=self.config, split=self.split)} diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 1e4e83b8..71723067 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -11,12 +11 @@ from pathlib import Path -from typing import ( - Any, - Callable, - List, - Literal, - Mapping, - Optional, - TypedDict, - TypeVar, - Union, - cast, -) +from typing import Any, Callable, List, Literal, Mapping, Optional, TypeVar, Union, cast @@ -30,2 +18,0 @@ from datasets import ( - get_dataset_config_names, - get_dataset_split_names, @@ -37,3 +24 @@ from libcommon.queue import JobInfo -from libcommon.simple_cache import DoesNotExist -from libcommon.simple_cache import SplitFullName as _SplitFullName -from libcommon.simple_cache import get_response +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response @@ -41 +25,0 @@ from libcommon.storage import StrPath -from libcommon.utils import orjson_dumps @@ -51,0 +36,6 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.utils import ( + SplitFirstRowsResponse, + create_truncated_row_items, + get_json_size, + to_features_list, +) @@ -192,20 +181,0 @@ Row = Mapping[str, Any] -class FeatureItem(TypedDict): - feature_idx: int - name: str - type: Mapping[str, Any] - - -class RowItem(TypedDict): - row_idx: int - row: Mapping[str, Any] - truncated_cells: List[str] - - -class SplitFirstRowsFromStreamingResponse(TypedDict): - dataset: str - config: str - split: str - features: List[FeatureItem] - rows: List[RowItem] - - @@ -244,124 +213,0 @@ def get_rows( -def get_json_size(obj: Any) -> int: - """Returns the size of an object in bytes once serialized as JSON - - Args: - obj (Any): the Python object - - Returns: - int: the size of the serialized object in bytes - """ - return len(orjson_dumps(obj)) - - -# from https://stackoverflow.com/a/43848928/7351594 -def utf8_lead_byte(b: int) -> bool: - """A UTF-8 intermediate byte starts with the bits 10xxxxxx.""" - return (b & 0xC0) != 0x80 - - -def utf8_byte_truncate(text: str, max_bytes: int) -> str: - """If text[max_bytes] is not a lead byte, back up until a lead byte is - found and truncate before that character.""" - utf8 = text.encode("utf8") - if len(utf8) <= max_bytes: - return text - i = max_bytes - while i > 0 and not utf8_lead_byte(utf8[i]): - i -= 1 - return utf8[:i].decode("utf8", "ignore") - - -# Mutates row_item, and returns it anyway -def truncate_row_item(row_item: RowItem, min_cell_bytes: int) -> RowItem: - row = {} - for column_name, cell in row_item["row"].items(): - # for now: all the cells above min_cell_bytes are truncated to min_cell_bytes - # it's done by replacing the cell (which can have any type) by a string with - # its JSON serialization, and then truncating it to min_cell_bytes - cell_json = orjson_dumps(cell) - if len(cell_json) <= min_cell_bytes: - row[column_name] = cell - else: - cell_json_str = cell_json.decode("utf8", "ignore") - row_item["truncated_cells"].append(column_name) - row[column_name] = utf8_byte_truncate(text=cell_json_str, max_bytes=min_cell_bytes) - row_item["row"] = row - return row_item - - -COMMA_SIZE = 1 # the comma "," is encoded with one byte in utf-8 - - -# Mutates row_items, and returns them anyway -def truncate_row_items(row_items: List[RowItem], min_cell_bytes: int, rows_max_bytes: int) -> List[RowItem]: - # compute the current size - rows_bytes = sum(get_json_size(row_item) for row_item in row_items) + COMMA_SIZE * (len(row_items) - 1) - - # Loop backwards, so that the last rows are truncated first - for row_item in reversed(row_items): - if rows_bytes < rows_max_bytes: - break - previous_size = get_json_size(row_item) + COMMA_SIZE - row_item = truncate_row_item(row_item=row_item, min_cell_bytes=min_cell_bytes) - new_size = get_json_size(row_item) + COMMA_SIZE - rows_bytes += new_size - previous_size - row_idx = row_item["row_idx"] - logging.debug(f"the size of the rows is now ({rows_bytes}) after truncating row idx={row_idx}") - return row_items - - -def to_row_item(row_idx: int, row: Row) -> RowItem: - return { - "row_idx": row_idx, - "row": row, - "truncated_cells": [], - } - - -def create_truncated_row_items( - rows: List[Row], - min_cell_bytes: int, - rows_max_bytes: int, - rows_min_number: int, -) -> List[RowItem]: - row_items = [] - rows_bytes = 0 - - # two restrictions must be enforced: - # - at least rows_min_number rows - # - at most rows_max_bytes bytes. Note that it's the limit to the sum of the rows sizes. The JSON response size - # will be greater, due to the other fields (row_idx, truncated_cells, features, etc.). - # To enforce this: - # 1. first get the first rows_min_number rows - for row_idx, row in enumerate(rows[:rows_min_number]): - row_item = to_row_item(row_idx=row_idx, row=row) - rows_bytes += get_json_size(row_item) + COMMA_SIZE - row_items.append(row_item) - - # 2. if the total is over the bytes limit, truncate the values, iterating backwards starting - # from the last rows, until getting under the threshold - # caveat: the truncation might not be enough to get under the threshold if: - # - the number of columns is too high - # - rows_max_bytes is too low (or even negative) - if rows_bytes >= rows_max_bytes: - logging.debug( - f"the size of the first {rows_min_number} rows ({rows_bytes}) is above the max number of bytes" - f" ({rows_max_bytes}), they will be truncated" - ) - return truncate_row_items(row_items=row_items, min_cell_bytes=min_cell_bytes, rows_max_bytes=rows_max_bytes) - - # 3. else: add the remaining rows until the end, or until the bytes threshold - for idx, row in enumerate(rows[rows_min_number:]): - row_idx = rows_min_number + idx - row_item = to_row_item(row_idx=row_idx, row=row) - rows_bytes += get_json_size(row_item) + COMMA_SIZE - if rows_bytes >= rows_max_bytes: - logging.debug( - f"the rows in the split have been truncated to {row_idx} row(s) to keep the size" - f" ({rows_bytes}) under the limit ({rows_max_bytes})" - ) - break - row_items.append(row_item) - return row_items - - @@ -396,34 +241,0 @@ def transform_rows( -# in JSON, dicts do not carry any order, so we need to return a list -# -# > An object is an *unordered* collection of zero or more name/value pairs, where a name is a string and a value -# is a string, number, boolean, null, object, or array. -# > An array is an *ordered* sequence of zero or more values. -# > The terms "object" and "array" come from the conventions of JavaScript. -# from https://stackoverflow.com/a/7214312/7351594 / https://www.rfc-editor.org/rfc/rfc7159.html -def to_features_list(features: Features) -> List[FeatureItem]: - features_dict = features.to_dict() - return [ - { - "feature_idx": idx, - "name": name, - "type": features_dict[name], - } - for idx, name in enumerate(features) - ] - - -class SplitFullName(TypedDict): - dataset: str - config: str - split: str - - -def get_dataset_split_full_names(dataset: str, use_auth_token: Union[bool, str, None] = False) -> List[SplitFullName]: - logging.info(f"get dataset '{dataset}' split full names") - return [ - {"dataset": dataset, "config": config, "split": split} - for config in get_dataset_config_names(path=dataset, use_auth_token=use_auth_token) - for split in get_dataset_split_names(path=dataset, config_name=config, use_auth_token=use_auth_token) - ] - - @@ -443 +255 @@ def compute_first_rows_response( -) -> SplitFirstRowsFromStreamingResponse: +) -> SplitFirstRowsResponse: @@ -478 +290 @@ def compute_first_rows_response( - [`SplitFirstRowsFromStreamingResponse`]: The list of first rows of the split. + [`SplitFirstRowsResponse`]: The list of first rows of the split. @@ -567 +379 @@ def compute_first_rows_response( - response_features_only: SplitFirstRowsFromStreamingResponse = { + response_features_only: SplitFirstRowsResponse = { @@ -703 +515 @@ class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): - def get_new_splits(self, _: Mapping[str, Any]) -> set[_SplitFullName]: + def get_new_splits(self, _: Mapping[str, Any]) -> set[SplitFullName]: @@ -707 +519 @@ class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): - return {_SplitFullName(dataset=self.dataset, config=self.config, split=self.split)} + return {SplitFullName(dataset=self.dataset, config=self.config, split=self.split)} diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 687fc272..a706ce3a 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -5,0 +6,3 @@ from typing import Any, List, Mapping, Optional, TypedDict +from datasets import Features +from libcommon.utils import orjson_dumps + @@ -37,0 +41,166 @@ class PreviousJob(TypedDict): + + +class FeatureItem(TypedDict): + feature_idx: int + name: str + type: Mapping[str, Any] + + +class RowItem(TypedDict): + row_idx: int + row: Mapping[str, Any] + truncated_cells: List[str] + + +class SplitFirstRowsResponse(TypedDict): + dataset: str + config: str + split: str + features: List[FeatureItem] + rows: List[RowItem] + + +# in JSON, dicts do not carry any order, so we need to return a list +# +# > An object is an *unordered* collection of zero or more name/value pairs, where a name is a string and a value +# is a string, number, boolean, null, object, or array. +# > An array is an *ordered* sequence of zero or more values. +# > The terms "object" and "array" come from the conventions of JavaScript. +# from https://stackoverflow.com/a/7214312/7351594 / https://www.rfc-editor.org/rfc/rfc7159.html +def to_features_list(features: Features) -> List[FeatureItem]: + features_dict = features.to_dict() + return [ + { + "feature_idx": idx, + "name": name, + "type": features_dict[name], + } + for idx, name in enumerate(features) + ] + + +def get_json_size(obj: Any) -> int: + """Returns the size of an object in bytes once serialized as JSON + + Args: + obj (Any): the Python object + + Returns: + int: the size of the serialized object in bytes + """ + return len(orjson_dumps(obj)) + + +# from https://stackoverflow.com/a/43848928/7351594 +def utf8_lead_byte(b: int) -> bool: + """A UTF-8 intermediate byte starts with the bits 10xxxxxx.""" + return (b & 0xC0) != 0x80 + + +def utf8_byte_truncate(text: str, max_bytes: int) -> str: + """If text[max_bytes] is not a lead byte, back up until a lead byte is + found and truncate before that character.""" + utf8 = text.encode("utf8") + if len(utf8) <= max_bytes: + return text + i = max_bytes + while i > 0 and not utf8_lead_byte(utf8[i]): + i -= 1 + return utf8[:i].decode("utf8", "ignore") + + +# Mutates row_item, and returns it anyway +def truncate_row_item(row_item: RowItem, min_cell_bytes: int) -> RowItem: + row = {} + for column_name, cell in row_item["row"].items(): + # for now: all the cells above min_cell_bytes are truncated to min_cell_bytes + # it's done by replacing the cell (which can have any type) by a string with + # its JSON serialization, and then truncating it to min_cell_bytes + cell_json = orjson_dumps(cell) + if len(cell_json) <= min_cell_bytes: + row[column_name] = cell + else: + cell_json_str = cell_json.decode("utf8", "ignore") + row_item["truncated_cells"].append(column_name) + row[column_name] = utf8_byte_truncate(text=cell_json_str, max_bytes=min_cell_bytes) + row_item["row"] = row + # row_idx = row_item["row_idx"] + # logging.debug(f"the size of the rows is now ({rows_bytes}) after truncating row idx={row_idx}") + return row_item + + +COMMA_SIZE = 1 # the comma "," is encoded with one byte in utf-8 + + +# Mutates row_items, and returns them anyway +def truncate_row_items(row_items: List[RowItem], min_cell_bytes: int, rows_max_bytes: int) -> List[RowItem]: + # compute the current size + rows_bytes = sum(get_json_size(row_item) for row_item in row_items) + COMMA_SIZE * (len(row_items) - 1) + + # Loop backwards, so that the last rows are truncated first + for row_item in reversed(row_items): + if rows_bytes < rows_max_bytes: + break + previous_size = get_json_size(row_item) + COMMA_SIZE + row_item = truncate_row_item(row_item=row_item, min_cell_bytes=min_cell_bytes) + new_size = get_json_size(row_item) + COMMA_SIZE + rows_bytes += new_size - previous_size + return row_items + + +Row = Mapping[str, Any] + + +def to_row_item(row_idx: int, row: Row) -> RowItem: + return { + "row_idx": row_idx, + "row": row, + "truncated_cells": [], + } + + +def create_truncated_row_items( + rows: List[Row], + min_cell_bytes: int, + rows_max_bytes: int, + rows_min_number: int, +) -> List[RowItem]: + row_items = [] + rows_bytes = 0 + + # two restrictions must be enforced: + # - at least rows_min_number rows + # - at most rows_max_bytes bytes. Note that it's the limit to the sum of the rows sizes. The JSON response size + # will be greater, due to the other fields (row_idx, truncated_cells, features, etc.). + # To enforce this: + # 1. first get the first rows_min_number rows + for row_idx, row in enumerate(rows[:rows_min_number]): + row_item = to_row_item(row_idx=row_idx, row=row) + rows_bytes += get_json_size(row_item) + COMMA_SIZE + row_items.append(row_item) + + # 2. if the total is over the bytes limit, truncate the values, iterating backwards starting + # from the last rows, until getting under the threshold + # caveat: the truncation might not be enough to get under the threshold if: + # - the number of columns is too high + # - rows_max_bytes is too low (or even negative) + if rows_bytes >= rows_max_bytes: + # logging.debug( + # f"the size of the first {rows_min_number} rows ({rows_bytes}) is above the max number of bytes" + # f" ({rows_max_bytes}), they will be truncated" + # ) + return truncate_row_items(row_items=row_items, min_cell_bytes=min_cell_bytes, rows_max_bytes=rows_max_bytes) + + # 3. else: add the remaining rows until the end, or until the bytes threshold + for idx, row in enumerate(rows[rows_min_number:]): + row_idx = rows_min_number + idx + row_item = to_row_item(row_idx=row_idx, row=row) + rows_bytes += get_json_size(row_item) + COMMA_SIZE + if rows_bytes >= rows_max_bytes: + # logging.debug( + # f"the rows in the split have been truncated to {row_idx} row(s) to keep the size" + # f" ({rows_bytes}) under the limit ({rows_max_bytes})" + # ) + break + row_items.append(row_item) + return row_items diff --git a/services/worker/tests/job_runners/split/config/dataset-split.parquet b/services/worker/tests/job_runners/split/config/dataset-split.parquet new file mode 100644 index 00000000..672ab11e Binary files /dev/null and b/services/worker/tests/job_runners/split/config/dataset-split.parquet differ diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py new file mode 100644 index 00000000..2e7cd272 --- /dev/null +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -0,0 +1,171 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import os +from dataclasses import replace +from http import HTTPStatus +from typing import Callable +from unittest.mock import patch + +import pytest +from libcommon.exceptions import CustomError +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import DoesNotExist, get_response, upsert_response +from libcommon.storage import StrPath +from pyarrow.fs import LocalFileSystem + +from worker.config import AppConfig, FirstRowsConfig +from worker.job_runners.split.first_rows_from_parquet import ( + SplitFirstRowsFromParquetJobRunner, +) +from worker.resources import LibrariesResource +from worker.utils import get_json_size + +from ...fixtures.hub import get_default_config_split + +GetJobRunner = Callable[[str, str, str, AppConfig, FirstRowsConfig, bool], SplitFirstRowsFromParquetJobRunner] + + [email protected] +def get_job_runner( + assets_directory: StrPath, + libraries_resource: LibrariesResource, + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + split: str, + app_config: AppConfig, + first_rows_config: FirstRowsConfig, + force: bool = False, + ) -> SplitFirstRowsFromParquetJobRunner: + return SplitFirstRowsFromParquetJobRunner( + job_info={ + "type": SplitFirstRowsFromParquetJobRunner.get_job_type(), + "dataset": dataset, + "config": config, + "split": split, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + app_config=app_config, + processing_step=ProcessingStep( + name=SplitFirstRowsFromParquetJobRunner.get_job_type(), + input_type="split", + requires=None, + required_by_dataset_viewer=True, + parent=None, + ancestors=[], + children=[], + job_runner_version=SplitFirstRowsFromParquetJobRunner.get_job_runner_version(), + ), + hf_datasets_cache=libraries_resource.hf_datasets_cache, + first_rows_config=first_rows_config, + assets_directory=assets_directory, + ) + + return _get_job_runner + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first_rows_config: FirstRowsConfig) -> None: + dataset = "doesnotexist" + dataset, config, split = get_default_config_split(dataset) + job_runner = get_job_runner(dataset, config, split, app_config, first_rows_config, False) + assert not job_runner.process() + with pytest.raises(DoesNotExist): + get_response(kind=job_runner.processing_step.cache_kind, dataset=dataset, config=config, split=split) + + [email protected]( + "rows_max_bytes,columns_max_number,error_code", + [ + (0, 10, "TooBigContentError"), # too small limit, even with truncation + (1_000, 1, "TooManyColumnsError"), # too small columns limit + (1_000, 10, None), + ], +) +def test_compute( + get_job_runner: GetJobRunner, + app_config: AppConfig, + first_rows_config: FirstRowsConfig, + rows_max_bytes: int, + columns_max_number: int, + error_code: str, +) -> None: + dataset, config, split = "dataset", "config", "split" + upsert_response( + kind="config-parquet", + dataset=dataset, + config=config, + content={ + "parquet_files": [ + { + "dataset": dataset, + "config": config, + "split": split, + "filename": f"{dataset}-{split}.parquet", + "size": 1000, + } + ] + }, + http_status=HTTPStatus.OK, + ) + + with patch("worker.job_runners.split.first_rows_from_parquet.get_parquet_fs") as mock_read: + initial_location = os.getcwd() + os.chdir("tests/job_runners/split") + # TODO: Make localsystem by relative path + fs = LocalFileSystem() + mock_read.return_value = fs + # ^ Mocking file system with local file + job_runner = get_job_runner( + dataset, + config, + split, + replace(app_config, common=replace(app_config.common, hf_token=None)), + replace( + first_rows_config, + max_number=1_000_000, + min_number=10, + max_bytes=rows_max_bytes, + min_cell_bytes=10, + columns_max_number=columns_max_number, + ), + False, + ) + + if error_code: + with pytest.raises(CustomError) as error_info: + job_runner.compute() + assert error_info.value.code == error_code + else: + response = job_runner.compute().content + assert get_json_size(response) <= rows_max_bytes + assert response + assert response["rows"] + assert response["features"] + assert len(response["rows"]) == 3 # testing file has 3 rows see config/dataset-split.parquet file + assert len(response["features"]) == 2 # testing file has 2 columns see config/dataset-split.parquet file + assert response["features"][0]["feature_idx"] == 0 + assert response["features"][0]["name"] == "col1" + assert response["features"][0]["type"]["_type"] == "Value" + assert response["features"][0]["type"]["dtype"] == "int32" + assert response["features"][1]["feature_idx"] == 1 + assert response["features"][1]["name"] == "col2" + assert response["features"][1]["type"]["_type"] == "Value" + assert response["features"][1]["type"]["dtype"] == "string" + assert response["rows"][0]["row_idx"] == 0 + assert response["rows"][0]["truncated_cells"] == [] + assert response["rows"][0]["row"] == {"col1": 1, "col2": "a"} + assert response["rows"][1]["row_idx"] == 1 + assert response["rows"][1]["truncated_cells"] == [] + assert response["rows"][1]["row"] == {"col1": 2, "col2": "b"} + assert response["rows"][2]["row_idx"] == 2 + assert response["rows"][2]["truncated_cells"] == [] + assert response["rows"][2]["row"] == {"col1": 3, "col2": "c"} + os.chdir(initial_location) diff --git a/services/worker/tests/job_runners/split/test_first_rows.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py similarity index 99% rename from services/worker/tests/job_runners/split/test_first_rows.py rename to services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index 9619c227..f734fcc9 100644 --- a/services/worker/tests/job_runners/split/test_first_rows.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -20 +19,0 @@ from worker.job_runners.split.first_rows_from_streaming import ( - get_json_size, @@ -22,0 +22 @@ from worker.resources import LibrariesResource +from worker.utils import get_json_size
153fc25691ccc61b2392d0c3fab5ae5cf0423b4c
Albert Villanova del Moral
2023-03-31T13:48:53
Remove assignee from Dataset Viewer Issue template (#993)
diff --git a/.github/ISSUE_TEMPLATE/hub-dataset-viewer.yml b/.github/ISSUE_TEMPLATE/hub-dataset-viewer.yml index fb48b3ce..31df50c6 100644 --- a/.github/ISSUE_TEMPLATE/hub-dataset-viewer.yml +++ b/.github/ISSUE_TEMPLATE/hub-dataset-viewer.yml @@ -5,2 +4,0 @@ labels: ["bug"] -assignees: - - severo
49e010aca716ba79fb1f1065b52ad6eebcfa8dfc
Andrea Francis Soria Jimenez
2023-03-30T19:20:45
check if /splits-name-from-* exists before processing it (#995)
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index 714a6a6b..4ffcdfed 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -206,0 +207 @@ class JobRunner(ABC): + _dataset_git_revision: Optional[str] = None @@ -287,3 +288,5 @@ class JobRunner(ABC): - return get_dataset_git_revision( - dataset=self.dataset, hf_endpoint=self.common_config.hf_endpoint, hf_token=self.common_config.hf_token - ) + if self._dataset_git_revision is None: + self._dataset_git_revision = get_dataset_git_revision( + dataset=self.dataset, hf_endpoint=self.common_config.hf_endpoint, hf_token=self.common_config.hf_token + ) + return self._dataset_git_revision diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index 8037079a..7d0a6493 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -8 +8,4 @@ from typing import Any, List, Literal, Mapping, Optional -from libcommon.constants import PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION +from libcommon.constants import ( + PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, +) @@ -10 +13,6 @@ from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response +from libcommon.simple_cache import ( + DoesNotExist, + SplitFullName, + get_response, + get_response_without_content, +) @@ -18,0 +27 @@ SplitNamesFromDatasetInfoJobRunnerErrorCode = Literal[ + "ResponseAlreadyComputedError", @@ -51,0 +61,7 @@ class PreviousStepFormatError(SplitNamesFromDatasetInfoJobRunnerError): +class ResponseAlreadyComputedError(SplitNamesFromDatasetInfoJobRunnerError): + """Raised when reponse has been already computed by /split-names-from-streaming job runner.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseAlreadyComputedError", cause, True) + + @@ -53,2 +69 @@ def compute_split_names_from_dataset_info_response( - dataset: str, - config: str, + dataset: str, config: str, dataset_git_revision: Optional[str] @@ -78,0 +94,2 @@ def compute_split_names_from_dataset_info_response( + - [`~job_runners.split_names_from_dataset_info.ResponseAlreadyComputedError`] + If reponse has been already computed by /split-names-from-streaming job runner. @@ -81,0 +99,16 @@ def compute_split_names_from_dataset_info_response( + try: + streaming_response = get_response_without_content( + kind="/split-names-from-streaming", dataset=dataset, config=config + ) + if ( + streaming_response["http_status"] == HTTPStatus.OK + and streaming_response["job_runner_version"] == PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION + and streaming_response["progress"] == 1.0 # completed response + and dataset_git_revision is not None + and streaming_response["dataset_git_revision"] == dataset_git_revision + ): + raise ResponseAlreadyComputedError( + "Response has already been computed by /split-names-from-streaming. Compute will be skipped." + ) + except DoesNotExist: + logging.debug("no cache found for /split-names-from-streaming, will proceed to compute from config-info") @@ -116,0 +150 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): + dataset_git_revision = self.get_dataset_git_revision() @@ -118 +152,3 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): - compute_split_names_from_dataset_info_response(dataset=self.dataset, config=self.config) + compute_split_names_from_dataset_info_response( + dataset=self.dataset, config=self.config, dataset_git_revision=dataset_git_revision + ) diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index ed1835f1..f9283ca5 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -10,2 +10,9 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError -from libcommon.constants import PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION -from libcommon.simple_cache import SplitFullName +from libcommon.constants import ( + PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, +) +from libcommon.simple_cache import ( + DoesNotExist, + SplitFullName, + get_response_without_content, +) @@ -19,0 +27 @@ SplitNamesFromStreamingJobRunnerErrorCode = Literal[ + "ResponseAlreadyComputedError", @@ -52,0 +61,7 @@ class EmptyDatasetError(SplitNamesFromStreamingJobRunnerError): +class ResponseAlreadyComputedError(SplitNamesFromStreamingJobRunnerError): + """Raised when reponse has been already computed by /split-names-from-dataset-info job runner.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseAlreadyComputedError", cause, True) + + @@ -55,0 +71 @@ def compute_split_names_from_streaming_response( + dataset_git_revision: Optional[str], @@ -83 +99 @@ def compute_split_names_from_streaming_response( - - [`~job_runners.split_names.EmptyDatasetError`] + - [`~job_runners.split_names_from_streaming.EmptyDatasetError`] @@ -85 +101 @@ def compute_split_names_from_streaming_response( - - [`~job_runners.split_names.SplitsNamesError`] + - [`~job_runners.split_names_from_streaming.SplitsNamesError`] @@ -86,0 +103,2 @@ def compute_split_names_from_streaming_response( + - [`~job_runners.split_names_from_streaming.ResponseAlreadyComputedError`] + If reponse has been already computed by /split-names-from-dataset-info job runner. @@ -89,0 +108,17 @@ def compute_split_names_from_streaming_response( + try: + dataset_info_response = get_response_without_content( + kind="/split-names-from-dataset-info", dataset=dataset, config=config + ) + if ( + dataset_info_response["http_status"] == HTTPStatus.OK + and dataset_info_response["job_runner_version"] == PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION + and dataset_info_response["progress"] == 1.0 # completed response + and dataset_git_revision is not None + and dataset_info_response["dataset_git_revision"] == dataset_git_revision + ): + raise ResponseAlreadyComputedError( + "Response has already been computed by /split-names-from-dataset-info. Compute will be skipped." + ) + except DoesNotExist: + logging.debug("no cache found for /split-names-from-dataset-info, will proceed to compute from streaming") + @@ -116,0 +152,2 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): + if self.dataset is None: + raise ValueError("dataset is required") @@ -118,0 +156 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): + dataset_git_revision = self.get_dataset_git_revision() @@ -121 +159,4 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): - dataset=self.dataset, config=self.config, hf_token=self.common_config.hf_token + dataset=self.dataset, + config=self.config, + dataset_git_revision=dataset_git_revision, + hf_token=self.common_config.hf_token, diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index 113099ed..ef2f31f8 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -5,0 +6 @@ from typing import Any, Callable +from unittest.mock import Mock @@ -7,0 +9 @@ import pytest +from libcommon.constants import PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION @@ -129,0 +132,2 @@ def test_compute( + job_runner.get_dataset_git_revision = Mock(return_value="1.0.0") # type: ignore + @@ -145,0 +150,37 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No + + [email protected]( + "streaming_response_status,dataset_git_revision,error_code,status_code", + [ + (HTTPStatus.OK, "CURRENT_GIT_REVISION", "ResponseAlreadyComputedError", HTTPStatus.INTERNAL_SERVER_ERROR), + (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "DatasetNotFoundError", HTTPStatus.NOT_FOUND), + (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "DatasetNotFoundError", HTTPStatus.NOT_FOUND), + ], +) +def test_response_already_computed( + app_config: AppConfig, + get_job_runner: GetJobRunner, + streaming_response_status: HTTPStatus, + dataset_git_revision: str, + error_code: str, + status_code: HTTPStatus, +) -> None: + dataset = "dataset" + config = "config" + current_dataset_git_revision = "CURRENT_GIT_REVISION" + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={}, + dataset_git_revision=dataset_git_revision, + job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, + progress=1.0, + http_status=streaming_response_status, + ) + job_runner = get_job_runner(dataset, config, app_config, False) + job_runner.get_dataset_git_revision = Mock(return_value=current_dataset_git_revision) # type: ignore + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.status_code == status_code + assert exc_info.value.code == error_code diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index 91855c8a..dc9f28af 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -6,0 +7 @@ from typing import Callable +from unittest.mock import Mock @@ -8,0 +10 @@ import pytest +from libcommon.constants import PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION @@ -13 +15 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import DoesNotExist, get_response +from libcommon.simple_cache import DoesNotExist, get_response, upsert_response @@ -67,0 +70 @@ def test_process(app_config: AppConfig, get_job_runner: GetJobRunner, hub_public + job_runner.get_dataset_git_revision = Mock(return_value="1.0.0") # type: ignore @@ -119,0 +123 @@ def test_compute_split_names_from_streaming_response( + job_runner.get_dataset_git_revision = Mock(return_value="1.0.0") # type: ignore @@ -136,0 +141,36 @@ def test_compute_split_names_from_streaming_response( + + [email protected]( + "dataset_info_response_status,dataset_git_revision,error_code", + [ + (HTTPStatus.OK, "CURRENT_GIT_REVISION", "ResponseAlreadyComputedError"), + (HTTPStatus.INTERNAL_SERVER_ERROR, "CURRENT_GIT_REVISION", "SplitNamesFromStreamingError"), + (HTTPStatus.OK, "DIFFERENT_GIT_REVISION", "SplitNamesFromStreamingError"), + ], +) +def test_response_already_computed( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset_info_response_status: HTTPStatus, + dataset_git_revision: str, + error_code: str, +) -> None: + dataset = "dataset" + config = "config" + current_dataset_git_revision = "CURRENT_GIT_REVISION" + upsert_response( + kind="/split-names-from-dataset-info", + dataset=dataset, + config=config, + content={}, + dataset_git_revision=dataset_git_revision, + job_runner_version=PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + progress=1.0, + http_status=dataset_info_response_status, + ) + job_runner = get_job_runner(dataset, config, app_config, False) + job_runner.get_dataset_git_revision = Mock(return_value=current_dataset_git_revision) # type: ignore + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.status_code == HTTPStatus.INTERNAL_SERVER_ERROR + assert exc_info.value.code == error_code
7b01da3f667370c51c0d753b483d2ebc109bcc49
Andrea Francis Soria Jimenez
2023-03-29T22:02:04
Updating dependencies in worker to support first rows from parquet (#996)
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 0934c65b..d769fb04 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. +# This file is automatically @generated by Poetry and should not be changed by hand. @@ -139,65 +138,0 @@ frozenlist = ">=1.1.0" -[[package]] -name = "apache-beam" -version = "2.41.0" -description = "Apache Beam SDK for Python" -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "apache-beam-2.41.0.zip", hash = "sha256:c1a0456a5b48c3481bf20dc904e4d812515144336873b322f17ba188e2fabd92"}, - {file = "apache_beam-2.41.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8fe21bf02554ad6ca1c1b19d37afc0c08b9d0676fb4b5a9b1f4a17303edd94d4"}, - {file = "apache_beam-2.41.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:2b3a06b0cb73a2e1b5ad892c3fab36bd8454ac8abee3cae23832c03ab1cc7121"}, - {file = "apache_beam-2.41.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:82faeb079d9612918b0ad7cbd12fa54ca56b6cb8175c043804446eb2744d965b"}, - {file = "apache_beam-2.41.0-cp37-cp37m-manylinux2010_i686.whl", hash = "sha256:ea0f500e16ffb0e5932c802abd301dc042e88ef27ded0935b9b8cac58113c43c"}, - {file = "apache_beam-2.41.0-cp37-cp37m-manylinux2010_x86_64.whl", hash = "sha256:e4c101f8c1427ced3c17525540c81d05137415dc4398d5d51df4c70608aee46d"}, - {file = "apache_beam-2.41.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:615612971a6aeb15eb41697945a25bb726f6f7410c71ce029fe3196cdf486edb"}, - {file = "apache_beam-2.41.0-cp37-cp37m-win32.whl", hash = "sha256:2f9395f2faefaa28306081e8fc7ebc5e7157f6734001bfcf9233c077e3d8b2ca"}, - {file = "apache_beam-2.41.0-cp37-cp37m-win_amd64.whl", hash = "sha256:bb216e5890279988c490428cfbce62a03fcdd90d6feca13ec29a074c8c3cfe36"}, - {file = "apache_beam-2.41.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:cbad1f7d207224efbe8e461ceacf8c21e21e4a5a011cde87e5881649e24b5e4e"}, - {file = "apache_beam-2.41.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:dab8ee4b15cc2608bf5a715167150210246924cca65ddc6847afa21f5211c22f"}, - {file = "apache_beam-2.41.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:cca61ef1cc417ce2eca8f331fa0f8f9bbceaf6e67460a048527ebd1c33562d16"}, - {file = "apache_beam-2.41.0-cp38-cp38-manylinux2010_i686.whl", hash = "sha256:c25ab457f4ebe356fe3726c5e7554ca29e975bf5df67ff20e339fc5e568ed550"}, - {file = "apache_beam-2.41.0-cp38-cp38-manylinux2010_x86_64.whl", hash = "sha256:b7c8ca34772c26cd3103b36b69f2c31fa834ac5bb85859ac9dd51b64a2100b5b"}, - {file = "apache_beam-2.41.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:9abcb01b85fd27eaac29ea90757da1d95e293ca2fdaf6b69192020ee05d71257"}, - {file = "apache_beam-2.41.0-cp38-cp38-win32.whl", hash = "sha256:7b6581739ea8d5a346b4a722d1d280adf748d74a5c31322288a0fa9ba3204645"}, - {file = "apache_beam-2.41.0-cp38-cp38-win_amd64.whl", hash = "sha256:7c2ab828a7a3b8973f5f01101fd7746a8562a20f5f390b07e301744afa2a83d7"}, - {file = "apache_beam-2.41.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:3e28758eec094b7f5585e92d8a6f9b5745a6b335d646b8fd58b6dd7f99109e67"}, - {file = "apache_beam-2.41.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:ea94024188e8aa1eb9774be66cf368d44c08cf3b34626fca4803bb33c353b72d"}, - {file = "apache_beam-2.41.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:06da861c4092f64ed9868375e8049ae26b3208c105f3f93268eacd3c7a35e1b9"}, - {file = "apache_beam-2.41.0-cp39-cp39-manylinux2010_i686.whl", hash = "sha256:e511f2cf7d767810ad51ddca72ab93992d0bbd310984f36d5a5659276f3e5e98"}, - {file = "apache_beam-2.41.0-cp39-cp39-manylinux2010_x86_64.whl", hash = "sha256:4777dbbb0ed371cf7c72d784acdefba5963d61bac11a3b62875b5817fad2d608"}, - {file = "apache_beam-2.41.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:bf3e5d838122d8fdf8fc1a1752ff2b661ad5a0641bb62dc227e433b557022acb"}, - {file = "apache_beam-2.41.0-cp39-cp39-win32.whl", hash = "sha256:fa9d2f4ce10662950fa9bec37295d8c1a50dbd4b558ea61693c6bd455d7db790"}, - {file = "apache_beam-2.41.0-cp39-cp39-win_amd64.whl", hash = "sha256:f978f7a815e2ee00c9bcec756b1aa7114ab4ba4c572978a48610f1bed6bc5e35"}, -] - -[package.dependencies] -cloudpickle = ">=2.1.0,<3" -crcmod = ">=1.7,<2.0" -dill = ">=0.3.1.1,<0.3.2" -fastavro = ">=0.23.6,<2" -grpcio = ">=1.33.1,<2" -hdfs = ">=2.1.0,<3.0.0" -httplib2 = ">=0.8,<0.21.0" -numpy = ">=1.14.3,<1.23.0" -orjson = "<4.0" -proto-plus = ">=1.7.1,<2" -protobuf = ">=3.12.2,<4" -pyarrow = ">=0.15.1,<8.0.0" -pydot = ">=1.2.0,<2" -pymongo = ">=3.8.0,<4.0.0" -python-dateutil = ">=2.8.0,<3" -pytz = ">=2018.3" -requests = ">=2.24.0,<3.0.0" -typing-extensions = ">=3.7.0" - -[package.extras] -aws = ["boto3 (>=1.9)"] -azure = ["azure-core (>=1.7.0)", "azure-storage-blob (>=12.3.2)"] -dataframe = ["pandas (>=1.0,<1.5)"] -docs = ["Sphinx (>=1.5.2,<2.0)", "docutils (==0.17.1)"] -gcp = ["cachetools (>=3.1.0,<5)", "google-api-core (!=2.8.2,<3)", "google-apitools (>=0.5.31,<0.5.32)", "google-auth (>=1.18.0,<3)", "google-auth-httplib2 (>=0.1.0,<0.2.0)", "google-cloud-bigquery (>=1.6.0,<3)", "google-cloud-bigquery-storage (>=2.6.3,<2.14)", "google-cloud-bigtable (>=0.31.1,<2)", "google-cloud-core (>=0.28.1,<3)", "google-cloud-datastore (>=1.8.0,<2)", "google-cloud-dlp (>=3.0.0,<4)", "google-cloud-language (>=1.3.0,<2)", "google-cloud-pubsub (>=2.1.0,<3)", "google-cloud-pubsublite (>=1.2.0,<2)", "google-cloud-recommendations-ai (>=0.1.0,<0.8.0)", "google-cloud-spanner (>=1.13.0,<2)", "google-cloud-videointelligence (>=1.8.0,<2)", "google-cloud-vision (>=0.38.0,<2)", "grpcio-gcp (>=0.2.2,<1)"] -interactive = ["facets-overview (>=1.0.0,<2)", "google-cloud-dataproc (>=3.0.0,<3.2.0)", "ipykernel (>=6,<7)", "ipython (>=7,<8)", "ipython (>=8,<9)", "ipywidgets (>=7.6.5,<8)", "jupyter-client (>=6.1.11,<6.1.13)", "timeloop (>=1.0.2,<2)"] -interactive-test = ["chromedriver-binary (>=100,<101)", "nbconvert (>=6.2.0,<7)", "nbformat (>=5.0.5,<6)", "needle (>=0.5.0,<1)", "pillow (>=7.1.1,<8)"] -test = ["cryptography (>=36.0.0)", "freezegun (>=0.3.12)", "joblib (>=1.0.1)", "mock (>=1.0.1,<3.0.0)", "pandas (<2.0.0)", "parameterized (>=0.7.1,<0.9.0)", "psycopg2-binary (>=2.8.5,<3.0.0)", "pyhamcrest (>=1.9,!=1.10.0,<2.0.0)", "pytest (>=4.4.0,<5.0)", "pytest-timeout (>=1.3.3,<2)", "pytest-xdist (>=1.29.0,<2)", "pyyaml (>=3.12,<7.0.0)", "requests-mock (>=1.7,<2.0)", "scikit-learn (>=0.20.0)", "sqlalchemy (>=1.3,<2.0)", "tenacity (>=5.0.2,<6.0)", "testcontainers[mysql] (>=3.0.3,<4.0.0)"] - @@ -735,12 +669,0 @@ colorama = {version = "*", markers = "platform_system == \"Windows\""} -[[package]] -name = "cloudpickle" -version = "2.2.1" -description = "Extended pickling support for Python objects" -category = "main" -optional = false -python-versions = ">=3.6" -files = [ - {file = "cloudpickle-2.2.1-py3-none-any.whl", hash = "sha256:61f594d1f4c295fa5cd9014ceb3a1fc4a70b0de1164b94fbc2d854ccba056f9f"}, - {file = "cloudpickle-2.2.1.tar.gz", hash = "sha256:d89684b8de9e34a2a43b3460fbca07d09d6e25ce858df4d5a44240403b6178f5"}, -] - @@ -923,11 +845,0 @@ files = [ -[[package]] -name = "crcmod" -version = "1.7" -description = "CRC Generator" -category = "main" -optional = false -python-versions = "*" -files = [ - {file = "crcmod-1.7.tar.gz", hash = "sha256:dc7051a0db5f2bd48665a990d3ec1cc305a466a77358ca4492826f41f283601e"}, -] - @@ -1040,11 +951,0 @@ idna = ["idna (>=2.1)"] -[[package]] -name = "docopt" -version = "0.6.2" -description = "Pythonic argument parser, that will make you smile" -category = "main" -optional = false -python-versions = "*" -files = [ - {file = "docopt-0.6.2.tar.gz", hash = "sha256:49b3a825280bd66b3aa83585ef59c4a8c82f2c8a522dbe754a8bc8d08c85c491"}, -] - @@ -1100,47 +1000,0 @@ test = ["pytest (>=6)"] -[[package]] -name = "fastavro" -version = "1.7.3" -description = "Fast read/write of AVRO files" -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "fastavro-1.7.3-cp310-cp310-macosx_11_0_x86_64.whl", hash = "sha256:10a5ac9d8c66d4ba24f25ad7313e2dab56d98ceebcf53ba9cfa88acdd135c794"}, - {file = "fastavro-1.7.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b2e6d8bb79e53dc39e620c777f14b5f7122f1bf21309a9fcf60085f8e062e49c"}, - {file = "fastavro-1.7.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9a0ba2f43844eb784f8abf5324a0c10474287beaecb14fb736e47136464e3044"}, - {file = "fastavro-1.7.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e20db96b092d7b6208f3063a424d35bb48c283e2d8b4e7ad4ee6541dc1fac2ed"}, - {file = "fastavro-1.7.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:52ba6bb5525561df577ebd94819784626caac9d8ad2ed167030403ba1bf73159"}, - {file = "fastavro-1.7.3-cp310-cp310-win_amd64.whl", hash = "sha256:22d6f3e73f471e2b4ba0785cb60df939792e8904db4ba93037ba6b7858f7d6f9"}, - {file = "fastavro-1.7.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:5dd5299cbc5bc2aa15f1c619f4cc55c054c6fe9ccd614f93eb1d6ab22cf314dd"}, - {file = "fastavro-1.7.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f4706a77038bf31ad2e8cc752a0c007894bd39ffb0b775c7824113743182c5f6"}, - {file = "fastavro-1.7.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b6b0e58e7dd34906d21738c3461cddef760de3b7845779169a378b2757afa693"}, - {file = "fastavro-1.7.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:04740e2dd27084b4155337d082f2a232cf1d801a1b009f772e50c8306a8f8aaf"}, - {file = "fastavro-1.7.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:2f95c767bf97e896640f24d58931b3a19df3d84ccaf0606c92e603c79de60f16"}, - {file = "fastavro-1.7.3-cp311-cp311-win_amd64.whl", hash = "sha256:a727e07007230267e25702d5f3738854cb315747fc58b84839699db30dedf490"}, - {file = "fastavro-1.7.3-cp37-cp37m-macosx_10_15_x86_64.whl", hash = "sha256:da71b9db7718f4682cc11e0f25b5e395d5f3bc17ddaf0224f39be3bac5309cfa"}, - {file = "fastavro-1.7.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d1a62c359f4c9472c3ebe2be478e203ff434cc1d6bebaf61181a4a121c0899a6"}, - {file = "fastavro-1.7.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9e2ba001864607b46fc2f6124d690731b19db215a84751c4b3b155e70b615d05"}, - {file = "fastavro-1.7.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:cd10bffc26457402da9727663de71c40dd717d90e8ab3d3b893bc227cad5e410"}, - {file = "fastavro-1.7.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:e5d0f16b85104aa0e2899a47c186be1082a10cecf6b331571afa92a4b8e6061a"}, - {file = "fastavro-1.7.3-cp37-cp37m-win_amd64.whl", hash = "sha256:b86d1c1188ec47aeb76d6195e36ab52665984e8e98f69a224ab550c82991fe07"}, - {file = "fastavro-1.7.3-cp38-cp38-macosx_10_15_x86_64.whl", hash = "sha256:a007151cc2a08e61dd5ea5b48989849d056a8d63b04d7e6799c36fdf0b702bf4"}, - {file = "fastavro-1.7.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7b176d5731f336c2c9c88d95225f71f862b2512c33ef917b1fe7f87379cc92fd"}, - {file = "fastavro-1.7.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c52e4a1b05306f82916eacf83c732a4637a5be748bc2ef2ff6fed1506535d692"}, - {file = "fastavro-1.7.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:207bc7663133ca766eaf9033806da4cf08071dacf2e9779aa9427df40815f846"}, - {file = "fastavro-1.7.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:cc98be4ad3d8fb9000abeeae0ecb0f8e62ec7898b791da5ec2f6de81dd2a73e2"}, - {file = "fastavro-1.7.3-cp38-cp38-win_amd64.whl", hash = "sha256:fb3879aaeb3b56ee5b3a22ffa11cbdf4ba65c04be4688ee8bd152aa6535a00ee"}, - {file = "fastavro-1.7.3-cp39-cp39-macosx_11_0_x86_64.whl", hash = "sha256:8c45f7fdfab351431d106f5981fdc2313a8cbfdb82d2b1172b2a144bfba376b7"}, - {file = "fastavro-1.7.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:127e928604753d845fa0f2ae758c1640215ff901a5ce20cdf7e9f154500c3212"}, - {file = "fastavro-1.7.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b82a71a6c014ec5d03293d8dc8b698220380266d5503779fd3712a94e4497069"}, - {file = "fastavro-1.7.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:08bfd749cce456f925203895d6732f6b68c973d63ff886733f27db3c2d3c0b9a"}, - {file = "fastavro-1.7.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:dc1e88d5db17e7ebc3fc764a1091f6c05a42e3cb0e2c8eaf49126743c7ca1bb5"}, - {file = "fastavro-1.7.3-cp39-cp39-win_amd64.whl", hash = "sha256:7d525f3f99cc49a5e245e08d7ab947195a18cbdd5c43af75c0989fbe14a32597"}, - {file = "fastavro-1.7.3.tar.gz", hash = "sha256:8b08bd3cba45830b64adda32ccc5b027a71b6941a99cc39f90d7019a7986cc19"}, -] - -[package.extras] -codecs = ["lz4", "python-snappy", "zstandard"] -lz4 = ["lz4"] -snappy = ["python-snappy"] -zstandard = ["zstandard"] - @@ -1532,3 +1386,3 @@ numpy = ">=1.14.5" -name = "hdfs" -version = "2.7.0" -description = "HdfsCLI: API and command line interface for HDFS." +name = "hffs" +version = "0.0.1.dev0" +description = "Filesystem interface over huggingface.co repositories" @@ -1537,5 +1391,3 @@ optional = false -python-versions = "*" -files = [ - {file = "hdfs-2.7.0-py3-none-any.whl", hash = "sha256:3428078ad1e83a2e2a11801c536ac2aa5094f5fabde5d1e7145bacbf4a599c1e"}, - {file = "hdfs-2.7.0.tar.gz", hash = "sha256:ecd4650c39bb4f9421641320f4931edd81cf7126ae4e5ec880215adf6435df3d"}, -] +python-versions = ">=3.7.0" +files = [] +develop = false @@ -1544,3 +1396,4 @@ files = [ -docopt = "*" -requests = ">=2.7.0" -six = ">=1.9.0" +fsspec = "*" +huggingface_hub = ">=0.12.0" +packaging = ">=20.9" +requests = "*" @@ -1549,3 +1402,9 @@ six = ">=1.9.0" -avro = ["fastavro (>=0.21.19)"] -dataframe = ["fastavro (>=0.21.19)", "pandas (>=0.14.1)"] -kerberos = ["requests-kerberos (>=0.7.0)"] +dev = ["black (>=23.1,<24.0)", "pytest", "ruff (>=0.0.241)"] +quality = ["black (>=23.1,<24.0)", "ruff (>=0.0.241)"] +tests = ["pytest"] + +[package.source] +type = "git" +url = "https://github.com/huggingface/hffs.git" +reference = "0e187e74d38e9436353691f4a7a26b15f0663f58" +resolved_reference = "0e187e74d38e9436353691f4a7a26b15f0663f58" @@ -1575,15 +1433,0 @@ lxml = ["lxml"] -[[package]] -name = "httplib2" -version = "0.20.4" -description = "A comprehensive HTTP client library." -category = "main" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" -files = [ - {file = "httplib2-0.20.4-py3-none-any.whl", hash = "sha256:8b6a905cb1c79eefd03f8669fd993c36dc341f7c558f056cb5a33b5c2f458543"}, - {file = "httplib2-0.20.4.tar.gz", hash = "sha256:58a98e45b4b1a48273073f905d2961666ecf0fbac4250ea5b47aef259eb5c585"}, -] - -[package.dependencies] -pyparsing = {version = ">=2.4.2,<3.0.0 || >3.0.0,<3.0.1 || >3.0.1,<3.0.2 || >3.0.2,<3.0.3 || >3.0.3,<4", markers = "python_version > \"3.0\""} - @@ -2881,0 +2726 @@ files = [ + {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -3108,18 +2952,0 @@ xxhash = ["xxhash (>=1.4.3)"] -[[package]] -name = "proto-plus" -version = "1.22.2" -description = "Beautiful, Pythonic protocol buffers." -category = "main" -optional = false -python-versions = ">=3.6" -files = [ - {file = "proto-plus-1.22.2.tar.gz", hash = "sha256:0e8cda3d5a634d9895b75c573c9352c16486cb75deb0e078b5fda34db4243165"}, - {file = "proto_plus-1.22.2-py3-none-any.whl", hash = "sha256:de34e52d6c9c6fcd704192f09767cb561bb4ee64e70eede20b0834d841f0be4d"}, -] - -[package.dependencies] -protobuf = ">=3.19.0,<5.0.0dev" - -[package.extras] -testing = ["google-api-core[grpc] (>=1.31.5)"] - @@ -3221 +3048 @@ name = "pyarrow" -version = "7.0.0" +version = "11.0.0" @@ -3227,30 +3054,25 @@ files = [ - {file = "pyarrow-7.0.0-cp310-cp310-macosx_10_13_universal2.whl", hash = "sha256:0f15213f380539c9640cb2413dc677b55e70f04c9e98cfc2e1d8b36c770e1036"}, - {file = "pyarrow-7.0.0-cp310-cp310-macosx_10_13_x86_64.whl", hash = "sha256:29c4e3b3be0b94d07ff4921a5e410fc690a3a066a850a302fc504de5fc638495"}, - {file = "pyarrow-7.0.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:8a9bfc8a016bcb8f9a8536d2fa14a890b340bc7a236275cd60fd4fb8b93ff405"}, - {file = "pyarrow-7.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:49d431ed644a3e8f53ae2bbf4b514743570b495b5829548db51610534b6eeee7"}, - {file = "pyarrow-7.0.0-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:aa6442a321c1e49480b3d436f7d631c895048a16df572cf71c23c6b53c45ed66"}, - {file = "pyarrow-7.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f6b01a23cb401750092c6f7c4dcae67cd8fd6b99ae710e26f654f23508f25f25"}, - {file = "pyarrow-7.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0f10928745c6ff66e121552731409803bed86c66ac79c64c90438b053b5242c5"}, - {file = "pyarrow-7.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:759090caa1474cafb5e68c93a9bd6cb45d8bb8e4f2cad2f1a0cc9439bae8ae88"}, - {file = "pyarrow-7.0.0-cp37-cp37m-macosx_10_13_x86_64.whl", hash = "sha256:e3fe34bcfc28d9c4a747adc3926d2307a04c5c50b89155946739515ccfe5eab0"}, - {file = "pyarrow-7.0.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:040dce5345603e4e621bcf4f3b21f18d557852e7b15307e559bb14c8951c8714"}, - {file = "pyarrow-7.0.0-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:ed4b647c3345ae3463d341a9d28d0260cd302fb92ecf4e2e3e0f1656d6e0e55c"}, - {file = "pyarrow-7.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e7fecd5d5604f47e003f50887a42aee06cb8b7bf8e8bf7dc543a22331d9ba832"}, - {file = "pyarrow-7.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1f2d00b892fe865e43346acb78761ba268f8bb1cbdba588816590abcb780ee3d"}, - {file = "pyarrow-7.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:f439f7d77201681fd31391d189aa6b1322d27c9311a8f2fce7d23972471b02b6"}, - {file = "pyarrow-7.0.0-cp38-cp38-macosx_10_13_x86_64.whl", hash = "sha256:3e06b0e29ce1e32f219c670c6b31c33d25a5b8e29c7828f873373aab78bf30a5"}, - {file = "pyarrow-7.0.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:13dc05bcf79dbc1bd2de1b05d26eb64824b85883d019d81ca3c2eca9b68b5a44"}, - {file = "pyarrow-7.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:06183a7ff2b0c030ec0413fc4dc98abad8cf336c78c280a0b7f4bcbebb78d125"}, - {file = "pyarrow-7.0.0-cp38-cp38-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:702c5a9f960b56d03569eaaca2c1a05e8728f05ea1a2138ef64234aa53cd5884"}, - {file = "pyarrow-7.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c7313038203df77ec4092d6363dbc0945071caa72635f365f2b1ae0dd7469865"}, - {file = "pyarrow-7.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e87d1f7dc7a0b2ecaeb0c7a883a85710f5b5626d4134454f905571c04bc73d5a"}, - {file = "pyarrow-7.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:ba69488ae25c7fde1a2ae9ea29daf04d676de8960ffd6f82e1e13ca945bb5861"}, - {file = "pyarrow-7.0.0-cp39-cp39-macosx_10_13_universal2.whl", hash = "sha256:11a591f11d2697c751261c9d57e6e5b0d38fdc7f0cc57f4fd6edc657da7737df"}, - {file = "pyarrow-7.0.0-cp39-cp39-macosx_10_13_x86_64.whl", hash = "sha256:6183c700877852dc0f8a76d4c0c2ffd803ba459e2b4a452e355c2d58d48cf39f"}, - {file = "pyarrow-7.0.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:d1748154714b543e6ae8452a68d4af85caf5298296a7e5d4d00f1b3021838ac6"}, - {file = "pyarrow-7.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcc8f934c7847a88f13ec35feecffb61fe63bb7a3078bd98dd353762e969ce60"}, - {file = "pyarrow-7.0.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:759f59ac77b84878dbd54d06cf6df74ff781b8e7cf9313eeffbb5ec97b94385c"}, - {file = "pyarrow-7.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d3e3f93ac2993df9c5e1922eab7bdea047b9da918a74e52145399bc1f0099a3"}, - {file = "pyarrow-7.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:306120af554e7e137895254a3b4741fad682875a5f6403509cd276de3fe5b844"}, - {file = "pyarrow-7.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:087769dac6e567d58d59b94c4f866b3356c00d3db5b261387ece47e7324c2150"}, - {file = "pyarrow-7.0.0.tar.gz", hash = "sha256:da656cad3c23a2ebb6a307ab01d35fce22f7850059cffafcb90d12590f8f4f38"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"}, + {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"}, + {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"}, + {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"}, + {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"}, + {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"}, + {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"}, + {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"}, @@ -3448,15 +3269,0 @@ files = [ -[[package]] -name = "pydot" -version = "1.4.2" -description = "Python interface to Graphviz's Dot" -category = "main" -optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" -files = [ - {file = "pydot-1.4.2-py2.py3-none-any.whl", hash = "sha256:66c98190c65b8d2e2382a441b4c0edfdb4f4c025ef9cb9874de478fb0793a451"}, - {file = "pydot-1.4.2.tar.gz", hash = "sha256:248081a39bcb56784deb018977e428605c1c758f10897a339fce1dd728ff007d"}, -] - -[package.dependencies] -pyparsing = ">=2.1.4" - @@ -3649 +3456 @@ description = "pyparsing module - Classes and methods to define and execute pars -category = "main" +category = "dev" @@ -5583 +5390 @@ python-versions = "3.9.15" -content-hash = "f3944a99f45c8d6f063ec218eb246c3c6ef4015d6f633350ec026948efd54549" +content-hash = "22687b53309b318053980c87bcec9741ce1be71bcbb792cf510a721d92919e5e" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index 17e16b5e..6a3d4378 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -12 +11,0 @@ aiohttp = "^3.8.4" -apache-beam = "2.41.0" # ^2 gives a InvalidWheelName error because it tries to install 2.42 that has not been released... @@ -15 +14 @@ conllu = "^4.5.2" -datasets = { extras = ["audio", "vision"], version = "~2.10.0" } +datasets = { extras = ["audio", "vision"], version = "^2.10.1" } @@ -17,0 +17 @@ gdown = "^4.6.3" +hffs = {git = "https://github.com/huggingface/hffs.git", rev="0e187e74d38e9436353691f4a7a26b15f0663f58"} @@ -25,0 +26 @@ nltk = "^3.8.1" +numpy = "~1.22.4" @@ -27,0 +29 @@ pdf2image = "^1.16.2" +pyarrow = "^11.0.0" @@ -87 +89,6 @@ strict = true -module = "datasets.*" +module = [ + "datasets.*", + "hffs.*", + "pyarrow.*", + "tqdm.*" +] diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 14eb69a8..982209ae 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -331 +331 @@ def create_dataset_info_response_for_csv(dataset: str, config: str) -> Any: -def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any: +def create_dataset_info_response_for_audio() -> Any: @@ -341 +341 @@ def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any: - "num_bytes": 1383, + "num_bytes": AUDIO_PARQUET_SIZE, @@ -345 +345 @@ def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any: - "download_size": 1383, + "download_size": AUDIO_PARQUET_SIZE, @@ -347 +347 @@ def create_dataset_info_response_for_audio(dataset: str, config: str) -> Any: - "size_in_bytes": 1442, + "size_in_bytes": 1443, @@ -359 +359 @@ def create_parquet_and_dataset_info_response(dataset: str, data_type: Literal["c - else create_dataset_info_response_for_audio(dataset, config) + else create_dataset_info_response_for_audio() @@ -378,2 +378,2 @@ def create_parquet_and_dataset_info_response(dataset: str, data_type: Literal["c -CSV_PARQUET_SIZE = 1_865 -AUDIO_PARQUET_SIZE = 1_383 +CSV_PARQUET_SIZE = 1_866 +AUDIO_PARQUET_SIZE = 1_384
3a98ba6166c1698a7b2ae959fbcdf5e95a4001a4
Steven Liu
2023-03-28T17:07:50
[docs] Pandas to Polars (#978)
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml index b35ddb65..67b9bc5d 100644 --- a/docs/source/_toctree.yml +++ b/docs/source/_toctree.yml @@ -16 +16 @@ - title: List parquet files + title: List Parquet files diff --git a/docs/source/parquet.mdx b/docs/source/parquet.mdx index eae81aff..94c13805 100644 --- a/docs/source/parquet.mdx +++ b/docs/source/parquet.mdx @@ -1 +1 @@ -# List parquet files +# List Parquet files @@ -3 +3 @@ -Datasets can be published in any format (CSV, JSONL, directories of images, etc.) on the Hub, and people generally use the [`datasets` library](https://huggingface.co/docs/datasets/) to access the data. To make it even easier, the datasets-server automatically converts every dataset to the [Parquet](https://parquet.apache.org/) format and publishes the parquet files on the Hub (in a specific branch: `ref/convert/parquet`). +Datasets can be published in any format (CSV, JSONL, directories of images, etc.) to the Hub, and they are easily accessed with the 🤗 [Datasets](https://huggingface.co/docs/datasets/) library. For a more performant experience (especially when it comes to large datasets), Datasets Server automatically converts every dataset to the [Parquet](https://parquet.apache.org/) format. The Parquet files are published to the Hub on a specific `ref/convert/parquet` branch (like this `amazon_polarity` [branch](https://huggingface.co/datasets/amazon_polarity/tree/refs%2Fconvert%2Fparquet) for example). @@ -5 +5 @@ Datasets can be published in any format (CSV, JSONL, directories of images, etc. -This guide shows you how to use Datasets Server's `/parquet` endpoint to retrieve the list of a dataset's parquet files programmatically. Feel free to also try it out with [Postman](https://www.postman.com/huggingface/workspace/hugging-face-apis/request/23242779-f0cde3b9-c2ee-4062-aaca-65c4cfdd96f8), [RapidAPI](https://rapidapi.com/hugging-face-hugging-face-default/api/hugging-face-datasets-api), or [ReDoc](https://redocly.github.io/redoc/?url=https://datasets-server.huggingface.co/openapi.json#operation/listSplits) +This guide shows you how to use Datasets Server's `/parquet` endpoint to retrieve a list of a dataset's files converted to Parquet. Feel free to also try it out with [Postman](https://www.postman.com/huggingface/workspace/hugging-face-apis/request/23242779-f0cde3b9-c2ee-4062-aaca-65c4cfdd96f8), [RapidAPI](https://rapidapi.com/hugging-face-hugging-face-default/api/hugging-face-datasets-api), or [ReDoc](https://redocly.github.io/redoc/?url=https://datasets-server.huggingface.co/openapi.json#operation/listSplits). @@ -49 +49,3 @@ curl https://datasets-server.huggingface.co/parquet?dataset=duorc \ -The endpoint response is a JSON containing a list of the dataset's parquet files. For example, the [duorc](https://huggingface.co/datasets/duorc) dataset has six parquet files, which corresponds to the `train`, `validation` and `test` splits of its two configurations (see the [/splits](./splits) guide): +The endpoint response is a JSON containing a list of the dataset's files in the Parquet format. For example, the [`duorc`](https://huggingface.co/datasets/duorc) dataset has six Parquet files, which corresponds to the `test`, `train` and `validation` splits of its two configurations, `ParaphraseRC` and `SelfRC` (see the [List splits and configurations](./splits) guide for more details about splits and configurations). + +The endpoint also gives the filename and size of each file: @@ -106,15 +108 @@ The endpoint response is a JSON containing a list of the dataset's parquet files -The dataset can then be accessed directly through the parquet files: - -```python -import pandas as pd -url = "https://huggingface.co/datasets/duorc/resolve/refs%2Fconvert%2Fparquet/ParaphraseRC/duorc-train.parquet" -pd.read_parquet(url).title.value_counts().head() -# Dracula 422 -# The Three Musketeers 412 -# Superman 193 -# Jane Eyre 190 -# The Thing 189 -# Name: title, dtype: int64 -``` - -## Sharded parquet files +## Sharded Parquet files @@ -122 +110 @@ pd.read_parquet(url).title.value_counts().head() -The big datasets are partitioned in parquet files (shards) of about 1 GiB. The file name gives the index of the shard and the total number of shards. For example, the `train` split of the [`alexandrainst/danish-wit`](https://datasets-server.huggingface.co/parquet?dataset=alexandrainst/danish-wit) dataset is partitioned into 9 shards, from `parquet-train-00000-of-00009.parquet` to `parquet-train-00008-of-00009.parquet`: +Big datasets are partitioned into Parquet files (shards) of about 1GB each. The filename contains the name of the dataset, the split, the shard index, and the total number of shards (`dataset-name-train-0000-of-0004.parquet`). For example, the `train` split of the [`amazon_polarity`](https://datasets-server.huggingface.co/parquet?dataset=amazon_polarity) dataset is partitioned into 4 shards: @@ -128,2 +116,2 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", + "dataset": "amazon_polarity", + "config": "amazon_polarity", @@ -131,39 +119,7 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-test.parquet", - "filename": "parquet-test.parquet", - "size": 48781933 - }, - { - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", - "split": "train", - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00000-of-00009.parquet", - "filename": "parquet-train-00000-of-00009.parquet", - "size": 937127291 - }, - { - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", - "split": "train", - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00001-of-00009.parquet", - "filename": "parquet-train-00001-of-00009.parquet", - "size": 925920565 - }, - { - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", - "split": "train", - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00002-of-00009.parquet", - "filename": "parquet-train-00002-of-00009.parquet", - "size": 940390661 - }, - { - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", - "split": "train", - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00003-of-00009.parquet", - "filename": "parquet-train-00003-of-00009.parquet", - "size": 934549621 - }, - { - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", + "url": "https://huggingface.co/datasets/amazon_polarity/resolve/refs%2Fconvert%2Fparquet/amazon_polarity/amazon_polarity-test.parquet", + "filename": "amazon_polarity-test.parquet", + "size": 117422359 + }, + { + "dataset": "amazon_polarity", + "config": "amazon_polarity", @@ -171,3 +127,3 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00004-of-00009.parquet", - "filename": "parquet-train-00004-of-00009.parquet", - "size": 493004154 + "url": "https://huggingface.co/datasets/amazon_polarity/resolve/refs%2Fconvert%2Fparquet/amazon_polarity/amazon_polarity-train-00000-of-00004.parquet", + "filename": "amazon_polarity-train-00000-of-00004.parquet", + "size": 320281121 @@ -176,2 +132,2 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", + "dataset": "amazon_polarity", + "config": "amazon_polarity", @@ -179,3 +135,3 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00005-of-00009.parquet", - "filename": "parquet-train-00005-of-00009.parquet", - "size": 942848888 + "url": "https://huggingface.co/datasets/amazon_polarity/resolve/refs%2Fconvert%2Fparquet/amazon_polarity/amazon_polarity-train-00001-of-00004.parquet", + "filename": "amazon_polarity-train-00001-of-00004.parquet", + "size": 320627716 @@ -184,2 +140,2 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", + "dataset": "amazon_polarity", + "config": "amazon_polarity", @@ -187,3 +143,3 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00006-of-00009.parquet", - "filename": "parquet-train-00006-of-00009.parquet", - "size": 933373843 + "url": "https://huggingface.co/datasets/amazon_polarity/resolve/refs%2Fconvert%2Fparquet/amazon_polarity/amazon_polarity-train-00002-of-00004.parquet", + "filename": "amazon_polarity-train-00002-of-00004.parquet", + "size": 320587882 @@ -192,2 +148,2 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", + "dataset": "amazon_polarity", + "config": "amazon_polarity", @@ -195,19 +151,3 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00007-of-00009.parquet", - "filename": "parquet-train-00007-of-00009.parquet", - "size": 936939176 - }, - { - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", - "split": "train", - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-train-00008-of-00009.parquet", - "filename": "parquet-train-00008-of-00009.parquet", - "size": 946933048 - }, - { - "dataset": "alexandrainst/danish-wit", - "config": "alexandrainst--danish-wit", - "split": "val", - "url": "https://huggingface.co/datasets/alexandrainst/danish-wit/resolve/refs%2Fconvert%2Fparquet/alexandrainst--danish-wit/parquet-val.parquet", - "filename": "parquet-val.parquet", - "size": 11437355 + "url": "https://huggingface.co/datasets/amazon_polarity/resolve/refs%2Fconvert%2Fparquet/amazon_polarity/amazon_polarity-train-00003-of-00004.parquet", + "filename": "amazon_polarity-train-00003-of-00004.parquet", + "size": 66515954 @@ -215,2 +155,3 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f - ] -} + ], + "pending": [], + "failed": []} @@ -219,18 +160 @@ The big datasets are partitioned in parquet files (shards) of about 1 GiB. The f -The shards can be concatenated: - -```python -import pandas as pd -import requests -r = requests.get("https://datasets-server.huggingface.co/parquet?dataset=alexandrainst/danish-wit") -j = r.json() -urls = [f['url'] for f in j['parquet_files'] if f['split'] == 'train'] -dfs = [pd.read_parquet(url) for url in urls] -df = pd.concat(dfs) -df.mime_type.value_counts().head() -# image/jpeg 140919 -# image/png 18608 -# image/svg+xml 6171 -# image/gif 1030 -# image/webp 1 -# Name: mime_type, dtype: int64 -``` +To read and query the Parquet files, take a look at the [Query datasets from Datasets Server](parquet_process) guide. \ No newline at end of file
a4665b4b1d21d974ce702bb7931647884e882a08
Quentin Lhoest
2023-03-27T18:00:12
support splits with dots (#990)
diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py index 25386508..7ffae0cb 100644 --- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py +++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py @@ -162 +162 @@ def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url -p = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>[\w]+?)(-[0-9]{5}-of-[0-9]{5})?.parquet") +p = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>\w+(\.\w+)*?)(-[0-9]{5}-of-[0-9]{5})?.parquet") diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index dbf2f292..b0c8c3f9 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -471,0 +472 @@ def test_compute_splits_response_simple_csv_error( + ("config/builder-split.with.dots-00000-of-00001.parquet", "split.with.dots", "config", False),
47886501d2b0e74be441643c22338665d54b3ac5
Sylvain Lesage
2023-03-27T08:27:35
Random rows (#875)
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml index 2b4e6b15..9c1af905 100644 --- a/.github/workflows/_quality-python.yml +++ b/.github/workflows/_quality-python.yml @@ -56 +56 @@ jobs: - run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^pymongo==/,+109 d' | sed '/^requests==2.28.2 ;/,+2 d' | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" + run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^pymongo==/,+109 d' | sed '/^requests==2.28.2 ;/,+2 d' | sed '/^kenlm @/d' | sed '/^fsspec==/,+2 d' | sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" diff --git a/e2e/tests/test_11_auth.py b/e2e/tests/test_11_auth.py index 3101e914..dc21d724 100644 --- a/e2e/tests/test_11_auth.py +++ b/e2e/tests/test_11_auth.py @@ -52,0 +53,25 @@ def test_auth_e2e( + + # ensure the /rows endpoint works as well + offset = 1 + limit = 10 + rows_response = poll_until_ready_and_assert( + relative_url=f"/rows?dataset={dataset}&config={config}&split={split}&offset={offset}&limit={limit}", + expected_status_code=expected_status_code, + expected_error_code=expected_error_code, + headers=headers, + ) + if not expected_error_code: + content = rows_response.json() + assert "rows" in content, rows_response + rows = content["rows"] + assert isinstance(rows, list), rows + assert len(rows) == 3, rows + assert rows[0] == {"row_idx": 1, "row": {"col_1": 1, "col_2": 1, "col_3": 1.0}, "truncated_cells": []}, rows[0] + assert "features" in content, rows_response + features = content["features"] + assert isinstance(features, list), features + assert features == [ + {"feature_idx": 0, "name": "col_1", "type": {"dtype": "int64", "_type": "Value"}}, + {"feature_idx": 1, "name": "col_2", "type": {"dtype": "int64", "_type": "Value"}}, + {"feature_idx": 2, "name": "col_3", "type": {"dtype": "float64", "_type": "Value"}}, + ], features diff --git a/e2e/tests/utils.py b/e2e/tests/utils.py index c65753f1..cbcdc83b 100644 --- a/e2e/tests/utils.py +++ b/e2e/tests/utils.py @@ -111 +111 @@ def poll_until_ready_and_assert( -) -> None: +) -> Any: @@ -128,0 +129 @@ def poll_until_ready_and_assert( + return response diff --git a/services/api/poetry.lock b/services/api/poetry.lock index 6b1a8955..e5409739 100644 --- a/services/api/poetry.lock +++ b/services/api/poetry.lock @@ -2,0 +3,124 @@ +[[package]] +name = "aiohttp" +version = "3.8.4" +description = "Async http client/server framework (asyncio)" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"}, + {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"}, + {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"}, + {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"}, + {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"}, + {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"}, + {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"}, + {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"}, + {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"}, + {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"}, + {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"}, + {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"}, +] + +[package.dependencies] +aiosignal = ">=1.1.2" +async-timeout = ">=4.0.0a3,<5.0" +attrs = ">=17.3.0" +charset-normalizer = ">=2.0,<4.0" +frozenlist = ">=1.1.1" +multidict = ">=4.5,<7.0" +yarl = ">=1.0,<2.0" + +[package.extras] +speedups = ["Brotli", "aiodns", "cchardet"] + +[[package]] +name = "aiosignal" +version = "1.3.1" +description = "aiosignal: a list of registered asynchronous callbacks" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, + {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, +] + +[package.dependencies] +frozenlist = ">=1.1.0" + @@ -35,0 +160,12 @@ files = [ +[[package]] +name = "async-timeout" +version = "4.0.2" +description = "Timeout context manager for asyncio programs" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"}, + {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"}, +] + @@ -476,0 +613,59 @@ toml = ">=0.10.0,<0.11.0" +[[package]] +name = "datasets" +version = "2.10.1" +description = "HuggingFace community-driven open-source library of datasets" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "datasets-2.10.1-py3-none-any.whl", hash = "sha256:bfde7253b31abfd075f9ad55961b13c135a5a9295cea04ae86c47b9c0e0466fd"}, + {file = "datasets-2.10.1.tar.gz", hash = "sha256:e2764c90aa3af96450a9747a934b8893b121f79f58d89e123cb1a7046bb8e81e"}, +] + +[package.dependencies] +aiohttp = "*" +dill = ">=0.3.0,<0.3.7" +fsspec = {version = ">=2021.11.1", extras = ["http"]} +huggingface-hub = ">=0.2.0,<1.0.0" +multiprocess = "*" +numpy = ">=1.17" +packaging = "*" +pandas = "*" +pyarrow = ">=6.0.0" +pyyaml = ">=5.1" +requests = ">=2.19.0" +responses = "<0.19" +tqdm = ">=4.62.1" +xxhash = "*" + +[package.extras] +apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] +audio = ["librosa"] +benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"] +docs = ["s3fs"] +jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "sqlalchemy (<2.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"] +s3 = ["s3fs"] +tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"] +tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "torchaudio (<0.12.0)", "transformers", "zstandard"] +torch = ["torch"] +vision = ["Pillow (>=6.2.1)"] + +[[package]] +name = "dill" +version = "0.3.6" +description = "serialize all of python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"}, + {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"}, +] + +[package.extras] +graph = ["objgraph (>=1.7.2)"] + @@ -562,0 +758,123 @@ pyflakes = ">=2.3.0,<2.4.0" +[[package]] +name = "frozenlist" +version = "1.3.3" +description = "A list-like structure which implements collections.abc.MutableSequence" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"}, + {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"}, + {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"}, + {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"}, + {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"}, + {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"}, + {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"}, + {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"}, + {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"}, + {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"}, + {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"}, +] + +[[package]] +name = "fsspec" +version = "2023.1.0" +description = "File-system specification" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "fsspec-2023.1.0-py3-none-any.whl", hash = "sha256:b833e2e541e9e8cde0ab549414187871243177feb3d344f9d27b25a93f5d8139"}, + {file = "fsspec-2023.1.0.tar.gz", hash = "sha256:fbae7f20ff801eb5f7d0bedf81f25c787c0dfac5e982d98fa3884a9cde2b5411"}, +] + +[package.dependencies] +aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""} +requests = {version = "*", optional = true, markers = "extra == \"http\""} + +[package.extras] +abfs = ["adlfs"] +adl = ["adlfs"] +arrow = ["pyarrow (>=1)"] +dask = ["dask", "distributed"] +dropbox = ["dropbox", "dropboxdrivefs", "requests"] +entrypoints = ["importlib-metadata"] +fuse = ["fusepy"] +gcs = ["gcsfs"] +git = ["pygit2"] +github = ["requests"] +gs = ["gcsfs"] +gui = ["panel"] +hdfs = ["pyarrow (>=1)"] +http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"] +libarchive = ["libarchive-c"] +oci = ["ocifs"] +s3 = ["s3fs"] +sftp = ["paramiko"] +smb = ["smbprotocol"] +ssh = ["paramiko"] +tqdm = ["tqdm"] + @@ -604,0 +923,27 @@ files = [ +[[package]] +name = "hffs" +version = "0.0.1.dev0" +description = "Filesystem interface over huggingface.co repositories" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [] +develop = false + +[package.dependencies] +fsspec = "*" +huggingface_hub = ">=0.12.0" +packaging = ">=20.9" +requests = "*" + +[package.extras] +dev = ["black (>=23.1,<24.0)", "pytest", "ruff (>=0.0.241)"] +quality = ["black (>=23.1,<24.0)", "ruff (>=0.0.241)"] +tests = ["pytest"] + +[package.source] +type = "git" +url = "https://github.com/huggingface/hffs.git" +reference = "0e187e74d38e9436353691f4a7a26b15f0663f58" +resolved_reference = "0e187e74d38e9436353691f4a7a26b15f0663f58" + @@ -1022,0 +1368,111 @@ files = [ +[[package]] +name = "multidict" +version = "6.0.4" +description = "multidict implementation" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"}, + {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"}, + {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"}, + {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"}, + {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"}, + {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"}, + {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"}, + {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"}, + {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"}, + {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"}, + {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"}, + {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"}, + {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"}, +] + +[[package]] +name = "multiprocess" +version = "0.70.14" +description = "better multiprocessing and multithreading in python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"}, + {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"}, + {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"}, + {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"}, + {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"}, + {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"}, +] + +[package.dependencies] +dill = ">=0.3.6" + @@ -1081,0 +1538,38 @@ files = [ +[[package]] +name = "numpy" +version = "1.24.2" +description = "Fundamental package for array computing in Python" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "numpy-1.24.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eef70b4fc1e872ebddc38cddacc87c19a3709c0e3e5d20bf3954c147b1dd941d"}, + {file = "numpy-1.24.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e8d2859428712785e8a8b7d2b3ef0a1d1565892367b32f915c4a4df44d0e64f5"}, + {file = "numpy-1.24.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6524630f71631be2dabe0c541e7675db82651eb998496bbe16bc4f77f0772253"}, + {file = "numpy-1.24.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a51725a815a6188c662fb66fb32077709a9ca38053f0274640293a14fdd22978"}, + {file = "numpy-1.24.2-cp310-cp310-win32.whl", hash = "sha256:2620e8592136e073bd12ee4536149380695fbe9ebeae845b81237f986479ffc9"}, + {file = "numpy-1.24.2-cp310-cp310-win_amd64.whl", hash = "sha256:97cf27e51fa078078c649a51d7ade3c92d9e709ba2bfb97493007103c741f1d0"}, + {file = "numpy-1.24.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7de8fdde0003f4294655aa5d5f0a89c26b9f22c0a58790c38fae1ed392d44a5a"}, + {file = "numpy-1.24.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4173bde9fa2a005c2c6e2ea8ac1618e2ed2c1c6ec8a7657237854d42094123a0"}, + {file = "numpy-1.24.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4cecaed30dc14123020f77b03601559fff3e6cd0c048f8b5289f4eeabb0eb281"}, + {file = "numpy-1.24.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9a23f8440561a633204a67fb44617ce2a299beecf3295f0d13c495518908e910"}, + {file = "numpy-1.24.2-cp311-cp311-win32.whl", hash = "sha256:e428c4fbfa085f947b536706a2fc349245d7baa8334f0c5723c56a10595f9b95"}, + {file = "numpy-1.24.2-cp311-cp311-win_amd64.whl", hash = "sha256:557d42778a6869c2162deb40ad82612645e21d79e11c1dc62c6e82a2220ffb04"}, + {file = "numpy-1.24.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d0a2db9d20117bf523dde15858398e7c0858aadca7c0f088ac0d6edd360e9ad2"}, + {file = "numpy-1.24.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:c72a6b2f4af1adfe193f7beb91ddf708ff867a3f977ef2ec53c0ffb8283ab9f5"}, + {file = "numpy-1.24.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c29e6bd0ec49a44d7690ecb623a8eac5ab8a923bce0bea6293953992edf3a76a"}, + {file = "numpy-1.24.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2eabd64ddb96a1239791da78fa5f4e1693ae2dadc82a76bc76a14cbb2b966e96"}, + {file = "numpy-1.24.2-cp38-cp38-win32.whl", hash = "sha256:e3ab5d32784e843fc0dd3ab6dcafc67ef806e6b6828dc6af2f689be0eb4d781d"}, + {file = "numpy-1.24.2-cp38-cp38-win_amd64.whl", hash = "sha256:76807b4063f0002c8532cfeac47a3068a69561e9c8715efdad3c642eb27c0756"}, + {file = "numpy-1.24.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4199e7cfc307a778f72d293372736223e39ec9ac096ff0a2e64853b866a8e18a"}, + {file = "numpy-1.24.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:adbdce121896fd3a17a77ab0b0b5eedf05a9834a18699db6829a64e1dfccca7f"}, + {file = "numpy-1.24.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:889b2cc88b837d86eda1b17008ebeb679d82875022200c6e8e4ce6cf549b7acb"}, + {file = "numpy-1.24.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f64bb98ac59b3ea3bf74b02f13836eb2e24e48e0ab0145bbda646295769bd780"}, + {file = "numpy-1.24.2-cp39-cp39-win32.whl", hash = "sha256:63e45511ee4d9d976637d11e6c9864eae50e12dc9598f531c035265991910468"}, + {file = "numpy-1.24.2-cp39-cp39-win_amd64.whl", hash = "sha256:a77d3e1163a7770164404607b7ba3967fb49b24782a6ef85d9b5f54126cc39e5"}, + {file = "numpy-1.24.2-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92011118955724465fb6853def593cf397b4a1367495e0b59a7e69d40c4eb71d"}, + {file = "numpy-1.24.2-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9006288bcf4895917d02583cf3411f98631275bc67cce355a7f39f8c14338fa"}, + {file = "numpy-1.24.2-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:150947adbdfeceec4e5926d956a06865c1c690f2fd902efede4ca6fe2e657c3f"}, + {file = "numpy-1.24.2.tar.gz", hash = "sha256:003a9f530e880cb2cd177cba1af7220b9aa42def9c4afc2a2fc3ee6be7eb2b22"}, +] + @@ -1163,0 +1658,45 @@ files = [ +[[package]] +name = "pandas" +version = "1.5.3" +description = "Powerful data structures for data analysis, time series, and statistics" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "pandas-1.5.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:3749077d86e3a2f0ed51367f30bf5b82e131cc0f14260c4d3e499186fccc4406"}, + {file = "pandas-1.5.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:972d8a45395f2a2d26733eb8d0f629b2f90bebe8e8eddbb8829b180c09639572"}, + {file = "pandas-1.5.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:50869a35cbb0f2e0cd5ec04b191e7b12ed688874bd05dd777c19b28cbea90996"}, + {file = "pandas-1.5.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c3ac844a0fe00bfaeb2c9b51ab1424e5c8744f89860b138434a363b1f620f354"}, + {file = "pandas-1.5.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7a0a56cef15fd1586726dace5616db75ebcfec9179a3a55e78f72c5639fa2a23"}, + {file = "pandas-1.5.3-cp310-cp310-win_amd64.whl", hash = "sha256:478ff646ca42b20376e4ed3fa2e8d7341e8a63105586efe54fa2508ee087f328"}, + {file = "pandas-1.5.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:6973549c01ca91ec96199e940495219c887ea815b2083722821f1d7abfa2b4dc"}, + {file = "pandas-1.5.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:c39a8da13cede5adcd3be1182883aea1c925476f4e84b2807a46e2775306305d"}, + {file = "pandas-1.5.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f76d097d12c82a535fda9dfe5e8dd4127952b45fea9b0276cb30cca5ea313fbc"}, + {file = "pandas-1.5.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e474390e60ed609cec869b0da796ad94f420bb057d86784191eefc62b65819ae"}, + {file = "pandas-1.5.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f2b952406a1588ad4cad5b3f55f520e82e902388a6d5a4a91baa8d38d23c7f6"}, + {file = "pandas-1.5.3-cp311-cp311-win_amd64.whl", hash = "sha256:bc4c368f42b551bf72fac35c5128963a171b40dce866fb066540eeaf46faa003"}, + {file = "pandas-1.5.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:14e45300521902689a81f3f41386dc86f19b8ba8dd5ac5a3c7010ef8d2932813"}, + {file = "pandas-1.5.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9842b6f4b8479e41968eced654487258ed81df7d1c9b7b870ceea24ed9459b31"}, + {file = "pandas-1.5.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:26d9c71772c7afb9d5046e6e9cf42d83dd147b5cf5bcb9d97252077118543792"}, + {file = "pandas-1.5.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5fbcb19d6fceb9e946b3e23258757c7b225ba450990d9ed63ccceeb8cae609f7"}, + {file = "pandas-1.5.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:565fa34a5434d38e9d250af3c12ff931abaf88050551d9fbcdfafca50d62babf"}, + {file = "pandas-1.5.3-cp38-cp38-win32.whl", hash = "sha256:87bd9c03da1ac870a6d2c8902a0e1fd4267ca00f13bc494c9e5a9020920e1d51"}, + {file = "pandas-1.5.3-cp38-cp38-win_amd64.whl", hash = "sha256:41179ce559943d83a9b4bbacb736b04c928b095b5f25dd2b7389eda08f46f373"}, + {file = "pandas-1.5.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:c74a62747864ed568f5a82a49a23a8d7fe171d0c69038b38cedf0976831296fa"}, + {file = "pandas-1.5.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c4c00e0b0597c8e4f59e8d461f797e5d70b4d025880516a8261b2817c47759ee"}, + {file = "pandas-1.5.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a50d9a4336a9621cab7b8eb3fb11adb82de58f9b91d84c2cd526576b881a0c5a"}, + {file = "pandas-1.5.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:dd05f7783b3274aa206a1af06f0ceed3f9b412cf665b7247eacd83be41cf7bf0"}, + {file = "pandas-1.5.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f69c4029613de47816b1bb30ff5ac778686688751a5e9c99ad8c7031f6508e5"}, + {file = "pandas-1.5.3-cp39-cp39-win32.whl", hash = "sha256:7cec0bee9f294e5de5bbfc14d0573f65526071029d036b753ee6507d2a21480a"}, + {file = "pandas-1.5.3-cp39-cp39-win_amd64.whl", hash = "sha256:dfd681c5dc216037e0b0a2c821f5ed99ba9f03ebcf119c7dac0e9a7b960b9ec9"}, + {file = "pandas-1.5.3.tar.gz", hash = "sha256:74a3fd7e5a7ec052f183273dc7b0acd3a863edf7520f5d3a1765c04ffdb3b0b1"}, +] + +[package.dependencies] +numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""} +python-dateutil = ">=2.8.1" +pytz = ">=2020.1" + +[package.extras] +test = ["hypothesis (>=5.5.3)", "pytest (>=6.0)", "pytest-xdist (>=1.31)"] + @@ -1337,0 +1877,38 @@ test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] +[[package]] +name = "pyarrow" +version = "11.0.0" +description = "Python library for Apache Arrow" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"}, + {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"}, + {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"}, + {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"}, + {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"}, + {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"}, + {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"}, + {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"}, +] + +[package.dependencies] +numpy = ">=1.16.6" + @@ -1652,0 +2230,15 @@ Werkzeug = ">=2.0.0" +[[package]] +name = "python-dateutil" +version = "2.8.2" +description = "Extensions to the standard Python datetime module" +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" +files = [ + {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"}, + {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"}, +] + +[package.dependencies] +six = ">=1.5" + @@ -1769,0 +2362,19 @@ test = ["commentjson", "packaging", "pytest"] +[[package]] +name = "responses" +version = "0.18.0" +description = "A utility library for mocking out the `requests` Python library." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, + {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, +] + +[package.dependencies] +requests = ">=2.0,<3.0" +urllib3 = ">=1.25.10" + +[package.extras] +tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] + @@ -1828 +2439 @@ description = "Python 2 and 3 compatibility utilities" -category = "dev" +category = "main" @@ -2139,0 +2751,196 @@ watchdog = ["watchdog"] +[[package]] +name = "xxhash" +version = "3.2.0" +description = "Python binding for xxHash" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"}, + {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"}, + {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"}, + {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"}, + {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"}, + {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"}, + {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"}, + {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"}, + {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"}, + {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"}, + {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"}, + {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"}, + {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"}, + {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"}, + {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"}, + {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"}, + {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"}, +] + +[[package]] +name = "yarl" +version = "1.8.2" +description = "Yet another URL library" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:bb81f753c815f6b8e2ddd2eef3c855cf7da193b82396ac013c661aaa6cc6b0a5"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:47d49ac96156f0928f002e2424299b2c91d9db73e08c4cd6742923a086f1c863"}, + {file = "yarl-1.8.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3fc056e35fa6fba63248d93ff6e672c096f95f7836938241ebc8260e062832fe"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58a3c13d1c3005dbbac5c9f0d3210b60220a65a999b1833aa46bd6677c69b08e"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:10b08293cda921157f1e7c2790999d903b3fd28cd5c208cf8826b3b508026996"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:de986979bbd87272fe557e0a8fcb66fd40ae2ddfe28a8b1ce4eae22681728fef"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c4fcfa71e2c6a3cb568cf81aadc12768b9995323186a10827beccf5fa23d4f8"}, + {file = "yarl-1.8.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae4d7ff1049f36accde9e1ef7301912a751e5bae0a9d142459646114c70ecba6"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:bf071f797aec5b96abfc735ab97da9fd8f8768b43ce2abd85356a3127909d146"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:74dece2bfc60f0f70907c34b857ee98f2c6dd0f75185db133770cd67300d505f"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:df60a94d332158b444301c7f569659c926168e4d4aad2cfbf4bce0e8fb8be826"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:63243b21c6e28ec2375f932a10ce7eda65139b5b854c0f6b82ed945ba526bff3"}, + {file = "yarl-1.8.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:cfa2bbca929aa742b5084fd4663dd4b87c191c844326fcb21c3afd2d11497f80"}, + {file = "yarl-1.8.2-cp310-cp310-win32.whl", hash = "sha256:b05df9ea7496df11b710081bd90ecc3a3db6adb4fee36f6a411e7bc91a18aa42"}, + {file = "yarl-1.8.2-cp310-cp310-win_amd64.whl", hash = "sha256:24ad1d10c9db1953291f56b5fe76203977f1ed05f82d09ec97acb623a7976574"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:2a1fca9588f360036242f379bfea2b8b44cae2721859b1c56d033adfd5893634"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:f37db05c6051eff17bc832914fe46869f8849de5b92dc4a3466cd63095d23dfd"}, + {file = "yarl-1.8.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:77e913b846a6b9c5f767b14dc1e759e5aff05502fe73079f6f4176359d832581"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0978f29222e649c351b173da2b9b4665ad1feb8d1daa9d971eb90df08702668a"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:388a45dc77198b2460eac0aca1efd6a7c09e976ee768b0d5109173e521a19daf"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2305517e332a862ef75be8fad3606ea10108662bc6fe08509d5ca99503ac2aee"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:42430ff511571940d51e75cf42f1e4dbdded477e71c1b7a17f4da76c1da8ea76"}, + {file = "yarl-1.8.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3150078118f62371375e1e69b13b48288e44f6691c1069340081c3fd12c94d5b"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:c15163b6125db87c8f53c98baa5e785782078fbd2dbeaa04c6141935eb6dab7a"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:4d04acba75c72e6eb90745447d69f84e6c9056390f7a9724605ca9c56b4afcc6"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:e7fd20d6576c10306dea2d6a5765f46f0ac5d6f53436217913e952d19237efc4"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:75c16b2a900b3536dfc7014905a128a2bea8fb01f9ee26d2d7d8db0a08e7cb2c"}, + {file = "yarl-1.8.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:6d88056a04860a98341a0cf53e950e3ac9f4e51d1b6f61a53b0609df342cc8b2"}, + {file = "yarl-1.8.2-cp311-cp311-win32.whl", hash = "sha256:fb742dcdd5eec9f26b61224c23baea46c9055cf16f62475e11b9b15dfd5c117b"}, + {file = "yarl-1.8.2-cp311-cp311-win_amd64.whl", hash = "sha256:8c46d3d89902c393a1d1e243ac847e0442d0196bbd81aecc94fcebbc2fd5857c"}, + {file = "yarl-1.8.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:ceff9722e0df2e0a9e8a79c610842004fa54e5b309fe6d218e47cd52f791d7ef"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3f6b4aca43b602ba0f1459de647af954769919c4714706be36af670a5f44c9c1"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1684a9bd9077e922300ecd48003ddae7a7474e0412bea38d4631443a91d61077"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ebb78745273e51b9832ef90c0898501006670d6e059f2cdb0e999494eb1450c2"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3adeef150d528ded2a8e734ebf9ae2e658f4c49bf413f5f157a470e17a4a2e89"}, + {file = "yarl-1.8.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:57a7c87927a468e5a1dc60c17caf9597161d66457a34273ab1760219953f7f4c"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:efff27bd8cbe1f9bd127e7894942ccc20c857aa8b5a0327874f30201e5ce83d0"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:a783cd344113cb88c5ff7ca32f1f16532a6f2142185147822187913eb989f739"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:705227dccbe96ab02c7cb2c43e1228e2826e7ead880bb19ec94ef279e9555b5b"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:34c09b43bd538bf6c4b891ecce94b6fa4f1f10663a8d4ca589a079a5018f6ed7"}, + {file = "yarl-1.8.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:a48f4f7fea9a51098b02209d90297ac324241bf37ff6be6d2b0149ab2bd51b37"}, + {file = "yarl-1.8.2-cp37-cp37m-win32.whl", hash = "sha256:0414fd91ce0b763d4eadb4456795b307a71524dbacd015c657bb2a39db2eab89"}, + {file = "yarl-1.8.2-cp37-cp37m-win_amd64.whl", hash = "sha256:d881d152ae0007809c2c02e22aa534e702f12071e6b285e90945aa3c376463c5"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5df5e3d04101c1e5c3b1d69710b0574171cc02fddc4b23d1b2813e75f35a30b1"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7a66c506ec67eb3159eea5096acd05f5e788ceec7b96087d30c7d2865a243918"}, + {file = "yarl-1.8.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2b4fa2606adf392051d990c3b3877d768771adc3faf2e117b9de7eb977741229"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1e21fb44e1eff06dd6ef971d4bdc611807d6bd3691223d9c01a18cec3677939e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:93202666046d9edadfe9f2e7bf5e0782ea0d497b6d63da322e541665d65a044e"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fc77086ce244453e074e445104f0ecb27530d6fd3a46698e33f6c38951d5a0f1"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:64dd68a92cab699a233641f5929a40f02a4ede8c009068ca8aa1fe87b8c20ae3"}, + {file = "yarl-1.8.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1b372aad2b5f81db66ee7ec085cbad72c4da660d994e8e590c997e9b01e44901"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:e6f3515aafe0209dd17fb9bdd3b4e892963370b3de781f53e1746a521fb39fc0"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:dfef7350ee369197106805e193d420b75467b6cceac646ea5ed3049fcc950a05"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:728be34f70a190566d20aa13dc1f01dc44b6aa74580e10a3fb159691bc76909d"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ff205b58dc2929191f68162633d5e10e8044398d7a45265f90a0f1d51f85f72c"}, + {file = "yarl-1.8.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:baf211dcad448a87a0d9047dc8282d7de59473ade7d7fdf22150b1d23859f946"}, + {file = "yarl-1.8.2-cp38-cp38-win32.whl", hash = "sha256:272b4f1599f1b621bf2aabe4e5b54f39a933971f4e7c9aa311d6d7dc06965165"}, + {file = "yarl-1.8.2-cp38-cp38-win_amd64.whl", hash = "sha256:326dd1d3caf910cd26a26ccbfb84c03b608ba32499b5d6eeb09252c920bcbe4f"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:f8ca8ad414c85bbc50f49c0a106f951613dfa5f948ab69c10ce9b128d368baf8"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:418857f837347e8aaef682679f41e36c24250097f9e2f315d39bae3a99a34cbf"}, + {file = "yarl-1.8.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:ae0eec05ab49e91a78700761777f284c2df119376e391db42c38ab46fd662b77"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:009a028127e0a1755c38b03244c0bea9d5565630db9c4cf9572496e947137a87"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3edac5d74bb3209c418805bda77f973117836e1de7c000e9755e572c1f7850d0"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:da65c3f263729e47351261351b8679c6429151ef9649bba08ef2528ff2c423b2"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ef8fb25e52663a1c85d608f6dd72e19bd390e2ecaf29c17fb08f730226e3a08"}, + {file = "yarl-1.8.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bcd7bb1e5c45274af9a1dd7494d3c52b2be5e6bd8d7e49c612705fd45420b12d"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:44ceac0450e648de86da8e42674f9b7077d763ea80c8ceb9d1c3e41f0f0a9951"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:97209cc91189b48e7cfe777237c04af8e7cc51eb369004e061809bcdf4e55220"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:48dd18adcf98ea9cd721a25313aef49d70d413a999d7d89df44f469edfb38a06"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:e59399dda559688461762800d7fb34d9e8a6a7444fd76ec33220a926c8be1516"}, + {file = "yarl-1.8.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d617c241c8c3ad5c4e78a08429fa49e4b04bedfc507b34b4d8dceb83b4af3588"}, + {file = "yarl-1.8.2-cp39-cp39-win32.whl", hash = "sha256:cb6d48d80a41f68de41212f3dfd1a9d9898d7841c8f7ce6696cf2fd9cb57ef83"}, + {file = "yarl-1.8.2-cp39-cp39-win_amd64.whl", hash = "sha256:6604711362f2dbf7160df21c416f81fac0de6dbcf0b5445a2ef25478ecc4c778"}, + {file = "yarl-1.8.2.tar.gz", hash = "sha256:49d43402c6e3013ad0978602bf6bf5328535c48d192304b91b97a3c6790b1562"}, +] + +[package.dependencies] +idna = ">=2.0" +multidict = ">=4.0" + @@ -2143 +2950 @@ python-versions = "3.9.15" -content-hash = "9f3e741a9858013b84eef040c1c7d4e02cf9843e3a6766d537147784bd05605d" +content-hash = "37daa2dafb9854811634de242057022f58b4e57ad36346f87be27f57aebf0277" diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml index 00f5afea..c96e59ad 100644 --- a/services/api/pyproject.toml +++ b/services/api/pyproject.toml @@ -8,0 +9 @@ license = "Apache-2.0" +datasets = "^2.10.1" @@ -9,0 +11 @@ environs = "^9.5.0" +hffs = {git = "https://github.com/huggingface/hffs.git", rev="0e187e74d38e9436353691f4a7a26b15f0663f58"} @@ -12,0 +15 @@ pyjwt = { extras = ["crypto"], version = "^2.6.0" } +pyarrow = "^11.0.0" @@ -59 +62 @@ disallow_untyped_calls = false -# ^ call to expected_algorithm.from_jwk force to set this to false +# ^ call to expected_algorithm.from_jwk forces to set this to false @@ -62,2 +65,7 @@ disallow_untyped_calls = false -module = "prometheus_client.*" -ignore_missing_imports = true +module = [ + "datasets.*", + "hffs.*", + "prometheus_client.*", + "pyarrow.*", + "tqdm.*" +] @@ -64,0 +73 @@ ignore_missing_imports = true +ignore_missing_imports = true diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index ec2d5494..b4b43f8b 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -19,0 +20 @@ from api.routes.healthcheck import healthcheck_endpoint +from api.routes.rows import create_rows_endpoint @@ -48,0 +50,4 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + parquet_processing_steps_by_input_type = endpoints_definition.steps_by_input_type_and_endpoint.get("/parquet") + if not parquet_processing_steps_by_input_type or not parquet_processing_steps_by_input_type["config"]: + raise RuntimeError("The parquet endpoint is not configured. Exiting.") + config_parquet_processing_steps = parquet_processing_steps_by_input_type["config"] @@ -120,0 +126,15 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + Route( + "/rows", + endpoint=create_rows_endpoint( + config_parquet_processing_steps=config_parquet_processing_steps, + init_processing_steps=init_processing_steps, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + hf_jwt_public_key=hf_jwt_public_key, + hf_jwt_algorithm=app_config.api.hf_jwt_algorithm, + external_auth_url=app_config.api.external_auth_url, + hf_timeout_seconds=app_config.api.hf_timeout_seconds, + max_age_long=app_config.api.max_age_long, + max_age_short=app_config.api.max_age_short, + ), + ), diff --git a/services/api/src/api/authentication.py b/services/api/src/api/authentication.py index fadabe51..4da03809 100644 --- a/services/api/src/api/authentication.py +++ b/services/api/src/api/authentication.py @@ -112 +112 @@ def auth_check( - f" {hf_timeout_seconds}" + f" {hf_timeout_seconds}, authorization: {auth.authorization}" diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py index 17eaf93f..93ed97f4 100644 --- a/services/api/src/api/routes/endpoint.py +++ b/services/api/src/api/routes/endpoint.py @@ -65 +65,8 @@ def get_cache_entry_from_steps( - ResponseNotReadyError: if no result is found. + - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the + dataset failed or timed out. + - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset + info failed or timed out. + - [`~libcommon.operations.PreviousStepError`]: a previous step has an error + - [`~libcommon.dataset.DatasetError`]: if the dataset could not be accessed or is not supported + - [`~api.utils.ResponseNotFoundError`]: if no result is found. + - [`~api.utils.ResponseNotReadyError`]: if the response is not ready yet. diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py new file mode 100644 index 00000000..0b837ed1 --- /dev/null +++ b/services/api/src/api/routes/rows.py @@ -0,0 +1,347 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from functools import lru_cache, partial +from typing import Any, List, Mapping, Optional, TypedDict + +import numpy as np +import pyarrow as pa +import pyarrow.parquet as pq +from datasets import Features +from hffs.fs import HfFileSystem +from libcommon.processing_graph import ProcessingStep +from starlette.requests import Request +from starlette.responses import Response +from tqdm.contrib.concurrent import thread_map + +from api.authentication import auth_check +from api.prometheus import StepProfiler +from api.routes.endpoint import get_cache_entry_from_steps +from api.utils import ( + ApiCustomError, + Endpoint, + InvalidParameterError, + MissingRequiredParameterError, + UnexpectedError, + are_valid_parameters, + get_json_api_error_response, + get_json_ok_response, +) + +MAX_ROWS = 100 + +PARQUET_REVISION = "refs/convert/parquet" + + +class FileSystemError(Exception): + pass + + +class ParquetResponseFormatError(Exception): + pass + + +class ParquetResponseEmptyError(Exception): + pass + + +# TODO: how to invalidate the cache when the parquet branch is created or deleted? +@lru_cache(maxsize=128) +def get_parquet_fs(dataset: str, hf_token: Optional[str]) -> HfFileSystem: + """Get the parquet filesystem for a dataset. + + The parquet files are stored in a separate branch of the dataset repository (see PARQUET_REVISION) + + Args: + dataset (str): The dataset name. + hf_token (Optional[str]): The token to access the filesystem. + + Returns: + HfFileSystem: The parquet filesystem. + """ + return HfFileSystem(dataset, repo_type="dataset", revision=PARQUET_REVISION, token=hf_token) + + +UNSUPPORTED_FEATURES_MAGIC_STRINGS = ["Image(", "Audio(", "'binary'"] + + +class RowsIndex: + def __init__( + self, + dataset: str, + config: str, + split: str, + config_parquet_processing_steps: List[ProcessingStep], + init_processing_steps: List[ProcessingStep], + hf_endpoint: str, + hf_token: Optional[str], + ): + self.dataset = dataset + self.config = config + self.split = split + self.__post_init__( + config_parquet_processing_steps=config_parquet_processing_steps, + init_processing_steps=init_processing_steps, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + ) + + def __post_init__( + self, + config_parquet_processing_steps: List[ProcessingStep], + init_processing_steps: List[ProcessingStep], + hf_endpoint: str, + hf_token: Optional[str], + ) -> None: + with StepProfiler(method="rows.index", step="all"): + # get the list of parquet files + with StepProfiler(method="rows.index", step="get list of parquet files for split"): + try: + result = get_cache_entry_from_steps( + processing_steps=config_parquet_processing_steps, + dataset=self.dataset, + config=self.config, + split=None, + init_processing_steps=init_processing_steps, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + ) + content = result["content"] + except ApiCustomError as e: + raise e + except Exception as e: + raise UnexpectedError("Could not get the list of parquet files to fetch the rows from.") from e + # ^ TODO: improve the error, depending on the case + try: + sources = sorted( + f"{self.config}/{parquet_file['filename']}" + for parquet_file in content["parquet_files"] + if parquet_file["split"] == self.split and parquet_file["config"] == self.config + ) + except Exception as e: + raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e + logging.debug( + f"Found {len(sources)} parquet files for dataset={self.dataset}, config={self.config}," + f" split={self.split}: {sources}" + ) + if not sources: + raise ParquetResponseEmptyError("No parquet files found.") + with StepProfiler(method="rows.index", step="get the Hub's dataset filesystem"): + fs = get_parquet_fs(dataset=self.dataset, hf_token=hf_token) + with StepProfiler(method="rows.index", step="get one parquet reader per parquet file"): + desc = f"{self.dataset}/{self.config}/{self.split}" + try: + parquet_files: List[pq.ParquetFile] = thread_map( + partial(pq.ParquetFile, filesystem=fs), sources, desc=desc, unit="pq", disable=True + ) + except Exception as e: + raise FileSystemError(f"Could not read the parquet files: {e}") from e + with StepProfiler(method="rows.index", step="get the dataset's features"): + self.features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema()) + + self.supported_columns, self.unsupported_columns = [], [] + for column, feature in self.features.items(): + str_feature = str(feature) + str_column = str(column) + if any(magic_string in str_feature for magic_string in UNSUPPORTED_FEATURES_MAGIC_STRINGS): + self.unsupported_columns.append(str_column) + else: + self.supported_columns.append(str_column) + + with StepProfiler(method="rows.index", step="create the row group offsets"): + self.row_group_offsets = np.cumsum( + [ + parquet_file.metadata.row_group(group_id).num_rows + for parquet_file in parquet_files + for group_id in range(parquet_file.metadata.num_row_groups) + ] + ) + with StepProfiler(method="rows.index", step="create the row group readers"): + self.row_group_readers = [ + partial(parquet_file.read_row_group, i=group_id, columns=self.supported_columns) + for parquet_file in parquet_files + for group_id in range(parquet_file.metadata.num_row_groups) + ] + + # note that this cache size is global for the class, not per instance + @lru_cache(maxsize=1024) + def query(self, offset: int, length: int) -> pa.Table: + """Query the parquet files + + Note that this implementation will always read at least one row group, to get the list of columns and always + have the same schema, even if the requested rows are invalid (out of range). + + Args: + offset (int): The first row to read. + length (int): The number of rows to read. + + Returns: + pa.Table: The requested rows. + """ + if (len(self.row_group_offsets) == 0) or (len(self.row_group_readers) == 0): + raise ParquetResponseEmptyError("No parquet files found.") + last_row_in_parquet = self.row_group_offsets[-1] - 1 + first_row = min(offset, last_row_in_parquet) + last_row = min(offset, offset + length - 1, last_row_in_parquet) + first_row_group_id, last_row_group_id = np.searchsorted( + self.row_group_offsets, [first_row, last_row], side="right" + ) + pa_table = pa.concat_tables( + [self.row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)] + ) + first_row_in_pa_table = self.row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0 + return pa_table.slice(offset - first_row_in_pa_table, length) + + +class Indexer: + def __init__( + self, + config_parquet_processing_steps: List[ProcessingStep], + init_processing_steps: List[ProcessingStep], + hf_endpoint: str, + hf_token: Optional[str] = None, + ): + self.config_parquet_processing_steps = config_parquet_processing_steps + self.init_processing_steps = init_processing_steps + self.hf_endpoint = hf_endpoint + self.hf_token = hf_token + + @lru_cache(maxsize=128) + def get_rows_index( + self, + dataset: str, + config: str, + split: str, + ) -> RowsIndex: + return RowsIndex( + dataset=dataset, + config=config, + split=split, + config_parquet_processing_steps=self.config_parquet_processing_steps, + init_processing_steps=self.init_processing_steps, + hf_endpoint=self.hf_endpoint, + hf_token=self.hf_token, + ) + + +class FeatureItem(TypedDict): + feature_idx: int + name: str + type: Mapping[str, Any] + + +# in JSON, dicts do not carry any order, so we need to return a list +# +# > An object is an *unordered* collection of zero or more name/value pairs, where a name is a string and a value +# is a string, number, boolean, null, object, or array. +# > An array is an *ordered* sequence of zero or more values. +# > The terms "object" and "array" come from the conventions of JavaScript. +# from https://stackoverflow.com/a/7214312/7351594 / https://www.rfc-editor.org/rfc/rfc7159.html +def to_features_list(features: Features) -> List[FeatureItem]: + features_dict = features.to_dict() + return [ + { + "feature_idx": idx, + "name": name, + "type": features_dict[name], + } + for idx, name in enumerate(features) + ] + + +class RowItem(TypedDict): + row_idx: int + row: Mapping[str, Any] + truncated_cells: List[str] + + +def to_rows_list(pa_table: pa.Table, offset: int, features: Features, unsupported_columns: List[str]) -> List[RowItem]: + num_rows = pa_table.num_rows + for idx, (column, feature) in enumerate(features.items()): + if column in unsupported_columns: + pa_table = pa_table.add_column(idx, column, pa.array([None] * num_rows)) + return [ + { + "row_idx": idx + offset, + "row": row, + "truncated_cells": unsupported_columns, + } + for idx, row in enumerate(pa_table.to_pylist()) + ] + + +def create_response(pa_table: pa.Table, offset: int, features: Features, unsupported_columns: List[str]) -> Any: + if set(pa_table.column_names).intersection(set(unsupported_columns)): + raise RuntimeError( + "The pyarrow table contains unsupported columns. They should have been ignored in the row group reader." + ) + return { + "features": to_features_list(features), + "rows": to_rows_list(pa_table, offset, features, unsupported_columns), + } + + +def create_rows_endpoint( + config_parquet_processing_steps: List[ProcessingStep], + init_processing_steps: List[ProcessingStep], + hf_endpoint: str, + hf_token: Optional[str] = None, + hf_jwt_public_key: Optional[str] = None, + hf_jwt_algorithm: Optional[str] = None, + external_auth_url: Optional[str] = None, + hf_timeout_seconds: Optional[float] = None, + max_age_long: int = 0, + max_age_short: int = 0, +) -> Endpoint: + indexer = Indexer( + config_parquet_processing_steps=config_parquet_processing_steps, + init_processing_steps=init_processing_steps, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + ) + + async def rows_endpoint(request: Request) -> Response: + with StepProfiler(method="rows_endpoint", step="all"): + try: + with StepProfiler(method="rows_endpoint", step="validate parameters"): + dataset = request.query_params.get("dataset") + config = request.query_params.get("config") + split = request.query_params.get("split") + if not dataset or not config or not split or not are_valid_parameters([dataset, config, split]): + raise MissingRequiredParameterError("Parameter 'dataset', 'config' and 'split' are required") + offset = int(request.query_params.get("offset", 0)) + if offset < 0: + raise InvalidParameterError(message="Offset must be positive") + length = int(request.query_params.get("length", MAX_ROWS)) + if length < 0: + raise InvalidParameterError("Length must be positive") + if length > MAX_ROWS: + raise InvalidParameterError(f"Length must be less than or equal to {MAX_ROWS}") + logging.info( + f"/rows, dataset={dataset}, config={config}, split={split}, offset={offset}, length={length}" + ) + with StepProfiler(method="rows_endpoint", step="check authentication"): + # if auth_check fails, it will raise an exception that will be caught below + auth_check( + dataset, + external_auth_url=external_auth_url, + request=request, + hf_jwt_public_key=hf_jwt_public_key, + hf_jwt_algorithm=hf_jwt_algorithm, + hf_timeout_seconds=hf_timeout_seconds, + ) + with StepProfiler(method="rows_endpoint", step="get row groups index"): + rows_index = indexer.get_rows_index(dataset=dataset, config=config, split=split) + with StepProfiler(method="rows_endpoint", step="query the rows"): + pa_table = rows_index.query(offset=offset, length=length) + with StepProfiler(method="rows_endpoint", step="transform to a list"): + response = create_response(pa_table, offset, rows_index.features, rows_index.unsupported_columns) + with StepProfiler(method="rows_endpoint", step="generate the OK response"): + return get_json_ok_response(content=response, max_age=max_age_long) + except Exception as e: + error = e if isinstance(e, ApiCustomError) else UnexpectedError("Unexpected error.", e) + with StepProfiler(method="rows_endpoint", step="generate API error response"): + return get_json_api_error_response(error=error, max_age=max_age_short) + + return rows_endpoint diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py index 09eb4a24..d0c5d164 100644 --- a/services/api/src/api/utils.py +++ b/services/api/src/api/utils.py @@ -14,0 +15 @@ ApiErrorCode = Literal[ + "InvalidParameter", @@ -46,0 +48,7 @@ class MissingRequiredParameterError(ApiCustomError): +class InvalidParameterError(ApiCustomError): + """Raised when a parameter has an invalid value.""" + + def __init__(self, message: str): + super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "InvalidParameter") + + diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py index d990a49c..04eb497e 100644 --- a/services/api/tests/conftest.py +++ b/services/api/tests/conftest.py @@ -48,0 +49 @@ def endpoint_config(monkeypatch_session: MonkeyPatch) -> EndpointConfig: + "/parquet": {"config": ["config-parquet"]}, diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index cdd42219..051fffe6 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -92 +92 @@ def test_get_cache_entry_from_steps() -> None: - step_whitout_error = graph.get_step(cache_without_error) + step_without_error = graph.get_step(cache_without_error) @@ -112 +112 @@ def test_get_cache_entry_from_steps() -> None: - [step_whitout_error, step_with_error], + [step_without_error, step_with_error], @@ -124 +124 @@ def test_get_cache_entry_from_steps() -> None: - [step_with_error, step_whitout_error], + [step_with_error, step_without_error], @@ -141 +141 @@ def test_get_cache_entry_from_steps() -> None: - # peding job thows exception + # pending job throws exception diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py index acd400cd..25386508 100644 --- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py +++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py @@ -162 +162 @@ def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url -p = re.compile(r"[\w]+-(?P<split>[\w]+?)(-[0-9]{5}-of-[0-9]{5})?.parquet") +p = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>[\w]+?)(-[0-9]{5}-of-[0-9]{5})?.parquet") diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index 4ae94cff..dbf2f292 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -468,0 +469 @@ def test_compute_splits_response_simple_csv_error( + ("config/builder-with-dashes-split.parquet", "split", "config", False), @@ -469,0 +471,7 @@ def test_compute_splits_response_simple_csv_error( + ("config/builder-with-dashes-split-00000-of-00001.parquet", "split", "config", False), + ( + "config/builder-with-dashes-caveat-asplitwithdashesisnotsupported-00000-of-00001.parquet", + "asplitwithdashesisnotsupported", + "config", + False, + ), @@ -471 +479,2 @@ def test_compute_splits_response_simple_csv_error( - ("config/builder-not-supported.parquet", "not-supported", "config", True), + ("plain_text/openwebtext-10k-train.parquet", "train", "plain_text", False), + ("plain_text/openwebtext-10k-train-00000-of-00001.parquet", "train", "plain_text", False), diff --git a/tools/Python.mk b/tools/Python.mk index f950eac5..18d1e236 100644 --- a/tools/Python.mk +++ b/tools/Python.mk @@ -31 +31 @@ pip-audit: - bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^pymongo==/,+109 d' | sed '/^requests==2.28.2 ;/,+2 d' | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" + bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^pymongo==/,+109 d' | sed '/^requests==2.28.2 ;/,+2 d' | sed '/^kenlm @/d' | sed '/^fsspec==/,+2 d' | sed '/^torch @/d' | sed '/^torchaudio @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed '/^hffs @/d')" diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index 3adf61f9..a5b22ac7 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -57 +57 @@ services: - API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://huggingface.co/api/keys/jwt} + API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://hub-ci.huggingface.co/api/keys/jwt}
2dc79dd55209a0d5813397b7c7ed465d84fa541a
Sylvain Lesage
2023-03-24T18:13:56
feat: 🎸 remove bigcode/the-stack from the blocked datasets (#986)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 487f3ba5..d03ecce9 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -87 +87 @@ parquetAndDatasetInfo: - blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,bigcode/the-stack,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,bigcode/the-stack-dedup,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member" + blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,bigcode/the-stack-dedup,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
91f1861b8754228114a80561ab613e317e7fd427
Sylvain Lesage
2023-03-24T17:58:02
Access docker host (#981)
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index 4b7cffce..6f4e91d8 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -12 +12 @@ services: - CACHE_MONGO_URL: ${CACHE_MONGO_URL-mongodb://mongodb} # use mongo container by default + CACHE_MONGO_URL: ${CACHE_MONGO_URL-mongodb://localhost:${MONGO_PORT-27017}} # use mongo container by default @@ -16 +16 @@ services: - QUEUE_MONGO_URL: ${QUEUE_MONGO_URL-mongodb://mongodb} # use mongo container by default + QUEUE_MONGO_URL: ${QUEUE_MONGO_URL-mongodb://localhost:${MONGO_PORT-27017}} # use mongo container by default diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index 8c68791c..3adf61f9 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -16,2 +16,4 @@ services: - URL_ADMIN: http://admin:${ADMIN_UVICORN_PORT-8081} - URL_API: http://api:${API_UVICORN_PORT-8080} + URL_ADMIN: http://host.docker.internal:${ADMIN_UVICORN_PORT-8081} + URL_API: http://host.docker.internal:${API_UVICORN_PORT-8080} + extra_hosts: + - "host.docker.internal:host-gateway" @@ -44,3 +46 @@ services: - ports: - # for debug - - ${ADMIN_UVICORN_PORT-8081}:${ADMIN_UVICORN_PORT-8081} + network_mode: host @@ -68,3 +68 @@ services: - ports: - # for debug - - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080} + network_mode: host @@ -77,0 +76,2 @@ services: + deploy: + replicas: 4 @@ -101,2 +101 @@ services: - depends_on: - - mongodb + network_mode: host @@ -109 +107,0 @@ services: - # for debug
4caf3a7629c8d1883199fb0aaa3ee0cb628d97cf
Polina Kazakova
2023-03-24T17:35:48
Make config-level `/split-names-from-dataset-info` dependent on `config-info` instead of `dataset-info` (#984)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 8721d0a0..05025322 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -158 +158 @@ class ProcessingGraphConfig: - "requires": "dataset-info", + "requires": "config-info", diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 7db1c9ad..e9cb544c 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -64 +64 @@ def test_default_graph() -> None: - assert config_info.children == [dataset_info] + assert config_info.children == [dataset_info, split_names_from_dataset_info] @@ -69 +69 @@ def test_default_graph() -> None: - assert dataset_info.children == [split_names_from_dataset_info] + assert dataset_info.children == [] @@ -73 +73 @@ def test_default_graph() -> None: - assert split_names_from_dataset_info.parent is dataset_info + assert split_names_from_dataset_info.parent is config_info @@ -75 +75 @@ def test_default_graph() -> None: - assert split_names_from_dataset_info.get_ancestors() == [parquet_and_dataset_info, config_info, dataset_info] + assert split_names_from_dataset_info.get_ancestors() == [parquet_and_dataset_info, config_info] @@ -98 +97,0 @@ def test_default_graph() -> None: - dataset_info, diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index 732be515..8037079a 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -83 +83 @@ def compute_split_names_from_dataset_info_response( - response = get_response(kind="dataset-info", dataset=dataset) + response = get_response(kind="config-info", dataset=dataset) @@ -85 +85 @@ def compute_split_names_from_dataset_info_response( - raise DatasetNotFoundError("No response found in previous step for this dataset.", e) from e + raise DatasetNotFoundError("No response found in previous step for this dataset: 'config-info'.", e) from e @@ -92 +92 @@ def compute_split_names_from_dataset_info_response( - splits_content = response["content"]["dataset_info"][config]["splits"] + splits_content = response["content"]["dataset_info"]["splits"] @@ -94 +94 @@ def compute_split_names_from_dataset_info_response( - raise PreviousStepFormatError("Previous step did not return the expected content.") from e + raise PreviousStepFormatError("Previous step 'config-info' did not return the expected content.") from e @@ -100 +100 @@ def compute_split_names_from_dataset_info_response( - return SplitsList({"splits": split_name_items}) + return SplitsList(splits=split_name_items) diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index 429a9b41..113099ed 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -73,7 +73,5 @@ def get_job_runner( - "config_name": { - "splits": { - "train": {"name": "train", "dataset_name": "ok"}, - "validation": {"name": "validation", "dataset_name": "ok"}, - "test": {"name": "test", "dataset_name": "ok"}, - }, - } + "splits": { + "train": {"name": "train", "dataset_name": "ok"}, + "validation": {"name": "validation", "dataset_name": "ok"}, + "test": {"name": "test", "dataset_name": "ok"}, + }, @@ -130 +128 @@ def test_compute( - upsert_response(kind="dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status) + upsert_response(kind="config-info", dataset=dataset, content=upstream_content, http_status=upstream_status)
38d6c6d35698e44112a0e9a36dde3555241d3e1a
Sylvain Lesage
2023-03-24T16:50:24
feat: 🎸 add bigcode/the-stack to the supported datasets (#983)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 67032e90..487f3ba5 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -87,0 +88 @@ parquetAndDatasetInfo: + supportedDatasets: "bigcode/the-stack"
f31723e0a395564c0c6948349c04e1a482da2b06
Polina Kazakova
2023-03-24T10:43:57
Get config names from /config-names instead of /parquet-and-dataset-info for dataset-level size and parquet (#982)
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 565d409b..384fc03d 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -14,0 +15,2 @@ from worker.job_runners.config.info import ConfigInfoJobRunner +from worker.job_runners.config.parquet import ConfigParquetJobRunner +from worker.job_runners.config.size import ConfigSizeJobRunner @@ -22,2 +23,0 @@ from worker.job_runners.config_names import ConfigNamesJobRunner -from worker.job_runners.config_parquet import ConfigParquetJobRunner -from worker.job_runners.config_size import ConfigSizeJobRunner @@ -24,0 +25,2 @@ from worker.job_runners.dataset.info import DatasetInfoJobRunner +from worker.job_runners.dataset.parquet import DatasetParquetJobRunner +from worker.job_runners.dataset.size import DatasetSizeJobRunner @@ -31,2 +32,0 @@ from worker.job_runners.dataset.split_names_from_streaming import ( -from worker.job_runners.dataset_parquet import DatasetParquetJobRunner -from worker.job_runners.dataset_size import DatasetSizeJobRunner diff --git a/services/worker/src/worker/job_runners/config_parquet.py b/services/worker/src/worker/job_runners/config/parquet.py similarity index 100% rename from services/worker/src/worker/job_runners/config_parquet.py rename to services/worker/src/worker/job_runners/config/parquet.py diff --git a/services/worker/src/worker/job_runners/config_size.py b/services/worker/src/worker/job_runners/config/size.py similarity index 100% rename from services/worker/src/worker/job_runners/config_size.py rename to services/worker/src/worker/job_runners/config/size.py diff --git a/services/worker/src/worker/job_runners/dataset_parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py similarity index 90% rename from services/worker/src/worker/job_runners/dataset_parquet.py rename to services/worker/src/worker/job_runners/dataset/parquet.py index 9ac56905..e7171db2 100644 --- a/services/worker/src/worker/job_runners/dataset_parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -13 +13 @@ from worker.job_runner import JobResult, JobRunner, JobRunnerError -from worker.job_runners.config_parquet import ConfigParquetResponse +from worker.job_runners.config.parquet import ConfigParquetResponse @@ -79 +79 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - response = get_response(kind="/parquet-and-dataset-info", dataset=dataset) + response = get_response(kind="/config-names", dataset=dataset) @@ -81,3 +81 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - raise DatasetNotFoundError( - "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e - ) from e + raise DatasetNotFoundError("No response for '/config-names' found for this dataset: .", e) from e @@ -86 +84 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - f"Previous step gave an error: {response['http_status']}. This job should not have been created." + f"Previous step raised an error: {response['http_status']}. This job should not have been created." @@ -88,0 +87,3 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] + if "config_names" not in content: + raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + @@ -94 +95,2 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - for config in content["dataset_info"].keys(): + for config_item in content["config_names"]: + config = config_item["config"] diff --git a/services/worker/src/worker/job_runners/dataset_size.py b/services/worker/src/worker/job_runners/dataset/size.py similarity index 91% rename from services/worker/src/worker/job_runners/dataset_size.py rename to services/worker/src/worker/job_runners/dataset/size.py index a5dde782..06753c55 100644 --- a/services/worker/src/worker/job_runners/dataset_size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -13 +13 @@ from worker.job_runner import JobResult, JobRunner, JobRunnerError -from worker.job_runners.config_size import ConfigSize, ConfigSizeResponse, SplitSize +from worker.job_runners.config.size import ConfigSize, ConfigSizeResponse, SplitSize @@ -92 +92 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - response = get_response(kind="/parquet-and-dataset-info", dataset=dataset) + response = get_response(kind="/config-names", dataset=dataset) @@ -94,3 +94 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - raise DatasetNotFoundError( - "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e - ) from e + raise DatasetNotFoundError("No response for '/config-names' found for this dataset: .", e) from e @@ -99 +97 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - f"Previous step gave an error: {response['http_status']}. This job should not have been created." + f"Previous step raised an error: {response['http_status']}. This job should not have been created." @@ -101,0 +100,3 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: + if "config_names" not in content: + raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + @@ -108 +109,2 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - for config in content["dataset_info"].keys(): + for config_item in content["config_names"]: + config = config_item["config"] diff --git a/services/worker/tests/job_runners/test_config_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py similarity index 98% rename from services/worker/tests/job_runners/test_config_parquet.py rename to services/worker/tests/job_runners/config/test_parquet.py index 820bb6ce..c48c44af 100644 --- a/services/worker/tests/job_runners/test_config_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -15 +15 @@ from worker.config import AppConfig -from worker.job_runners.config_parquet import ( +from worker.job_runners.config.parquet import ( diff --git a/services/worker/tests/job_runners/test_config_size.py b/services/worker/tests/job_runners/config/test_size.py similarity index 99% rename from services/worker/tests/job_runners/test_config_size.py rename to services/worker/tests/job_runners/config/test_size.py index 546a2829..a257b3b5 100644 --- a/services/worker/tests/job_runners/test_config_size.py +++ b/services/worker/tests/job_runners/config/test_size.py @@ -15 +15 @@ from worker.config import AppConfig -from worker.job_runners.config_size import ( +from worker.job_runners.config.size import ( diff --git a/services/worker/tests/job_runners/test_dataset_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py similarity index 81% rename from services/worker/tests/job_runners/test_dataset_parquet.py rename to services/worker/tests/job_runners/dataset/test_parquet.py index 37cb0006..dc0f643d 100644 --- a/services/worker/tests/job_runners/test_dataset_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -15,2 +15,2 @@ from worker.config import AppConfig -from worker.job_runners.config_parquet import ConfigParquetResponse -from worker.job_runners.dataset_parquet import ( +from worker.job_runners.config.parquet import ConfigParquetResponse +from worker.job_runners.dataset.parquet import ( @@ -22,4 +22 @@ from worker.job_runners.dataset_parquet import ( -from worker.job_runners.parquet_and_dataset_info import ( - ParquetAndDatasetInfoResponse, - ParquetFileItem, -) +from worker.job_runners.parquet_and_dataset_info import ParquetFileItem @@ -27 +24 @@ from worker.job_runners.parquet_and_dataset_info import ( -from .utils import UpstreamResponse +from ..utils import UpstreamResponse @@ -83 +80 @@ def get_job_runner( - kind="/parquet-and-dataset-info", + kind="/config-names", @@ -87,18 +84,4 @@ def get_job_runner( - content=ParquetAndDatasetInfoResponse( - parquet_files=[ - ParquetFileItem( - dataset="ok", - config="config_1", - split="train", - url="url1", - filename="filename1", - size=0, - ), - ParquetFileItem( - dataset="ok", - config="config_2", - split="train", - url="url2", - filename="filename2", - size=0, - ), + content={ + "config_names": [ + {"dataset": "dataset_ok", "config": "config_1"}, + {"dataset": "dataset_ok", "config": "config_2"}, @@ -106,2 +89 @@ def get_job_runner( - dataset_info={"config_1": "value", "config_2": "value"}, - ), + }, @@ -165 +147 @@ def get_job_runner( - kind="/parquet-and-dataset-info", + kind="/config-names", @@ -180 +162 @@ def get_job_runner( - kind="/parquet-and-dataset-info", + kind="/config-names", diff --git a/services/worker/tests/job_runners/test_dataset_size.py b/services/worker/tests/job_runners/dataset/test_size.py similarity index 66% rename from services/worker/tests/job_runners/test_dataset_size.py rename to services/worker/tests/job_runners/dataset/test_size.py index 3413bd0d..668d2576 100644 --- a/services/worker/tests/job_runners/test_dataset_size.py +++ b/services/worker/tests/job_runners/dataset/test_size.py @@ -15 +15 @@ from worker.config import AppConfig -from worker.job_runners.dataset_size import ( +from worker.job_runners.dataset.size import ( @@ -21 +21 @@ from worker.job_runners.dataset_size import ( -from .utils import UpstreamResponse +from ..utils import UpstreamResponse @@ -77 +77 @@ def get_job_runner( - kind="/parquet-and-dataset-info", + kind="/config-names", @@ -82,6 +82,3 @@ def get_job_runner( - "parquet_files": [ - {"dataset": "dataset_ok", "config": "config_1", "split": "train", "size": 14281188}, - {"dataset": "dataset_ok", "config": "config_1", "split": "test", "size": 2383903}, - {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 1234}, - {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 6789}, - {"dataset": "dataset_ok", "config": "config_2", "split": "test", "size": 2383903}, + "config_names": [ + {"dataset": "dataset_ok", "config": "config_1"}, + {"dataset": "dataset_ok", "config": "config_2"}, @@ -89,79 +85,0 @@ def get_job_runner( - "dataset_info": { - "config_1": { - "features": { - "image": {"_type": "Image"}, - "label": { - "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], - "_type": "ClassLabel", - }, - }, - "splits": { - "train": { - "name": "train", - "num_bytes": 17470800, - "num_examples": 60000, - "dataset_name": "dataset_ok", - }, - "test": { - "name": "test", - "num_bytes": 2916432, - "num_examples": 10000, - "dataset_name": "dataset_ok", - }, - }, - "download_checksums": { - "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": { - "num_bytes": 9912422, - "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609", - }, - "https://storage.googleapis.com/cvdf-datasets/mnist/train-labels-idx1-ubyte.gz": { - "num_bytes": 28881, - "checksum": "3552534a0a558bbed6aed32b30c495cca23d567ec52cac8be1a0730e8010255c", - }, - "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-images-idx3-ubyte.gz": { - "num_bytes": 1648877, - "checksum": "8d422c7b0a1c1c79245a5bcf07fe86e33eeafee792b84584aec276f5a2dbc4e6", - }, - "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-labels-idx1-ubyte.gz": { - "num_bytes": 4542, - "checksum": "f7ae60f92e00ec6debd23a6088c31dbd2371eca3ffa0defaefb259924204aec6", - }, - }, - "download_size": 11594722, - "dataset_size": 20387232, - "size_in_bytes": 31981954, - }, - "config_2": { - "features": { - "image": {"_type": "Image"}, - "image2": {"_type": "Image"}, - "label": { - "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], - "_type": "ClassLabel", - }, - }, - "splits": { - "train": { - "name": "train", - "num_bytes": 5678, - "num_examples": 3000, - "dataset_name": "dataset_ok", - }, - "test": { - "name": "test", - "num_bytes": 1234, - "num_examples": 1000, - "dataset_name": "dataset_ok", - }, - }, - "download_checksums": { - "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": { - "num_bytes": 9912422, - "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609", - }, - }, - "download_size": 9912422, - "dataset_size": 6912, - "size_in_bytes": 9919334, - }, - }, @@ -327 +245 @@ def get_job_runner( - kind="/parquet-and-dataset-info", + kind="/config-names", @@ -342 +260 @@ def get_job_runner( - kind="/parquet-and-dataset-info", + kind="/config-names",
340f5ba4d572c257650e03ae4054f07d743e97fa
Polina Kazakova
2023-03-23T18:40:34
Config-level dataset info (#962)
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 13683ff7..39b97b92 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -188 +188 @@ workers: - workerOnlyJobTypes: "config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" + workerOnlyJobTypes: "config-parquet,dataset-parquet,config-info,dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 2dea2353..67032e90 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -216 +216 @@ workers: - workerOnlyJobTypes: "/config-names,config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" + workerOnlyJobTypes: "/config-names,config-parquet,dataset-parquet,config-info,dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" diff --git a/chart/values.yaml b/chart/values.yaml index 4cfdcd32..57d59865 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -333 +333 @@ workers: - workerOnlyJobTypes: "config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" + workerOnlyJobTypes: "config-parquet,dataset-parquet,config-info,dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 4e1c820f..053469d8 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -37,0 +38,6 @@ from mongodb_migration.migrations._20230320165700_queue_first_rows_from_streamin +from mongodb_migration.migrations._20230323155000_cache_dataset_info import ( + MigrationCacheUpdateDatasetInfo, +) +from mongodb_migration.migrations._20230323160000_queue_dataset_info import ( + MigrationQueueUpdateDatasetInfo, +) @@ -85,0 +92,8 @@ class MigrationsCollector: + MigrationCacheUpdateDatasetInfo( + version="20230323155000", + description="update 'kind' field in cache from '/dataset-info' to 'dataset-info'", + ), + MigrationQueueUpdateDatasetInfo( + version="20230323160000", + description="update 'type' and 'unicity_id' fields in job from /dataset-info to dataset-info", + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323155000_cache_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323155000_cache_dataset_info.py new file mode 100644 index 00000000..53d3929a --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323155000_cache_dataset_info.py @@ -0,0 +1,33 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from libcommon.simple_cache import CachedResponse +from mongoengine.connection import get_db + +from mongodb_migration.check import check_documents +from mongodb_migration.migration import Migration + +dataset_info = "/dataset-info" +dataset_info_updated = "dataset-info" +db_name = "cache" + + +class MigrationCacheUpdateDatasetInfo(Migration): + def up(self) -> None: + logging.info(f"Rename cache_kind field from {dataset_info} to {dataset_info_updated}") + db = get_db(db_name) + + # update existing documents with the old kind + db["cachedResponsesBlue"].update_many({"kind": dataset_info}, {"$set": {"kind": dataset_info_updated}}) + + def down(self) -> None: + logging.info(f"Rollback cache_kind field from {dataset_info_updated} to {dataset_info}") + db = get_db(db_name) + db["cachedResponsesBlue"].update_many({"kind": dataset_info_updated}, {"$set": {"kind": dataset_info}}) + + def validate(self) -> None: + logging.info("Validate modified documents") + + check_documents(DocCls=CachedResponse, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323160000_queue_dataset_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323160000_queue_dataset_info.py new file mode 100644 index 00000000..19e73670 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230323160000_queue_dataset_info.py @@ -0,0 +1,71 @@ +import logging + +from libcommon.queue import Job +from mongoengine.connection import get_db + +from mongodb_migration.check import check_documents +from mongodb_migration.migration import Migration + +dataset_info = "/dataset-info" +dataset_info_updated = "dataset-info" +db_name = "queue" + + +# connection already occurred in the main.py (caveat: we use globals) +class MigrationQueueUpdateDatasetInfo(Migration): + def up(self) -> None: + logging.info( + f"Rename unicity_id field from Job[{dataset_info}][<dataset>][<config>][split] to" + f" Job[{dataset_info_updated}][<dataset>][<config>][split] and change type from {dataset_info} to" + f" {dataset_info_updated}" + ) + + db = get_db(db_name) + db["jobsBlue"].update_many( + {"type": dataset_info}, + [ + { + "$set": { + "unicity_id": { + "$replaceOne": { + "input": "$unicity_id", + "find": f"Job[{dataset_info}]", + "replacement": f"Job[{dataset_info_updated}]", + } + }, + "type": dataset_info_updated, + } + }, + ], # type: ignore + ) + + def down(self) -> None: + logging.info( + f"Rename unicity_id field from Job[{dataset_info_updated}][<dataset>][<config>][split] to" + f" Job[{dataset_info}][<dataset>][<config>][split] and change type from {dataset_info_updated} to" + f" {dataset_info}" + ) + + db = get_db(db_name) + db["jobsBlue"].update_many( + {"type": dataset_info_updated}, + [ + { + "$set": { + "unicity_id": { + "$replaceOne": { + "input": "$unicity_id", + "find": f"Job[{dataset_info_updated}]", + "replacement": f"Job[{dataset_info}]", + } + }, + "type": dataset_info_updated, + } + }, + ], # type: ignore + ) + + def validate(self) -> None: + logging.info("Validate modified documents") + + check_documents(DocCls=Job, sample_size=10) diff --git a/jobs/mongodb_migration/tests/migrations/test_20230323155000_cache_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230323155000_cache_dataset_info.py new file mode 100644 index 00000000..5d1f3b1b --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230323155000_cache_dataset_info.py @@ -0,0 +1,29 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230323155000_cache_dataset_info import ( + MigrationCacheUpdateDatasetInfo, +) + + +def test_cache_update_dataset_info_kind(mongo_host: str) -> None: + old_kind, new_kind = "/dataset-info", "dataset-info" + with MongoResource(database="test_cache_update_dataset_info_kind", host=mongo_host, mongoengine_alias="cache"): + db = get_db("cache") + db["cachedResponsesBlue"].insert_many([{"kind": old_kind, "dataset": "dataset", "http_status": 200}]) + assert db["cachedResponsesBlue"].find_one({"kind": old_kind}) # Ensure there is at least one record to update + + migration = MigrationCacheUpdateDatasetInfo( + version="20230323155000", + description=f"update 'kind' field in cache from {old_kind} to {new_kind}", + ) + migration.up() + + assert not db["cachedResponsesBlue"].find_one({"kind": old_kind}) # Ensure 0 records with old kind + + assert db["cachedResponsesBlue"].find_one({"kind": new_kind}) + + db["cachedResponsesBlue"].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230323160000_queue_dataset_info.py b/jobs/mongodb_migration/tests/migrations/test_20230323160000_queue_dataset_info.py new file mode 100644 index 00000000..423bef04 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230323160000_queue_dataset_info.py @@ -0,0 +1,41 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230323160000_queue_dataset_info import ( + MigrationQueueUpdateDatasetInfo, +) + + +def test_queue_update_dataset_info_type_and_unicity_id(mongo_host: str) -> None: + old_kind, new_kind = "/dataset-info", "dataset-info" + with MongoResource( + database="test_queue_update_dataset_info_type_and_unicity_id", host=mongo_host, mongoengine_alias="queue" + ): + db = get_db("queue") + db["jobsBlue"].insert_many( + [ + { + "type": old_kind, + "unicity_id": f"Job[{old_kind}][dataset][config][split]", + "dataset": "dataset", + "http_status": 200, + } + ] + ) + assert db["jobsBlue"].find_one({"type": old_kind}) # Ensure there is at least one record to update + + migration = MigrationQueueUpdateDatasetInfo( + version="20230323160000", + description=f"update 'type' and 'unicity_id' fields in job from {old_kind} to {new_kind}", + ) + migration.up() + + assert not db["jobsBlue"].find_one({"type": old_kind}) # Ensure 0 records with old type + + result = db["jobsBlue"].find_one({"type": new_kind}) + assert result + assert result["unicity_id"] == f"Job[{new_kind}][dataset][config][split]" + db["jobsBlue"].drop() diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 74b666ee..8721d0a0 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -11,0 +12 @@ from libcommon.constants import ( + PROCESSING_STEP_CONFIG_INFO_VERSION, @@ -145,2 +146,2 @@ class ProcessingGraphConfig: - "/dataset-info": { - "input_type": "dataset", + "config-info": { + "input_type": "config", @@ -147,0 +149,5 @@ class ProcessingGraphConfig: + "job_runner_version": PROCESSING_STEP_CONFIG_INFO_VERSION, + }, + "dataset-info": { + "input_type": "dataset", + "requires": "config-info", @@ -152 +158 @@ class ProcessingGraphConfig: - "requires": "/dataset-info", + "requires": "dataset-info", diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index dc51ce04..f1a67f3e 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -11,0 +12 @@ PROCESSING_STEP_CONFIG_SIZE_VERSION = 1 +PROCESSING_STEP_CONFIG_INFO_VERSION = 1 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index f9122608..7db1c9ad 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -19 +19,2 @@ def test_default_graph() -> None: - dataset_info = graph.get_step("/dataset-info") + config_info = graph.get_step("config-info") + dataset_info = graph.get_step("dataset-info") @@ -48 +49 @@ def test_default_graph() -> None: - assert parquet_and_dataset_info.children == [config_parquet, dataset_info, config_size] + assert parquet_and_dataset_info.children == [config_parquet, config_info, config_size] @@ -60,0 +62,5 @@ def test_default_graph() -> None: + assert config_info is not None + assert config_info.parent is parquet_and_dataset_info + assert config_info.children == [dataset_info] + assert config_info.get_ancestors() == [parquet_and_dataset_info] + @@ -62 +68 @@ def test_default_graph() -> None: - assert dataset_info.parent is parquet_and_dataset_info + assert dataset_info.parent is config_info @@ -64 +70 @@ def test_default_graph() -> None: - assert dataset_info.get_ancestors() == [parquet_and_dataset_info] + assert dataset_info.get_ancestors() == [parquet_and_dataset_info, config_info] @@ -69 +75 @@ def test_default_graph() -> None: - assert split_names_from_dataset_info.get_ancestors() == [parquet_and_dataset_info, dataset_info] + assert split_names_from_dataset_info.get_ancestors() == [parquet_and_dataset_info, config_info, dataset_info] @@ -90,0 +97 @@ def test_default_graph() -> None: + config_info, diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index b8c0cab9..1628c9e2 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -125 +125 @@ class EndpointConfig: - "/dataset-info": {"dataset": ["/dataset-info"]}, + "/dataset-info": {"dataset": ["dataset-info"], "config": ["config-info"]}, diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index 2fa834f5..cdd42219 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -29,0 +30 @@ def test_endpoints_definition() -> None: + assert sorted(list(config_names)) == ["dataset"] @@ -34,0 +36 @@ def test_endpoints_definition() -> None: + assert sorted(list(splits)) == ["config", "dataset"] @@ -35,0 +38 @@ def test_endpoints_definition() -> None: + assert splits["config"] is not None @@ -40,0 +44 @@ def test_endpoints_definition() -> None: + assert sorted(list(first_rows)) == ["split"] @@ -45,0 +50 @@ def test_endpoints_definition() -> None: + assert sorted(list(parquet_and_dataset_info)) == ["dataset"] @@ -50,0 +56 @@ def test_endpoints_definition() -> None: + assert sorted(list(parquet)) == ["config", "dataset"] @@ -51,0 +58 @@ def test_endpoints_definition() -> None: + assert parquet["config"] is not None @@ -52,0 +60 @@ def test_endpoints_definition() -> None: + assert len(parquet["config"]) == 1 # Only has one processing step @@ -55,0 +64 @@ def test_endpoints_definition() -> None: + assert sorted(list(dataset_info)) == ["config", "dataset"] @@ -56,0 +66 @@ def test_endpoints_definition() -> None: + assert dataset_info["config"] is not None @@ -57,0 +68 @@ def test_endpoints_definition() -> None: + assert len(dataset_info["config"]) == 1 # Only has one processing step @@ -60,0 +72 @@ def test_endpoints_definition() -> None: + assert sorted(list(size)) == ["config", "dataset"] diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index bc155d6d..565d409b 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -13,0 +14 @@ from worker.job_runner import JobRunner +from worker.job_runners.config.info import ConfigInfoJobRunner @@ -22,0 +24 @@ from worker.job_runners.config_size import ConfigSizeJobRunner +from worker.job_runners.dataset.info import DatasetInfoJobRunner @@ -29 +30,0 @@ from worker.job_runners.dataset.split_names_from_streaming import ( -from worker.job_runners.dataset_info import DatasetInfoJobRunner @@ -131,0 +133,7 @@ class JobRunnerFactory(BaseJobRunnerFactory): + if job_type == ConfigInfoJobRunner.get_job_type(): + return ConfigInfoJobRunner( + job_info=job_info, + common_config=self.app_config.common, + worker_config=self.app_config.worker, + processing_step=processing_step, + ) @@ -175,0 +184 @@ class JobRunnerFactory(BaseJobRunnerFactory): + ConfigInfoJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py new file mode 100644 index 00000000..fe913f81 --- /dev/null +++ b/services/worker/src/worker/job_runners/config/info.py @@ -0,0 +1,146 @@ +import logging +from http import HTTPStatus +from typing import Any, Dict, Literal, Mapping, Optional, Set, TypedDict + +from libcommon.constants import PROCESSING_STEP_CONFIG_INFO_VERSION +from libcommon.dataset import DatasetNotFoundError +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response + +from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError +from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoResponse + +ConfigInfoJobRunnerErrorCode = Literal[ + "PreviousStepStatusError", + "PreviousStepFormatError", + "MissingInfoForConfigError", +] + + +class ConfigInfoJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: ConfigInfoJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class PreviousStepStatusError(ConfigInfoJobRunnerError): + """Raised when the previous step gave an error. The job should not have been created.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) + + +class PreviousStepFormatError(ConfigInfoJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +class MissingInfoForConfigError(ConfigInfoJobRunnerError): + """Raised when the dataset info from the parquet export is missing the requested dataset configuration.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "MissingInfoForConfigError", cause, False) + + +class ConfigInfoResponse(TypedDict): + dataset_info: Dict[str, Any] + + +def compute_config_info_response(dataset: str, config: str) -> ConfigInfoResponse: + """ + Get the response of config-info for one specific config of a specific dataset on huggingface.co. + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + Returns: + `ConfigInfoResponse`: An object with the dataset_info response for requested config. + <Tip> + Raises the following errors: + - [`~job_runners.config_info.PreviousStepStatusError`] + `If the previous step gave an error. + - [`~job_runners.config_info.PreviousStepFormatError`] + If the content of the previous step doesn't have the expected format + - [`~job_runners.config_info.MissingInfoForConfigError`] + If the dataset info from the parquet export doesn't have the requested dataset configuration + - [`~libcommon.dataset.DatasetNotFoundError`] + If the dataset does not exist, or if the + token does not give the sufficient access to the dataset, or if the dataset is private + (private datasets are not supported by the datasets server) + </Tip> + """ + logging.info(f"get dataset_info for {dataset=} and {config=}") + + try: + response = get_response(kind="/parquet-and-dataset-info", dataset=dataset) + except DoesNotExist as e: + raise DatasetNotFoundError( + "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e + ) from e + + if response["http_status"] != HTTPStatus.OK: + raise PreviousStepStatusError(f"Previous step raised an error: {response['http_status']}..") + + try: + content = ParquetAndDatasetInfoResponse( + parquet_files=response["content"]["parquet_files"], dataset_info=response["content"]["dataset_info"] + ) + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content: 'dataset_info'.", e) from e + + if config not in content["dataset_info"]: + if not isinstance(content["dataset_info"], dict): + raise PreviousStepFormatError( + "Previous step did not return the expected content.", + TypeError(f"dataset_info should be a dict, but got {type(content['dataset_info'])}"), + ) + raise MissingInfoForConfigError( + f"Dataset configuration '{config}' is missing in the dataset info from the parquet export. " + f"Available configurations: {', '.join(list(content['dataset_info'])[:10])}" + + f"... ({len(content['dataset_info']) - 10})" + if len(content["dataset_info"]) > 10 + else "" + ) + try: + config_info = content["dataset_info"][config] + + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + return ConfigInfoResponse(dataset_info=config_info) + + +class ConfigInfoJobRunner(JobRunner): + @staticmethod + def get_job_type() -> str: + return "config-info" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_CONFIG_INFO_VERSION + + def compute(self) -> CompleteJobResult: + if self.dataset is None: + raise ValueError("dataset is required") + if self.config is None: + raise ValueError("config is required") + return CompleteJobResult(compute_config_info_response(dataset=self.dataset, config=self.config)) + + # TODO: is it needed? + def get_new_splits(self, content: Mapping[str, Any]) -> Set[SplitFullName]: + """Get the set of new splits, from the content created by the compute.""" + return { + SplitFullName(dataset=self.dataset, config=self.config, split=split) + for split in content["dataset_info"]["splits"] + } diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index eebeab16..732be515 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -58 +58 @@ def compute_split_names_from_dataset_info_response( - computed from cached response in /dataset-info step. + computed from cached response in dataset-info step. @@ -61 +61 @@ def compute_split_names_from_dataset_info_response( - like the size or number of samples. See /dataset-info or dataset-size for that. + like the size or number of samples. See dataset-info or dataset-size for that. @@ -83 +83 @@ def compute_split_names_from_dataset_info_response( - response = get_response(kind="/dataset-info", dataset=dataset) + response = get_response(kind="dataset-info", dataset=dataset) diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index 82e1fa98..ed1835f1 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -69 +69 @@ def compute_split_names_from_streaming_response( - like the size or number of samples. See /dataset-info or dataset-size for that. + like the size or number of samples. See dataset-info or dataset-size for that. diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py new file mode 100644 index 00000000..5d3d1b39 --- /dev/null +++ b/services/worker/src/worker/job_runners/dataset/info.py @@ -0,0 +1,155 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from http import HTTPStatus +from typing import Any, Dict, List, Literal, Mapping, Optional, Tuple, TypedDict + +from libcommon.constants import PROCESSING_STEP_DATASET_INFO_VERSION +from libcommon.dataset import DatasetNotFoundError +from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response + +from worker.job_runner import JobResult, JobRunner, JobRunnerError +from worker.utils import PreviousJob + +DatasetInfoJobRunnerErrorCode = Literal[ + "PreviousStepStatusError", + "PreviousStepFormatError", +] + + +class DatasetInfoResponse(TypedDict): + dataset_info: Dict[str, Any] + pending: List[PreviousJob] + failed: List[PreviousJob] + + +class DatasetInfoJobRunnerError(JobRunnerError): + """Base class for exceptions in this module.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: DatasetInfoJobRunnerErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class PreviousStepStatusError(DatasetInfoJobRunnerError): + """Raised when the previous step gave an error. The job should not have been created.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) + + +class PreviousStepFormatError(DatasetInfoJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + +def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, float]: + """ + Get the response of dataset-info for one specific dataset on huggingface.co. + Args: + dataset (`str`): + A namespace (user or an organization) and a repo name separated + by a `/`. + Returns: + (`DatasetInfoResponse`, `float`): Tuple of an object with the dataset_info response and + progress float value from 0. to 1. which corresponds to the percentage of dataset configs + correctly processed and included in current response (some configs might not exist in cache yet + or raise errors). + <Tip> + Raises the following errors: + - [`~job_runners.dataset_info.PreviousStepStatusError`] + If the previous step gave an error. + - [`~job_runners.dataset_info.PreviousStepFormatError`] + If the content of the previous step doesn't have the expected format. + - [`~libcommon.dataset.DatasetNotFoundError`] + If the dataset does not exist, or if the + token does not give the sufficient access to the dataset, or if the dataset is private + (private datasets are not supported by the datasets server) + </Tip> + """ + logging.info(f"get dataset_info for {dataset=}") + + try: + response = get_response(kind="/config-names", dataset=dataset) + except DoesNotExist as e: + raise DatasetNotFoundError("No response for '/config-names' found for this dataset: .", e) from e + if response["http_status"] != HTTPStatus.OK: + raise PreviousStepStatusError( + f"Previous step raised an error: {response['http_status']}. This job should not have been created." + ) + content = response["content"] + if "config_names" not in content: + raise PreviousStepFormatError("'/config-names' did not return the expected content: 'config_names'.") + + try: + config_infos: Dict[str, Any] = {} + total = 0 + pending, failed = [], [] + for config_item in content["config_names"]: + config = config_item["config"] + total += 1 + try: + config_response = get_response(kind="config-info", dataset=dataset, config=config) + except DoesNotExist: + logging.debug(f"No response found in previous step for {dataset=} {config=}: 'config-info'.") + pending.append( + PreviousJob( + kind="config-info", + dataset=dataset, + config=config, + split=None, + ) + ) + continue + if config_response["http_status"] != HTTPStatus.OK: + logging.debug(f"Previous step gave an error: {response['http_status']}") + failed.append( + PreviousJob( + kind="config-info", + dataset=dataset, + config=config, + split=None, + ) + ) + continue + config_infos[config] = config_response["content"]["dataset_info"] + + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + progress = (total - len(pending)) / total if total else 1.0 + + return DatasetInfoResponse(dataset_info=config_infos, pending=pending, failed=failed), progress + + +class DatasetInfoJobRunner(JobRunner): + @staticmethod + def get_job_type() -> str: + return "dataset-info" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_DATASET_INFO_VERSION + + def compute(self) -> JobResult: + response_content, progress = compute_dataset_info_response(dataset=self.dataset) + return JobResult(response_content, progress=progress) + + def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]: + """Get the set of new splits, from the content created by self.compute()""" + return { + SplitFullName(dataset=self.dataset, config=config, split=split) + for config in content["dataset_info"] + for split in content["dataset_info"][config]["splits"] + } diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py index 65a41e6a..9a8f69a3 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py @@ -81 +81 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - response = get_response(kind="/dataset-info", dataset=dataset) + response = get_response(kind="dataset-info", dataset=dataset) @@ -84 +84 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - raise DatasetNotFoundError("No response found in previous step for this dataset: '/dataset-info'.", e) from e + raise DatasetNotFoundError("No response found in previous step for this dataset: 'dataset-info'.", e) from e @@ -86 +86 @@ def compute_dataset_split_names_from_dataset_info_response(dataset: str) -> Tupl - raise PreviousStepFormatError("Previous step '/dataset-info' did not return the expected content.") from e + raise PreviousStepFormatError("Previous step 'dataset-info' did not return the expected content.") from e diff --git a/services/worker/src/worker/job_runners/dataset_info.py b/services/worker/src/worker/job_runners/dataset_info.py deleted file mode 100644 index 00be4cb2..00000000 --- a/services/worker/src/worker/job_runners/dataset_info.py +++ /dev/null @@ -1,109 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import logging -from http import HTTPStatus -from typing import Any, Literal, Mapping, Optional, TypedDict - -from libcommon.constants import PROCESSING_STEP_DATASET_INFO_VERSION -from libcommon.dataset import DatasetNotFoundError -from libcommon.simple_cache import DoesNotExist, SplitFullName, get_response - -from worker.job_runner import CompleteJobResult, JobRunner, JobRunnerError - -DatasetInfoJobRunnerErrorCode = Literal[ - "PreviousStepStatusError", - "PreviousStepFormatError", -] - - -class DatasetInfoResponse(TypedDict): - dataset_info: dict[str, Any] - - -class DatasetInfoJobRunnerError(JobRunnerError): - """Base class for exceptions in this module.""" - - def __init__( - self, - message: str, - status_code: HTTPStatus, - code: DatasetInfoJobRunnerErrorCode, - cause: Optional[BaseException] = None, - disclose_cause: bool = False, - ): - super().__init__( - message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause - ) - - -class PreviousStepStatusError(DatasetInfoJobRunnerError): - """Raised when the previous step gave an error. The job should not have been created.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) - - -class PreviousStepFormatError(DatasetInfoJobRunnerError): - """Raised when the content of the previous step has not the expected format.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) - - -def compute_dataset_info_response(dataset: str) -> DatasetInfoResponse: - """ - Get the response of /dataset-info for one specific dataset on huggingface.co. - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - Returns: - `DatasetInfoResponse`: An object with the dataset_info response. - <Tip> - Raises the following errors: - - [`~job_runners.dataset_info.PreviousStepStatusError`] - If the previous step gave an error. - - [`~job_runners.dataset_info.PreviousStepFormatError`] - If the content of the previous step has not the expected format - </Tip> - """ - logging.info(f"get dataset_info for dataset={dataset}") - - try: - response = get_response(kind="/parquet-and-dataset-info", dataset=dataset) - except DoesNotExist as e: - raise DatasetNotFoundError( - "No response found in previous step for this dataset: '/parquet-and-dataset-info'.", e - ) from e - if response["http_status"] != HTTPStatus.OK: - raise PreviousStepStatusError( - f"Previous step gave an error: {response['http_status']}. This job should not have been created." - ) - content = response["content"] - if "dataset_info" not in content: - raise PreviousStepFormatError("Previous step did not return the expected content: 'dataset_info'.") - return { - "dataset_info": content["dataset_info"], - } - - -class DatasetInfoJobRunner(JobRunner): - @staticmethod - def get_job_type() -> str: - return "/dataset-info" - - @staticmethod - def get_job_runner_version() -> int: - return PROCESSING_STEP_DATASET_INFO_VERSION - - def compute(self) -> CompleteJobResult: - return CompleteJobResult(compute_dataset_info_response(dataset=self.dataset)) - - def get_new_splits(self, content: Mapping[str, Any]) -> set[SplitFullName]: - """Get the set of new splits, from the content created by the compute.""" - return { - SplitFullName(dataset=self.dataset, config=config, split=split) - for config in content["dataset_info"].keys() - for split in content["dataset_info"][config]["splits"].keys() - } diff --git a/services/worker/src/worker/job_runners/dataset_parquet.py b/services/worker/src/worker/job_runners/dataset_parquet.py index 4820b779..9ac56905 100644 --- a/services/worker/src/worker/job_runners/dataset_parquet.py +++ b/services/worker/src/worker/job_runners/dataset_parquet.py @@ -14,0 +15 @@ from worker.job_runners.parquet_and_dataset_info import ParquetFileItem +from worker.utils import PreviousJob @@ -22,7 +22,0 @@ SizesJobRunnerErrorCode = Literal[ -class PreviousJob(TypedDict): - kind: str - dataset: str - config: Optional[str] - split: Optional[str] - - diff --git a/services/worker/src/worker/job_runners/dataset_size.py b/services/worker/src/worker/job_runners/dataset_size.py index d606fc62..a5dde782 100644 --- a/services/worker/src/worker/job_runners/dataset_size.py +++ b/services/worker/src/worker/job_runners/dataset_size.py @@ -13,0 +14 @@ from worker.job_runners.config_size import ConfigSize, ConfigSizeResponse, Split +from worker.utils import PreviousJob @@ -35,7 +35,0 @@ class DatasetSizeContent(TypedDict): -class PreviousJob(TypedDict): - kind: str - dataset: str - config: Optional[str] - split: Optional[str] - - diff --git a/services/worker/src/worker/job_runners/splits.py b/services/worker/src/worker/job_runners/splits.py index 8914ed5b..cc9d611b 100644 --- a/services/worker/src/worker/job_runners/splits.py +++ b/services/worker/src/worker/job_runners/splits.py @@ -98 +98 @@ def compute_splits_response( - of samples. See /dataset-info or dataset-size for that. + of samples. See dataset-info or dataset-size for that. diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 906f7c91..687fc272 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -30,0 +31,7 @@ class DatasetSplitNamesResponse(TypedDict): + + +class PreviousJob(TypedDict): + kind: str + dataset: str + config: Optional[str] + split: Optional[str] diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py new file mode 100644 index 00000000..7da2a824 --- /dev/null +++ b/services/worker/tests/job_runners/config/test_info.py @@ -0,0 +1,238 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Callable + +import pytest +from libcommon.dataset import DatasetNotFoundError +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response + +from worker.config import AppConfig +from worker.job_runners.config.info import ( + ConfigInfoJobRunner, + PreviousStepFormatError, + PreviousStepStatusError, +) + + [email protected](autouse=True) +def prepare_and_clean_mongo(app_config: AppConfig) -> None: + # prepare the database before each test, and clean it afterwards + pass + + +GetJobRunner = Callable[[str, str, AppConfig, bool], ConfigInfoJobRunner] + + +CONFIG_INFO_1 = { + "description": "_DESCRIPTION", + "citation": "_CITATION", + "homepage": "_HOMEPAGE", + "license": "_LICENSE", + "features": { + "image": {"_type": "Image"}, + "label": { + "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], + "_type": "ClassLabel", + }, + }, + "splits": { + "train": { + "name": "train", + "num_bytes": 17470800, + "num_examples": 60000, + "dataset_name": "dataset_ok", + }, + "test": { + "name": "test", + "num_bytes": 2916432, + "num_examples": 10000, + "dataset_name": "dataset_ok", + }, + }, + "builder_name": "dataset_ok", + "config_name": "config_1", + "version": {"version_str": "0.0.0", "major": 0, "minor": 0, "patch": 0}, + "download_checksums": { + "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": { + "num_bytes": 9912422, + "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609", + }, + "https://storage.googleapis.com/cvdf-datasets/mnist/train-labels-idx1-ubyte.gz": { + "num_bytes": 28881, + "checksum": "3552534a0a558bbed6aed32b30c495cca23d567ec52cac8be1a0730e8010255c", + }, + "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-images-idx3-ubyte.gz": { + "num_bytes": 1648877, + "checksum": "8d422c7b0a1c1c79245a5bcf07fe86e33eeafee792b84584aec276f5a2dbc4e6", + }, + "https://storage.googleapis.com/cvdf-datasets/mnist/t10k-labels-idx1-ubyte.gz": { + "num_bytes": 4542, + "checksum": "f7ae60f92e00ec6debd23a6088c31dbd2371eca3ffa0defaefb259924204aec6", + }, + }, + "download_size": 11594722, + "dataset_size": 20387232, + "size_in_bytes": 31981954, +} + +CONFIG_INFO_2 = { + "description": "_DESCRIPTION", + "citation": "_CITATION", + "homepage": "_HOMEPAGE", + "license": "_LICENSE", + "features": { + "image": {"_type": "Image"}, + "image2": {"_type": "Image"}, + "label": { + "names": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], + "_type": "ClassLabel", + }, + }, + "splits": { + "train": { + "name": "train", + "num_bytes": 5678, + "num_examples": 3000, + "dataset_name": "dataset_ok", + }, + "test": { + "name": "test", + "num_bytes": 1234, + "num_examples": 1000, + "dataset_name": "dataset_ok", + }, + }, + "builder_name": "dataset_ok", + "config_name": "config_2", + "version": {"version_str": "0.0.0", "major": 0, "minor": 0, "patch": 0}, + "download_checksums": { + "https://storage.googleapis.com/cvdf-datasets/mnist/train-images-idx3-ubyte.gz": { + "num_bytes": 9912422, + "checksum": "440fcabf73cc546fa21475e81ea370265605f56be210a4024d2ca8f203523609", + }, + }, + "download_size": 9912422, + "dataset_size": 6912, + "size_in_bytes": 9919334, +} + +DATASET_INFO_OK = { + "config_1": CONFIG_INFO_1, + "config_2": CONFIG_INFO_2, +} + +PARQUET_FILES = [ + {"dataset": "dataset_ok", "config": "config_1", "split": "train", "size": 14281188}, + {"dataset": "dataset_ok", "config": "config_1", "split": "test", "size": 2383903}, + {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 1234}, + {"dataset": "dataset_ok", "config": "config_2", "split": "train", "size": 6789}, + {"dataset": "dataset_ok", "config": "config_2", "split": "test", "size": 2383903}, +] + + [email protected] +def get_job_runner( + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + app_config: AppConfig, + force: bool = False, + ) -> ConfigInfoJobRunner: + return ConfigInfoJobRunner( + job_info={ + "type": ConfigInfoJobRunner.get_job_type(), + "dataset": dataset, + "config": config, + "split": None, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + common_config=app_config.common, + worker_config=app_config.worker, + processing_step=ProcessingStep( + name=ConfigInfoJobRunner.get_job_type(), + input_type="dataset", + requires=None, + required_by_dataset_viewer=False, + parent=None, + ancestors=[], + children=[], + job_runner_version=ConfigInfoJobRunner.get_job_runner_version(), + ), + ) + + return _get_job_runner + + [email protected]( + "dataset,config,upstream_status,upstream_content,expected_error_code,expected_content,should_raise", + [ + ( + "dataset_ok", + "config_1", + HTTPStatus.OK, + { + "parquet_files": PARQUET_FILES, + "dataset_info": DATASET_INFO_OK, + }, + None, + {"dataset_info": CONFIG_INFO_1}, + False, + ), + ( + "status_error", + "config_1", + HTTPStatus.NOT_FOUND, + {"error": "error"}, + PreviousStepStatusError.__name__, + None, + True, + ), + ( + "format_error", + "config_1", + HTTPStatus.OK, + {"not_dataset_info": "wrong_format"}, + PreviousStepFormatError.__name__, + None, + True, + ), + ], +) +def test_compute( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset: str, + config: str, + upstream_status: HTTPStatus, + upstream_content: Any, + expected_error_code: str, + expected_content: Any, + should_raise: bool, +) -> None: + upsert_response( + kind="/parquet-and-dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status + ) + job_runner = get_job_runner(dataset, config, app_config, False) + if should_raise: + with pytest.raises(Exception) as e: + job_runner.compute() + assert e.type.__name__ == expected_error_code + else: + assert job_runner.compute().content == expected_content + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = config = "doesnotexist" + job_runner = get_job_runner(dataset, config, app_config, False) + with pytest.raises(DatasetNotFoundError): + job_runner.compute() diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index d9f31437..429a9b41 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -130 +130 @@ def test_compute( - upsert_response(kind="/dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status) + upsert_response(kind="dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status) diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py new file mode 100644 index 00000000..0b72cf4b --- /dev/null +++ b/services/worker/tests/job_runners/dataset/test_info.py @@ -0,0 +1,243 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Callable, List + +import pytest +from libcommon.dataset import DatasetNotFoundError +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response + +from worker.config import AppConfig +from worker.job_runners.dataset.info import ( + DatasetInfoJobRunner, + PreviousStepFormatError, + PreviousStepStatusError, +) +from worker.utils import PreviousJob + +from ..config.test_info import CONFIG_INFO_1, CONFIG_INFO_2, DATASET_INFO_OK +from ..utils import UpstreamResponse + + [email protected](autouse=True) +def prepare_and_clean_mongo(app_config: AppConfig) -> None: + # prepare the database before each test, and clean it afterwards + pass + + +GetJobRunner = Callable[[str, AppConfig, bool], DatasetInfoJobRunner] + + +UPSTREAM_RESPONSE_CONFIG_NAMES: UpstreamResponse = UpstreamResponse( + kind="/config-names", + dataset="dataset_ok", + config=None, + http_status=HTTPStatus.OK, + content={ + "config_names": [ + {"dataset": "dataset_ok", "config": "config_1"}, + {"dataset": "dataset_ok", "config": "config_2"}, + ], + }, +) + +UPSTREAM_RESPONSE_CONFIG_INFO_1: UpstreamResponse = UpstreamResponse( + kind="config-info", + dataset="dataset_ok", + config="config_1", + http_status=HTTPStatus.OK, + content={"dataset_info": CONFIG_INFO_1}, +) + +UPSTREAM_RESPONSE_CONFIG_INFO_2: UpstreamResponse = UpstreamResponse( + kind="config-info", + dataset="dataset_ok", + config="config_2", + http_status=HTTPStatus.OK, + content={"dataset_info": CONFIG_INFO_2}, +) + +EXPECTED_OK = ( + { + "dataset_info": DATASET_INFO_OK, + "pending": [], + "failed": [], + }, + 1.0, +) + +EXPECTED_PARTIAL_PENDING = ( + { + "dataset_info": { + "config_1": CONFIG_INFO_1, + }, + "pending": [ + PreviousJob( + kind="config-info", + dataset="dataset_ok", + config="config_2", + split=None, + ) + ], + "failed": [], + }, + 0.5, +) + +EXPECTED_PARTIAL_FAILED = ( + { + "dataset_info": { + "config_1": CONFIG_INFO_1, + }, + "pending": [], + "failed": [ + PreviousJob( + kind="config-info", + dataset="dataset_ok", + config="config_2", + split=None, + ) + ], + }, + 1.0, +) + + [email protected] +def get_job_runner( + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + app_config: AppConfig, + force: bool = False, + ) -> DatasetInfoJobRunner: + return DatasetInfoJobRunner( + job_info={ + "type": DatasetInfoJobRunner.get_job_type(), + "dataset": dataset, + "config": None, + "split": None, + "job_id": "job_id", + "force": force, + "priority": Priority.NORMAL, + }, + common_config=app_config.common, + worker_config=app_config.worker, + processing_step=ProcessingStep( + name=DatasetInfoJobRunner.get_job_type(), + input_type="dataset", + requires=None, + required_by_dataset_viewer=False, + parent=None, + ancestors=[], + children=[], + job_runner_version=DatasetInfoJobRunner.get_job_runner_version(), + ), + ) + + return _get_job_runner + + [email protected]( + "dataset,upstream_responses,expected_error_code,expected,should_raise", + [ + ( + "dataset_ok", + [ + UPSTREAM_RESPONSE_CONFIG_NAMES, + UPSTREAM_RESPONSE_CONFIG_INFO_1, + UPSTREAM_RESPONSE_CONFIG_INFO_2, + ], + None, + EXPECTED_OK, + False, + ), + ( + "dataset_ok", + [UPSTREAM_RESPONSE_CONFIG_NAMES, UPSTREAM_RESPONSE_CONFIG_INFO_1], + None, + EXPECTED_PARTIAL_PENDING, + False, + ), + ( + "dataset_ok", + [ + UPSTREAM_RESPONSE_CONFIG_NAMES, + UPSTREAM_RESPONSE_CONFIG_INFO_1, + UpstreamResponse( + kind="config-info", + dataset="dataset_ok", + config="config_2", + http_status=HTTPStatus.NOT_FOUND, + content={"error": "error"}, + ), + ], + None, + EXPECTED_PARTIAL_FAILED, + False, + ), + ( + "status_error", + [ + UpstreamResponse( + kind="/config-names", + dataset="status_error", + config=None, + http_status=HTTPStatus.NOT_FOUND, + content={"error": "error"}, + ) + ], + PreviousStepStatusError.__name__, + None, + True, + ), + ( + "format_error", + [ + UpstreamResponse( + kind="/config-names", + dataset="format_error", + config=None, + http_status=HTTPStatus.OK, + content={"not_dataset_info": "wrong_format"}, + ) + ], + PreviousStepFormatError.__name__, + None, + True, + ), + ], +) +def test_compute( + app_config: AppConfig, + get_job_runner: GetJobRunner, + dataset: str, + upstream_responses: List[UpstreamResponse], + expected_error_code: str, + expected: Any, + should_raise: bool, +) -> None: + for upstream_response in upstream_responses: + upsert_response(**upstream_response) + job_runner = get_job_runner(dataset, app_config, False) + if should_raise: + with pytest.raises(Exception) as e: + job_runner.compute() + assert e.type.__name__ == expected_error_code + else: + compute_result = job_runner.compute() + assert compute_result.content == expected[0] + assert compute_result.progress == expected[1] + + +def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset = "doesnotexist" + job_runner = get_job_runner(dataset, app_config, False) + with pytest.raises(DatasetNotFoundError): + job_runner.compute() diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py index 127a15bf..0c0fccea 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py @@ -149 +149 @@ def test_compute_progress( - kind="/dataset-info", + kind="dataset-info", @@ -172 +172 @@ def test_compute_error(app_config: AppConfig, get_job_runner: GetJobRunner) -> N - kind="/dataset-info", + kind="dataset-info", @@ -198 +198 @@ def test_compute_format_error(app_config: AppConfig, get_job_runner: GetJobRunne - kind="/dataset-info", + kind="dataset-info", diff --git a/services/worker/tests/job_runners/test_dataset_info.py b/services/worker/tests/job_runners/test_dataset_info.py deleted file mode 100644 index ff2d4138..00000000 --- a/services/worker/tests/job_runners/test_dataset_info.py +++ /dev/null @@ -1,116 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from http import HTTPStatus -from typing import Any, Callable - -import pytest -from libcommon.dataset import DatasetNotFoundError -from libcommon.processing_graph import ProcessingStep -from libcommon.queue import Priority -from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import upsert_response - -from worker.config import AppConfig -from worker.job_runners.dataset_info import ( - DatasetInfoJobRunner, - PreviousStepFormatError, - PreviousStepStatusError, -) - - [email protected](autouse=True) -def prepare_and_clean_mongo(app_config: AppConfig) -> None: - # prepare the database before each test, and clean it afterwards - pass - - -GetJobRunner = Callable[[str, AppConfig, bool], DatasetInfoJobRunner] - - [email protected] -def get_job_runner( - cache_mongo_resource: CacheMongoResource, - queue_mongo_resource: QueueMongoResource, -) -> GetJobRunner: - def _get_job_runner( - dataset: str, - app_config: AppConfig, - force: bool = False, - ) -> DatasetInfoJobRunner: - return DatasetInfoJobRunner( - job_info={ - "type": DatasetInfoJobRunner.get_job_type(), - "dataset": dataset, - "config": None, - "split": None, - "job_id": "job_id", - "force": force, - "priority": Priority.NORMAL, - }, - common_config=app_config.common, - worker_config=app_config.worker, - processing_step=ProcessingStep( - name=DatasetInfoJobRunner.get_job_type(), - input_type="dataset", - requires=None, - required_by_dataset_viewer=False, - parent=None, - ancestors=[], - children=[], - job_runner_version=DatasetInfoJobRunner.get_job_runner_version(), - ), - ) - - return _get_job_runner - - [email protected]( - "dataset,upstream_status,upstream_content,expected_error_code,expected_content,should_raise", - [ - ( - "ok", - HTTPStatus.OK, - {"parquet_files": [{"key": "value"}], "dataset_info": {"key": "value"}}, - None, - {"dataset_info": {"key": "value"}}, - False, - ), - ("status_error", HTTPStatus.NOT_FOUND, {"error": "error"}, PreviousStepStatusError.__name__, None, True), - ( - "format_error", - HTTPStatus.OK, - {"not_dataset_info": "wrong_format"}, - PreviousStepFormatError.__name__, - None, - True, - ), - ], -) -def test_compute( - app_config: AppConfig, - get_job_runner: GetJobRunner, - dataset: str, - upstream_status: HTTPStatus, - upstream_content: Any, - expected_error_code: str, - expected_content: Any, - should_raise: bool, -) -> None: - upsert_response( - kind="/parquet-and-dataset-info", dataset=dataset, content=upstream_content, http_status=upstream_status - ) - job_runner = get_job_runner(dataset, app_config, False) - if should_raise: - with pytest.raises(Exception) as e: - job_runner.compute() - assert e.type.__name__ == expected_error_code - else: - assert job_runner.compute().content == expected_content - - -def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset = "doesnotexist" - job_runner = get_job_runner(dataset, app_config, False) - with pytest.raises(DatasetNotFoundError): - job_runner.compute() diff --git a/services/worker/tests/job_runners/test_dataset_parquet.py b/services/worker/tests/job_runners/test_dataset_parquet.py index 9d77b955..37cb0006 100644 --- a/services/worker/tests/job_runners/test_dataset_parquet.py +++ b/services/worker/tests/job_runners/test_dataset_parquet.py @@ -5 +5 @@ from http import HTTPStatus -from typing import Any, Callable, Mapping, Optional, TypedDict +from typing import Any, Callable @@ -26,0 +27,2 @@ from worker.job_runners.parquet_and_dataset_info import ( +from .utils import UpstreamResponse + @@ -37,8 +38,0 @@ GetJobRunner = Callable[[str, AppConfig, bool], DatasetParquetJobRunner] -class UpstreamResponse(TypedDict): - kind: str - dataset: str - config: Optional[str] - http_status: HTTPStatus - content: Mapping[str, Any] - - diff --git a/services/worker/tests/job_runners/test_dataset_size.py b/services/worker/tests/job_runners/test_dataset_size.py index 8b4bc1a1..3413bd0d 100644 --- a/services/worker/tests/job_runners/test_dataset_size.py +++ b/services/worker/tests/job_runners/test_dataset_size.py @@ -5 +5 @@ from http import HTTPStatus -from typing import Any, Callable, Mapping, Optional, TypedDict +from typing import Any, Callable @@ -20,0 +21,2 @@ from worker.job_runners.dataset_size import ( +from .utils import UpstreamResponse + @@ -31,8 +32,0 @@ GetJobRunner = Callable[[str, AppConfig, bool], DatasetSizeJobRunner] -class UpstreamResponse(TypedDict): - kind: str - dataset: str - config: Optional[str] - http_status: HTTPStatus - content: Mapping[str, Any] - - diff --git a/services/worker/tests/job_runners/utils.py b/services/worker/tests/job_runners/utils.py new file mode 100644 index 00000000..30e58fdf --- /dev/null +++ b/services/worker/tests/job_runners/utils.py @@ -0,0 +1,13 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus +from typing import Any, Mapping, Optional, TypedDict + + +class UpstreamResponse(TypedDict): + kind: str + dataset: str + config: Optional[str] + http_status: HTTPStatus + content: Mapping[str, Any] diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py index f9044e6f..3e07c047 100644 --- a/services/worker/tests/test_job_runner_factory.py +++ b/services/worker/tests/test_job_runner_factory.py @@ -30 +30,2 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph: - ("/dataset-info", "DatasetInfoJobRunner"), + ("config-info", "ConfigInfoJobRunner"), + ("dataset-info", "DatasetInfoJobRunner"),
28d339e40889d091fd5c34796fd0eee37e55ecbf
Quentin Lhoest
2023-03-23T13:16:30
Kill long jobs (#977)
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index 11b01b55..fc3cd97c 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -12,0 +13,2 @@ +- name: WORKER_MAX_JOB_DURATION_SECONDS + value: {{ .Values.worker.maxJobDurationSeconds | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index 8d50e5a7..4cfdcd32 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -123,0 +124,2 @@ worker: + # the maximum duration of a job before it gets stopped for exceeded the maximum duration + maxJobDurationSeconds: 1200 diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index b50bd11b..492a80d5 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -683,0 +684,13 @@ class Queue: + def kill_long_job(self, long_job: JobInfo) -> int: + """Kill the long job in the queue, setting its status to ERROR. + It does nothing if the input job has already been updated and + is not in the STARTED status anymore. + + Returns: number of killed long jobs. + """ + long_job_id = long_job["job_id"] + logging.info("Killing a long job. Job id = " + long_job_id) + return Job.objects(pk=long_job_id, status=Status.STARTED).update( + status=Status.ERROR, finished_at=get_datetime() + ) + diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index 3b6a320d..d388b37a 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -18,0 +19 @@ WORKER_HEARTBEAT_INTERVAL_SECONDS = 60 +WORKER_KILL_LONG_JOB_INTERVAL_SECONDS = 60 @@ -20,0 +22 @@ WORKER_MAX_DISK_USAGE_PCT = 90 +WORKER_MAX_JOB_DURATION_SECONDS = 20 * 60 @@ -44,0 +47,2 @@ class WorkerConfig: + max_job_duration_seconds: int = WORKER_MAX_JOB_DURATION_SECONDS + kill_long_job_interval_seconds: int = WORKER_KILL_LONG_JOB_INTERVAL_SECONDS @@ -67,0 +72,6 @@ class WorkerConfig: + max_job_duration_seconds=env.int( + name="MAX_JOB_DURATION_SECONDS", default=WORKER_MAX_JOB_DURATION_SECONDS + ), + kill_long_job_interval_seconds=env.int( + name="KILL_LONG_JOB_INTERVAL_SECONDS", default=WORKER_KILL_LONG_JOB_INTERVAL_SECONDS + ), diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py index bba453b6..aea8a152 100644 --- a/services/worker/src/worker/executor.py +++ b/services/worker/src/worker/executor.py @@ -4 +3,0 @@ import asyncio -import json @@ -7,0 +7 @@ import sys +from datetime import datetime, timedelta @@ -9,0 +10 @@ from typing import Any, Callable, Optional +import orjson @@ -11 +12 @@ from filelock import FileLock -from libcommon.queue import Queue +from libcommon.queue import Queue, get_datetime @@ -57,0 +59 @@ class WorkerExecutor: + exceptions = [] @@ -60,0 +63,10 @@ class WorkerExecutor: + def custom_exception_handler(loop: asyncio.AbstractEventLoop, context: dict[str, Any]) -> None: + nonlocal exceptions + # first, handle with default handler + loop.default_exception_handler(context) + + exception = context.get("exception") + if exception: + exceptions.append(repr(exception)) + loop.stop() + @@ -61,0 +74 @@ class WorkerExecutor: + loop.set_exception_handler(custom_exception_handler) @@ -64,0 +78,7 @@ class WorkerExecutor: + loop.create_task( + every( + self.kill_long_job, + worker_loop_executor=worker_loop_executor, + seconds=self.app_config.worker.kill_long_job_interval_seconds, + ) + ) @@ -67,0 +88,2 @@ class WorkerExecutor: + if exceptions: + raise RuntimeError(f"Some async tasks failed: {exceptions}") @@ -69 +91 @@ class WorkerExecutor: - def get_state(self) -> WorkerState: + def get_state(self) -> Optional[WorkerState]: @@ -72 +94 @@ class WorkerExecutor: - return WorkerState(current_job_info=None) + return None @@ -75,4 +97,7 @@ class WorkerExecutor: - with open(worker_state_file_path, "r") as worker_state_f: - worker_state = json.load(worker_state_f) - return WorkerState(current_job_info=worker_state.get("current_job_info")) - except json.JSONDecodeError as err: + with open(worker_state_file_path, "rb") as worker_state_f: + worker_state = orjson.loads(worker_state_f.read()) + return WorkerState( + current_job_info=worker_state.get("current_job_info"), + last_updated=datetime.fromisoformat(worker_state["last_updated"]), + ) + except (orjson.JSONDecodeError, KeyError) as err: @@ -83 +108 @@ class WorkerExecutor: - if worker_state["current_job_info"]: + if worker_state and worker_state["current_job_info"]: @@ -94,0 +120,19 @@ class WorkerExecutor: + def kill_long_job(self, worker_loop_executor: OutputExecutor) -> None: + worker_state = self.get_state() + if worker_state and worker_state["current_job_info"]: + long_job = worker_state["current_job_info"] + last_updated = worker_state["last_updated"] + if last_updated + timedelta(seconds=self.app_config.worker.max_job_duration_seconds) <= get_datetime(): + _duration_seconds = int((get_datetime() - last_updated).total_seconds()) + logging.warning( + f"Job {long_job} exceeded maximum duration of" + f" {self.app_config.worker.max_job_duration_seconds} seconds ({_duration_seconds} seconds)." + ) + try: + worker_loop_executor.stop() # raises an error if the worker returned exit code 1 + finally: + Queue().kill_long_job(long_job) + job_runner = self.job_runner_factory.create_job_runner(long_job) + message = "Job runner was killed while running this job (job exceeded maximum duration)." + job_runner.set_exceeded_maximum_duration(message=message) + diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index d4e5e53e..714a6a6b 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -34,0 +35 @@ GeneralJobRunnerErrorCode = Literal[ + "JobRunnerExceededMaximumDurationError", @@ -167,0 +169,13 @@ class JobRunnerCrashedError(GeneralJobRunnerError): +class JobRunnerExceededMaximumDurationError(GeneralJobRunnerError): + """Raised when the job runner was killed because the job exceeded the maximum duration.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__( + message=message, + status_code=HTTPStatus.NOT_IMPLEMENTED, + code="JobRunnerExceededMaximumDurationError", + cause=cause, + disclose_cause=False, + ) + + @@ -496,0 +511,19 @@ class JobRunner(ABC): + + def set_exceeded_maximum_duration(self, message: str, cause: Optional[BaseException] = None) -> None: + error = JobRunnerExceededMaximumDurationError(message=message, cause=cause) + upsert_response( + kind=self.processing_step.cache_kind, + dataset=self.dataset, + config=self.config, + split=self.split, + content=dict(error.as_response()), + http_status=error.status_code, + error_code=error.code, + details=dict(error.as_response_with_cause()), + job_runner_version=self.get_job_runner_version(), + dataset_git_revision=self.get_dataset_git_revision(), + ) + logging.debug( + f"response for dataset={self.dataset} config={self.config} split={self.split} had an error (exceeded" + " maximum duration), cache updated" + ) diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py index 86ab375f..e961fdde 100644 --- a/services/worker/src/worker/loop.py +++ b/services/worker/src/worker/loop.py @@ -4 +3,0 @@ -import json @@ -8,0 +8 @@ from dataclasses import dataclass, field +from datetime import datetime @@ -10,0 +11 @@ from typing import Optional, TypedDict +import orjson @@ -12 +13 @@ from filelock import FileLock -from libcommon.queue import EmptyQueueError, JobInfo, Queue +from libcommon.queue import EmptyQueueError, JobInfo, Queue, get_datetime @@ -24,0 +26 @@ class WorkerState(TypedDict): + last_updated: datetime @@ -147 +149 @@ class Loop: - worker_state: WorkerState = {"current_job_info": current_job_info} + worker_state: WorkerState = {"current_job_info": current_job_info, "last_updated": get_datetime()} @@ -149,2 +151,2 @@ class Loop: - with open(self.state_file_path, "w") as worker_state_f: - json.dump(worker_state, worker_state_f) + with open(self.state_file_path, "wb") as worker_state_f: + worker_state_f.write(orjson.dumps(worker_state)) diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 47163ea9..38156f40 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -68,0 +69 @@ def set_env_vars( + mp.setenv("WORKER_KILL_LONG_JOBS_INTERVAL_SECONDS", "1") diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py index a6c06a6c..f09eadc9 100644 --- a/services/worker/tests/test_executor.py +++ b/services/worker/tests/test_executor.py @@ -1 +0,0 @@ -import json @@ -10,0 +10 @@ from unittest.mock import patch +import orjson @@ -15 +15,9 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.queue import DoesNotExist, Job, JobInfo, Priority, Status, get_datetime +from libcommon.queue import ( + DoesNotExist, + Job, + JobInfo, + Priority, + Queue, + Status, + get_datetime, +) @@ -28 +36 @@ from worker.resources import LibrariesResource -_TIME = int(time.time() * 10e3) +_TIME = int(os.environ.get("WORKER_TEST_TIME", int(time.time() * 10e3))) @@ -47,2 +55,2 @@ def write_worker_state(worker_state: WorkerState, worker_state_file_path: str) - - with open(worker_state_file_path, "w") as worker_state_f: - json.dump(worker_state, worker_state_f) + with open(worker_state_file_path, "wb") as worker_state_f: + worker_state_f.write(orjson.dumps(worker_state)) @@ -58 +66 @@ def start_worker_loop() -> None: - worker_state = WorkerState(current_job_info=current_job_info) + worker_state = WorkerState(current_job_info=current_job_info, last_updated=get_datetime()) @@ -74,0 +83,19 @@ def start_worker_loop_that_times_out() -> None: +def start_worker_loop_with_long_job() -> None: + app_config = AppConfig.from_env() + if not app_config.worker.state_file_path: + raise ValueError("Failed to get worker state because 'state_file_path' is missing.") + if "--print-worker-state-path" in sys.argv: + print(app_config.worker.state_file_path, flush=True) + current_job_info = get_job_info("long") + with QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url): + current_job = Job.objects(pk=current_job_info["job_id"]).get() + assert current_job.started_at is not None + worker_state = WorkerState( + current_job_info=current_job_info, last_updated=pytz.UTC.localize(current_job.started_at) + ) + if current_job.status == Status.STARTED: + write_worker_state(worker_state, app_config.worker.state_file_path) + time.sleep(20) + Queue().finish_job(current_job_info["job_id"], finished_status=Status.SUCCESS) + + @@ -78 +105 @@ def set_worker_state(worker_state_file_path: str) -> Iterator[WorkerState]: - worker_state = WorkerState(current_job_info=job_info) + worker_state = WorkerState(current_job_info=job_info, last_updated=get_datetime()) @@ -198 +225 @@ def test_executor_get_empty_state( - assert executor.get_state() == WorkerState(current_job_info=None) + assert executor.get_state() is None @@ -220 +246,0 @@ def test_executor_kill_zombies( - job_runner_factory: JobRunnerFactory, @@ -250,0 +277,2 @@ def test_executor_start( + tmp_dataset_repo_factory: Callable[[str], str], + cache_mongo_resource: CacheMongoResource, @@ -253,0 +282,3 @@ def test_executor_start( + zombie = set_zombie_job_in_queue + tmp_dataset_repo_factory(zombie.dataset) + # tmp_dataset_repo_factory(zombie.dataset) @@ -256 +287,3 @@ def test_executor_start( - with patch("worker.executor.START_WORKER_LOOP_PATH", __file__): + with patch("worker.executor.START_WORKER_LOOP_PATH", __file__), patch.dict( + os.environ, {"WORKER_TEST_TIME": str(_TIME)} + ): @@ -279 +312,3 @@ def test_executor_raises_on_bad_worker( - with patch("worker.executor.START_WORKER_LOOP_PATH", __file__): + with patch("worker.executor.START_WORKER_LOOP_PATH", __file__), patch.dict( + os.environ, {"WORKER_TEST_TIME": str(_TIME)} + ): @@ -283,0 +319,48 @@ def test_executor_raises_on_bad_worker( +def test_executor_stops_on_long_job( + executor: WorkerExecutor, + queue_mongo_resource: QueueMongoResource, + cache_mongo_resource: CacheMongoResource, + tmp_dataset_repo_factory: Callable[[str], str], + set_long_running_job_in_queue: Job, + set_just_started_job_in_queue: Job, +) -> None: + if not queue_mongo_resource.is_available(): + raise RuntimeError("Mongo resource is not available") + long_job = set_long_running_job_in_queue + normal_job = set_just_started_job_in_queue + tmp_dataset_repo_factory(long_job.dataset) + try: + with patch.dict(os.environ, {"WORKER_LOOP_TYPE": "start_worker_loop_with_long_job"}): + with patch.object(executor, "max_seconds_without_heartbeat_for_zombies", -1): # don't kill normal_job + with patch("worker.executor.START_WORKER_LOOP_PATH", __file__), patch.dict( + os.environ, {"WORKER_TEST_TIME": str(_TIME)} + ): + executor.start() + + assert long_job is not None + assert str(long_job.pk) == get_job_info("long")["job_id"] + + long_job.reload() + assert long_job.status == Status.ERROR, "must be an error because too long" + + responses = CachedResponse.objects() + assert len(responses) == 1 + response = responses[0] + expected_error = { + "error": "Job runner was killed while running this job (job exceeded maximum duration).", + } + assert response.http_status == HTTPStatus.NOT_IMPLEMENTED + assert response.error_code == "JobRunnerExceededMaximumDurationError" + assert response.dataset == long_job.dataset + assert response.config == long_job.config + assert response.split == long_job.split + assert response.content == expected_error + assert response.details == expected_error + + normal_job.reload() + assert normal_job.status == Status.STARTED, "must stay untouched" + + finally: + CachedResponse.objects().delete() + + @@ -289,0 +373,2 @@ if __name__ == "__main__": + elif worker_loop_type == "start_worker_loop_with_long_job": + start_worker_loop_with_long_job() diff --git a/tools/docker-compose-base.yml b/tools/docker-compose-base.yml index 7da735f2..9b7b6284 100644 --- a/tools/docker-compose-base.yml +++ b/tools/docker-compose-base.yml @@ -20,0 +21 @@ services: + WORKER_MAX_JOB_DURATION_SECONDS: ${WORKER_MAX_JOB_DURATION_SECONDS-1200} diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index e3a60b41..4b7cffce 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -20,0 +21 @@ services: + WORKER_MAX_JOB_DURATION_SECONDS: ${WORKER_MAX_JOB_DURATION_SECONDS-1200}
dc51be62a0313a690f65b644a267c565621df0c0
Andrea Francis Soria Jimenez
2023-03-23T12:00:53
/first-rows to first-rows-from-streaming (#967)
diff --git a/e2e/tests/test_31_admin_metrics.py b/e2e/tests/test_31_admin_metrics.py index 84f548b8..c3ef9de7 100644 --- a/e2e/tests/test_31_admin_metrics.py +++ b/e2e/tests/test_31_admin_metrics.py @@ -33,2 +33,2 @@ def test_metrics() -> None: - for queue in ["/splits", "/first-rows", "dataset-parquet"]: - # eg. 'queue_jobs_total{pid="10",queue="/first-rows",status="started"}' + for queue in ["/splits", "split-first-rows-from-streaming", "dataset-parquet"]: + # eg. 'queue_jobs_total{pid="10",queue="split-first-rows-from-streaming",status="started"}' @@ -40 +40 @@ def test_metrics() -> None: - for cache_kind in ["/splits", "/first-rows", "dataset-parquet"]: + for cache_kind in ["/splits", "split-first-rows-from-streaming", "dataset-parquet"]: diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py index e5724f67..065ea00a 100644 --- a/front/admin_ui/app.py +++ b/front/admin_ui/app.py @@ -67 +67 @@ with gr.Blocks() as demo: - refresh_type = gr.Textbox(label="Processing step type", placeholder="first-rows") + refresh_type = gr.Textbox(label="Processing step type", placeholder="split-first-rows-from-streaming") diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index ec2d79f7..4e1c820f 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -31,0 +32,6 @@ from mongodb_migration.migrations._20230313164200_cache_remove_worker_version im +from mongodb_migration.migrations._20230320163700_cache_first_rows_from_streaming import ( + MigrationCacheUpdateFirstRows, +) +from mongodb_migration.migrations._20230320165700_queue_first_rows_from_streaming import ( + MigrationQueueUpdateFirstRows, +) @@ -69,0 +76,10 @@ class MigrationsCollector: + MigrationCacheUpdateFirstRows( + version="20230320163700", + description="update 'kind' field in cache from /first-rows to split-first-rows-from-streaming", + ), + MigrationQueueUpdateFirstRows( + version="20230320165700", + description=( + "update 'type' and 'unicity_id' fields in job from /first-rows to split-first-rows-from-streaming" + ), + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320163700_cache_first_rows_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320163700_cache_first_rows_from_streaming.py new file mode 100644 index 00000000..37a9ecb5 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320163700_cache_first_rows_from_streaming.py @@ -0,0 +1,38 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from libcommon.simple_cache import CachedResponse +from mongoengine.connection import get_db + +from mongodb_migration.check import check_documents +from mongodb_migration.migration import Migration + +first_rows = "/first-rows" +split_first_rows_from_streaming = "split-first-rows-from-streaming" +db_name = "cache" + + +# connection already occurred in the main.py (caveat: we use globals) +class MigrationCacheUpdateFirstRows(Migration): + def up(self) -> None: + logging.info(f"Rename cache_kind field from {first_rows} to {split_first_rows_from_streaming}") + db = get_db(db_name) + + # update existing documents with the old kind + db["cachedResponsesBlue"].update_many( + {"kind": first_rows}, {"$set": {"kind": split_first_rows_from_streaming}} + ) + + def down(self) -> None: + logging.info(f"Rollback cache_kind field from {split_first_rows_from_streaming} to {first_rows}") + db = get_db(db_name) + db["cachedResponsesBlue"].update_many( + {"kind": split_first_rows_from_streaming}, {"$set": {"kind": first_rows}} + ) + + def validate(self) -> None: + logging.info("Validate modified documents") + + check_documents(DocCls=CachedResponse, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320165700_queue_first_rows_from_streaming.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320165700_queue_first_rows_from_streaming.py new file mode 100644 index 00000000..5eeaf1f7 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230320165700_queue_first_rows_from_streaming.py @@ -0,0 +1,74 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from libcommon.queue import Job +from mongoengine.connection import get_db + +from mongodb_migration.check import check_documents +from mongodb_migration.migration import Migration + +first_rows = "/first-rows" +split_first_rows_from_streaming = "split-first-rows-from-streaming" +db_name = "queue" + + +# connection already occurred in the main.py (caveat: we use globals) +class MigrationQueueUpdateFirstRows(Migration): + def up(self) -> None: + logging.info( + f"Rename unicity_id field from Job[{first_rows}][<dataset>][<config>][split] to" + f" Job[{split_first_rows_from_streaming}][<dataset>][<config>][split] and change type from {first_rows} to" + f" {split_first_rows_from_streaming}" + ) + + db = get_db("queue") + db["jobsBlue"].update_many( + {"type": first_rows}, + [ + { + "$set": { + "unicity_id": { + "$replaceOne": { + "input": "$unicity_id", + "find": f"Job[{first_rows}]", + "replacement": f"Job[{split_first_rows_from_streaming}]", + } + }, + "type": split_first_rows_from_streaming, + } + }, + ], # type: ignore + ) + + def down(self) -> None: + logging.info( + f"Rename unicity_id field from Job[{split_first_rows_from_streaming}][<dataset>][<config>][split] to" + f" Job[{first_rows}][<dataset>][<config>][split] and change type from {split_first_rows_from_streaming} to" + f" {first_rows}" + ) + + db = get_db("queue") + db["jobsBlue"].update_many( + {"type": split_first_rows_from_streaming}, + [ + { + "$set": { + "unicity_id": { + "$replaceOne": { + "input": "$unicity_id", + "find": f"Job[{split_first_rows_from_streaming}]", + "replacement": f"Job[{first_rows}]", + } + }, + "type": split_first_rows_from_streaming, + } + }, + ], # type: ignore + ) + + def validate(self) -> None: + logging.info("Validate modified documents") + + check_documents(DocCls=Job, sample_size=10) diff --git a/jobs/mongodb_migration/tests/migrations/test_20230320163700_cache_first_rows_from_streaming.py b/jobs/mongodb_migration/tests/migrations/test_20230320163700_cache_first_rows_from_streaming.py new file mode 100644 index 00000000..7c3b5bca --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230320163700_cache_first_rows_from_streaming.py @@ -0,0 +1,29 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230320163700_cache_first_rows_from_streaming import ( + MigrationCacheUpdateFirstRows, +) + + +def test_cache_update_first_rows_kind(mongo_host: str) -> None: + with MongoResource(database="test_cache_update_first_rows_kind", host=mongo_host, mongoengine_alias="cache"): + db = get_db("cache") + db["cachedResponsesBlue"].insert_many([{"kind": "/first-rows", "dataset": "dataset", "http_status": 200}]) + assert db["cachedResponsesBlue"].find_one( + {"kind": "/first-rows"} + ) # Ensure there is at least one record to update + + migration = MigrationCacheUpdateFirstRows( + version="20230320163700", + description="update 'kind' field in cache from /first-rows to split-first-rows-from-streaming", + ) + migration.up() + + assert not db["cachedResponsesBlue"].find_one({"kind": "/first-rows"}) # Ensure 0 records with old kind + + assert db["cachedResponsesBlue"].find_one({"kind": "split-first-rows-from-streaming"}) + + db["cachedResponsesBlue"].drop() diff --git a/jobs/mongodb_migration/tests/migrations/test_20230320165700_queue_first_rows_from_streaming.py b/jobs/mongodb_migration/tests/migrations/test_20230320165700_queue_first_rows_from_streaming.py new file mode 100644 index 00000000..4ee385da --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230320165700_queue_first_rows_from_streaming.py @@ -0,0 +1,41 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230320165700_queue_first_rows_from_streaming import ( + MigrationQueueUpdateFirstRows, +) + + +def test_queue_update_first_rows_type_and_unicity_id(mongo_host: str) -> None: + with MongoResource( + database="test_queue_update_first_rows_type_and_unicity_id", host=mongo_host, mongoengine_alias="queue" + ): + db = get_db("queue") + db["jobsBlue"].insert_many( + [ + { + "type": "/first-rows", + "unicity_id": "Job[/first-rows][dataset][config][split]", + "dataset": "dataset", + "http_status": 200, + } + ] + ) + assert db["jobsBlue"].find_one({"type": "/first-rows"}) # Ensure there is at least one record to update + + migration = MigrationQueueUpdateFirstRows( + version="20230320165700", + description=( + "update 'type' and 'unicity_id' fields in job from /first-rows to split-first-rows-from-streaming" + ), + ) + migration.up() + + assert not db["jobsBlue"].find_one({"type": "/first-rows"}) # Ensure 0 records with old type + + result = db["jobsBlue"].find_one({"type": "split-first-rows-from-streaming"}) + assert result + assert result["unicity_id"] == "Job[split-first-rows-from-streaming][dataset][config][split]" + db["jobsBlue"].drop() diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 0e95033e..74b666ee 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -20 +19,0 @@ from libcommon.constants import ( - PROCESSING_STEP_FIRST_ROWS_VERSION, @@ -21,0 +21 @@ from libcommon.constants import ( + PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION, @@ -125 +125 @@ class ProcessingGraphConfig: - "/first-rows": { + "split-first-rows-from-streaming": { @@ -129 +129 @@ class ProcessingGraphConfig: - "job_runner_version": PROCESSING_STEP_FIRST_ROWS_VERSION, + "job_runner_version": PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 4875ef66..dc51ce04 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -17 +16,0 @@ PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION = 1 -PROCESSING_STEP_FIRST_ROWS_VERSION = 2 @@ -18,0 +18 @@ PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION = 1 +PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION = 2 diff --git a/libs/libcommon/tests/test_processing_steps.py b/libs/libcommon/tests/test_processing_steps.py index 30c0be9b..f9122608 100644 --- a/libs/libcommon/tests/test_processing_steps.py +++ b/libs/libcommon/tests/test_processing_steps.py @@ -15 +15 @@ def test_default_graph() -> None: - first_rows = graph.get_step("/first-rows") + split_first_rows_from_streaming = graph.get_step("split-first-rows-from-streaming") @@ -33 +33 @@ def test_default_graph() -> None: - assert split_names_from_streaming.children == [first_rows, dataset_split_names_from_streaming] + assert split_names_from_streaming.children == [split_first_rows_from_streaming, dataset_split_names_from_streaming] @@ -41,4 +41,4 @@ def test_default_graph() -> None: - assert first_rows is not None - assert first_rows.parent is split_names_from_streaming - assert first_rows.children == [] - assert first_rows.get_ancestors() == [config_names, split_names_from_streaming] + assert split_first_rows_from_streaming is not None + assert split_first_rows_from_streaming.parent is split_names_from_streaming + assert split_first_rows_from_streaming.children == [] + assert split_first_rows_from_streaming.get_ancestors() == [config_names, split_names_from_streaming] @@ -96 +96 @@ def test_default_graph() -> None: - assert graph.get_steps_required_by_dataset_viewer() == [splits, first_rows] + assert graph.get_steps_required_by_dataset_viewer() == [splits, split_first_rows_from_streaming] diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index eb87e1d2..b8c0cab9 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -119 +119 @@ class EndpointConfig: - "/first-rows": {"split": ["/first-rows"]}, + "/first-rows": {"split": ["split-first-rows-from-streaming"]}, diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py index 0784eda0..d990a49c 100644 --- a/services/api/tests/conftest.py +++ b/services/api/tests/conftest.py @@ -48 +48 @@ def endpoint_config(monkeypatch_session: MonkeyPatch) -> EndpointConfig: - "/first-rows": {"split": ["/first-rows"]}, + "/first-rows": {"split": ["split-first-rows-from-streaming"]}, diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index c9ad8081..bc155d6d 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -32 +31,0 @@ from worker.job_runners.dataset_size import DatasetSizeJobRunner -from worker.job_runners.first_rows import FirstRowsJobRunner @@ -33,0 +33,3 @@ from worker.job_runners.parquet_and_dataset_info import ParquetAndDatasetInfoJob +from worker.job_runners.split.first_rows_from_streaming import ( + SplitFirstRowsFromStreamingJobRunner, +) @@ -91 +93 @@ class JobRunnerFactory(BaseJobRunnerFactory): - if job_type == FirstRowsJobRunner.get_job_type(): + if job_type == SplitFirstRowsFromStreamingJobRunner.get_job_type(): @@ -93 +95 @@ class JobRunnerFactory(BaseJobRunnerFactory): - return FirstRowsJobRunner( + return SplitFirstRowsFromStreamingJobRunner( @@ -169 +171 @@ class JobRunnerFactory(BaseJobRunnerFactory): - FirstRowsJobRunner.get_job_type(), + SplitFirstRowsFromStreamingJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/split/__init__.py b/services/worker/src/worker/job_runners/split/__init__.py new file mode 100644 index 00000000..1e9d0c5a --- /dev/null +++ b/services/worker/src/worker/job_runners/split/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py similarity index 89% rename from services/worker/src/worker/job_runners/first_rows.py rename to services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index c8631024..1e4e83b8 100644 --- a/services/worker/src/worker/job_runners/first_rows.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -34,2 +34 @@ from datasets import ( -from datasets.data_files import EmptyDatasetError as _EmptyDatasetError -from libcommon.constants import PROCESSING_STEP_FIRST_ROWS_VERSION +from libcommon.constants import PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION @@ -37,0 +37 @@ from libcommon.queue import JobInfo +from libcommon.simple_cache import DoesNotExist @@ -38,0 +39 @@ from libcommon.simple_cache import SplitFullName as _SplitFullName +from libcommon.simple_cache import get_response @@ -52 +53 @@ from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner -FirstRowsJobRunnerErrorCode = Literal[ +SplitFirstRowsFromStreamingJobRunnerErrorCode = Literal[ @@ -61,0 +63,2 @@ FirstRowsJobRunnerErrorCode = Literal[ + "PreviousStepStatusError", + "PreviousStepFormatError", @@ -65 +68 @@ FirstRowsJobRunnerErrorCode = Literal[ -class FirstRowsJobRunnerError(JobRunnerError): +class SplitFirstRowsFromStreamingJobRunnerError(JobRunnerError): @@ -72 +75 @@ class FirstRowsJobRunnerError(JobRunnerError): - code: FirstRowsJobRunnerErrorCode, + code: SplitFirstRowsFromStreamingJobRunnerErrorCode, @@ -81 +84 @@ class FirstRowsJobRunnerError(JobRunnerError): -class SplitsNamesError(FirstRowsJobRunnerError): +class SplitsNamesError(SplitFirstRowsFromStreamingJobRunnerError): @@ -88 +91 @@ class SplitsNamesError(FirstRowsJobRunnerError): -class EmptyDatasetError(FirstRowsJobRunnerError): +class EmptyDatasetError(SplitFirstRowsFromStreamingJobRunnerError): @@ -95 +98 @@ class EmptyDatasetError(FirstRowsJobRunnerError): -class InfoError(FirstRowsJobRunnerError): +class InfoError(SplitFirstRowsFromStreamingJobRunnerError): @@ -102 +105 @@ class InfoError(FirstRowsJobRunnerError): -class FeaturesError(FirstRowsJobRunnerError): +class FeaturesError(SplitFirstRowsFromStreamingJobRunnerError): @@ -109 +112 @@ class FeaturesError(FirstRowsJobRunnerError): -class StreamingRowsError(FirstRowsJobRunnerError): +class StreamingRowsError(SplitFirstRowsFromStreamingJobRunnerError): @@ -116 +119 @@ class StreamingRowsError(FirstRowsJobRunnerError): -class NormalRowsError(FirstRowsJobRunnerError): +class NormalRowsError(SplitFirstRowsFromStreamingJobRunnerError): @@ -123 +126 @@ class NormalRowsError(FirstRowsJobRunnerError): -class RowsPostProcessingError(FirstRowsJobRunnerError): +class RowsPostProcessingError(SplitFirstRowsFromStreamingJobRunnerError): @@ -130 +133 @@ class RowsPostProcessingError(FirstRowsJobRunnerError): -class TooManyColumnsError(FirstRowsJobRunnerError): +class TooManyColumnsError(SplitFirstRowsFromStreamingJobRunnerError): @@ -137 +140 @@ class TooManyColumnsError(FirstRowsJobRunnerError): -class TooBigContentError(FirstRowsJobRunnerError): +class TooBigContentError(SplitFirstRowsFromStreamingJobRunnerError): @@ -143,0 +147,14 @@ class TooBigContentError(FirstRowsJobRunnerError): +class PreviousStepStatusError(SplitFirstRowsFromStreamingJobRunnerError): + """Raised when the previous step gave an error. The job should not have been created.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepStatusError", cause, False) + + +class PreviousStepFormatError(SplitFirstRowsFromStreamingJobRunnerError): + """Raised when the content of the previous step has not the expected format.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "PreviousStepFormatError", cause, False) + + @@ -187 +204 @@ class RowItem(TypedDict): -class FirstRowsResponse(TypedDict): +class SplitFirstRowsFromStreamingResponse(TypedDict): @@ -426 +443 @@ def compute_first_rows_response( -) -> FirstRowsResponse: +) -> SplitFirstRowsFromStreamingResponse: @@ -461 +478 @@ def compute_first_rows_response( - [`FirstRowsResponse`]: The list of first rows of the split. + [`SplitFirstRowsFromStreamingResponse`]: The list of first rows of the split. @@ -487 +503,0 @@ def compute_first_rows_response( - # get the list of splits @@ -489,7 +505,3 @@ def compute_first_rows_response( - split_full_names = get_dataset_split_full_names(dataset=dataset, use_auth_token=use_auth_token) - except _EmptyDatasetError as err: - raise EmptyDatasetError("The dataset is empty.", cause=err) from err - except Exception as err: - raise SplitsNamesError("Cannot get the split names for the dataset.", cause=err) from err - # ^ can raise DatasetNotFoundError or SplitsNamesError - if config not in [split_full_name["config"] for split_full_name in split_full_names]: + upstream_response = get_response(kind="/split-names-from-streaming", dataset=dataset, config=config) + splits_content = upstream_response["content"]["splits"] + except DoesNotExist: @@ -497,8 +509,9 @@ def compute_first_rows_response( - if {"dataset": dataset, "config": config, "split": split} not in [ - { - "dataset": split_full_name["dataset"], - "config": split_full_name["config"], - "split": split_full_name["split"], - } - for split_full_name in split_full_names - ]: + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.") from e + + if upstream_response["http_status"] != HTTPStatus.OK: + raise PreviousStepStatusError( + f"Previous step gave an error: {upstream_response['http_status']}. This job should not have been created." + ) + + if split not in [split_item["split"] for split_item in splits_content]: @@ -554 +567 @@ def compute_first_rows_response( - response_features_only: FirstRowsResponse = { + response_features_only: SplitFirstRowsFromStreamingResponse = { @@ -640 +653 @@ def compute_first_rows_response( -class FirstRowsJobRunner(DatasetsBasedJobRunner): +class SplitFirstRowsFromStreamingJobRunner(DatasetsBasedJobRunner): @@ -646 +659 @@ class FirstRowsJobRunner(DatasetsBasedJobRunner): - return "/first-rows" + return "split-first-rows-from-streaming" @@ -650 +663 @@ class FirstRowsJobRunner(DatasetsBasedJobRunner): - return PROCESSING_STEP_FIRST_ROWS_VERSION + return PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 1760fae2..14eb69a8 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -518,0 +519,14 @@ def hub_datasets( + "does_not_exist_config": { + "name": "does_not_exist_config", + "config_names_response": None, + "splits_response": None, + "first_rows_response": None, + "parquet_and_dataset_info_response": None, + }, + "does_not_exist_split": { + "name": "does_not_exist_split", + "config_names_response": None, + "splits_response": None, + "first_rows_response": None, + "parquet_and_dataset_info_response": None, + }, diff --git a/services/worker/tests/job_runners/split/__init__.py b/services/worker/tests/job_runners/split/__init__.py new file mode 100644 index 00000000..1e9d0c5a --- /dev/null +++ b/services/worker/tests/job_runners/split/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/split/test_first_rows.py similarity index 75% rename from services/worker/tests/job_runners/test_first_rows.py rename to services/worker/tests/job_runners/split/test_first_rows.py index 15ab3186..9619c227 100644 --- a/services/worker/tests/job_runners/test_first_rows.py +++ b/services/worker/tests/job_runners/split/test_first_rows.py @@ -14 +14 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import DoesNotExist, get_response +from libcommon.simple_cache import DoesNotExist, get_response, upsert_response @@ -18 +18,4 @@ from worker.config import AppConfig, FirstRowsConfig -from worker.job_runners.first_rows import FirstRowsJobRunner, get_json_size +from worker.job_runners.split.first_rows_from_streaming import ( + SplitFirstRowsFromStreamingJobRunner, + get_json_size, +) @@ -21 +24 @@ from worker.resources import LibrariesResource -from ..fixtures.hub import HubDatasets, get_default_config_split +from ...fixtures.hub import HubDatasets, get_default_config_split @@ -23 +26 @@ from ..fixtures.hub import HubDatasets, get_default_config_split -GetJobRunner = Callable[[str, str, str, AppConfig, FirstRowsConfig, bool], FirstRowsJobRunner] +GetJobRunner = Callable[[str, str, str, AppConfig, FirstRowsConfig, bool], SplitFirstRowsFromStreamingJobRunner] @@ -40,2 +43,2 @@ def get_job_runner( - ) -> FirstRowsJobRunner: - return FirstRowsJobRunner( + ) -> SplitFirstRowsFromStreamingJobRunner: + return SplitFirstRowsFromStreamingJobRunner( @@ -43 +46 @@ def get_job_runner( - "type": FirstRowsJobRunner.get_job_type(), + "type": SplitFirstRowsFromStreamingJobRunner.get_job_type(), @@ -53 +56 @@ def get_job_runner( - name=FirstRowsJobRunner.get_job_type(), + name=SplitFirstRowsFromStreamingJobRunner.get_job_type(), @@ -60 +63 @@ def get_job_runner( - job_runner_version=FirstRowsJobRunner.get_job_runner_version(), + job_runner_version=SplitFirstRowsFromStreamingJobRunner.get_job_runner_version(), @@ -76,0 +80,7 @@ def test_should_skip_job( + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={"splits": [{"dataset": dataset, "config": config, "split": split}]}, + http_status=HTTPStatus.OK, + ) @@ -87,0 +98,7 @@ def test_compute( + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={"splits": [{"dataset": dataset, "config": config, "split": split}]}, + http_status=HTTPStatus.OK, + ) @@ -124 +141 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first - ("empty", False, "EmptyDatasetError", "EmptyDatasetError"), + ("does_not_exist_config", False, "ConfigNotFoundError", None), @@ -127,3 +144,3 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner, first - ("does_not_exist", False, "SplitsNamesError", "FileNotFoundError"), - ("gated", False, "SplitsNamesError", "FileNotFoundError"), - ("private", False, "SplitsNamesError", "FileNotFoundError"), + ("does_not_exist_split", False, "SplitNotFoundError", None), + ("gated", False, "InfoError", "FileNotFoundError"), + ("private", False, "InfoError", "FileNotFoundError"), @@ -159,0 +177,7 @@ def test_number_rows( + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={"splits": [{"dataset": dataset, "config": config, "split": split}]}, + http_status=HTTPStatus.OK, + ) @@ -162,0 +187,17 @@ def test_number_rows( + elif error_code == "SplitNotFoundError": + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={"splits": [{"dataset": dataset, "config": config, "split": "other_split"}]}, + http_status=HTTPStatus.OK, + ) + elif error_code in ("InfoError", "SplitsNamesError"): + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={"splits": [{"dataset": dataset, "config": config, "split": split}]}, + http_status=HTTPStatus.OK, + ) + @@ -216,0 +258,8 @@ def test_truncation( + upsert_response( + kind="/split-names-from-streaming", + dataset=dataset, + config=config, + content={"splits": [{"dataset": dataset, "config": config, "split": split}]}, + http_status=HTTPStatus.OK, + ) + diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py index cadd352d..f9044e6f 100644 --- a/services/worker/tests/test_job_runner_factory.py +++ b/services/worker/tests/test_job_runner_factory.py @@ -26 +26 @@ def processing_graph(app_config: AppConfig) -> ProcessingGraph: - ("/first-rows", "FirstRowsJobRunner"), + ("split-first-rows-from-streaming", "SplitFirstRowsFromStreamingJobRunner"),
9bcae7ac67f6893f7e5008fa0ef7d17b9a42a48d
Andrea Francis Soria Jimenez
2023-03-21T16:43:17
Disable cache refresh job on values and dev yaml files (#974)
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 7d9b7950..13683ff7 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -30,5 +30,6 @@ images: - cacheRefresh: - registry: huggingface - useGlobalRegistry: false - repository: datasets-server-jobs-cache_refresh - tag: sha-fb3399a + # cacheRefresh: + # registry: huggingface + # useGlobalRegistry: false + # repository: datasets-server-jobs-cache_refresh + # tag: sha-fb3399a + # --- temporary disabling this job --- diff --git a/chart/templates/jobs/cache-refresh/job.yaml b/chart/templates/jobs/cache-refresh/job.yaml deleted file mode 100644 index 9a24f3ec..00000000 --- a/chart/templates/jobs/cache-refresh/job.yaml +++ /dev/null @@ -1,24 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -{{- if .Values.images.jobs.cacheRefresh }} -apiVersion: batch/v1 -kind: Job -metadata: - labels: {{ include "labels.cacheRefresh" . | nindent 4 }} - name: "{{ include "name" . }}-job-cache-refresh" - namespace: {{ .Release.Namespace }} -spec: - ttlSecondsAfterFinished: 300 - template: - metadata: - labels: {{ include "labels.cacheRefresh" . | nindent 8 }} - spec: - restartPolicy: Never - {{- include "image.imagePullSecrets" . | nindent 6 }} - containers: {{ include "containerCacheRefresh" . | nindent 8 }} - nodeSelector: {{ toYaml .Values.cacheRefresh.nodeSelector | nindent 8 }} - tolerations: {{ toYaml .Values.cacheRefresh.tolerations | nindent 8 }} - securityContext: {{ include "securityContext" . | nindent 8 }} - backoffLimit: 0 -{{- end}} diff --git a/chart/values.yaml b/chart/values.yaml index 94a3abb6..8d50e5a7 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -35,5 +35,6 @@ images: - cacheRefresh: - registry: huggingface - useGlobalRegistry: false - repository: datasets-server-jobs-cache_refresh - tag: sha-fb3399a + # cacheRefresh: + # registry: huggingface + # useGlobalRegistry: false + # repository: datasets-server-jobs-cache_refresh + # tag: sha-fb3399a + # --- temporary disabling this job ---
5c0d8534eb3a93289603c80a9f79bcf272df4f47
Andrea Francis Soria Jimenez
2023-03-21T15:05:43
Reduce prod resources and remove cache refresh job (temporal change) (#972)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 3c99fb35..2dea2353 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -29,5 +29,6 @@ images: - cacheRefresh: - registry: huggingface - useGlobalRegistry: false - repository: datasets-server-jobs-cache_refresh - tag: sha-fb3399a + # cacheRefresh: + # registry: huggingface + # useGlobalRegistry: false + # repository: datasets-server-jobs-cache_refresh + # tag: sha-fb3399a + # --- temporary disabling this job --- @@ -203 +204 @@ workers: - replicas: 100 + replicas: 40
5c0e59f79d4a1f2f09b0b2ddef02b0b30a85f226
Andrea Francis Soria Jimenez
2023-03-21T13:16:14
Removing hook annotations (#971)
diff --git a/chart/templates/jobs/cache-refresh/job.yaml b/chart/templates/jobs/cache-refresh/job.yaml index 29b24427..9a24f3ec 100644 --- a/chart/templates/jobs/cache-refresh/job.yaml +++ b/chart/templates/jobs/cache-refresh/job.yaml @@ -11,4 +10,0 @@ metadata: - annotations: - "helm.sh/hook": pre-upgrade - "helm.sh/hook-weight": "-1" - "helm.sh/hook-delete-policy": before-hook-creation
19d20a266da5423e671708d72cbf534b367450f4
Andrea Francis Soria Jimenez
2023-03-21T12:43:32
Increase resources for 'all' worker (#970)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index edae846e..3c99fb35 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -203 +203 @@ workers: - replicas: 22 + replicas: 100
b1973fee9ee0c8d3654fa8db5c2ed11bcb34c642
Andrea Francis Soria Jimenez
2023-03-20T19:45:11
Refresh when cache is outdated because of job_runner_version (#912)
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml index d2c0f629..42f13df5 100644 --- a/.github/workflows/cd.yml +++ b/.github/workflows/cd.yml @@ -25,0 +26,2 @@ jobs: + - directory: jobs + project: cache_refresh @@ -101,0 +104,2 @@ jobs: + cacheRefresh: + tag: sha-${{ steps.vars.outputs.sha_short }} diff --git a/.github/workflows/j-cache-refresh.yml b/.github/workflows/j-cache-refresh.yml new file mode 100644 index 00000000..35ef8434 --- /dev/null +++ b/.github/workflows/j-cache-refresh.yml @@ -0,0 +1,33 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +name: jobs/cache_refresh +on: + workflow_dispatch: + push: + branches: + - main + paths: + - 'jobs/cache_refresh/**' + - 'libs/libcommon/**' + - '.github/workflows/j-cache-refresh.yml' + - '.github/workflows/_quality-python.yml' + - '.github/workflows/_unit-tests-python.yml' + - 'tools/docker-compose-mongo.yml' + pull_request: + paths: + - 'jobs/cache_refresh/**' + - 'libs/libcommon/**' + - '.github/workflows/j-cache-refresh.yml' + - '.github/workflows/_quality-python.yml' + - '.github/workflows/_unit-tests-python.yml' + - 'tools/docker-compose-mongo.yml' +jobs: + quality: + uses: ./.github/workflows/_quality-python.yml + with: + working-directory: jobs/cache_refresh + unit-tests: + uses: ./.github/workflows/_unit-tests-python.yml + with: + working-directory: jobs/cache_refresh diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace index ad6493aa..604b61a1 100644 --- a/.vscode/monorepo.code-workspace +++ b/.vscode/monorepo.code-workspace @@ -14,0 +15,4 @@ + { + "name": "jobs/cache_refresh", + "path": "../jobs/cache_refresh" + }, diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 06d69c0e..84411992 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.5.0 +version: 1.6.0 diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 77fedca1..7d9b7950 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -29,0 +30,5 @@ images: + cacheRefresh: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-jobs-cache_refresh + tag: sha-fb3399a diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 54129648..edae846e 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -28,0 +29,5 @@ images: + cacheRefresh: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-jobs-cache_refresh + tag: sha-fb3399a diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index 56f838af..4a601cf6 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -36,0 +37,4 @@ Docker image management +{{- define "jobs.cacheRefresh.image" -}} +{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.jobs.cacheRefresh "global" .Values.global.huggingface) }} +{{- end -}} + @@ -71,0 +76,5 @@ app.kubernetes.io/component: "{{ include "name" . }}-mongodb-migration" +{{- define "labels.cacheRefresh" -}} +{{ include "hf.labels.commons" . }} +app.kubernetes.io/component: "{{ include "name" . }}-cache-refresh" +{{- end -}} + diff --git a/chart/templates/jobs/cache-refresh/_container.tpl b/chart/templates/jobs/cache-refresh/_container.tpl new file mode 100644 index 00000000..758baf57 --- /dev/null +++ b/chart/templates/jobs/cache-refresh/_container.tpl @@ -0,0 +1,15 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- define "containerCacheRefresh" -}} +- name: "{{ include "name" . }}-cache-refresh" + image: {{ include "jobs.cacheRefresh.image" . }} + imagePullPolicy: {{ .Values.images.pullPolicy }} + env: + {{ include "envCache" . | nindent 2 }} + {{ include "envQueue" . | nindent 2 }} + {{ include "envCommon" . | nindent 2 }} + securityContext: + allowPrivilegeEscalation: false + resources: {{ toYaml .Values.cacheRefresh.resources | nindent 4 }} +{{- end -}} diff --git a/chart/templates/jobs/cache-refresh/job.yaml b/chart/templates/jobs/cache-refresh/job.yaml new file mode 100644 index 00000000..29b24427 --- /dev/null +++ b/chart/templates/jobs/cache-refresh/job.yaml @@ -0,0 +1,28 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +{{- if .Values.images.jobs.cacheRefresh }} +apiVersion: batch/v1 +kind: Job +metadata: + labels: {{ include "labels.cacheRefresh" . | nindent 4 }} + name: "{{ include "name" . }}-job-cache-refresh" + namespace: {{ .Release.Namespace }} + annotations: + "helm.sh/hook": pre-upgrade + "helm.sh/hook-weight": "-1" + "helm.sh/hook-delete-policy": before-hook-creation +spec: + ttlSecondsAfterFinished: 300 + template: + metadata: + labels: {{ include "labels.cacheRefresh" . | nindent 8 }} + spec: + restartPolicy: Never + {{- include "image.imagePullSecrets" . | nindent 6 }} + containers: {{ include "containerCacheRefresh" . | nindent 8 }} + nodeSelector: {{ toYaml .Values.cacheRefresh.nodeSelector | nindent 8 }} + tolerations: {{ toYaml .Values.cacheRefresh.tolerations | nindent 8 }} + securityContext: {{ include "securityContext" . | nindent 8 }} + backoffLimit: 0 +{{- end}} diff --git a/chart/values.yaml b/chart/values.yaml index 2bfcca2a..94a3abb6 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -34,0 +35,5 @@ images: + cacheRefresh: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-jobs-cache_refresh + tag: sha-fb3399a @@ -178,0 +184,9 @@ mongodbMigration: +cacheRefresh: + nodeSelector: {} + resources: + requests: + cpu: 0 + limits: + cpu: 0 + tolerations: [] + diff --git a/jobs/cache_refresh/.flake8 b/jobs/cache_refresh/.flake8 new file mode 100644 index 00000000..f7d6157c --- /dev/null +++ b/jobs/cache_refresh/.flake8 @@ -0,0 +1,5 @@ +[flake8] +# Recommend matching the black line length (119), +# rather than using the flake8 default of 79: +max-line-length = 119 +extend-ignore = "E203" diff --git a/jobs/cache_refresh/.python-version b/jobs/cache_refresh/.python-version new file mode 100644 index 00000000..b326afbc --- /dev/null +++ b/jobs/cache_refresh/.python-version @@ -0,0 +1 @@ +3.9.15 diff --git a/jobs/cache_refresh/Dockerfile b/jobs/cache_refresh/Dockerfile new file mode 100644 index 00000000..0a9188b2 --- /dev/null +++ b/jobs/cache_refresh/Dockerfile @@ -0,0 +1,33 @@ +# build with +# docker build -t some_tag_admin -f Dockerfile ../.. +FROM python:3.9.15-slim + +ENV PYTHONFAULTHANDLER=1 \ + PYTHONUNBUFFERED=1 \ + PYTHONHASHSEED=random \ + PIP_NO_CACHE_DIR=1 \ + PIP_DISABLE_PIP_VERSION_CHECK=on \ + PIP_DEFAULT_TIMEOUT=100 \ + POETRY_NO_INTERACTION=1 \ + # Versions: + POETRY_VERSION=1.3.2 \ + POETRY_VIRTUALENVS_IN_PROJECT=true \ + PATH="$PATH:/root/.local/bin" + +# System deps: +RUN apt-get update \ + && apt-get install -y build-essential python3-dev make \ + && rm -rf /var/lib/apt/lists/* +RUN pip install -U pip +RUN pip install "poetry==$POETRY_VERSION" + +WORKDIR /src +COPY jobs/cache_refresh/poetry.lock ./jobs/cache_refresh/poetry.lock +COPY jobs/cache_refresh/pyproject.toml ./jobs/cache_refresh/pyproject.toml +COPY libs/libcommon ./libs/libcommon +WORKDIR /src/jobs/cache_refresh/ +RUN poetry install --no-cache +COPY jobs/cache_refresh/src ./src +RUN poetry install --no-cache + +ENTRYPOINT ["poetry", "run", "python", "src/cache_refresh/main.py"] diff --git a/jobs/cache_refresh/Makefile b/jobs/cache_refresh/Makefile new file mode 100644 index 00000000..f421f1ee --- /dev/null +++ b/jobs/cache_refresh/Makefile @@ -0,0 +1,15 @@ +# environment variables for the commands (docker compose, poetry) +export COMPOSE_PROJECT_NAME := cache_refresh +export MONGO_PORT := 27080 +export CACHE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +# makefile variables +DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml + +include ../../tools/Python.mk +include ../../tools/PythonTest.mk +include ../../tools/Docker.mk + +.PHONY: run +run: + poetry run python src/cache_refresh/main.py diff --git a/jobs/cache_refresh/poetry.lock b/jobs/cache_refresh/poetry.lock new file mode 100644 index 00000000..6c387d05 --- /dev/null +++ b/jobs/cache_refresh/poetry.lock @@ -0,0 +1,1578 @@ +# This file is automatically @generated by Poetry and should not be changed by hand. + +[[package]] +name = "appdirs" +version = "1.4.4" +description = "A small Python module for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "appdirs-1.4.4-py2.py3-none-any.whl", hash = "sha256:a841dacd6b99318a741b166adb07e19ee71a274450e68237b4650ca1055ab128"}, + {file = "appdirs-1.4.4.tar.gz", hash = "sha256:7d5d0167b2b1ba821647616af46a749d1c653740dd0d2415100fe26e27afdf41"}, +] + +[[package]] +name = "attrs" +version = "22.2.0" +description = "Classes Without Boilerplate" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "attrs-22.2.0-py3-none-any.whl", hash = "sha256:29e95c7f6778868dbd49170f98f8818f78f3dc5e0e37c0b1f474e3561b240836"}, + {file = "attrs-22.2.0.tar.gz", hash = "sha256:c9227bfc2f01993c03f68db37d1d15c9690188323c067c641f1a35ca58185f99"}, +] + +[package.extras] +cov = ["attrs[tests]", "coverage-enable-subprocess", "coverage[toml] (>=5.3)"] +dev = ["attrs[docs,tests]"] +docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope.interface"] +tests = ["attrs[tests-no-zope]", "zope.interface"] +tests-no-zope = ["cloudpickle", "cloudpickle", "hypothesis", "hypothesis", "mypy (>=0.971,<0.990)", "mypy (>=0.971,<0.990)", "pympler", "pympler", "pytest (>=4.3.0)", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-mypy-plugins", "pytest-xdist[psutil]", "pytest-xdist[psutil]"] + +[[package]] +name = "bandit" +version = "1.7.4" +description = "Security oriented static analyser for python code." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "bandit-1.7.4-py3-none-any.whl", hash = "sha256:412d3f259dab4077d0e7f0c11f50f650cc7d10db905d98f6520a95a18049658a"}, + {file = "bandit-1.7.4.tar.gz", hash = "sha256:2d63a8c573417bae338962d4b9b06fbc6080f74ecd955a092849e1e65c717bd2"}, +] + +[package.dependencies] +colorama = {version = ">=0.3.9", markers = "platform_system == \"Windows\""} +GitPython = ">=1.0.1" +PyYAML = ">=5.3.1" +stevedore = ">=1.20.0" + +[package.extras] +test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "toml"] +toml = ["toml"] +yaml = ["PyYAML"] + +[[package]] +name = "black" +version = "22.12.0" +description = "The uncompromising code formatter." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "black-22.12.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eedd20838bd5d75b80c9f5487dbcb06836a43833a37846cf1d8c1cc01cef59d"}, + {file = "black-22.12.0-cp310-cp310-win_amd64.whl", hash = "sha256:159a46a4947f73387b4d83e87ea006dbb2337eab6c879620a3ba52699b1f4351"}, + {file = "black-22.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d30b212bffeb1e252b31dd269dfae69dd17e06d92b87ad26e23890f3efea366f"}, + {file = "black-22.12.0-cp311-cp311-win_amd64.whl", hash = "sha256:7412e75863aa5c5411886804678b7d083c7c28421210180d67dfd8cf1221e1f4"}, + {file = "black-22.12.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c116eed0efb9ff870ded8b62fe9f28dd61ef6e9ddd28d83d7d264a38417dcee2"}, + {file = "black-22.12.0-cp37-cp37m-win_amd64.whl", hash = "sha256:1f58cbe16dfe8c12b7434e50ff889fa479072096d79f0a7f25e4ab8e94cd8350"}, + {file = "black-22.12.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77d86c9f3db9b1bf6761244bc0b3572a546f5fe37917a044e02f3166d5aafa7d"}, + {file = "black-22.12.0-cp38-cp38-win_amd64.whl", hash = "sha256:82d9fe8fee3401e02e79767016b4907820a7dc28d70d137eb397b92ef3cc5bfc"}, + {file = "black-22.12.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:101c69b23df9b44247bd88e1d7e90154336ac4992502d4197bdac35dd7ee3320"}, + {file = "black-22.12.0-cp39-cp39-win_amd64.whl", hash = "sha256:559c7a1ba9a006226f09e4916060982fd27334ae1998e7a38b3f33a37f7a2148"}, + {file = "black-22.12.0-py3-none-any.whl", hash = "sha256:436cc9167dd28040ad90d3b404aec22cedf24a6e4d7de221bec2730ec0c97bcf"}, + {file = "black-22.12.0.tar.gz", hash = "sha256:229351e5a18ca30f447bf724d007f890f97e13af070bb6ad4c0a441cd7596a2f"}, +] + +[package.dependencies] +click = ">=8.0.0" +mypy-extensions = ">=0.4.3" +pathspec = ">=0.9.0" +platformdirs = ">=2" +tomli = {version = ">=1.1.0", markers = "python_full_version < \"3.11.0a7\""} +typing-extensions = {version = ">=3.10.0.0", markers = "python_version < \"3.10\""} + +[package.extras] +colorama = ["colorama (>=0.4.3)"] +d = ["aiohttp (>=3.7.4)"] +jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"] +uvloop = ["uvloop (>=0.15.2)"] + +[[package]] +name = "cachecontrol" +version = "0.12.11" +description = "httplib2 caching for requests" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "CacheControl-0.12.11-py2.py3-none-any.whl", hash = "sha256:2c75d6a8938cb1933c75c50184549ad42728a27e9f6b92fd677c3151aa72555b"}, + {file = "CacheControl-0.12.11.tar.gz", hash = "sha256:a5b9fcc986b184db101aa280b42ecdcdfc524892596f606858e0b7a8b4d9e144"}, +] + +[package.dependencies] +lockfile = {version = ">=0.9", optional = true, markers = "extra == \"filecache\""} +msgpack = ">=0.5.2" +requests = "*" + +[package.extras] +filecache = ["lockfile (>=0.9)"] +redis = ["redis (>=2.10.5)"] + +[[package]] +name = "certifi" +version = "2022.12.7" +description = "Python package for providing Mozilla's CA Bundle." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "certifi-2022.12.7-py3-none-any.whl", hash = "sha256:4ad3232f5e926d6718ec31cfc1fcadfde020920e278684144551c91769c7bc18"}, + {file = "certifi-2022.12.7.tar.gz", hash = "sha256:35824b4c3a97115964b408844d64aa14db1cc518f6562e8d7261699d1350a9e3"}, +] + +[[package]] +name = "charset-normalizer" +version = "3.1.0" +description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"}, + {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"}, +] + +[[package]] +name = "click" +version = "8.1.3" +description = "Composable command line interface toolkit" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "click-8.1.3-py3-none-any.whl", hash = "sha256:bb4d8133cb15a609f44e8213d9b391b0809795062913b383c62be0ee95b1db48"}, + {file = "click-8.1.3.tar.gz", hash = "sha256:7682dc8afb30297001674575ea00d1814d808d6a36af415a82bd481d37ba7b8e"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[[package]] +name = "colorama" +version = "0.4.6" +description = "Cross-platform colored terminal text." +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7" +files = [ + {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"}, + {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"}, +] + +[[package]] +name = "coverage" +version = "7.2.1" +description = "Code coverage measurement for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "coverage-7.2.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:49567ec91fc5e0b15356da07a2feabb421d62f52a9fff4b1ec40e9e19772f5f8"}, + {file = "coverage-7.2.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d2ef6cae70168815ed91388948b5f4fcc69681480a0061114db737f957719f03"}, + {file = "coverage-7.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3004765bca3acd9e015794e5c2f0c9a05587f5e698127ff95e9cfba0d3f29339"}, + {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cca7c0b7f5881dfe0291ef09ba7bb1582cb92ab0aeffd8afb00c700bf692415a"}, + {file = "coverage-7.2.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2167d116309f564af56f9aa5e75ef710ef871c5f9b313a83050035097b56820"}, + {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:cb5f152fb14857cbe7f3e8c9a5d98979c4c66319a33cad6e617f0067c9accdc4"}, + {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:87dc37f16fb5e3a28429e094145bf7c1753e32bb50f662722e378c5851f7fdc6"}, + {file = "coverage-7.2.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e191a63a05851f8bce77bc875e75457f9b01d42843f8bd7feed2fc26bbe60833"}, + {file = "coverage-7.2.1-cp310-cp310-win32.whl", hash = "sha256:e3ea04b23b114572b98a88c85379e9e9ae031272ba1fb9b532aa934c621626d4"}, + {file = "coverage-7.2.1-cp310-cp310-win_amd64.whl", hash = "sha256:0cf557827be7eca1c38a2480484d706693e7bb1929e129785fe59ec155a59de6"}, + {file = "coverage-7.2.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:570c21a29493b350f591a4b04c158ce1601e8d18bdcd21db136fbb135d75efa6"}, + {file = "coverage-7.2.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:9e872b082b32065ac2834149dc0adc2a2e6d8203080501e1e3c3c77851b466f9"}, + {file = "coverage-7.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fac6343bae03b176e9b58104a9810df3cdccd5cfed19f99adfa807ffbf43cf9b"}, + {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:abacd0a738e71b20e224861bc87e819ef46fedba2fb01bc1af83dfd122e9c319"}, + {file = "coverage-7.2.1-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9256d4c60c4bbfec92721b51579c50f9e5062c21c12bec56b55292464873508"}, + {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:80559eaf6c15ce3da10edb7977a1548b393db36cbc6cf417633eca05d84dd1ed"}, + {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0bd7e628f6c3ec4e7d2d24ec0e50aae4e5ae95ea644e849d92ae4805650b4c4e"}, + {file = "coverage-7.2.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:09643fb0df8e29f7417adc3f40aaf379d071ee8f0350ab290517c7004f05360b"}, + {file = "coverage-7.2.1-cp311-cp311-win32.whl", hash = "sha256:1b7fb13850ecb29b62a447ac3516c777b0e7a09ecb0f4bb6718a8654c87dfc80"}, + {file = "coverage-7.2.1-cp311-cp311-win_amd64.whl", hash = "sha256:617a94ada56bbfe547aa8d1b1a2b8299e2ec1ba14aac1d4b26a9f7d6158e1273"}, + {file = "coverage-7.2.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8649371570551d2fd7dee22cfbf0b61f1747cdfb2b7587bb551e4beaaa44cb97"}, + {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d2b9b5e70a21474c105a133ba227c61bc95f2ac3b66861143ce39a5ea4b3f84"}, + {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae82c988954722fa07ec5045c57b6d55bc1a0890defb57cf4a712ced65b26ddd"}, + {file = "coverage-7.2.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:861cc85dfbf55a7a768443d90a07e0ac5207704a9f97a8eb753292a7fcbdfcfc"}, + {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0339dc3237c0d31c3b574f19c57985fcbe494280153bbcad33f2cdf469f4ac3e"}, + {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:5928b85416a388dd557ddc006425b0c37e8468bd1c3dc118c1a3de42f59e2a54"}, + {file = "coverage-7.2.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:8d3843ca645f62c426c3d272902b9de90558e9886f15ddf5efe757b12dd376f5"}, + {file = "coverage-7.2.1-cp37-cp37m-win32.whl", hash = "sha256:6a034480e9ebd4e83d1aa0453fd78986414b5d237aea89a8fdc35d330aa13bae"}, + {file = "coverage-7.2.1-cp37-cp37m-win_amd64.whl", hash = "sha256:6fce673f79a0e017a4dc35e18dc7bb90bf6d307c67a11ad5e61ca8d42b87cbff"}, + {file = "coverage-7.2.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:7f099da6958ddfa2ed84bddea7515cb248583292e16bb9231d151cd528eab657"}, + {file = "coverage-7.2.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:97a3189e019d27e914ecf5c5247ea9f13261d22c3bb0cfcfd2a9b179bb36f8b1"}, + {file = "coverage-7.2.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a81dbcf6c6c877986083d00b834ac1e84b375220207a059ad45d12f6e518a4e3"}, + {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:78d2c3dde4c0b9be4b02067185136b7ee4681978228ad5ec1278fa74f5ca3e99"}, + {file = "coverage-7.2.1-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a209d512d157379cc9ab697cbdbb4cfd18daa3e7eebaa84c3d20b6af0037384"}, + {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f3d07edb912a978915576a776756069dede66d012baa503022d3a0adba1b6afa"}, + {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8dca3c1706670297851bca1acff9618455122246bdae623be31eca744ade05ec"}, + {file = "coverage-7.2.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b1991a6d64231a3e5bbe3099fb0dd7c9aeaa4275ad0e0aeff4cb9ef885c62ba2"}, + {file = "coverage-7.2.1-cp38-cp38-win32.whl", hash = "sha256:22c308bc508372576ffa3d2dbc4824bb70d28eeb4fcd79d4d1aed663a06630d0"}, + {file = "coverage-7.2.1-cp38-cp38-win_amd64.whl", hash = "sha256:b0c0d46de5dd97f6c2d1b560bf0fcf0215658097b604f1840365296302a9d1fb"}, + {file = "coverage-7.2.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4dd34a935de268a133e4741827ae951283a28c0125ddcdbcbba41c4b98f2dfef"}, + {file = "coverage-7.2.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0f8318ed0f3c376cfad8d3520f496946977abde080439d6689d7799791457454"}, + {file = "coverage-7.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:834c2172edff5a08d78e2f53cf5e7164aacabeb66b369f76e7bb367ca4e2d993"}, + {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e4d70c853f0546855f027890b77854508bdb4d6a81242a9d804482e667fff6e6"}, + {file = "coverage-7.2.1-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a6450da4c7afc4534305b2b7d8650131e130610cea448ff240b6ab73d7eab63"}, + {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:99f4dd81b2bb8fc67c3da68b1f5ee1650aca06faa585cbc6818dbf67893c6d58"}, + {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bdd3f2f285ddcf2e75174248b2406189261a79e7fedee2ceeadc76219b6faa0e"}, + {file = "coverage-7.2.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:f29351393eb05e6326f044a7b45ed8e38cb4dcc38570d12791f271399dc41431"}, + {file = "coverage-7.2.1-cp39-cp39-win32.whl", hash = "sha256:e2b50ebc2b6121edf352336d503357321b9d8738bb7a72d06fc56153fd3f4cd8"}, + {file = "coverage-7.2.1-cp39-cp39-win_amd64.whl", hash = "sha256:bd5a12239c0006252244f94863f1c518ac256160cd316ea5c47fb1a11b25889a"}, + {file = "coverage-7.2.1-pp37.pp38.pp39-none-any.whl", hash = "sha256:436313d129db7cf5b4ac355dd2bd3f7c7e5294af077b090b85de75f8458b8616"}, + {file = "coverage-7.2.1.tar.gz", hash = "sha256:c77f2a9093ccf329dd523a9b2b3c854c20d2a3d968b6def3b820272ca6732242"}, +] + +[package.extras] +toml = ["tomli"] + +[[package]] +name = "cyclonedx-python-lib" +version = "3.1.5" +description = "A library for producing CycloneDX SBOM (Software Bill of Materials) files." +category = "dev" +optional = false +python-versions = ">=3.6,<4.0" +files = [ + {file = "cyclonedx_python_lib-3.1.5-py3-none-any.whl", hash = "sha256:8981ca462fba91469c268d684a03f72c89c7a807674d884f83a28d8c2822a9b6"}, + {file = "cyclonedx_python_lib-3.1.5.tar.gz", hash = "sha256:1ccd482024a30b95c4fffb3fe567a9df97b705f34c1075f8abde8537867600c3"}, +] + +[package.dependencies] +packageurl-python = ">=0.9" +setuptools = ">=47.0.0" +sortedcontainers = ">=2.4.0,<3.0.0" +toml = ">=0.10.0,<0.11.0" + +[[package]] +name = "dnspython" +version = "1.16.0" +description = "DNS toolkit" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "dnspython-1.16.0-py2.py3-none-any.whl", hash = "sha256:f69c21288a962f4da86e56c4905b49d11aba7938d3d740e80d9e366ee4f1632d"}, + {file = "dnspython-1.16.0.zip", hash = "sha256:36c5e8e38d4369a08b6780b7f27d790a292b2b08eea01607865bf0936c558e01"}, +] + +[package.extras] +dnssec = ["ecdsa (>=0.13)", "pycryptodome"] +idna = ["idna (>=2.1)"] + +[[package]] +name = "environs" +version = "9.5.0" +description = "simplified environment variable parsing" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "environs-9.5.0-py2.py3-none-any.whl", hash = "sha256:1e549569a3de49c05f856f40bce86979e7d5ffbbc4398e7f338574c220189124"}, + {file = "environs-9.5.0.tar.gz", hash = "sha256:a76307b36fbe856bdca7ee9161e6c466fd7fcffc297109a118c59b54e27e30c9"}, +] + +[package.dependencies] +marshmallow = ">=3.0.0" +python-dotenv = "*" + +[package.extras] +dev = ["dj-database-url", "dj-email-url", "django-cache-url", "flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)", "pytest", "tox"] +django = ["dj-database-url", "dj-email-url", "django-cache-url"] +lint = ["flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)"] +tests = ["dj-database-url", "dj-email-url", "django-cache-url", "pytest"] + +[[package]] +name = "exceptiongroup" +version = "1.1.0" +description = "Backport of PEP 654 (exception groups)" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "exceptiongroup-1.1.0-py3-none-any.whl", hash = "sha256:327cbda3da756e2de031a3107b81ab7b3770a602c4d16ca618298c526f4bec1e"}, + {file = "exceptiongroup-1.1.0.tar.gz", hash = "sha256:bcb67d800a4497e1b404c2dd44fca47d3b7a5e5433dbab67f96c1a685cdfdf23"}, +] + +[package.extras] +test = ["pytest (>=6)"] + +[[package]] +name = "filelock" +version = "3.9.0" +description = "A platform independent file lock." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "filelock-3.9.0-py3-none-any.whl", hash = "sha256:f58d535af89bb9ad5cd4df046f741f8553a418c01a7856bf0d173bbc9f6bd16d"}, + {file = "filelock-3.9.0.tar.gz", hash = "sha256:7b319f24340b51f55a2bf7a12ac0755a9b03e718311dac567a0f4f7fabd2f5de"}, +] + +[package.extras] +docs = ["furo (>=2022.12.7)", "sphinx (>=5.3)", "sphinx-autodoc-typehints (>=1.19.5)"] +testing = ["covdefaults (>=2.2.2)", "coverage (>=7.0.1)", "pytest (>=7.2)", "pytest-cov (>=4)", "pytest-timeout (>=2.1)"] + +[[package]] +name = "flake8" +version = "3.9.2" +description = "the modular source code checker: pep8 pyflakes and co" +category = "dev" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7" +files = [ + {file = "flake8-3.9.2-py2.py3-none-any.whl", hash = "sha256:bf8fd333346d844f616e8d47905ef3a3384edae6b4e9beb0c5101e25e3110907"}, + {file = "flake8-3.9.2.tar.gz", hash = "sha256:07528381786f2a6237b061f6e96610a4167b226cb926e2aa2b6b1d78057c576b"}, +] + +[package.dependencies] +mccabe = ">=0.6.0,<0.7.0" +pycodestyle = ">=2.7.0,<2.8.0" +pyflakes = ">=2.3.0,<2.4.0" + +[[package]] +name = "gitdb" +version = "4.0.10" +description = "Git Object Database" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "gitdb-4.0.10-py3-none-any.whl", hash = "sha256:c286cf298426064079ed96a9e4a9d39e7f3e9bf15ba60701e95f5492f28415c7"}, + {file = "gitdb-4.0.10.tar.gz", hash = "sha256:6eb990b69df4e15bad899ea868dc46572c3f75339735663b81de79b06f17eb9a"}, +] + +[package.dependencies] +smmap = ">=3.0.1,<6" + +[[package]] +name = "gitpython" +version = "3.1.31" +description = "GitPython is a Python library used to interact with Git repositories" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"}, + {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"}, +] + +[package.dependencies] +gitdb = ">=4.0.1,<5" + +[[package]] +name = "html5lib" +version = "1.1" +description = "HTML parser based on the WHATWG HTML specification" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "html5lib-1.1-py2.py3-none-any.whl", hash = "sha256:0d78f8fde1c230e99fe37986a60526d7049ed4bf8a9fadbad5f00e22e58e041d"}, + {file = "html5lib-1.1.tar.gz", hash = "sha256:b2e5b40261e20f354d198eae92afc10d750afb487ed5e50f9c4eaf07c184146f"}, +] + +[package.dependencies] +six = ">=1.9" +webencodings = "*" + +[package.extras] +all = ["chardet (>=2.2)", "genshi", "lxml"] +chardet = ["chardet (>=2.2)"] +genshi = ["genshi"] +lxml = ["lxml"] + +[[package]] +name = "huggingface-hub" +version = "0.12.1" +description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "huggingface_hub-0.12.1-py3-none-any.whl", hash = "sha256:867586cc8543fe1bd43a219fedbea7d71690021ad80f0c46f35c4751069278d7"}, + {file = "huggingface_hub-0.12.1.tar.gz", hash = "sha256:6f960f6246ef9c3446d0d6275e853485515682c350917fdaf2a59705f8b9ebb3"}, +] + +[package.dependencies] +filelock = "*" +packaging = ">=20.9" +pyyaml = ">=5.1" +requests = "*" +tqdm = ">=4.42.1" +typing-extensions = ">=3.7.4.3" + +[package.extras] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +cli = ["InquirerPy (==0.3.4)"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "jedi", "mypy (==0.982)", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] +fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] +quality = ["black (==22.3)", "flake8 (>=3.8.3)", "flake8-bugbear", "isort (>=5.5.4)", "mypy (==0.982)"] +tensorflow = ["graphviz", "pydot", "tensorflow"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "isort (>=5.5.4)", "jedi", "pytest", "pytest-cov", "pytest-env", "pytest-xdist", "soundfile"] +torch = ["torch"] +typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] + +[[package]] +name = "idna" +version = "3.4" +description = "Internationalized Domain Names in Applications (IDNA)" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"}, + {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"}, +] + +[[package]] +name = "iniconfig" +version = "2.0.0" +description = "brain-dead simple config-ini parsing" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "iniconfig-2.0.0-py3-none-any.whl", hash = "sha256:b6a85871a79d2e3b22d2d1b94ac2824226a63c6b741c88f7ae975f18b6778374"}, + {file = "iniconfig-2.0.0.tar.gz", hash = "sha256:2d91e135bf72d31a410b17c16da610a82cb55f6b0477d1a902134b24a455b8b3"}, +] + +[[package]] +name = "isort" +version = "5.12.0" +description = "A Python utility / library to sort Python imports." +category = "dev" +optional = false +python-versions = ">=3.8.0" +files = [ + {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"}, + {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"}, +] + +[package.extras] +colors = ["colorama (>=0.4.3)"] +pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"] +plugins = ["setuptools"] +requirements-deprecated-finder = ["pip-api", "pipreqs"] + +[[package]] +name = "libcommon" +version = "0.6.8" +description = "Library for utils common to all the services" +category = "main" +optional = false +python-versions = "3.9.15" +files = [] +develop = true + +[package.dependencies] +appdirs = "^1.4.4" +environs = "^9.5.0" +huggingface-hub = "^0.12.0" +mongo-types = "0.15.1" +mongoengine = "^0.24.2" +orjson = "^3.8.6" +psutil = "^5.9.4" +pymongo = {version = "^3.13.0", extras = ["srv"]} +pytz = "^2020.1" +requests = "^2.28.2" + +[package.source] +type = "directory" +url = "../../libs/libcommon" + +[[package]] +name = "lockfile" +version = "0.12.2" +description = "Platform-independent file locking module" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "lockfile-0.12.2-py2.py3-none-any.whl", hash = "sha256:6c3cb24f344923d30b2785d5ad75182c8ea7ac1b6171b08657258ec7429d50fa"}, + {file = "lockfile-0.12.2.tar.gz", hash = "sha256:6aed02de03cba24efabcd600b30540140634fc06cfa603822d508d5361e9f799"}, +] + +[[package]] +name = "markdown-it-py" +version = "2.2.0" +description = "Python port of markdown-it. Markdown parsing, done right!" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "markdown-it-py-2.2.0.tar.gz", hash = "sha256:7c9a5e412688bc771c67432cbfebcdd686c93ce6484913dccf06cb5a0bea35a1"}, + {file = "markdown_it_py-2.2.0-py3-none-any.whl", hash = "sha256:5a35f8d1870171d9acc47b99612dc146129b631baf04970128b568f190d0cc30"}, +] + +[package.dependencies] +mdurl = ">=0.1,<1.0" + +[package.extras] +benchmarking = ["psutil", "pytest", "pytest-benchmark"] +code-style = ["pre-commit (>=3.0,<4.0)"] +compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"] +linkify = ["linkify-it-py (>=1,<3)"] +plugins = ["mdit-py-plugins"] +profiling = ["gprof2dot"] +rtd = ["attrs", "myst-parser", "pyyaml", "sphinx", "sphinx-copybutton", "sphinx-design", "sphinx_book_theme"] +testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"] + +[[package]] +name = "marshmallow" +version = "3.19.0" +description = "A lightweight library for converting complex datatypes to and from native Python datatypes." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "marshmallow-3.19.0-py3-none-any.whl", hash = "sha256:93f0958568da045b0021ec6aeb7ac37c81bfcccbb9a0e7ed8559885070b3a19b"}, + {file = "marshmallow-3.19.0.tar.gz", hash = "sha256:90032c0fd650ce94b6ec6dc8dfeb0e3ff50c144586462c389b81a07205bedb78"}, +] + +[package.dependencies] +packaging = ">=17.0" + +[package.extras] +dev = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)", "pytest", "pytz", "simplejson", "tox"] +docs = ["alabaster (==0.7.12)", "autodocsumm (==0.2.9)", "sphinx (==5.3.0)", "sphinx-issues (==3.0.1)", "sphinx-version-warning (==1.1.2)"] +lint = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)"] +tests = ["pytest", "pytz", "simplejson"] + +[[package]] +name = "mccabe" +version = "0.6.1" +description = "McCabe checker, plugin for flake8" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "mccabe-0.6.1-py2.py3-none-any.whl", hash = "sha256:ab8a6258860da4b6677da4bd2fe5dc2c659cff31b3ee4f7f5d64e79735b80d42"}, + {file = "mccabe-0.6.1.tar.gz", hash = "sha256:dd8d182285a0fe56bace7f45b5e7d1a6ebcbf524e8f3bd87eb0f125271b8831f"}, +] + +[[package]] +name = "mdurl" +version = "0.1.2" +description = "Markdown URL utilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8"}, + {file = "mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba"}, +] + +[[package]] +name = "mongo-types" +version = "0.15.1" +description = "Type stubs for mongoengine w/ basic support for bson and pymongo" +category = "main" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "mongo-types-0.15.1.tar.gz", hash = "sha256:0a9deeb7733ea7da5db3711d92e22d93556b522f860bbff82e5df44c53bd06a9"}, + {file = "mongo_types-0.15.1-py3-none-any.whl", hash = "sha256:9417ae5b9a759c09630b5ec7d66904cc333c2d2fcfe75e2760a332ed5e267309"}, +] + +[[package]] +name = "mongoengine" +version = "0.24.2" +description = "MongoEngine is a Python Object-Document Mapper for working with MongoDB." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "mongoengine-0.24.2-py3-none-any.whl", hash = "sha256:f5c4e1b206b2ccffe4adc7a6283ed26dd799bd115a5fb1d2e885a075132cdb88"}, + {file = "mongoengine-0.24.2.tar.gz", hash = "sha256:c76d49658575bb995682e2e77c8ef7cda63faf939415b32ee923745d120f8b02"}, +] + +[package.dependencies] +pymongo = ">=3.4,<5.0" + +[[package]] +name = "mongomock" +version = "4.1.2" +description = "Fake pymongo stub for testing simple MongoDB-dependent code" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "mongomock-4.1.2-py2.py3-none-any.whl", hash = "sha256:08a24938a05c80c69b6b8b19a09888d38d8c6e7328547f94d46cadb7f47209f2"}, + {file = "mongomock-4.1.2.tar.gz", hash = "sha256:f06cd62afb8ae3ef63ba31349abd220a657ef0dd4f0243a29587c5213f931b7d"}, +] + +[package.dependencies] +packaging = "*" +sentinels = "*" + +[[package]] +name = "msgpack" +version = "1.0.4" +description = "MessagePack serializer" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:4ab251d229d10498e9a2f3b1e68ef64cb393394ec477e3370c457f9430ce9250"}, + {file = "msgpack-1.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:112b0f93202d7c0fef0b7810d465fde23c746a2d482e1e2de2aafd2ce1492c88"}, + {file = "msgpack-1.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:002b5c72b6cd9b4bafd790f364b8480e859b4712e91f43014fe01e4f957b8467"}, + {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:35bc0faa494b0f1d851fd29129b2575b2e26d41d177caacd4206d81502d4c6a6"}, + {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4733359808c56d5d7756628736061c432ded018e7a1dff2d35a02439043321aa"}, + {file = "msgpack-1.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:eb514ad14edf07a1dbe63761fd30f89ae79b42625731e1ccf5e1f1092950eaa6"}, + {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c23080fdeec4716aede32b4e0ef7e213c7b1093eede9ee010949f2a418ced6ba"}, + {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:49565b0e3d7896d9ea71d9095df15b7f75a035c49be733051c34762ca95bbf7e"}, + {file = "msgpack-1.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:aca0f1644d6b5a73eb3e74d4d64d5d8c6c3d577e753a04c9e9c87d07692c58db"}, + {file = "msgpack-1.0.4-cp310-cp310-win32.whl", hash = "sha256:0dfe3947db5fb9ce52aaea6ca28112a170db9eae75adf9339a1aec434dc954ef"}, + {file = "msgpack-1.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dea20515f660aa6b7e964433b1808d098dcfcabbebeaaad240d11f909298075"}, + {file = "msgpack-1.0.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:e83f80a7fec1a62cf4e6c9a660e39c7f878f603737a0cdac8c13131d11d97f52"}, + {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3c11a48cf5e59026ad7cb0dc29e29a01b5a66a3e333dc11c04f7e991fc5510a9"}, + {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1276e8f34e139aeff1c77a3cefb295598b504ac5314d32c8c3d54d24fadb94c9"}, + {file = "msgpack-1.0.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6c9566f2c39ccced0a38d37c26cc3570983b97833c365a6044edef3574a00c08"}, + {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:fcb8a47f43acc113e24e910399376f7277cf8508b27e5b88499f053de6b115a8"}, + {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:76ee788122de3a68a02ed6f3a16bbcd97bc7c2e39bd4d94be2f1821e7c4a64e6"}, + {file = "msgpack-1.0.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0a68d3ac0104e2d3510de90a1091720157c319ceeb90d74f7b5295a6bee51bae"}, + {file = "msgpack-1.0.4-cp36-cp36m-win32.whl", hash = "sha256:85f279d88d8e833ec015650fd15ae5eddce0791e1e8a59165318f371158efec6"}, + {file = "msgpack-1.0.4-cp36-cp36m-win_amd64.whl", hash = "sha256:c1683841cd4fa45ac427c18854c3ec3cd9b681694caf5bff04edb9387602d661"}, + {file = "msgpack-1.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:a75dfb03f8b06f4ab093dafe3ddcc2d633259e6c3f74bb1b01996f5d8aa5868c"}, + {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9667bdfdf523c40d2511f0e98a6c9d3603be6b371ae9a238b7ef2dc4e7a427b0"}, + {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11184bc7e56fd74c00ead4f9cc9a3091d62ecb96e97653add7a879a14b003227"}, + {file = "msgpack-1.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ac5bd7901487c4a1dd51a8c58f2632b15d838d07ceedaa5e4c080f7190925bff"}, + {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1e91d641d2bfe91ba4c52039adc5bccf27c335356055825c7f88742c8bb900dd"}, + {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:2a2df1b55a78eb5f5b7d2a4bb221cd8363913830145fad05374a80bf0877cb1e"}, + {file = "msgpack-1.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:545e3cf0cf74f3e48b470f68ed19551ae6f9722814ea969305794645da091236"}, + {file = "msgpack-1.0.4-cp37-cp37m-win32.whl", hash = "sha256:2cc5ca2712ac0003bcb625c96368fd08a0f86bbc1a5578802512d87bc592fe44"}, + {file = "msgpack-1.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:eba96145051ccec0ec86611fe9cf693ce55f2a3ce89c06ed307de0e085730ec1"}, + {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:7760f85956c415578c17edb39eed99f9181a48375b0d4a94076d84148cf67b2d"}, + {file = "msgpack-1.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:449e57cc1ff18d3b444eb554e44613cffcccb32805d16726a5494038c3b93dab"}, + {file = "msgpack-1.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d603de2b8d2ea3f3bcb2efe286849aa7a81531abc52d8454da12f46235092bcb"}, + {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:48f5d88c99f64c456413d74a975bd605a9b0526293218a3b77220a2c15458ba9"}, + {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6916c78f33602ecf0509cc40379271ba0f9ab572b066bd4bdafd7434dee4bc6e"}, + {file = "msgpack-1.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:81fc7ba725464651190b196f3cd848e8553d4d510114a954681fd0b9c479d7e1"}, + {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:d5b5b962221fa2c5d3a7f8133f9abffc114fe218eb4365e40f17732ade576c8e"}, + {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:77ccd2af37f3db0ea59fb280fa2165bf1b096510ba9fe0cc2bf8fa92a22fdb43"}, + {file = "msgpack-1.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:b17be2478b622939e39b816e0aa8242611cc8d3583d1cd8ec31b249f04623243"}, + {file = "msgpack-1.0.4-cp38-cp38-win32.whl", hash = "sha256:2bb8cdf50dd623392fa75525cce44a65a12a00c98e1e37bf0fb08ddce2ff60d2"}, + {file = "msgpack-1.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:26b8feaca40a90cbe031b03d82b2898bf560027160d3eae1423f4a67654ec5d6"}, + {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:462497af5fd4e0edbb1559c352ad84f6c577ffbbb708566a0abaaa84acd9f3ae"}, + {file = "msgpack-1.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:2999623886c5c02deefe156e8f869c3b0aaeba14bfc50aa2486a0415178fce55"}, + {file = "msgpack-1.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f0029245c51fd9473dc1aede1160b0a29f4a912e6b1dd353fa6d317085b219da"}, + {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ed6f7b854a823ea44cf94919ba3f727e230da29feb4a99711433f25800cf747f"}, + {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0df96d6eaf45ceca04b3f3b4b111b86b33785683d682c655063ef8057d61fd92"}, + {file = "msgpack-1.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6a4192b1ab40f8dca3f2877b70e63799d95c62c068c84dc028b40a6cb03ccd0f"}, + {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:0e3590f9fb9f7fbc36df366267870e77269c03172d086fa76bb4eba8b2b46624"}, + {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:1576bd97527a93c44fa856770197dec00d223b0b9f36ef03f65bac60197cedf8"}, + {file = "msgpack-1.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:63e29d6e8c9ca22b21846234913c3466b7e4ee6e422f205a2988083de3b08cae"}, + {file = "msgpack-1.0.4-cp39-cp39-win32.whl", hash = "sha256:fb62ea4b62bfcb0b380d5680f9a4b3f9a2d166d9394e9bbd9666c0ee09a3645c"}, + {file = "msgpack-1.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:4d5834a2a48965a349da1c5a79760d94a1a0172fbb5ab6b5b33cbf8447e109ce"}, + {file = "msgpack-1.0.4.tar.gz", hash = "sha256:f5d869c18f030202eb412f08b28d2afeea553d6613aee89e200d7aca7ef01f5f"}, +] + +[[package]] +name = "mypy" +version = "1.1.1" +description = "Optional static typing for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mypy-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39c7119335be05630611ee798cc982623b9e8f0cff04a0b48dfc26100e0b97af"}, + {file = "mypy-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:61bf08362e93b6b12fad3eab68c4ea903a077b87c90ac06c11e3d7a09b56b9c1"}, + {file = "mypy-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dbb19c9f662e41e474e0cff502b7064a7edc6764f5262b6cd91d698163196799"}, + {file = "mypy-1.1.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:315ac73cc1cce4771c27d426b7ea558fb4e2836f89cb0296cbe056894e3a1f78"}, + {file = "mypy-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:5cb14ff9919b7df3538590fc4d4c49a0f84392237cbf5f7a816b4161c061829e"}, + {file = "mypy-1.1.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:26cdd6a22b9b40b2fd71881a8a4f34b4d7914c679f154f43385ca878a8297389"}, + {file = "mypy-1.1.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5b5f81b40d94c785f288948c16e1f2da37203c6006546c5d947aab6f90aefef2"}, + {file = "mypy-1.1.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21b437be1c02712a605591e1ed1d858aba681757a1e55fe678a15c2244cd68a5"}, + {file = "mypy-1.1.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d809f88734f44a0d44959d795b1e6f64b2bbe0ea4d9cc4776aa588bb4229fc1c"}, + {file = "mypy-1.1.1-cp311-cp311-win_amd64.whl", hash = "sha256:a380c041db500e1410bb5b16b3c1c35e61e773a5c3517926b81dfdab7582be54"}, + {file = "mypy-1.1.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:b7c7b708fe9a871a96626d61912e3f4ddd365bf7f39128362bc50cbd74a634d5"}, + {file = "mypy-1.1.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c1c10fa12df1232c936830839e2e935d090fc9ee315744ac33b8a32216b93707"}, + {file = "mypy-1.1.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:0a28a76785bf57655a8ea5eb0540a15b0e781c807b5aa798bd463779988fa1d5"}, + {file = "mypy-1.1.1-cp37-cp37m-win_amd64.whl", hash = "sha256:ef6a01e563ec6a4940784c574d33f6ac1943864634517984471642908b30b6f7"}, + {file = "mypy-1.1.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:d64c28e03ce40d5303450f547e07418c64c241669ab20610f273c9e6290b4b0b"}, + {file = "mypy-1.1.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:64cc3afb3e9e71a79d06e3ed24bb508a6d66f782aff7e56f628bf35ba2e0ba51"}, + {file = "mypy-1.1.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce61663faf7a8e5ec6f456857bfbcec2901fbdb3ad958b778403f63b9e606a1b"}, + {file = "mypy-1.1.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:2b0c373d071593deefbcdd87ec8db91ea13bd8f1328d44947e88beae21e8d5e9"}, + {file = "mypy-1.1.1-cp38-cp38-win_amd64.whl", hash = "sha256:2888ce4fe5aae5a673386fa232473014056967f3904f5abfcf6367b5af1f612a"}, + {file = "mypy-1.1.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:19ba15f9627a5723e522d007fe708007bae52b93faab00f95d72f03e1afa9598"}, + {file = "mypy-1.1.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:59bbd71e5c58eed2e992ce6523180e03c221dcd92b52f0e792f291d67b15a71c"}, + {file = "mypy-1.1.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9401e33814cec6aec8c03a9548e9385e0e228fc1b8b0a37b9ea21038e64cdd8a"}, + {file = "mypy-1.1.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4b398d8b1f4fba0e3c6463e02f8ad3346f71956b92287af22c9b12c3ec965a9f"}, + {file = "mypy-1.1.1-cp39-cp39-win_amd64.whl", hash = "sha256:69b35d1dcb5707382810765ed34da9db47e7f95b3528334a3c999b0c90fe523f"}, + {file = "mypy-1.1.1-py3-none-any.whl", hash = "sha256:4e4e8b362cdf99ba00c2b218036002bdcdf1e0de085cdb296a49df03fb31dfc4"}, + {file = "mypy-1.1.1.tar.gz", hash = "sha256:ae9ceae0f5b9059f33dbc62dea087e942c0ccab4b7a003719cb70f9b8abfa32f"}, +] + +[package.dependencies] +mypy-extensions = ">=1.0.0" +tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""} +typing-extensions = ">=3.10" + +[package.extras] +dmypy = ["psutil (>=4.0)"] +install-types = ["pip"] +python2 = ["typed-ast (>=1.4.0,<2)"] +reports = ["lxml"] + +[[package]] +name = "mypy-extensions" +version = "1.0.0" +description = "Type system extensions for programs checked with the mypy type checker." +category = "dev" +optional = false +python-versions = ">=3.5" +files = [ + {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"}, + {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"}, +] + +[[package]] +name = "orjson" +version = "3.8.7" +description = "Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "orjson-3.8.7-cp310-cp310-macosx_10_7_x86_64.whl", hash = "sha256:f98c82850b7b4b7e27785ca43706fa86c893cdb88d54576bbb9b0d9c1070e421"}, + {file = "orjson-3.8.7-cp310-cp310-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:1dee503c6c1a0659c5b46f5f39d9ca9d3657b11ca8bb4af8506086df416887d9"}, + {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cc4fa83831f42ce5c938f8cefc2e175fa1df6f661fdeaba3badf26d2b8cfcf73"}, + {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9e432c6c9c8b97ad825276d5795286f7cc9689f377a97e3b7ecf14918413303f"}, + {file = "orjson-3.8.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee519964a5a0efb9633f38b1129fd242807c5c57162844efeeaab1c8de080051"}, + {file = "orjson-3.8.7-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:109b539ce5bf60a121454d008fa67c3b67e5a3249e47d277012645922cf74bd0"}, + {file = "orjson-3.8.7-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ad4d441fbde4133af6fee37f67dbf23181b9c537ecc317346ec8c3b4c8ec7705"}, + {file = "orjson-3.8.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:89dc786419e1ce2588345f58dd6a434e6728bce66b94989644234bcdbe39b603"}, + {file = "orjson-3.8.7-cp310-none-win_amd64.whl", hash = "sha256:697abde7350fb8076d44bcb6b4ab3ce415ae2b5a9bb91efc460e5ab0d96bb5d3"}, + {file = "orjson-3.8.7-cp311-cp311-macosx_10_7_x86_64.whl", hash = "sha256:1c19f47b35b9966a3abadf341b18ee4a860431bf2b00fd8d58906d51cf78aa70"}, + {file = "orjson-3.8.7-cp311-cp311-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:3ffaabb380cd0ee187b4fc362516df6bf739808130b1339445c7d8878fca36e7"}, + {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5d88837002c5a8af970745b8e0ca1b0fdb06aafbe7f1279e110d338ea19f3d23"}, + {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ff60187d1b7e0bfab376b6002b08c560b7de06c87cf3a8ac639ecf58f84c5f3b"}, + {file = "orjson-3.8.7-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0110970aed35dec293f30ed1e09f8604afd5d15c5ef83de7f6c427619b3ba47b"}, + {file = "orjson-3.8.7-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:51b275475d4e36118b65ad56f9764056a09d985c5d72e64579bf8816f1356a5e"}, + {file = "orjson-3.8.7-cp311-none-win_amd64.whl", hash = "sha256:63144d27735f3b60f079f247ac9a289d80dfe49a7f03880dfa0c0ba64d6491d5"}, + {file = "orjson-3.8.7-cp37-cp37m-macosx_10_7_x86_64.whl", hash = "sha256:a16273d77db746bb1789a2bbfded81148a60743fd6f9d5185e02d92e3732fa18"}, + {file = "orjson-3.8.7-cp37-cp37m-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:5bb32259ea22cc9dd47a6fdc4b8f9f1e2f798fcf56c7c1122a7df0f4c5d33bf3"}, + {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ad02e9102d4ba67db30a136e631e32aeebd1dce26c9f5942a457b02df131c5d0"}, + {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:dbcfcec2b7ac52deb7be3685b551addc28ee8fa454ef41f8b714df6ba0e32a27"}, + {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e1a0e5504a5fc86083cc210c6946e8d61e13fe9f1d7a7bf81b42f7050a49d4fb"}, + {file = "orjson-3.8.7-cp37-cp37m-manylinux_2_28_x86_64.whl", hash = "sha256:7bd4fd37adb03b1f2a1012d43c9f95973a02164e131dfe3ff804d7e180af5653"}, + {file = "orjson-3.8.7-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:188ed9f9a781333ad802af54c55d5a48991e292239aef41bd663b6e314377eb8"}, + {file = "orjson-3.8.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:cc52f58c688cb10afd810280e450f56fbcb27f52c053463e625c8335c95db0dc"}, + {file = "orjson-3.8.7-cp37-none-win_amd64.whl", hash = "sha256:403c8c84ac8a02c40613b0493b74d5256379e65196d39399edbf2ed3169cbeb5"}, + {file = "orjson-3.8.7-cp38-cp38-macosx_10_7_x86_64.whl", hash = "sha256:7d6ac5f8a2a17095cd927c4d52abbb38af45918e0d3abd60fb50cfd49d71ae24"}, + {file = "orjson-3.8.7-cp38-cp38-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:0295a7bfd713fa89231fd0822c995c31fc2343c59a1d13aa1b8b6651335654f5"}, + {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:feb32aaaa34cf2f891eb793ad320d4bb6731328496ae59b6c9eb1b620c42b529"}, + {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7a3ab1a473894e609b6f1d763838c6689ba2b97620c256a32c4d9f10595ac179"}, + {file = "orjson-3.8.7-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e8c430d82b532c5ab95634e034bbf6ca7432ffe175a3e63eadd493e00b3a555"}, + {file = "orjson-3.8.7-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:366cc75f7e09106f9dac95a675aef413367b284f25507d21e55bd7f45f445e80"}, + {file = "orjson-3.8.7-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:84d154d07e8b17d97e990d5d710b719a031738eb1687d8a05b9089f0564ff3e0"}, + {file = "orjson-3.8.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:06180014afcfdc167ca984b312218aa62ce20093965c437c5f9166764cb65ef7"}, + {file = "orjson-3.8.7-cp38-none-win_amd64.whl", hash = "sha256:41244431ba13f2e6ef22b52c5cf0202d17954489f4a3c0505bd28d0e805c3546"}, + {file = "orjson-3.8.7-cp39-cp39-macosx_10_7_x86_64.whl", hash = "sha256:b20f29fa8371b8023f1791df035a2c3ccbd98baa429ac3114fc104768f7db6f8"}, + {file = "orjson-3.8.7-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl", hash = "sha256:226bfc1da2f21ee74918cee2873ea9a0fec1a8830e533cb287d192d593e99d02"}, + {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e75c11023ac29e29fd3e75038d0e8dd93f9ea24d7b9a5e871967a8921a88df24"}, + {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:78604d3acfd7cd502f6381eea0c42281fe2b74755b334074ab3ebc0224100be1"}, + {file = "orjson-3.8.7-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e7129a6847f0494aa1427167486ef6aea2e835ba05f6c627df522692ee228f65"}, + {file = "orjson-3.8.7-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:1a1a8f4980059f48483782c608145b0f74538c266e01c183d9bcd9f8b71dbada"}, + {file = "orjson-3.8.7-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:d60304172a33705ce4bd25a6261ab84bed2dab0b3d3b79672ea16c7648af4832"}, + {file = "orjson-3.8.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4f733062d84389c32c0492e5a4929056fac217034a94523debe0430bcc602cda"}, + {file = "orjson-3.8.7-cp39-none-win_amd64.whl", hash = "sha256:010e2970ec9e826c332819e0da4b14b29b19641da0f1a6af4cec91629ef9b988"}, + {file = "orjson-3.8.7.tar.gz", hash = "sha256:8460c8810652dba59c38c80d27c325b5092d189308d8d4f3e688dbd8d4f3b2dc"}, +] + +[[package]] +name = "packageurl-python" +version = "0.10.4" +description = "A purl aka. Package URL parser and builder" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "packageurl-python-0.10.4.tar.gz", hash = "sha256:5c91334f942cd55d45eb0c67dd339a535ef90e25f05b9ec016ad188ed0ef9048"}, + {file = "packageurl_python-0.10.4-py3-none-any.whl", hash = "sha256:bf8a1ffe755634776f6563904d792fb0aa13b377fc86115c36fe17f69b6e59db"}, +] + +[package.extras] +build = ["wheel"] +test = ["black", "isort", "pytest"] + +[[package]] +name = "packaging" +version = "23.0" +description = "Core utilities for Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "packaging-23.0-py3-none-any.whl", hash = "sha256:714ac14496c3e68c99c29b00845f7a2b85f3bb6f1078fd9f72fd20f0570002b2"}, + {file = "packaging-23.0.tar.gz", hash = "sha256:b6ad297f8907de0fa2fe1ccbd26fdaf387f5f47c7275fedf8cce89f99446cf97"}, +] + +[[package]] +name = "pathspec" +version = "0.11.0" +description = "Utility library for gitignore style pattern matching of file paths." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pathspec-0.11.0-py3-none-any.whl", hash = "sha256:3a66eb970cbac598f9e5ccb5b2cf58930cd8e3ed86d393d541eaf2d8b1705229"}, + {file = "pathspec-0.11.0.tar.gz", hash = "sha256:64d338d4e0914e91c1792321e6907b5a593f1ab1851de7fc269557a21b30ebbc"}, +] + +[[package]] +name = "pbr" +version = "5.11.1" +description = "Python Build Reasonableness" +category = "dev" +optional = false +python-versions = ">=2.6" +files = [ + {file = "pbr-5.11.1-py2.py3-none-any.whl", hash = "sha256:567f09558bae2b3ab53cb3c1e2e33e726ff3338e7bae3db5dc954b3a44eef12b"}, + {file = "pbr-5.11.1.tar.gz", hash = "sha256:aefc51675b0b533d56bb5fd1c8c6c0522fe31896679882e1c4c63d5e4a0fccb3"}, +] + +[[package]] +name = "pip" +version = "23.0.1" +description = "The PyPA recommended tool for installing Python packages." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-23.0.1-py3-none-any.whl", hash = "sha256:236bcb61156d76c4b8a05821b988c7b8c35bf0da28a4b614e8d6ab5212c25c6f"}, + {file = "pip-23.0.1.tar.gz", hash = "sha256:cd015ea1bfb0fcef59d8a286c1f8bebcb983f6317719d415dc5351efb7cd7024"}, +] + +[[package]] +name = "pip-api" +version = "0.0.30" +description = "An unofficial, importable pip API" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-api-0.0.30.tar.gz", hash = "sha256:a05df2c7aa9b7157374bcf4273544201a0c7bae60a9c65bcf84f3959ef3896f3"}, + {file = "pip_api-0.0.30-py3-none-any.whl", hash = "sha256:2a0314bd31522eb9ffe8a99668b0d07fee34ebc537931e7b6483001dbedcbdc9"}, +] + +[package.dependencies] +pip = "*" + +[[package]] +name = "pip-audit" +version = "2.4.14" +description = "A tool for scanning Python environments for known vulnerabilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip_audit-2.4.14-py3-none-any.whl", hash = "sha256:f9632b9f67bcf3fda78ef7651a03c8ed926d1eaeda474dcbdcb26a5518dd6ffc"}, + {file = "pip_audit-2.4.14.tar.gz", hash = "sha256:1259629fe24302e257052e977146f56bebf34927740d5efd184aaafa3b1b3b38"}, +] + +[package.dependencies] +CacheControl = {version = ">=0.12.10", extras = ["filecache"]} +cyclonedx-python-lib = ">=2.0.0,<2.5.0 || >2.5.0" +html5lib = ">=1.1" +packaging = ">=23.0.0" +pip-api = ">=0.0.28" +pip-requirements-parser = ">=32.0.0" +resolvelib = ">=0.8.0" +rich = ">=12.4" +toml = ">=0.10" + +[package.extras] +dev = ["build", "bump (>=1.3.2)", "pip-audit[doc,lint,test]"] +doc = ["pdoc"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.228)", "types-html5lib", "types-requests", "types-toml"] +test = ["coverage[toml]", "pretend", "pytest", "pytest-cov"] + +[[package]] +name = "pip-requirements-parser" +version = "32.0.1" +description = "pip requirements parser - a mostly correct pip requirements parsing library because it uses pip's own code." +category = "dev" +optional = false +python-versions = ">=3.6.0" +files = [ + {file = "pip-requirements-parser-32.0.1.tar.gz", hash = "sha256:b4fa3a7a0be38243123cf9d1f3518da10c51bdb165a2b2985566247f9155a7d3"}, + {file = "pip_requirements_parser-32.0.1-py3-none-any.whl", hash = "sha256:4659bc2a667783e7a15d190f6fccf8b2486685b6dba4c19c3876314769c57526"}, +] + +[package.dependencies] +packaging = "*" +pyparsing = "*" + +[package.extras] +docs = ["Sphinx (>=3.3.1)", "doc8 (>=0.8.1)", "sphinx-rtd-theme (>=0.5.0)"] +testing = ["aboutcode-toolkit (>=6.0.0)", "black", "pytest (>=6,!=7.0.0)", "pytest-xdist (>=2)"] + +[[package]] +name = "platformdirs" +version = "3.1.0" +description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "platformdirs-3.1.0-py3-none-any.whl", hash = "sha256:13b08a53ed71021350c9e300d4ea8668438fb0046ab3937ac9a29913a1a1350a"}, + {file = "platformdirs-3.1.0.tar.gz", hash = "sha256:accc3665857288317f32c7bebb5a8e482ba717b474f3fc1d18ca7f9214be0cef"}, +] + +[package.extras] +docs = ["furo (>=2022.12.7)", "proselint (>=0.13)", "sphinx (>=6.1.3)", "sphinx-autodoc-typehints (>=1.22,!=1.23.4)"] +test = ["appdirs (==1.4.4)", "covdefaults (>=2.2.2)", "pytest (>=7.2.1)", "pytest-cov (>=4)", "pytest-mock (>=3.10)"] + +[[package]] +name = "pluggy" +version = "1.0.0" +description = "plugin and hook calling mechanisms for python" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pluggy-1.0.0-py2.py3-none-any.whl", hash = "sha256:74134bbf457f031a36d68416e1509f34bd5ccc019f0bcc952c7b909d06b37bd3"}, + {file = "pluggy-1.0.0.tar.gz", hash = "sha256:4224373bacce55f955a878bf9cfa763c1e360858e330072059e10bad68531159"}, +] + +[package.extras] +dev = ["pre-commit", "tox"] +testing = ["pytest", "pytest-benchmark"] + +[[package]] +name = "psutil" +version = "5.9.4" +description = "Cross-platform lib for process and system monitoring in Python." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "psutil-5.9.4-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:c1ca331af862803a42677c120aff8a814a804e09832f166f226bfd22b56feee8"}, + {file = "psutil-5.9.4-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:68908971daf802203f3d37e78d3f8831b6d1014864d7a85937941bb35f09aefe"}, + {file = "psutil-5.9.4-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:3ff89f9b835100a825b14c2808a106b6fdcc4b15483141482a12c725e7f78549"}, + {file = "psutil-5.9.4-cp27-cp27m-win32.whl", hash = "sha256:852dd5d9f8a47169fe62fd4a971aa07859476c2ba22c2254d4a1baa4e10b95ad"}, + {file = "psutil-5.9.4-cp27-cp27m-win_amd64.whl", hash = "sha256:9120cd39dca5c5e1c54b59a41d205023d436799b1c8c4d3ff71af18535728e94"}, + {file = "psutil-5.9.4-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:6b92c532979bafc2df23ddc785ed116fced1f492ad90a6830cf24f4d1ea27d24"}, + {file = "psutil-5.9.4-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:efeae04f9516907be44904cc7ce08defb6b665128992a56957abc9b61dca94b7"}, + {file = "psutil-5.9.4-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:54d5b184728298f2ca8567bf83c422b706200bcbbfafdc06718264f9393cfeb7"}, + {file = "psutil-5.9.4-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16653106f3b59386ffe10e0bad3bb6299e169d5327d3f187614b1cb8f24cf2e1"}, + {file = "psutil-5.9.4-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:54c0d3d8e0078b7666984e11b12b88af2db11d11249a8ac8920dd5ef68a66e08"}, + {file = "psutil-5.9.4-cp36-abi3-win32.whl", hash = "sha256:149555f59a69b33f056ba1c4eb22bb7bf24332ce631c44a319cec09f876aaeff"}, + {file = "psutil-5.9.4-cp36-abi3-win_amd64.whl", hash = "sha256:fd8522436a6ada7b4aad6638662966de0d61d241cb821239b2ae7013d41a43d4"}, + {file = "psutil-5.9.4-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:6001c809253a29599bc0dfd5179d9f8a5779f9dffea1da0f13c53ee568115e1e"}, + {file = "psutil-5.9.4.tar.gz", hash = "sha256:3d7f9739eb435d4b1338944abe23f49584bde5395f27487d2ee25ad9a8774a62"}, +] + +[package.extras] +test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] + +[[package]] +name = "pycodestyle" +version = "2.7.0" +description = "Python style guide checker" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycodestyle-2.7.0-py2.py3-none-any.whl", hash = "sha256:514f76d918fcc0b55c6680472f0a37970994e07bbb80725808c17089be302068"}, + {file = "pycodestyle-2.7.0.tar.gz", hash = "sha256:c389c1d06bf7904078ca03399a4816f974a1d590090fecea0c63ec26ebaf1cef"}, +] + +[[package]] +name = "pyflakes" +version = "2.3.1" +description = "passive checker of Python programs" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pyflakes-2.3.1-py2.py3-none-any.whl", hash = "sha256:7893783d01b8a89811dd72d7dfd4d84ff098e5eed95cfa8905b22bbffe52efc3"}, + {file = "pyflakes-2.3.1.tar.gz", hash = "sha256:f5bc8ecabc05bb9d291eb5203d6810b49040f6ff446a756326104746cc00c1db"}, +] + +[[package]] +name = "pygments" +version = "2.14.0" +description = "Pygments is a syntax highlighting package written in Python." +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "Pygments-2.14.0-py3-none-any.whl", hash = "sha256:fa7bd7bd2771287c0de303af8bfdfc731f51bd2c6a47ab69d117138893b82717"}, + {file = "Pygments-2.14.0.tar.gz", hash = "sha256:b3ed06a9e8ac9a9aae5a6f5dbe78a8a58655d17b43b93c078f094ddc476ae297"}, +] + +[package.extras] +plugins = ["importlib-metadata"] + +[[package]] +name = "pymongo" +version = "3.13.0" +description = "Python driver for MongoDB <http://www.mongodb.org>" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl", hash = "sha256:3ad3a3df830f7df7e0856c2bdb54d19f5bf188bd7420985e18643b8e4d2a075f"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:b96e0e9d2d48948240b510bac81614458fc10adcd3a93240c2fd96448b4efd35"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9f592b202d77923498b32ddc5b376e5fa9ba280d3e16ed56cb8c932fe6d6a478"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:851f2bb52b5cb2f4711171ca925e0e05344a8452972a748a8a8ffdda1e1d72a7"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:1c9d23f62a3fa7523d849c4942acc0d9ff7081ebc00c808ee7cfdc070df0687f"}, + {file = "pymongo-3.13.0-cp27-cp27m-win32.whl", hash = "sha256:a17b81f22398e3e0f72bdf938e98c810286994b2bcc0a125cd5ad8fd4ea54ad7"}, + {file = "pymongo-3.13.0-cp27-cp27m-win_amd64.whl", hash = "sha256:4f6dd55dab77adf60b445c11f426ee5cdfa1b86f6d54cb937bfcbf09572333ab"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:776f90bf2252f90a4ae838e7917638894c6356bef7265f424592e2fd1f577d05"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:50b99f4d3eee6f03778fe841d6f470e6c18e744dc665156da6da3bc6e65b398d"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50a81b2d9f188c7909e0a1084fa969bb92a788076809c437ac1ae80393f46df9"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c7c45a8a1a752002b0a7c81ab3a4c5e3b6f67f9826b16fbe3943f5329f565f24"}, + {file = "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1037097708498bdc85f23c8798a5c46c7bce432d77d23608ff14e0d831f1a971"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl", hash = "sha256:b5b733694e7df22d5c049581acfc487695a6ff813322318bed8dd66f79978636"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:d7c91747ec8dde51440dd594603158cc98abb3f7df84b2ed8a836f138285e4fb"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl", hash = "sha256:f4175fcdddf764d371ee52ec4505a40facee2533e84abf2953cda86d050cfa1f"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl", hash = "sha256:93d4e9a02c17813b34e4bd9f6fbf07310c140c8f74341537c24d07c1cdeb24d1"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl", hash = "sha256:3b261d593f2563299062733ae003a925420a86ff4ddda68a69097d67204e43f3"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl", hash = "sha256:172db03182a22e9002157b262c1ea3b0045c73d4ff465adc152ce5b4b0e7b8d4"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09de3bfc995ae8cb955abb0c9ae963c134dba1b5622be3bcc527b89b0fd4091c"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c0379447587ee4b8f983ba183202496e86c0358f47c45612619d634d1fcd82bd"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30245a8747dc90019a3c9ad9df987e0280a3ea632ad36227cde7d1d8dcba0830"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65b6fddf6a7b91da044f202771a38e71bbb9bf42720a406b26b25fe2256e7102"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5831a377d15a626fbec10890ffebc4c6abcd37e4126737932cd780a171eabdc1"}, + {file = "pymongo-3.13.0-cp310-cp310-win32.whl", hash = "sha256:944249aa83dee314420c37d0f40c30a8f6dc4a3877566017b87062e53af449f4"}, + {file = "pymongo-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea8824ebc9a1a5c8269e8f1e3989b5a6bec876726e2f3c33ebd036cb488277f0"}, + {file = "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:bdd34c57b4da51a7961beb33645646d197e41f8517801dc76b37c1441e7a4e10"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26f9cc42a162faa241c82e117ac85734ae9f14343dc2df1c90c6b2181f791b22"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a82a1c10f5608e6494913faa169e213d703194bfca0aa710901f303be212414"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8927f22ef6a16229da7f18944deac8605bdc2c0858be5184259f2f7ce7fd4459"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e6f8191a282ef77e526f8f8f63753a437e4aa4bc78f5edd8b6b6ed0eaebd5363"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4d9ed67c987bf9ac2ac684590ba3d2599cdfb0f331ee3db607f9684469b3b59d"}, + {file = "pymongo-3.13.0-cp311-cp311-win32.whl", hash = "sha256:e8f6979664ff477cd61b06bf8aba206df7b2334209815ab3b1019931dab643d6"}, + {file = "pymongo-3.13.0-cp311-cp311-win_amd64.whl", hash = "sha256:174fd1000e896d0dfbc7f6d7e6a1992a4868796c7dec31679e38218c78d6a942"}, + {file = "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl", hash = "sha256:d1ee773fb72ba024e7e3bb6ea8907fe52bccafcb5184aaced6bad995bd30ea20"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:28565e3dbd69fe5fe35a210067064dbb6ed5abe997079f653c19c873c3896fe6"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:5c1db7d366004d6c699eb08c716a63ae0a3e946d061cbebea65d7ce361950265"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e1956f3338c10308e2f99c2c9ff46ae412035cbcd7aaa76c39ccdb806854a247"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:10f0fddc1d63ba3d4a4bffcc7720184c1b7efd570726ad5e2f55818da320239f"}, + {file = "pymongo-3.13.0-cp35-cp35m-win32.whl", hash = "sha256:570ae3365b23d4fd8c669cb57613b1a90b2757e993588d3370ef90945dbeec4b"}, + {file = "pymongo-3.13.0-cp35-cp35m-win_amd64.whl", hash = "sha256:79f777eaf3f5b2c6d81f9ef00d87837001d7063302503bbcbfdbf3e9bc27c96f"}, + {file = "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl", hash = "sha256:d42eb29ba314adfd9c11234b4b646f61b0448bf9b00f14db4b317e6e4b947e77"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:e5e87c0eb774561c546f979342a8ff36ebee153c60a0b6c6b03ba989ceb9538c"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:0f2c5a5984599a88d087a15859860579b825098b473d8c843f1979a83d159f2e"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:59c98e86c5e861032b71e6e5b65f23e6afaacea6e82483b66f1191a5021a7b4f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:70b67390e27e58876853efbb87e43c85252de2515e2887f7dd901b4fa3d21973"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:42ba8606492d76e6f9e4c7a458ed4bc712603be393259a52450345f0945da2cf"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:0e5536994cf2d8488c6fd9dea71df3c4dbb3e0d2ba5e695da06d9142a29a0969"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:fe8194f107f0fa3cabd14e9e809f174eca335993c1db72d1e74e0f496e7afe1f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d593d50815771f517d3ac4367ff716e3f3c78edae51d98e1e25791459f8848ff"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5136ebe8da6a1604998a8eb96be55935aa5f7129c41cc7bddc400d48e8df43be"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a424bdedfd84454d2905a861e0d4bb947cc5bd024fdeb3600c1a97d2be0f4255"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5161167b3840e9c84c80f2534ea6a099f51749d5673b662a3dd248be17c3208"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:644470442beaf969df99c4e00367a817eee05f0bba5d888f1ba6fe97b5e1c102"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2406df90b2335371706c59b7d79e9633b81ed2a7ecd48c1faf8584552bdf2d90"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:222591b828de10ac90064047b5d4916953f38c38b155009c4b8b5e0d33117c2b"}, + {file = "pymongo-3.13.0-cp36-cp36m-win32.whl", hash = "sha256:7cb987b199fa223ad78eebaa9fbc183d5a5944bfe568a9d6f617316ca1c1f32f"}, + {file = "pymongo-3.13.0-cp36-cp36m-win_amd64.whl", hash = "sha256:a6cbb73d9fc2282677e2b7a137d13da987bd0b13abd88ed27bba5534c226db06"}, + {file = "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl", hash = "sha256:b1223b826acbef07a7f5eb9bf37247b0b580119916dca9eae19d92b1290f5855"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:398fb86d374dc351a4abc2e24cd15e5e14b2127f6d90ce0df3fdf2adcc55ac1b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:9c3d07ea19cd2856d9943dce37e75d69ecbb5baf93c3e4c82f73b6075c481292"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:2943d739715f265a2983ac43747595b6af3312d0a370614040959fd293763adf"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:c3b70ed82f20d18d22eafc9bda0ea656605071762f7d31f3c5afc35c59d3393b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:7ec2bb598847569ae34292f580842d37619eea3e546005042f485e15710180d5"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:8cc37b437cba909bef06499dadd91a39c15c14225e8d8c7870020049f8a549fe"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:65a063970e15a4f338f14b820561cf6cdaf2839691ac0adb2474ddff9d0b8b0b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02f0e1a75d3bc0e16c7e15daf9c56185642be055e425f3b34888fc6eb1b22401"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16e74b9c2aca2734c7f49f00fe68d6830a30d26df60e2ace7fe40ccb92087b94"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:24e954be35ad4537840f20bbc8d75320ae647d3cb4fab12cb8fcd2d55f408e76"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a149377d1ff766fd618500798d0d94637f66d0ae222bb6d28f41f3e15c626297"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61660710b054ae52c8fc10368e91d74719eb05554b631d7f8ca93d21d2bff2e6"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4bbc0d27dfef7689285e54f2e0a224f0c7cd9d5c46d2638fabad5500b951c92f"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:9b2ed9c3b30f11cd4a3fbfc22167af7987b01b444215c2463265153fe7cf66d6"}, + {file = "pymongo-3.13.0-cp37-cp37m-win32.whl", hash = "sha256:1c2c5e2b00e2fadcd590c0b2e293d71215e98ed1cb635cfca2be4998d197e534"}, + {file = "pymongo-3.13.0-cp37-cp37m-win_amd64.whl", hash = "sha256:32eac95bbb030b2376ffd897376c6f870222a3457f01a9ce466b9057876132f8"}, + {file = "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a796ef39dadf9d73af05d24937644d386495e43a7d13617aa3651d836da542c8"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:b6793baf4639c72a500698a49e9250b293e17ae1faf11ac1699d8141194786fe"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:80d8576b04d0824f63bf803190359c0d3bcb6e7fa63fefbd4bc0ceaa7faae38c"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:db2e11507fe9cc2a722be21ccc62c1b1295398fe9724c1f14900cdc7166fc0d7"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:b01ce58eec5edeededf1992d2dce63fb8565e437be12d6f139d75b15614c4d08"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:d1a19d6c5098f1f4e11430cd74621699453cbc534dd7ade9167e582f50814b19"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:7219b1a726ced3bacecabef9bd114529bbb69477901373e800d7d0140baadc95"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:2dae3b353a10c3767e0aa1c1492f2af388f1012b08117695ab3fd1f219e5814e"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:12721d926d43d33dd3318e58dce9b0250e8a9c6e1093fa8e09f4805193ff4b43"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6af0a4b17faf26779d5caee8542a4f2cba040cea27d3bffc476cbc6ccbd4c8ee"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:09b9d0f5a445c7e0ddcc021b09835aa6556f0166afc498f57dfdd72cdf6f02ad"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db5b4f8ad8607a3d612da1d4c89a84e4cf5c88f98b46365820d9babe5884ba45"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:34dbf5fecf653c152edb75a35a8b15dfdc4549473484ee768aeb12c97983cead"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:34cd48df7e1fc69222f296d8f69e3957eb7c6b5aa0709d3467184880ed7538c0"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c8f755ff1f4ab4ca790d1d6d3229006100b301475948021b6b2757822e0d6c97"}, + {file = "pymongo-3.13.0-cp38-cp38-win32.whl", hash = "sha256:b0746d0d4535f56bbaa63a8f6da362f330804d578e66e126b226eebe76c2bf00"}, + {file = "pymongo-3.13.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ad0515abb132f52ce9d8abd1a29681a1e65dba7b7fe13ea01e1a8db5715bf80"}, + {file = "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:3c5cb6c93c94df76a879bad4b89db0104b01806d17c2b803c1316ba50962b6d6"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2e0854170813238f0c3131050c67cb1fb1ade75c93bf6cd156c1bd9a16095528"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:1410faa51ce835cc1234c99ec42e98ab4f3c6f50d92d86a2d4f6e11c97ee7a4e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:d7910135f5de1c5c3578e61d6f4b087715b15e365f11d4fa51a9cee92988b2bd"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:028175dd8d2979a889153a2308e8e500b3df7d9e3fd1c33ca7fdeadf61cc87a2"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:2bfc39276c0e6d07c95bd1088b5003f049e986e089509f7dbd68bb7a4b1e65ac"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:4092b660ec720d44d3ca81074280dc25c7a3718df1b6c0fe9fe36ac6ed2833e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:5bdeb71a610a7b801416268e500e716d0fe693fb10d809e17f0fb3dac5be5a34"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aa3bca8e76f5c00ed2bb4325e0e383a547d71595926d5275d7c88175aaf7435e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7c7cab8155f430ca460a6fc7ae8a705b34f3e279a57adb5f900eb81943ec777c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4a32f3dfcca4a4816373bdb6256c18c78974ebb3430e7da988516cd95b2bd6e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30ed2788a6ec68743e2040ab1d16573d7d9f6e7333e45070ce9268cbc93d148c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21e61a536ffed84d10376c21c13a6ed1ebefb61989a844952547c229d6aeedf3"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0665412dce26b2318092a33bd2d2327d487c4490cfcde158d6946d39b1e28d78"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:64ed1a5ce5e5926727eb0f87c698c4d9a7a9f7b0953683a65e9ce2b7cc5f8e91"}, + {file = "pymongo-3.13.0-cp39-cp39-win32.whl", hash = "sha256:7593cb1214185a0c5b43b96effc51ce82ddc933298ee36db7dc2bd45d61b4adc"}, + {file = "pymongo-3.13.0-cp39-cp39-win_amd64.whl", hash = "sha256:3cfc9bc1e8b5667bc1f3dbe46d2f85b3f24ff7533893bdc1203058012db2c046"}, + {file = "pymongo-3.13.0.tar.gz", hash = "sha256:e22d6cf5802cd09b674c307cc9e03870b8c37c503ebec3d25b86f2ce8c535dc7"}, +] + +[package.dependencies] +dnspython = {version = ">=1.16.0,<1.17.0", optional = true, markers = "extra == \"srv\""} + +[package.extras] +aws = ["pymongo-auth-aws (<2.0.0)"] +encryption = ["pymongocrypt (>=1.1.0,<2.0.0)"] +gssapi = ["pykerberos"] +ocsp = ["certifi", "pyopenssl (>=17.2.0)", "requests (<3.0.0)", "service-identity (>=18.1.0)"] +snappy = ["python-snappy"] +srv = ["dnspython (>=1.16.0,<1.17.0)"] +tls = ["ipaddress"] +zstd = ["zstandard"] + +[[package]] +name = "pyparsing" +version = "3.0.9" +description = "pyparsing module - Classes and methods to define and execute parsing grammars" +category = "dev" +optional = false +python-versions = ">=3.6.8" +files = [ + {file = "pyparsing-3.0.9-py3-none-any.whl", hash = "sha256:5026bae9a10eeaefb61dab2f09052b9f4307d44aee4eda64b309723d8d206bbc"}, + {file = "pyparsing-3.0.9.tar.gz", hash = "sha256:2b020ecf7d21b687f219b71ecad3631f644a47f01403fa1d1036b0c6416d70fb"}, +] + +[package.extras] +diagrams = ["jinja2", "railroad-diagrams"] + +[[package]] +name = "pytest" +version = "7.2.2" +description = "pytest: simple powerful testing with Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pytest-7.2.2-py3-none-any.whl", hash = "sha256:130328f552dcfac0b1cec75c12e3f005619dc5f874f0a06e8ff7263f0ee6225e"}, + {file = "pytest-7.2.2.tar.gz", hash = "sha256:c99ab0c73aceb050f68929bc93af19ab6db0558791c6a0715723abe9d0ade9d4"}, +] + +[package.dependencies] +attrs = ">=19.2.0" +colorama = {version = "*", markers = "sys_platform == \"win32\""} +exceptiongroup = {version = ">=1.0.0rc8", markers = "python_version < \"3.11\""} +iniconfig = "*" +packaging = "*" +pluggy = ">=0.12,<2.0" +tomli = {version = ">=1.0.0", markers = "python_version < \"3.11\""} + +[package.extras] +testing = ["argcomplete", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.7.2)", "requests", "xmlschema"] + +[[package]] +name = "pytest-cov" +version = "2.12.1" +description = "Pytest plugin for measuring coverage." +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"}, + {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"}, +] + +[package.dependencies] +coverage = ">=5.2.1" +pytest = ">=4.6" +toml = "*" + +[package.extras] +testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"] + +[[package]] +name = "python-dotenv" +version = "1.0.0" +description = "Read key-value pairs from a .env file and set them as environment variables" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"}, + {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"}, +] + +[package.extras] +cli = ["click (>=5.0)"] + +[[package]] +name = "pytz" +version = "2020.5" +description = "World timezone definitions, modern and historical" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"}, + {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"}, +] + +[[package]] +name = "pyyaml" +version = "6.0" +description = "YAML parser and emitter for Python" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d4db7c7aef085872ef65a8fd7d6d09a14ae91f691dec3e87ee5ee0539d516f53"}, + {file = "PyYAML-6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9df7ed3b3d2e0ecfe09e14741b857df43adb5a3ddadc919a2d94fbdf78fea53c"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77f396e6ef4c73fdc33a9157446466f1cff553d979bd00ecb64385760c6babdc"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a80a78046a72361de73f8f395f1f1e49f956c6be882eed58505a15f3e430962b"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f84fbc98b019fef2ee9a1cb3ce93e3187a6df0b2538a651bfb890254ba9f90b5"}, + {file = "PyYAML-6.0-cp310-cp310-win32.whl", hash = "sha256:2cd5df3de48857ed0544b34e2d40e9fac445930039f3cfe4bcc592a1f836d513"}, + {file = "PyYAML-6.0-cp310-cp310-win_amd64.whl", hash = "sha256:daf496c58a8c52083df09b80c860005194014c3698698d1a57cbcfa182142a3a"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d4b0ba9512519522b118090257be113b9468d804b19d63c71dbcf4a48fa32358"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:81957921f441d50af23654aa6c5e5eaf9b06aba7f0a19c18a538dc7ef291c5a1"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:afa17f5bc4d1b10afd4466fd3a44dc0e245382deca5b3c353d8b757f9e3ecb8d"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dbad0e9d368bb989f4515da330b88a057617d16b6a8245084f1b05400f24609f"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:432557aa2c09802be39460360ddffd48156e30721f5e8d917f01d31694216782"}, + {file = "PyYAML-6.0-cp311-cp311-win32.whl", hash = "sha256:bfaef573a63ba8923503d27530362590ff4f576c626d86a9fed95822a8255fd7"}, + {file = "PyYAML-6.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b45c0191e6d66c470b6cf1b9531a771a83c1c4208272ead47a3ae4f2f603bf"}, + {file = "PyYAML-6.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:897b80890765f037df3403d22bab41627ca8811ae55e9a722fd0392850ec4d86"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50602afada6d6cbfad699b0c7bb50d5ccffa7e46a3d738092afddc1f9758427f"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:48c346915c114f5fdb3ead70312bd042a953a8ce5c7106d5bfb1a5254e47da92"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:98c4d36e99714e55cfbaaee6dd5badbc9a1ec339ebfc3b1f52e293aee6bb71a4"}, + {file = "PyYAML-6.0-cp36-cp36m-win32.whl", hash = "sha256:0283c35a6a9fbf047493e3a0ce8d79ef5030852c51e9d911a27badfde0605293"}, + {file = "PyYAML-6.0-cp36-cp36m-win_amd64.whl", hash = "sha256:07751360502caac1c067a8132d150cf3d61339af5691fe9e87803040dbc5db57"}, + {file = "PyYAML-6.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:819b3830a1543db06c4d4b865e70ded25be52a2e0631ccd2f6a47a2822f2fd7c"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:473f9edb243cb1935ab5a084eb238d842fb8f404ed2193a915d1784b5a6b5fc0"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0ce82d761c532fe4ec3f87fc45688bdd3a4c1dc5e0b4a19814b9009a29baefd4"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:231710d57adfd809ef5d34183b8ed1eeae3f76459c18fb4a0b373ad56bedcdd9"}, + {file = "PyYAML-6.0-cp37-cp37m-win32.whl", hash = "sha256:c5687b8d43cf58545ade1fe3e055f70eac7a5a1a0bf42824308d868289a95737"}, + {file = "PyYAML-6.0-cp37-cp37m-win_amd64.whl", hash = "sha256:d15a181d1ecd0d4270dc32edb46f7cb7733c7c508857278d3d378d14d606db2d"}, + {file = "PyYAML-6.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b4624f379dab24d3725ffde76559cff63d9ec94e1736b556dacdfebe5ab6d4b"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:213c60cd50106436cc818accf5baa1aba61c0189ff610f64f4a3e8c6726218ba"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9fa600030013c4de8165339db93d182b9431076eb98eb40ee068700c9c813e34"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:277a0ef2981ca40581a47093e9e2d13b3f1fbbeffae064c1d21bfceba2030287"}, + {file = "PyYAML-6.0-cp38-cp38-win32.whl", hash = "sha256:d4eccecf9adf6fbcc6861a38015c2a64f38b9d94838ac1810a9023a0609e1b78"}, + {file = "PyYAML-6.0-cp38-cp38-win_amd64.whl", hash = "sha256:1e4747bc279b4f613a09eb64bba2ba602d8a6664c6ce6396a4d0cd413a50ce07"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:055d937d65826939cb044fc8c9b08889e8c743fdc6a32b33e2390f66013e449b"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e61ceaab6f49fb8bdfaa0f92c4b57bcfbea54c09277b1b4f7ac376bfb7a7c174"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d67d839ede4ed1b28a4e8909735fc992a923cdb84e618544973d7dfc71540803"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba8c411ef271aa037d7357a2bc8f9ee8b58b9965831d9e51baf703280dc73d3"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:40527857252b61eacd1d9af500c3337ba8deb8fc298940291486c465c8b46ec0"}, + {file = "PyYAML-6.0-cp39-cp39-win32.whl", hash = "sha256:b5b9eccad747aabaaffbc6064800670f0c297e52c12754eb1d976c57e4f74dcb"}, + {file = "PyYAML-6.0-cp39-cp39-win_amd64.whl", hash = "sha256:b3d267842bf12586ba6c734f89d1f5b871df0273157918b0ccefa29deb05c21c"}, + {file = "PyYAML-6.0.tar.gz", hash = "sha256:68fb519c14306fec9720a2a5b45bc9f0c8d1b9c72adf45c37baedfcd949c35a2"}, +] + +[[package]] +name = "requests" +version = "2.28.2" +description = "Python HTTP for Humans." +category = "main" +optional = false +python-versions = ">=3.7, <4" +files = [ + {file = "requests-2.28.2-py3-none-any.whl", hash = "sha256:64299f4909223da747622c030b781c0d7811e359c37124b4bd368fb8c6518baa"}, + {file = "requests-2.28.2.tar.gz", hash = "sha256:98b1b2782e3c6c4904938b84c0eb932721069dfdb9134313beff7c83c2df24bf"}, +] + +[package.dependencies] +certifi = ">=2017.4.17" +charset-normalizer = ">=2,<4" +idna = ">=2.5,<4" +urllib3 = ">=1.21.1,<1.27" + +[package.extras] +socks = ["PySocks (>=1.5.6,!=1.5.7)"] +use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] + +[[package]] +name = "resolvelib" +version = "0.9.0" +description = "Resolve abstract dependencies into concrete ones" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "resolvelib-0.9.0-py2.py3-none-any.whl", hash = "sha256:597adcbdf81d62d0cde55d90faa8e79187ec0f18e5012df30bd7a751b26343ae"}, + {file = "resolvelib-0.9.0.tar.gz", hash = "sha256:40ab05117c3281b1b160105e10075094c5ab118315003c922b77673a365290e1"}, +] + +[package.extras] +examples = ["html5lib", "packaging", "pygraphviz", "requests"] +lint = ["black", "flake8", "isort", "mypy", "types-requests"] +release = ["build", "towncrier", "twine"] +test = ["commentjson", "packaging", "pytest"] + +[[package]] +name = "rich" +version = "13.3.2" +description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal" +category = "dev" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "rich-13.3.2-py3-none-any.whl", hash = "sha256:a104f37270bf677148d8acb07d33be1569eeee87e2d1beb286a4e9113caf6f2f"}, + {file = "rich-13.3.2.tar.gz", hash = "sha256:91954fe80cfb7985727a467ca98a7618e5dd15178cc2da10f553b36a93859001"}, +] + +[package.dependencies] +markdown-it-py = ">=2.2.0,<3.0.0" +pygments = ">=2.13.0,<3.0.0" + +[package.extras] +jupyter = ["ipywidgets (>=7.5.1,<9)"] + +[[package]] +name = "sentinels" +version = "1.0.0" +description = "Various objects to denote special meanings in python" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "sentinels-1.0.0.tar.gz", hash = "sha256:7be0704d7fe1925e397e92d18669ace2f619c92b5d4eb21a89f31e026f9ff4b1"}, +] + +[[package]] +name = "setuptools" +version = "67.5.1" +description = "Easily download, build, install, upgrade, and uninstall Python packages" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "setuptools-67.5.1-py3-none-any.whl", hash = "sha256:1c39d42bda4cb89f7fdcad52b6762e3c309ec8f8715b27c684176b7d71283242"}, + {file = "setuptools-67.5.1.tar.gz", hash = "sha256:15136a251127da2d2e77ac7a1bc231eb504654f7e3346d93613a13f2e2787535"}, +] + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-hoverxref (<2)", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (==0.8.3)", "sphinx-reredirects", "sphinxcontrib-towncrier"] +testing = ["build[virtualenv]", "filelock (>=3.4.0)", "flake8 (<5)", "flake8-2020", "ini2toml[lite] (>=0.9)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pip (>=19.1)", "pip-run (>=8.8)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-flake8", "pytest-mypy (>=0.9.1)", "pytest-perf", "pytest-timeout", "pytest-xdist", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel"] +testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pytest", "pytest-enabler", "pytest-xdist", "tomli", "virtualenv (>=13.0.0)", "wheel"] + +[[package]] +name = "six" +version = "1.16.0" +description = "Python 2 and 3 compatibility utilities" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "six-1.16.0-py2.py3-none-any.whl", hash = "sha256:8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254"}, + {file = "six-1.16.0.tar.gz", hash = "sha256:1e61c37477a1626458e36f7b1d82aa5c9b094fa4802892072e49de9c60c4c926"}, +] + +[[package]] +name = "smmap" +version = "5.0.0" +description = "A pure Python implementation of a sliding window memory map manager" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "smmap-5.0.0-py3-none-any.whl", hash = "sha256:2aba19d6a040e78d8b09de5c57e96207b09ed71d8e55ce0959eeee6c8e190d94"}, + {file = "smmap-5.0.0.tar.gz", hash = "sha256:c840e62059cd3be204b0c9c9f74be2c09d5648eddd4580d9314c3ecde0b30936"}, +] + +[[package]] +name = "sortedcontainers" +version = "2.4.0" +description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"}, + {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"}, +] + +[[package]] +name = "stevedore" +version = "5.0.0" +description = "Manage dynamic plugins for Python applications" +category = "dev" +optional = false +python-versions = ">=3.8" +files = [ + {file = "stevedore-5.0.0-py3-none-any.whl", hash = "sha256:bd5a71ff5e5e5f5ea983880e4a1dd1bb47f8feebbb3d95b592398e2f02194771"}, + {file = "stevedore-5.0.0.tar.gz", hash = "sha256:2c428d2338976279e8eb2196f7a94910960d9f7ba2f41f3988511e95ca447021"}, +] + +[package.dependencies] +pbr = ">=2.0.0,<2.1.0 || >2.1.0" + +[[package]] +name = "toml" +version = "0.10.2" +description = "Python Library for Tom's Obvious, Minimal Language" +category = "dev" +optional = false +python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b"}, + {file = "toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f"}, +] + +[[package]] +name = "tomli" +version = "2.0.1" +description = "A lil' TOML parser" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tomli-2.0.1-py3-none-any.whl", hash = "sha256:939de3e7a6161af0c887ef91b7d41a53e7c5a1ca976325f429cb46ea9bc30ecc"}, + {file = "tomli-2.0.1.tar.gz", hash = "sha256:de526c12914f0c550d15924c62d72abc48d6fe7364aa87328337a31007fe8a4f"}, +] + +[[package]] +name = "tqdm" +version = "4.65.0" +description = "Fast, Extensible Progress Meter" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"}, + {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[package.extras] +dev = ["py-make (>=0.1.0)", "twine", "wheel"] +notebook = ["ipywidgets (>=6)"] +slack = ["slack-sdk"] +telegram = ["requests"] + +[[package]] +name = "typing-extensions" +version = "4.5.0" +description = "Backported and Experimental Type Hints for Python 3.7+" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "typing_extensions-4.5.0-py3-none-any.whl", hash = "sha256:fb33085c39dd998ac16d1431ebc293a8b3eedd00fd4a32de0ff79002c19511b4"}, + {file = "typing_extensions-4.5.0.tar.gz", hash = "sha256:5cb5f4a79139d699607b3ef622a1dedafa84e115ab0024e0d9c044a9479ca7cb"}, +] + +[[package]] +name = "urllib3" +version = "1.26.14" +description = "HTTP library with thread-safe connection pooling, file post, and more." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*, !=3.5.*" +files = [ + {file = "urllib3-1.26.14-py2.py3-none-any.whl", hash = "sha256:75edcdc2f7d85b137124a6c3c9fc3933cdeaa12ecb9a6a959f22797a0feca7e1"}, + {file = "urllib3-1.26.14.tar.gz", hash = "sha256:076907bf8fd355cde77728471316625a4d2f7e713c125f51953bb5b3eecf4f72"}, +] + +[package.extras] +brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)", "brotlipy (>=0.6.0)"] +secure = ["certifi", "cryptography (>=1.3.4)", "idna (>=2.0.0)", "ipaddress", "pyOpenSSL (>=0.14)", "urllib3-secure-extra"] +socks = ["PySocks (>=1.5.6,!=1.5.7,<2.0)"] + +[[package]] +name = "webencodings" +version = "0.5.1" +description = "Character encoding aliases for legacy web content" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78"}, + {file = "webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923"}, +] + +[metadata] +lock-version = "2.0" +python-versions = "3.9.15" +content-hash = "32e1ab78f1abed143644c7bb574041748bb74ff3e8763051f48b5e253a29db56" diff --git a/jobs/cache_refresh/poetry.toml b/jobs/cache_refresh/poetry.toml new file mode 100644 index 00000000..5fcef8cd --- /dev/null +++ b/jobs/cache_refresh/poetry.toml @@ -0,0 +1,3 @@ +[virtualenvs] +in-project = true +prefer-active-python = true diff --git a/jobs/cache_refresh/pyproject.toml b/jobs/cache_refresh/pyproject.toml new file mode 100644 index 00000000..17fcdbbf --- /dev/null +++ b/jobs/cache_refresh/pyproject.toml @@ -0,0 +1,44 @@ +[tool.poetry] +authors = ["Andrea Soria <[email protected]>"] +description = "Refresh outdated cache entries based on job runner version" +name = "cache_refresh" +version = "0.0.1" +license = "Apache-2.0" + +[tool.poetry.dependencies] +environs = "^9.5.0" +libcommon = {path = "../../libs/libcommon", develop = true} +python = "3.9.15" + +[tool.poetry.group.dev.dependencies] +bandit = "^1.7.4" +black = "^22.12.0" +flake8 = "^3.9.2" +isort = "^5.12.0" +mypy = "^1.0.0" +pip-audit = "^2.4.14" +pytest = "^7.2.1" +pytest-cov = "^2.12.1" + +[build-system] +build-backend = "poetry.core.masonry.api" +requires = ["poetry-core>=1.0.0"] + +[tool.pytest.ini_options] +filterwarnings = ["ignore::DeprecationWarning"] +markers = [ + "wip: tests being developed" +] + +[tool.coverage.run] +source = ["cache_refresh"] + +[tool.isort] +profile = "black" + +[tool.black] +line-length = 119 +preview = true + +[tool.mypy] +strict = true diff --git a/jobs/cache_refresh/src/cache_refresh/__init__.py b/jobs/cache_refresh/src/cache_refresh/__init__.py new file mode 100644 index 00000000..1e9d0c5a --- /dev/null +++ b/jobs/cache_refresh/src/cache_refresh/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. diff --git a/jobs/cache_refresh/src/cache_refresh/config.py b/jobs/cache_refresh/src/cache_refresh/config.py new file mode 100644 index 00000000..4daeb631 --- /dev/null +++ b/jobs/cache_refresh/src/cache_refresh/config.py @@ -0,0 +1,21 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from dataclasses import dataclass, field + +from libcommon.config import CacheConfig, CommonConfig, QueueConfig + + +@dataclass(frozen=True) +class JobConfig: + cache: CacheConfig = field(default_factory=CacheConfig) + common: CommonConfig = field(default_factory=CommonConfig) + queue: QueueConfig = field(default_factory=QueueConfig) + + @classmethod + def from_env(cls) -> "JobConfig": + return cls( + cache=CacheConfig.from_env(), + common=CommonConfig.from_env(), + queue=QueueConfig.from_env(), + ) diff --git a/jobs/cache_refresh/src/cache_refresh/main.py b/jobs/cache_refresh/src/cache_refresh/main.py new file mode 100644 index 00000000..e9e7e7f8 --- /dev/null +++ b/jobs/cache_refresh/src/cache_refresh/main.py @@ -0,0 +1,52 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +import sys + +from libcommon.config import ProcessingGraphConfig +from libcommon.log import init_logging +from libcommon.processing_graph import ProcessingGraph +from libcommon.resources import CacheMongoResource, QueueMongoResource + +from cache_refresh.config import JobConfig +from cache_refresh.outdated_cache import refresh_cache + + +def run_job() -> None: + job_config = JobConfig.from_env() + + init_logging(log_level=job_config.common.log_level) + # ^ set first to have logs as soon as possible + + with ( + CacheMongoResource( + database=job_config.cache.mongo_database, host=job_config.cache.mongo_url + ) as cache_resource, + QueueMongoResource( + database=job_config.queue.mongo_database, host=job_config.queue.mongo_url + ) as queue_resource, + ): + if not cache_resource.is_available(): + logging.warning( + "The connection to the cache database could not be established. The cache refresh job is skipped." + ) + return + if not queue_resource.is_available(): + logging.warning( + "The connection to the queue database could not be established. The cache refresh job is skipped." + ) + return + + processing_graph_config = ProcessingGraphConfig.from_env() + processing_graph = ProcessingGraph(processing_graph_config.specification) + processing_steps = list(processing_graph.steps.values()) + refresh_cache(processing_steps) + + +if __name__ == "__main__": + try: + run_job() + sys.exit(0) + except Exception: + sys.exit(1) diff --git a/jobs/cache_refresh/src/cache_refresh/outdated_cache.py b/jobs/cache_refresh/src/cache_refresh/outdated_cache.py new file mode 100644 index 00000000..9fee3cd7 --- /dev/null +++ b/jobs/cache_refresh/src/cache_refresh/outdated_cache.py @@ -0,0 +1,42 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from typing import List + +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Priority, Queue +from libcommon.simple_cache import get_outdated_split_full_names_for_step + + +def refresh_cache(processing_steps: List[ProcessingStep]) -> None: + queue = Queue() + for processing_step in processing_steps: + current_version = processing_step.job_runner_version + cache_records = get_outdated_split_full_names_for_step(processing_step.cache_kind, current_version) + logging.info( + ( + f"processing_step={processing_step.cache_kind} ", + ( + f"{len(cache_records)} cache entries are outdated for" + f" processing_step={processing_step.cache_kind} (version {current_version}). Creating jobs to" + " update them." + ), + ) + ) + + for cache_info in cache_records: + logging.debug( + ( + f"upsert_job for processing_step={processing_step.job_type}", + f" dataset={cache_info.dataset} config={cache_info.config} split={cache_info.split}", + ), + ) + queue.upsert_job( + job_type=processing_step.job_type, + dataset=cache_info.dataset, + config=cache_info.config, + split=cache_info.split, + force=False, + priority=Priority.LOW, + ) diff --git a/jobs/cache_refresh/tests/__init__.py b/jobs/cache_refresh/tests/__init__.py new file mode 100644 index 00000000..1e9d0c5a --- /dev/null +++ b/jobs/cache_refresh/tests/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. diff --git a/jobs/cache_refresh/tests/conftest.py b/jobs/cache_refresh/tests/conftest.py new file mode 100644 index 00000000..793b4e18 --- /dev/null +++ b/jobs/cache_refresh/tests/conftest.py @@ -0,0 +1,43 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from typing import Iterator + +from libcommon.queue import _clean_queue_database +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import _clean_cache_database +from pytest import MonkeyPatch, fixture + +from cache_refresh.config import JobConfig + + +# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200 +@fixture(scope="session") +def monkeypatch_session() -> Iterator[MonkeyPatch]: + monkeypatch_session = MonkeyPatch() + monkeypatch_session.setenv("CACHE_MONGO_DATABASE", "datasets_server_cache_test") + monkeypatch_session.setenv("QUEUE_MONGO_DATABASE", "datasets_server_queue_test") + yield monkeypatch_session + monkeypatch_session.undo() + + +@fixture(scope="session") +def job_config(monkeypatch_session: MonkeyPatch) -> JobConfig: + job_config = JobConfig.from_env() + if "test" not in job_config.cache.mongo_database or "test" not in job_config.queue.mongo_database: + raise ValueError("Test must be launched on a test mongo database") + return job_config + + +@fixture(autouse=True) +def cache_mongo_resource(job_config: JobConfig) -> Iterator[CacheMongoResource]: + with CacheMongoResource(database=job_config.cache.mongo_database, host=job_config.cache.mongo_url) as resource: + yield resource + _clean_cache_database() + + +@fixture(autouse=True) +def queue_mongo_resource(job_config: JobConfig) -> Iterator[QueueMongoResource]: + with QueueMongoResource(database=job_config.queue.mongo_database, host=job_config.queue.mongo_url) as resource: + yield resource + _clean_queue_database() diff --git a/jobs/cache_refresh/tests/test_outdated_cache.py b/jobs/cache_refresh/tests/test_outdated_cache.py new file mode 100644 index 00000000..fe0c1f52 --- /dev/null +++ b/jobs/cache_refresh/tests/test_outdated_cache.py @@ -0,0 +1,48 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from http import HTTPStatus + +import pytest +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import Queue +from libcommon.simple_cache import upsert_response + +from cache_refresh.outdated_cache import refresh_cache + + [email protected]( + "job_runner_version, upserts_job", + [ + (0, True), + (1, False), + (2, False), + ], +) +def test_refresh_cache( + job_runner_version: int, + upserts_job: bool, +) -> None: + processing_steps = [ + ProcessingStep( + name="/config-names", + input_type="dataset", + requires=None, + required_by_dataset_viewer=False, + parent=None, + ancestors=[], + children=[], + job_runner_version=1, + ) + ] + queue = Queue() + assert not queue.is_job_in_process(job_type="/config-names", dataset="dataset") + upsert_response( + kind="/config-names", + dataset="dataset", + content={}, + http_status=HTTPStatus.OK, + job_runner_version=job_runner_version, + ) + refresh_cache(processing_steps=processing_steps) + assert queue.is_job_in_process(job_type="/config-names", dataset="dataset") == upserts_job diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index a511b792..0e95033e 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -10,0 +11,15 @@ from environs import Env +from libcommon.constants import ( + PROCESSING_STEP_CONFIG_NAMES_VERSION, + PROCESSING_STEP_CONFIG_PARQUET_VERSION, + PROCESSING_STEP_CONFIG_SIZE_VERSION, + PROCESSING_STEP_DATASET_INFO_VERSION, + PROCESSING_STEP_DATASET_PARQUET_VERSION, + PROCESSING_STEP_DATASET_SIZE_VERSION, + PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION, + PROCESSING_STEP_FIRST_ROWS_VERSION, + PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION, + PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, + PROCESSING_STEP_SPLITS_VERSION, +) @@ -99,3 +114,11 @@ class ProcessingGraphConfig: - "/config-names": {"input_type": "dataset"}, - "/split-names-from-streaming": {"input_type": "config", "requires": "/config-names"}, - "/splits": {"input_type": "dataset", "required_by_dataset_viewer": True}, # to be deprecated + "/config-names": {"input_type": "dataset", "job_runner_version": PROCESSING_STEP_CONFIG_NAMES_VERSION}, + "/split-names-from-streaming": { + "input_type": "config", + "requires": "/config-names", + "job_runner_version": PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION, + }, + "/splits": { + "input_type": "dataset", + "required_by_dataset_viewer": True, + "job_runner_version": PROCESSING_STEP_SPLITS_VERSION, + }, # to be deprecated @@ -105,0 +129,35 @@ class ProcessingGraphConfig: + "job_runner_version": PROCESSING_STEP_FIRST_ROWS_VERSION, + }, + "/parquet-and-dataset-info": { + "input_type": "dataset", + "job_runner_version": PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION, + }, + "config-parquet": { + "input_type": "config", + "requires": "/parquet-and-dataset-info", + "job_runner_version": PROCESSING_STEP_CONFIG_PARQUET_VERSION, + }, + "dataset-parquet": { + "input_type": "dataset", + "requires": "config-parquet", + "job_runner_version": PROCESSING_STEP_DATASET_PARQUET_VERSION, + }, + "/dataset-info": { + "input_type": "dataset", + "requires": "/parquet-and-dataset-info", + "job_runner_version": PROCESSING_STEP_DATASET_INFO_VERSION, + }, + "/split-names-from-dataset-info": { + "input_type": "config", + "requires": "/dataset-info", + "job_runner_version": PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, + }, + "config-size": { + "input_type": "config", + "requires": "/parquet-and-dataset-info", + "job_runner_version": PROCESSING_STEP_CONFIG_SIZE_VERSION, + }, + "dataset-size": { + "input_type": "dataset", + "requires": "config-size", + "job_runner_version": PROCESSING_STEP_DATASET_SIZE_VERSION, @@ -107,7 +164,0 @@ class ProcessingGraphConfig: - "/parquet-and-dataset-info": {"input_type": "dataset"}, - "config-parquet": {"input_type": "config", "requires": "/parquet-and-dataset-info"}, - "dataset-parquet": {"input_type": "dataset", "requires": "config-parquet"}, - "/dataset-info": {"input_type": "dataset", "requires": "/parquet-and-dataset-info"}, - "/split-names-from-dataset-info": {"input_type": "config", "requires": "/dataset-info"}, - "config-size": {"input_type": "config", "requires": "/parquet-and-dataset-info"}, - "dataset-size": {"input_type": "dataset", "requires": "config-size"}, @@ -116,0 +168 @@ class ProcessingGraphConfig: + "job_runner_version": PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION, @@ -120,0 +173 @@ class ProcessingGraphConfig: + "job_runner_version": PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index a9ea4c95..4875ef66 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -7,0 +8,14 @@ QUEUE_TTL_SECONDS = 604_800 # 7 days + +PROCESSING_STEP_CONFIG_NAMES_VERSION = 1 +PROCESSING_STEP_CONFIG_PARQUET_VERSION = 3 +PROCESSING_STEP_CONFIG_SIZE_VERSION = 1 +PROCESSING_STEP_DATASET_INFO_VERSION = 1 +PROCESSING_STEP_DATASET_PARQUET_VERSION = 1 +PROCESSING_STEP_DATASET_SIZE_VERSION = 1 +PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION = 1 +PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION = 1 +PROCESSING_STEP_FIRST_ROWS_VERSION = 2 +PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION = 1 +PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION = 2 +PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION = 2 +PROCESSING_STEP_SPLITS_VERSION = 2 diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index 55d8adc4..b6ba8308 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -19,0 +20 @@ class ProcessingStepSpecification(_ProcessingStepSpecification, total=False): + job_runner_version: int @@ -42,0 +44 @@ class ProcessingStep: + job_runner_version: int @@ -102,0 +105 @@ class ProcessingGraph: + job_runner_version=specification["job_runner_version"], diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index b54933e8..37287402 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -368,0 +369,9 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo +def get_outdated_split_full_names_for_step(kind: str, current_version: int) -> List[SplitFullName]: + responses = CachedResponse.objects(kind=kind, job_runner_version__lt=current_version).only( + "dataset", "config", "split" + ) + return [ + SplitFullName(dataset=response.dataset, config=response.config, split=response.split) for response in responses + ] + + diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index addace5b..b939fa39 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -26,0 +27 @@ from libcommon.simple_cache import ( + get_outdated_split_full_names_for_step, @@ -618,0 +620,27 @@ def test_stress_get_cache_reports(num_entries: int) -> None: + + +def test_get_outdated_split_full_names_for_step() -> None: + kind = "kind" + current_version = 2 + minor_version = 1 + + result = get_outdated_split_full_names_for_step(kind=kind, current_version=current_version) + upsert_response( + kind=kind, + dataset="dataset_with_current_version", + content={}, + http_status=HTTPStatus.OK, + job_runner_version=current_version, + ) + assert not result + + upsert_response( + kind=kind, + dataset="dataset_with_minor_version", + content={}, + http_status=HTTPStatus.OK, + job_runner_version=minor_version, + ) + result = get_outdated_split_full_names_for_step(kind=kind, current_version=current_version) + assert result + assert len(result) == 1 diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py index 6afb9357..816151ce 100644 --- a/services/api/tests/routes/test_valid.py +++ b/services/api/tests/routes/test_valid.py @@ -18,0 +19 @@ dataset_step = ProcessingStep( + job_runner_version=1, @@ -27,0 +29 @@ config_step = ProcessingStep( + job_runner_version=1, @@ -36,0 +39 @@ split_step = ProcessingStep( + job_runner_version=1, diff --git a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py index a94762b7..eebeab16 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_dataset_info.py @@ -7,0 +8 @@ from typing import Any, List, Literal, Mapping, Optional +from libcommon.constants import PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION @@ -109 +110 @@ class SplitNamesFromDatasetInfoJobRunner(DatasetsBasedJobRunner): - return 2 + return PROCESSING_STEP_SPLIT_NAMES_FROM_DATASET_INFO_VERSION diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index 0d2c7008..82e1fa98 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -9,0 +10 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError +from libcommon.constants import PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION @@ -113 +114 @@ class SplitNamesFromStreamingJobRunner(DatasetsBasedJobRunner): - return 2 + return PROCESSING_STEP_SPLIT_NAMES_FROM_STREAMING_VERSION diff --git a/services/worker/src/worker/job_runners/config_names.py b/services/worker/src/worker/job_runners/config_names.py index 3ecccf11..df6442b4 100644 --- a/services/worker/src/worker/job_runners/config_names.py +++ b/services/worker/src/worker/job_runners/config_names.py @@ -9,0 +10 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError +from libcommon.constants import PROCESSING_STEP_CONFIG_NAMES_VERSION @@ -105 +106 @@ class ConfigNamesJobRunner(DatasetsBasedJobRunner): - return 1 + return PROCESSING_STEP_CONFIG_NAMES_VERSION diff --git a/services/worker/src/worker/job_runners/config_parquet.py b/services/worker/src/worker/job_runners/config_parquet.py index 5f0b3030..6f3a2803 100644 --- a/services/worker/src/worker/job_runners/config_parquet.py +++ b/services/worker/src/worker/job_runners/config_parquet.py @@ -7,0 +8 @@ from typing import Any, List, Literal, Mapping, Optional, TypedDict +from libcommon.constants import PROCESSING_STEP_CONFIG_PARQUET_VERSION @@ -133 +134 @@ class ConfigParquetJobRunner(JobRunner): - return 3 + return PROCESSING_STEP_CONFIG_PARQUET_VERSION diff --git a/services/worker/src/worker/job_runners/config_size.py b/services/worker/src/worker/job_runners/config_size.py index 933de199..95b23b3b 100644 --- a/services/worker/src/worker/job_runners/config_size.py +++ b/services/worker/src/worker/job_runners/config_size.py @@ -7,0 +8 @@ from typing import Any, Literal, Mapping, Optional, TypedDict +from libcommon.constants import PROCESSING_STEP_CONFIG_SIZE_VERSION @@ -194 +195 @@ class ConfigSizeJobRunner(JobRunner): - return 1 + return PROCESSING_STEP_CONFIG_SIZE_VERSION diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py index 700fb262..65a41e6a 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py +++ b/services/worker/src/worker/job_runners/dataset/split_names_from_dataset_info.py @@ -7,0 +8,3 @@ from typing import Any, List, Literal, Mapping, Optional, Tuple +from libcommon.constants import ( + PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION, +) @@ -145 +148 @@ class DatasetSplitNamesFromDatasetInfoJobRunner(JobRunner): - return 1 + return PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_DATASET_INFO_VERSION diff --git a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py index ac9ea187..0d02287f 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/dataset/split_names_from_streaming.py @@ -7,0 +8,3 @@ from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict +from libcommon.constants import ( + PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION, +) @@ -158 +161 @@ class DatasetSplitNamesFromStreamingJobRunner(JobRunner): - return 1 + return PROCESSING_STEP_DATASET_SPLIT_NAMES_FROM_STREAMING_VERSION diff --git a/services/worker/src/worker/job_runners/dataset_info.py b/services/worker/src/worker/job_runners/dataset_info.py index 0d45dc81..00be4cb2 100644 --- a/services/worker/src/worker/job_runners/dataset_info.py +++ b/services/worker/src/worker/job_runners/dataset_info.py @@ -7,0 +8 @@ from typing import Any, Literal, Mapping, Optional, TypedDict +from libcommon.constants import PROCESSING_STEP_DATASET_INFO_VERSION @@ -97 +98 @@ class DatasetInfoJobRunner(JobRunner): - return 1 + return PROCESSING_STEP_DATASET_INFO_VERSION diff --git a/services/worker/src/worker/job_runners/dataset_parquet.py b/services/worker/src/worker/job_runners/dataset_parquet.py index aa69cce8..4820b779 100644 --- a/services/worker/src/worker/job_runners/dataset_parquet.py +++ b/services/worker/src/worker/job_runners/dataset_parquet.py @@ -7,0 +8 @@ from typing import Any, List, Literal, Mapping, Optional, Tuple, TypedDict +from libcommon.constants import PROCESSING_STEP_DATASET_PARQUET_VERSION @@ -153 +154 @@ class DatasetParquetJobRunner(JobRunner): - return 1 + return PROCESSING_STEP_DATASET_PARQUET_VERSION diff --git a/services/worker/src/worker/job_runners/dataset_size.py b/services/worker/src/worker/job_runners/dataset_size.py index 49d3b495..d606fc62 100644 --- a/services/worker/src/worker/job_runners/dataset_size.py +++ b/services/worker/src/worker/job_runners/dataset_size.py @@ -7,0 +8 @@ from typing import Any, Literal, Mapping, Optional, Tuple, TypedDict +from libcommon.constants import PROCESSING_STEP_DATASET_SIZE_VERSION @@ -181 +182 @@ class DatasetSizeJobRunner(JobRunner): - return 1 + return PROCESSING_STEP_DATASET_SIZE_VERSION diff --git a/services/worker/src/worker/job_runners/first_rows.py b/services/worker/src/worker/job_runners/first_rows.py index b21202c0..c8631024 100644 --- a/services/worker/src/worker/job_runners/first_rows.py +++ b/services/worker/src/worker/job_runners/first_rows.py @@ -34,0 +35 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError +from libcommon.constants import PROCESSING_STEP_FIRST_ROWS_VERSION @@ -649 +650 @@ class FirstRowsJobRunner(DatasetsBasedJobRunner): - return 2 + return PROCESSING_STEP_FIRST_ROWS_VERSION diff --git a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py index 0cd5189f..acd400cd 100644 --- a/services/worker/src/worker/job_runners/parquet_and_dataset_info.py +++ b/services/worker/src/worker/job_runners/parquet_and_dataset_info.py @@ -41,0 +42 @@ from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFo +from libcommon.constants import PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION @@ -873 +874 @@ class ParquetAndDatasetInfoJobRunner(DatasetsBasedJobRunner): - return 1 + return PROCESSING_STEP_PARQUET_AND_DATASET_INFO_VERSION diff --git a/services/worker/src/worker/job_runners/splits.py b/services/worker/src/worker/job_runners/splits.py index 7751a912..8914ed5b 100644 --- a/services/worker/src/worker/job_runners/splits.py +++ b/services/worker/src/worker/job_runners/splits.py @@ -9,0 +10 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError +from libcommon.constants import PROCESSING_STEP_SPLITS_VERSION @@ -134 +135 @@ class SplitsJobRunner(DatasetsBasedJobRunner): - return 2 + return PROCESSING_STEP_SPLITS_VERSION diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 49c949a1..47163ea9 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -124,0 +125 @@ def test_processing_step() -> ProcessingStep: + job_runner_version=1, diff --git a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py index bfd6dbfa..d9f31437 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_dataset_info.py @@ -56,0 +57 @@ def get_job_runner( + job_runner_version=SplitNamesFromDatasetInfoJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index e457e9d6..91855c8a 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -56,0 +57 @@ def get_job_runner( + job_runner_version=SplitNamesFromStreamingJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py index 01adbf29..127a15bf 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_dataset_info.py @@ -52,0 +53 @@ def get_job_runner( + job_runner_version=DatasetSplitNamesFromDatasetInfoJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py index 48490bd3..19ca85ad 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/dataset/test_split_names_from_streaming.py @@ -52,0 +53 @@ def get_job_runner( + job_runner_version=DatasetSplitNamesFromStreamingJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__datasets_based_worker.py index 38059247..9af31da0 100644 --- a/services/worker/tests/job_runners/test__datasets_based_worker.py +++ b/services/worker/tests/job_runners/test__datasets_based_worker.py @@ -77,0 +78 @@ def get_job_runner( + job_runner_version=DummyJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_config_names.py b/services/worker/tests/job_runners/test_config_names.py index 29c53611..28dd2955 100644 --- a/services/worker/tests/job_runners/test_config_names.py +++ b/services/worker/tests/job_runners/test_config_names.py @@ -53,0 +54 @@ def get_job_runner( + job_runner_version=ConfigNamesJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_config_parquet.py b/services/worker/tests/job_runners/test_config_parquet.py index 6850a899..820bb6ce 100644 --- a/services/worker/tests/job_runners/test_config_parquet.py +++ b/services/worker/tests/job_runners/test_config_parquet.py @@ -66,0 +67 @@ def get_job_runner( + job_runner_version=ConfigParquetJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_config_size.py b/services/worker/tests/job_runners/test_config_size.py index 35add26e..546a2829 100644 --- a/services/worker/tests/job_runners/test_config_size.py +++ b/services/worker/tests/job_runners/test_config_size.py @@ -61,0 +62 @@ def get_job_runner( + job_runner_version=ConfigSizeJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_dataset_info.py b/services/worker/tests/job_runners/test_dataset_info.py index 739ca2b6..ff2d4138 100644 --- a/services/worker/tests/job_runners/test_dataset_info.py +++ b/services/worker/tests/job_runners/test_dataset_info.py @@ -60,0 +61 @@ def get_job_runner( + job_runner_version=DatasetInfoJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_dataset_parquet.py b/services/worker/tests/job_runners/test_dataset_parquet.py index ad435715..9d77b955 100644 --- a/services/worker/tests/job_runners/test_dataset_parquet.py +++ b/services/worker/tests/job_runners/test_dataset_parquet.py @@ -74,0 +75 @@ def get_job_runner( + job_runner_version=DatasetParquetJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_dataset_size.py b/services/worker/tests/job_runners/test_dataset_size.py index 72b53aa3..8b4bc1a1 100644 --- a/services/worker/tests/job_runners/test_dataset_size.py +++ b/services/worker/tests/job_runners/test_dataset_size.py @@ -68,0 +69 @@ def get_job_runner( + job_runner_version=DatasetSizeJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_first_rows.py b/services/worker/tests/job_runners/test_first_rows.py index 9fae0ae7..15ab3186 100644 --- a/services/worker/tests/job_runners/test_first_rows.py +++ b/services/worker/tests/job_runners/test_first_rows.py @@ -59,0 +60 @@ def get_job_runner( + job_runner_version=FirstRowsJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py index 39f55cd6..4ae94cff 100644 --- a/services/worker/tests/job_runners/test_parquet_and_dataset_info.py +++ b/services/worker/tests/job_runners/test_parquet_and_dataset_info.py @@ -95,0 +96 @@ def get_job_runner( + job_runner_version=ParquetAndDatasetInfoJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/job_runners/test_splits.py b/services/worker/tests/job_runners/test_splits.py index b68f8800..25d4370b 100644 --- a/services/worker/tests/job_runners/test_splits.py +++ b/services/worker/tests/job_runners/test_splits.py @@ -53,0 +54 @@ def get_job_runner( + job_runner_version=SplitsJobRunner.get_job_runner_version(), diff --git a/services/worker/tests/test_job_runner.py b/services/worker/tests/test_job_runner.py index af0cee4e..2ac22706 100644 --- a/services/worker/tests/test_job_runner.py +++ b/services/worker/tests/test_job_runner.py @@ -230,0 +231 @@ def test_check_type( + job_runner_version=1, @@ -252,4 +253,4 @@ def test_create_children_jobs() -> None: - "/dummy": {"input_type": "dataset"}, - "/child-dataset": {"input_type": "dataset", "requires": "/dummy"}, - "/child-config": {"input_type": "config", "requires": "/dummy"}, - "/child-split": {"input_type": "split", "requires": "/dummy"}, + "/dummy": {"input_type": "dataset", "job_runner_version": 1}, + "/child-dataset": {"input_type": "dataset", "requires": "/dummy", "job_runner_version": 1}, + "/child-config": {"input_type": "config", "requires": "/dummy", "job_runner_version": 1}, + "/child-split": {"input_type": "split", "requires": "/dummy", "job_runner_version": 1},
6212793896a15d537f7ba1ff508197c02e7c8e8a
Sylvain Lesage
2023-03-20T09:11:40
feat: 🎸 reduce the number of allowed jobs for one namespace (#961)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index c104cf7d..54129648 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -194 +194 @@ workers: - maxJobsPerNamespace: 20 + maxJobsPerNamespace: 5 @@ -209,2 +209,2 @@ workers: - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" + maxJobsPerNamespace: 2 + workerOnlyJobTypes: "/config-names,config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info"
2a5a2b7d3e01b4b11d7a4161d8387469838f9c0a
Sylvain Lesage
2023-03-20T08:51:22
feat: 🎸 authenticate webhooks sent by the Hub (#955)
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 09be9749..77fedca1 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -57,0 +58,4 @@ secrets: + hfWebhookSecret: + fromSecret: false + secretName: "webhook-secret" + value: "" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 60eacd2d..c104cf7d 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -56,0 +57,3 @@ secrets: + hfWebhookSecret: + fromSecret: true + secretName: "webhook-secret" diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl index 08a088cf..35f357c0 100644 --- a/chart/templates/services/api/_container.tpl +++ b/chart/templates/services/api/_container.tpl @@ -20,0 +21,10 @@ + - name: API_HF_WEBHOOK_SECRET + {{- if .Values.secrets.hfWebhookSecret.fromSecret }} + valueFrom: + secretKeyRef: + name: {{ .Values.secrets.hfWebhookSecret.secretName | quote }} + key: WEBHOOK_SECRET + optional: false + {{- else }} + value: {{ .Values.secrets.hfWebhookSecret.value }} + {{- end }} diff --git a/chart/values.yaml b/chart/values.yaml index 1e1c6394..2bfcca2a 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -73,0 +74,4 @@ secrets: + hfWebhookSecret: + fromSecret: false + secretName: "webhook-secret" + value: "" diff --git a/e2e/tests/test_11_auth.py b/e2e/tests/test_11_auth.py index 6c693ddf..3101e914 100644 --- a/e2e/tests/test_11_auth.py +++ b/e2e/tests/test_11_auth.py @@ -16,2 +15,0 @@ from .utils import get_default_config_split, poll_until_ready_and_assert - # gated: webhook_status_code is 200 because the access is asked for the app token, not the user token - # (which is not passed to the webhook request) @@ -21,2 +18,0 @@ from .utils import get_default_config_split, poll_until_ready_and_assert - # private: webhook_status_code is 400 because the access is asked for the app token, which has no - # access to the private datasets. As a consequence, no data in the cache diff --git a/e2e/tests/utils.py b/e2e/tests/utils.py index 176e60be..c65753f1 100644 --- a/e2e/tests/utils.py +++ b/e2e/tests/utils.py @@ -8 +8 @@ from pathlib import Path -from typing import Any, Literal, Mapping, Optional, Tuple +from typing import Any, Mapping, Optional, Tuple @@ -101,9 +100,0 @@ def get_default_config_split(dataset: str) -> Tuple[str, str, str]: -def post_webhook( - dataset: str, event: Literal["add", "remove", "update", "move"], movedTo: Optional[str] = None -) -> Response: - json: dict[str, Any] = {"event": event, "repo": {"type": "dataset", "name": dataset}} - if movedTo is not None: - json["movedTo"] = movedTo - return post("/webhook", json=json) - - diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py index fd59a273..4c05ec6f 100644 --- a/libs/libcommon/src/libcommon/operations.py +++ b/libs/libcommon/src/libcommon/operations.py @@ -79,50 +78,0 @@ def delete_dataset(dataset: str) -> None: -def move_dataset( - from_dataset: str, - to_dataset: str, - init_processing_steps: List[ProcessingStep], - hf_endpoint: str, - hf_token: Optional[str] = None, - force: bool = False, - priority: Priority = Priority.NORMAL, - hf_timeout_seconds: Optional[float] = None, -) -> None: - """ - Move a dataset - - Note that the implementation is simply to add or update the new dataset, then delete the old one in case of - success. - - Args: - from_dataset (str): the dataset to move - to_dataset (str): the destination dataset - init_processing_steps (List[ProcessingStep]): the processing steps that must be run when updating a dataset - hf_endpoint (str): the HF endpoint - hf_token (Optional[str], optional): The HF token. Defaults to None. - force (bool, optional): Force the update. Defaults to False. - priority (Priority, optional): The priority of the job. Defaults to Priority.NORMAL. - hf_timeout_seconds (Optional[float], optional): The timeout for requests to the hub. None means no timeout. - Defaults to None. - - Returns: None. - - Raises: - - [`~libcommon.dataset.AskAccessHubRequestError`]: if the request to the Hub to get access to the - dataset failed or timed out. - - [`~libcommon.dataset.DatasetInfoHubRequestError`]: if the request to the Hub to get the dataset - info failed or timed out. - - [`~libcommon.dataset.DatasetError`]: if the dataset could not be accessed or is not supported - """ - logging.debug(f"move dataset '{from_dataset}' to '{to_dataset}'") - update_dataset( - dataset=to_dataset, - init_processing_steps=init_processing_steps, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - force=force, - priority=priority, - hf_timeout_seconds=hf_timeout_seconds, - ) - # ^ can raise - delete_dataset(dataset=from_dataset) - - diff --git a/services/api/README.md b/services/api/README.md index 2b9d5b8b..3a147851 100644 --- a/services/api/README.md +++ b/services/api/README.md @@ -16,0 +17 @@ Set environment variables to configure the application (`API_` prefix): +- `API_HF_WEBHOOK_SECRET`: a shared secret sent by the Hub in the "X-Webhook-Secret" header of POST requests sent to /webhook, to authenticate the originator and bypass some validation of the content (avoiding roundtrip to the Hub). If not set, all the validations are done. Defaults to empty. diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index 5d594000..ec2d5494 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -114,0 +115 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + hf_webhook_secret=app_config.api.hf_webhook_secret, diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 346cc3a6..eb87e1d2 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -42,0 +43 @@ API_HF_TIMEOUT_SECONDS = 0.2 +API_HF_WEBHOOK_SECRET = None @@ -53,0 +55 @@ class ApiConfig: + hf_webhook_secret: Optional[str] = API_HF_WEBHOOK_SECRET @@ -68,0 +71 @@ class ApiConfig: + hf_webhook_secret=env.str(name="HF_WEBHOOK_SECRET", default=API_HF_WEBHOOK_SECRET), diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py index 8c5498b6..5ec9d42f 100644 --- a/services/api/src/api/routes/webhook.py +++ b/services/api/src/api/routes/webhook.py @@ -9 +9 @@ from libcommon.dataset import DatasetError -from libcommon.operations import delete_dataset, move_dataset, update_dataset +from libcommon.operations import delete_dataset, update_dataset @@ -63,0 +64 @@ def process_payload( + trust_sender: bool = False, @@ -79,0 +81 @@ def process_payload( + do_check_support=False, # always create a job, even if the dataset is not supported @@ -81,16 +83,16 @@ def process_payload( - elif event == "remove": - delete_dataset(dataset=dataset) - elif event == "move": - moved_to = payload["movedTo"] - if moved_to is None: - return - move_dataset( - from_dataset=dataset, - to_dataset=moved_to, - init_processing_steps=init_processing_steps, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - force=False, - priority=Priority.NORMAL, - hf_timeout_seconds=hf_timeout_seconds, - ) + elif trust_sender: + # destructive actions (delete, move) require a trusted sender + if event == "move" and (moved_to := payload["movedTo"]): + update_dataset( + dataset=moved_to, + init_processing_steps=init_processing_steps, + hf_token=hf_token, + hf_endpoint=hf_endpoint, + force=False, + priority=Priority.NORMAL, + hf_timeout_seconds=hf_timeout_seconds, + do_check_support=False, + ) + delete_dataset(dataset=dataset) + elif event == "remove": + delete_dataset(dataset=dataset) @@ -102,0 +105 @@ def create_webhook_endpoint( + hf_webhook_secret: Optional[str] = None, @@ -115 +118 @@ def create_webhook_endpoint( - with StepProfiler(method="webhook_endpoint", step="parse payload"): + with StepProfiler(method="webhook_endpoint", step="parse payload and headers"): @@ -127,0 +131,9 @@ def create_webhook_endpoint( + HEADER = "x-webhook-secret" + trust_sender = ( + hf_webhook_secret is not None + and (secret := request.headers.get(HEADER)) is not None + and secret == hf_webhook_secret + ) + if not trust_sender: + logging.info(f"/webhook: the sender is not trusted. JSON: {json}") + @@ -135,0 +148 @@ def create_webhook_endpoint( + trust_sender=trust_sender,
c8b2b696067335b017b9e16f3ac79d8d39e79a29
Andrea Francis Soria Jimenez
2023-03-17T16:43:25
Adding ligth worker (#945)
diff --git a/chart/env/dev.yaml b/chart/env/dev.yaml index 57ff300e..09be9749 100644 --- a/chart/env/dev.yaml +++ b/chart/env/dev.yaml @@ -162 +162 @@ workers: - deployName: "generic" + deployName: "all" @@ -176 +176 @@ workers: - deployName: "first-rows" + deployName: "light" @@ -178,15 +178 @@ workers: - workerOnlyJobTypes: "/first-rows" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 100m - memory: "512Mi" - limits: - cpu: 1 - memory: "4Gi" - tolerations: [] - - - deployName: "parquet-and-dataset-info" - maxJobsPerNamespace: 1 - workerOnlyJobTypes: "/parquet-and-dataset-info" + workerOnlyJobTypes: "config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 77bd813b..60eacd2d 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -190 +190 @@ workers: - deployName: "generic" + deployName: "all" @@ -195 +195 @@ workers: - replicas: 12 + replicas: 22 @@ -205 +205 @@ workers: - deployName: "config-names" + deployName: "light" @@ -207,121 +207 @@ workers: - workerOnlyJobTypes: "/config-names" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 1 - memory: "2Gi" - limits: - cpu: 2 - memory: "8Gi" - tolerations: [] - - - deployName: "split-names-from-streaming" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "/split-names-from-streaming" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 1 - memory: "2Gi" - limits: - cpu: 2 - memory: "8Gi" - tolerations: [] - - - deployName: "splits" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "/splits" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 1 - memory: "2Gi" - limits: - cpu: 2 - memory: "8Gi" - tolerations: [] - - - deployName: "first-rows" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "/first-rows" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 1 - memory: "2Gi" - limits: - cpu: 2 - memory: "8Gi" - tolerations: [] - - - deployName: "parquet-and-dataset-info" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "/parquet-and-dataset-info" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 1 - memory: "2Gi" - limits: - cpu: 2 - memory: "8Gi" - tolerations: [] - - - deployName: "parquet" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "dataset-parquet,config-parquet" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 200m - memory: "100Mi" - limits: - cpu: 2 - memory: "1Gi" - tolerations: [] - - - deployName: "dataset-info" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "/dataset-info" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 200m - memory: "100Mi" - limits: - cpu: 2 - memory: "1Gi" - tolerations: [] - - - deployName: "split-names-from-dataset-info" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "/split-names-from-dataset-info" - nodeSelector: - role-datasets-server-worker: "true" - replicas: 2 - resources: - requests: - cpu: 200m - memory: "100Mi" - limits: - cpu: 2 - memory: "1Gi" - tolerations: [] - - - deployName: "sizes" - maxJobsPerNamespace: 20 - workerOnlyJobTypes: "config-size,dataset-size" + workerOnlyJobTypes: "config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info" diff --git a/chart/values.yaml b/chart/values.yaml index 744663c0..1e1c6394 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -293 +293 @@ workers: - deployName: "generic" + deployName: "all" @@ -308 +308 @@ workers: - deployName: "config-names" + deployName: "light" @@ -312,121 +312 @@ workers: - workerOnlyJobTypes: "/config-names" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 1 - limits: - cpu: 1 - tolerations: [] - - - # name of the deployment - deployName: "split-names-from-streaming" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "/split-names-from-streaming" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 1 - limits: - cpu: 1 - tolerations: [] - - - # name of the deployment - deployName: "splits" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "/splits" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - - - # name of the deployment - deployName: "first-rows" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "/first-rows" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - - - # name of the deployment - deployName: "parquet-and-dataset-info" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "/parquet-and-dataset-info" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - - - # name of the deployment - deployName: "parquet" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "dataset-parquet,config-parquet" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - - - # name of the deployment - deployName: "dataset-info" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "/dataset-info" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - - - # name of the deployment - deployName: "split-names-from-dataset-info" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "/split-names-from-dataset-info" - nodeSelector: {} - replicas: 1 - resources: - requests: - cpu: 0 - limits: - cpu: 0 - tolerations: [] - - - # name of the deployment - deployName: "sizes" - # Maximum number of jobs running at the same time for the same namespace - maxJobsPerNamespace: 1 - # job types that this worker can process - workerOnlyJobTypes: "config-size,dataset-size" + workerOnlyJobTypes: "config-parquet,dataset-parquet,/dataset-info,/split-names-from-dataset-info,config-size,dataset-size,dataset-split-names-from-streaming,dataset-split-names-from-dataset-info"