hash
stringlengths
40
40
authorName
stringclasses
42 values
authorEmail
stringclasses
41 values
date
timestamp[ms]date
2021-07-26 09:52:55
2025-07-18 10:19:56
subject
stringlengths
11
116
diff
stringlengths
0
987k
6215b9d5fa0155ad23fca5151e69d2428c43ad53
Sylvain Lesage
2023-07-05T15:05:59
feat: 🎸 add "difficulty" field to JobDocument (#1489)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 24ea7c61..b8e32aef 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.15.0 +version: 1.15.1 diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 61bdd330..ab27b0c7 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -294,0 +295,2 @@ workers: + workerDifficultyMax: "" + workerDifficultyMin: "" @@ -309,0 +312,2 @@ workers: + workerDifficultyMax: "" + workerDifficultyMin: "" diff --git a/chart/env/staging.yaml b/chart/env/staging.yaml index 5027dfd9..64848c3e 100644 --- a/chart/env/staging.yaml +++ b/chart/env/staging.yaml @@ -235,0 +236,2 @@ workers: + workerDifficultyMax: "" + workerDifficultyMin: "" @@ -249,0 +252,2 @@ workers: + workerDifficultyMax: "" + workerDifficultyMin: "" diff --git a/chart/templates/worker/_container.tpl b/chart/templates/worker/_container.tpl index f9b86817..2bfc9ab0 100644 --- a/chart/templates/worker/_container.tpl +++ b/chart/templates/worker/_container.tpl @@ -18,0 +19,4 @@ + - name: WORKER_DIFFICULTY_MAX + value: {{ .workerValues.workerDifficultyMax | quote }} + - name: WORKER_DIFFICULTY_MIN + value: {{ .workerValues.workerDifficultyMin | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index e8f01d0b..e9736b2e 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -457,0 +458,4 @@ workers: + # max difficulty of the jobs that this worker will process + workerDifficultyMax: "" + # min difficulty of the jobs that this worker will process + workerDifficultyMin: "" diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py index aed47435..8811b02f 100644 --- a/jobs/cache_maintenance/tests/test_collect_metrics.py +++ b/jobs/cache_maintenance/tests/test_collect_metrics.py @@ -28 +28,6 @@ def test_collect_metrics() -> None: - job_type=processing_step.job_type, dataset="dataset", revision="revision", config="config", split="split" + job_type=processing_step.job_type, + dataset="dataset", + revision="revision", + config="config", + split="split", + difficulty=50, diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index eaf51193..7430144b 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -52,0 +53,3 @@ from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_con +from mongodb_migration.migrations._20230705160600_queue_job_add_difficulty import ( + MigrationQueueAddDifficultyToJob, +) @@ -243,0 +247 @@ class MigrationsCollector: + MigrationQueueAddDifficultyToJob(version="20230705160600", description="add 'difficulty' field to jobs"), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230705160600_queue_job_add_difficulty.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230705160600_queue_job_add_difficulty.py new file mode 100644 index 00000000..f02ed2fe --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230705160600_queue_job_add_difficulty.py @@ -0,0 +1,44 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from libcommon.config import ProcessingGraphConfig +from libcommon.constants import ( + DEFAULT_DIFFICULTY, + QUEUE_COLLECTION_JOBS, + QUEUE_MONGOENGINE_ALIAS, +) +from libcommon.queue import JobDocument +from mongoengine.connection import get_db + +from mongodb_migration.check import check_documents +from mongodb_migration.migration import Migration + + +# connection already occurred in the main.py (caveat: we use globals) +class MigrationQueueAddDifficultyToJob(Migration): + def up(self) -> None: + logging.info("If missing, add the difficulty with a value that depends on the job type, else 50") + db = get_db(QUEUE_MONGOENGINE_ALIAS) + processing_graph_config = ProcessingGraphConfig() + for job_type, spec in processing_graph_config.specification.items(): + difficulty = spec.get("difficulty", DEFAULT_DIFFICULTY) + db[QUEUE_COLLECTION_JOBS].update_many( + {"type": job_type, "difficulty": {"$exists": False}}, + {"$set": {"difficulty": difficulty}}, + ) + db[QUEUE_COLLECTION_JOBS].update_many( + {"difficulty": {"$exists": False}}, + {"$set": {"difficulty": DEFAULT_DIFFICULTY}}, + ) + + def down(self) -> None: + logging.info("Remove the difficulty field from all the jobs") + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].update_many({}, {"$unset": {"difficulty": ""}}) + + def validate(self) -> None: + logging.info("Ensure that a random selection of jobs have the 'difficulty' field set") + + check_documents(DocCls=JobDocument, sample_size=10) diff --git a/jobs/mongodb_migration/tests/migrations/__init__.py b/jobs/mongodb_migration/tests/migrations/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py index 7fced24f..52bfd45c 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py @@ -27,0 +28 @@ def test_queue_delete_indexes_with_force(mongo_host: str) -> None: + difficulty=50, diff --git a/jobs/mongodb_migration/tests/migrations/test_20230705160600_queue_job_add_difficulty.py b/jobs/mongodb_migration/tests/migrations/test_20230705160600_queue_job_add_difficulty.py new file mode 100644 index 00000000..247f9f61 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230705160600_queue_job_add_difficulty.py @@ -0,0 +1,55 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230705160600_queue_job_add_difficulty import ( + MigrationQueueAddDifficultyToJob, +) + + +def test_queue_add_difficulty_to_jobs(mongo_host: str) -> None: + with MongoResource(database="test_queue_add_difficulty_to_jobs", host=mongo_host, mongoengine_alias="queue"): + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_JOBS].insert_many( + [ + { + "type": "test", + "dataset": "test", + "revision": "123", + "unicity_id": "test", + "namespace": "test", + "created_at": "2022-01-01T00:00:00.000000Z", + }, + { + "type": "dataset-is-valid", + "dataset": "test", + "revision": "123", + "unicity_id": "test", + "namespace": "test", + "created_at": "2022-01-01T00:00:00.000000Z", + }, + ] + ) + + migration = MigrationQueueAddDifficultyToJob( + version="20230705160600", + description="add difficulty field to jobs", + ) + migration.up() + + results = list(db[QUEUE_COLLECTION_JOBS].find({"type": "test"})) + assert len(results) == 1 + assert results[0]["difficulty"] == 50 + results = list(db[QUEUE_COLLECTION_JOBS].find({"type": "dataset-is-valid"})) + assert len(results) == 1 + assert results[0]["difficulty"] == 20 + + migration.down() + results = list(db[QUEUE_COLLECTION_JOBS].find({"dataset": "test"})) + assert len(results) == 2 + assert all("difficulty" not in result for result in results) + + db[QUEUE_COLLECTION_JOBS].drop() diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index aa8b7027..ba56516c 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -127,0 +128 @@ def test_queue_delete_ttl_index(mongo_host: str) -> None: + difficulty=50, diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 59006344..472fa14f 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -210,0 +211 @@ class ProcessingGraphConfig: + "difficulty": 50, @@ -216,0 +218 @@ class ProcessingGraphConfig: + "difficulty": 60, @@ -222,0 +225 @@ class ProcessingGraphConfig: + "difficulty": 70, @@ -227,0 +231 @@ class ProcessingGraphConfig: + "difficulty": 70, @@ -233,0 +238 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -239,0 +245 @@ class ProcessingGraphConfig: + "difficulty": 50, @@ -245,0 +252 @@ class ProcessingGraphConfig: + "difficulty": 40, @@ -250,0 +258 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -255,0 +264 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -260,0 +270 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -266,0 +277 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -272,0 +284 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -277,0 +290 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -286,0 +300 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -296,0 +311 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -301,0 +317 @@ class ProcessingGraphConfig: + "difficulty": 40, @@ -306,0 +323 @@ class ProcessingGraphConfig: + "difficulty": 70, @@ -311,0 +329 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -320,0 +339 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -325,0 +345 @@ class ProcessingGraphConfig: + "difficulty": 20, @@ -334,0 +355 @@ class ProcessingGraphConfig: + "difficulty": 70, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 5bb181a5..19dc97e7 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -17,0 +18 @@ QUEUE_TTL_SECONDS = 600 # 10 minutes +DEFAULT_DIFFICULTY = 50 diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py index 1a5669b0..24be4e57 100644 --- a/libs/libcommon/src/libcommon/orchestrator.py +++ b/libs/libcommon/src/libcommon/orchestrator.py @@ -296,0 +297 @@ class AfterJobPlan(Plan): + "difficulty": next_processing_step.difficulty, @@ -565,0 +567 @@ class DatasetBackfillPlan(Plan): + "difficulty": artifact_state.processing_step.difficulty, diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index f6f1e636..0a04a70c 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -22 +22,5 @@ import networkx as nx -from libcommon.constants import DEFAULT_INPUT_TYPE, DEFAULT_JOB_RUNNER_VERSION +from libcommon.constants import ( + DEFAULT_DIFFICULTY, + DEFAULT_INPUT_TYPE, + DEFAULT_JOB_RUNNER_VERSION, +) @@ -56,0 +61 @@ class ProcessingStepSpecification(TypedDict, total=False): + difficulty: int @@ -82,0 +88 @@ class ProcessingStep: + difficulty: int @@ -100,0 +107 @@ class ProcessingStep: + difficulty=self.difficulty, @@ -194,0 +202 @@ class ProcessingGraph: + difficulty=specification.get("difficulty", DEFAULT_DIFFICULTY), diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 57031369..a9af875b 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -20 +20 @@ from mongoengine.errors import DoesNotExist, NotUniqueError -from mongoengine.fields import DateTimeField, EnumField, StringField +from mongoengine.fields import DateTimeField, EnumField, IntField, StringField @@ -71,0 +72 @@ class JobDict(TypedDict): + difficulty: int @@ -110,0 +112,7 @@ class NoWaitingJobError(Exception): +class JobQueryFilters(TypedDict, total=False): + type__nin: List[str] + type__in: List[str] + difficulty__gte: int + difficulty__lte: int + + @@ -130,0 +139 @@ class JobDocument(Document): + difficulty (`int`): The difficulty of the job: 0=easy, 100=hard as a convention. @@ -170,0 +180 @@ class JobDocument(Document): + difficulty = IntField(required=True) @@ -186,0 +197 @@ class JobDocument(Document): + "difficulty": self.difficulty, @@ -206,0 +218 @@ class JobDocument(Document): + "difficulty": self.difficulty, @@ -227,0 +240 @@ class JobDocument(Document): + "difficulty": self.difficulty, @@ -336,2 +349,2 @@ class Queue: - - a job can be in the queue only once (unicity_id) in the "started" or "waiting" state - - a job can be in the queue multiple times in the other states (success, error, cancelled) + - a job can be in the queue only once (unicity_id) in the "started" state + - a job can be in the queue multiple times in the other states @@ -338,0 +352 @@ class Queue: + - a job has a difficulty (from 0: easy to 100: hard, as a convention) @@ -347,0 +362 @@ class Queue: + difficulty: int, @@ -360,0 +376 @@ class Queue: + difficulty (`int`): The difficulty of the job. @@ -377,0 +394 @@ class Queue: + difficulty=difficulty, @@ -408,0 +426 @@ class Queue: + difficulty=job_info["difficulty"], @@ -437,0 +456,2 @@ class Queue: + difficulty_min: Optional[int] = None, + difficulty_max: Optional[int] = None, @@ -449,0 +470,2 @@ class Queue: + difficulty_min: if not None, only jobs with a difficulty greater or equal to this value are considered. + difficulty_max: if not None, only jobs with a difficulty lower or equal to this value are considered. @@ -462 +484 @@ class Queue: - filters = {} + filters: JobQueryFilters = {} @@ -466,0 +489,4 @@ class Queue: + if difficulty_min: + filters["difficulty__gte"] = difficulty_min + if difficulty_max: + filters["difficulty__lte"] = difficulty_max @@ -523 +549,5 @@ class Queue: - self, job_types_blocked: Optional[list[str]] = None, job_types_only: Optional[list[str]] = None + self, + difficulty_min: Optional[int] = None, + difficulty_max: Optional[int] = None, + job_types_blocked: Optional[list[str]] = None, + job_types_only: Optional[list[str]] = None, @@ -533,0 +564,2 @@ class Queue: + difficulty_min: if not None, only jobs with a difficulty greater or equal to this value are considered. + difficulty_max: if not None, only jobs with a difficulty lower or equal to this value are considered. @@ -545 +577,5 @@ class Queue: - priority=priority, job_types_blocked=job_types_blocked, job_types_only=job_types_only + priority=priority, + job_types_blocked=job_types_blocked, + job_types_only=job_types_only, + difficulty_min=difficulty_min, + difficulty_max=difficulty_max, @@ -607 +643,5 @@ class Queue: - self, job_types_blocked: Optional[list[str]] = None, job_types_only: Optional[list[str]] = None + self, + difficulty_min: Optional[int] = None, + difficulty_max: Optional[int] = None, + job_types_blocked: Optional[list[str]] = None, + job_types_only: Optional[list[str]] = None, @@ -614,0 +655,2 @@ class Queue: + difficulty_min: if not None, only jobs with a difficulty greater or equal to this value are considered. + difficulty_max: if not None, only jobs with a difficulty lower or equal to this value are considered. @@ -629 +671,4 @@ class Queue: - job_types_blocked=job_types_blocked, job_types_only=job_types_only + job_types_blocked=job_types_blocked, + job_types_only=job_types_only, + difficulty_min=difficulty_min, + difficulty_max=difficulty_max, @@ -767 +812,6 @@ class Queue: - job_type=job.type, dataset=job.dataset, revision=job.revision, config=job.config, split=job.split + job_type=job.type, + dataset=job.dataset, + revision=job.revision, + config=job.config, + split=job.split, + difficulty=job.difficulty, diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py index ac79f831..1fe129e7 100644 --- a/libs/libcommon/src/libcommon/utils.py +++ b/libs/libcommon/src/libcommon/utils.py @@ -38,0 +39 @@ class JobInfo(TypedDict): + difficulty: int @@ -49,0 +51 @@ class FlatJobInfo(TypedDict): + difficulty: int diff --git a/libs/libcommon/tests/test_backfill.py b/libs/libcommon/tests/test_backfill.py index d69c231a..ddb58f60 100644 --- a/libs/libcommon/tests/test_backfill.py +++ b/libs/libcommon/tests/test_backfill.py @@ -35,0 +36 @@ from .utils import ( + DIFFICULTY, @@ -815 +816,3 @@ def test_delete_jobs( - job = queue.add_job(job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority) + job = queue.add_job( + job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority, difficulty=DIFFICULTY + ) diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py index 32a687c4..8563cad5 100644 --- a/libs/libcommon/tests/test_orchestrator.py +++ b/libs/libcommon/tests/test_orchestrator.py @@ -27,0 +28 @@ from .utils import ( + DIFFICULTY, @@ -153,0 +155 @@ def test_finish_job( + difficulty=DIFFICULTY, diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 792905ee..eaf064b3 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -37,0 +38 @@ def test_add_job() -> None: + test_difficulty = 50 @@ -41 +42 @@ def test_add_job() -> None: - job1 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + job1 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty) @@ -43 +44 @@ def test_add_job() -> None: - job2 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + job2 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty) @@ -58 +59 @@ def test_add_job() -> None: - job3 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + job3 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty) @@ -92,0 +94 @@ def test_cancel_jobs_by_job_id( + test_difficulty = 50 @@ -98 +100 @@ def test_cancel_jobs_by_job_id( - job = queue.add_job(job_type=test_type, dataset=job_id, revision="test_revision") + job = queue.add_job(job_type=test_type, dataset=job_id, revision="test_revision", difficulty=test_difficulty) @@ -126,0 +129 @@ def test_priority_logic_creation_order() -> None: + test_difficulty = 50 @@ -128,2 +131,16 @@ def test_priority_logic_creation_order() -> None: - queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1") - queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2") + queue.add_job( + job_type=test_type, + dataset="dataset1", + revision=test_revision, + config="config", + split="split1", + difficulty=test_difficulty, + ) + queue.add_job( + job_type=test_type, + dataset="dataset1", + revision=test_revision, + config="config", + split="split2", + difficulty=test_difficulty, + ) @@ -138,0 +156 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None: + test_difficulty = 50 @@ -140,3 +158,24 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None: - queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1") - queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2") - queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1") + queue.add_job( + job_type=test_type, + dataset="dataset1", + revision=test_revision, + config="config", + split="split1", + difficulty=test_difficulty, + ) + queue.add_job( + job_type=test_type, + dataset="dataset1", + revision=test_revision, + config="config", + split="split2", + difficulty=test_difficulty, + ) + queue.add_job( + job_type=test_type, + dataset="dataset2", + revision=test_revision, + config="config", + split="split1", + difficulty=test_difficulty, + ) @@ -153,0 +193 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None: + test_difficulty = 50 @@ -155,3 +194,0 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None: - queue.add_job(job_type=test_type, dataset="org1/dataset1", revision=test_revision, config="config", split="split1") - queue.add_job(job_type=test_type, dataset="org1/dataset2", revision=test_revision, config="config", split="split1") - queue.add_job(job_type=test_type, dataset="org2/dataset2", revision=test_revision, config="config", split="split1") @@ -159 +196,30 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None: - job_type=test_type, dataset="no_org_dataset3", revision=test_revision, config="config", split="split1" + job_type=test_type, + dataset="org1/dataset1", + revision=test_revision, + config="config", + split="split1", + difficulty=test_difficulty, + ) + queue.add_job( + job_type=test_type, + dataset="org1/dataset2", + revision=test_revision, + config="config", + split="split1", + difficulty=test_difficulty, + ) + queue.add_job( + job_type=test_type, + dataset="org2/dataset2", + revision=test_revision, + config="config", + split="split1", + difficulty=test_difficulty, + ) + queue.add_job( + job_type=test_type, + dataset="no_org_dataset3", + revision=test_revision, + config="config", + split="split1", + difficulty=test_difficulty, @@ -172,0 +239 @@ def test_priority_logic_priority_order() -> None: + test_difficulty = 50 @@ -179,0 +247 @@ def test_priority_logic_priority_order() -> None: + difficulty=test_difficulty, @@ -187,0 +256 @@ def test_priority_logic_priority_order() -> None: + difficulty=test_difficulty, @@ -197 +266 @@ def test_priority_logic_priority_order() -> None: - "job_type,job_types_blocked,job_types_only,should_raise", + "job_types_blocked,job_types_only,should_raise", @@ -199,10 +268,10 @@ def test_priority_logic_priority_order() -> None: - ("test_type", None, None, False), - ("test_type", None, ["test_type"], False), - ("test_type", ["other_type"], None, False), - ("test_type", ["other_type"], ["test_type"], False), - ("test_type", None, ["other_type"], True), - ("test_type", ["test_type"], None, True), - ("test_type", ["test_type"], ["test_type"], True), - ("test_type", ["other_type", "test_type"], None, True), - ("test_type", ["other_type"], ["other_type"], True), - ("test_type", ["other_type", "test_type"], ["other_type", "test_type"], True), + (None, None, False), + (None, ["test_type"], False), + (["other_type"], None, False), + (["other_type"], ["test_type"], False), + (None, ["other_type"], True), + (["test_type"], None, True), + (["test_type"], ["test_type"], True), + (["other_type", "test_type"], None, True), + (["other_type"], ["other_type"], True), + (["other_type", "test_type"], ["other_type", "test_type"], True), @@ -212 +281 @@ def test_job_types_only( - job_type: str, job_types_blocked: Optional[list[str]], job_types_only: Optional[list[str]], should_raise: bool + job_types_blocked: Optional[list[str]], job_types_only: Optional[list[str]], should_raise: bool @@ -213,0 +283 @@ def test_job_types_only( + job_type = "test_type" @@ -215,0 +286 @@ def test_job_types_only( + test_difficulty = 50 @@ -217 +288,8 @@ def test_job_types_only( - queue.add_job(job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None) + queue.add_job( + job_type=job_type, + dataset=test_dataset, + revision=test_revision, + config=None, + split=None, + difficulty=test_difficulty, + ) @@ -228,0 +307,39 @@ def test_job_types_only( [email protected]( + "difficulty_min,difficulty_max,should_raise", + [ + (None, None, False), + (None, 60, False), + (40, None, False), + (40, 60, False), + (50, 50, False), + (None, 40, True), + (60, None, True), + (60, 60, True), + (40, 40, True), + ], +) +def test_difficulty(difficulty_min: Optional[int], difficulty_max: Optional[int], should_raise: bool) -> None: + job_type = "test_type" + test_dataset = "test_dataset" + test_revision = "test_revision" + test_difficulty = 50 + queue = Queue() + queue.add_job( + job_type=job_type, + dataset=test_dataset, + revision=test_revision, + config=None, + split=None, + difficulty=test_difficulty, + ) + assert queue.is_job_in_process( + job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None + ) + if should_raise: + with pytest.raises(EmptyQueueError): + queue.start_job(difficulty_max=difficulty_max, difficulty_min=difficulty_min) + else: + job_info = queue.start_job(difficulty_max=difficulty_max, difficulty_min=difficulty_min) + assert job_info["params"]["dataset"] == test_dataset + + @@ -233,0 +351 @@ def test_count_by_status() -> None: + test_difficulty = 50 @@ -242 +360 @@ def test_count_by_status() -> None: - queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty) @@ -247 +365 @@ def test_count_by_status() -> None: - queue.add_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision) + queue.add_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty) @@ -255,0 +374 @@ def test_get_dataset_pending_jobs_for_type() -> None: + test_difficulty = 50 @@ -266 +385,8 @@ def test_get_dataset_pending_jobs_for_type() -> None: - queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) + queue.add_job( + job_type=job_type, + dataset=dataset, + revision=test_revision, + config=config, + split=None, + difficulty=test_difficulty, + ) @@ -272 +398,8 @@ def test_get_dataset_pending_jobs_for_type() -> None: - queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) + queue.add_job( + job_type=job_type, + dataset=dataset, + revision=test_revision, + config=config, + split=None, + difficulty=test_difficulty, + ) @@ -277 +410,8 @@ def test_get_dataset_pending_jobs_for_type() -> None: - queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) + queue.add_job( + job_type=job_type, + dataset=dataset, + revision=test_revision, + config=config, + split=None, + difficulty=test_difficulty, + ) @@ -287,0 +428 @@ def test_queue_heartbeat() -> None: + test_difficulty = 50 @@ -289 +430,8 @@ def test_queue_heartbeat() -> None: - job = queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1") + job = queue.add_job( + job_type=job_type, + dataset="dataset1", + revision="revision", + config="config", + split="split1", + difficulty=test_difficulty, + ) @@ -300,0 +449 @@ def test_queue_get_zombies() -> None: + test_difficulty = 50 @@ -304 +453,6 @@ def test_queue_get_zombies() -> None: - job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1" + job_type=job_type, + dataset="dataset1", + revision="revision", + config="config", + split="split1", + difficulty=test_difficulty, @@ -307 +461,8 @@ def test_queue_get_zombies() -> None: - queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split2") + queue.add_job( + job_type=job_type, + dataset="dataset1", + revision="revision", + config="config", + split="split2", + difficulty=test_difficulty, + ) diff --git a/libs/libcommon/tests/utils.py b/libs/libcommon/tests/utils.py index e4addd22..3ee60d34 100644 --- a/libs/libcommon/tests/utils.py +++ b/libs/libcommon/tests/utils.py @@ -34,0 +35 @@ JOB_TYPE = "job_type" +DIFFICULTY = 50 @@ -363,0 +365 @@ def artifact_id_to_job_info(artifact_id: str) -> JobInfo: + difficulty=DIFFICULTY, diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py index 33855962..df11c6f8 100644 --- a/services/admin/src/admin/app.py +++ b/services/admin/src/admin/app.py @@ -131,0 +132 @@ def create_app() -> Starlette: + difficulty=processing_step.difficulty, diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py index 3f37f937..f54814b1 100644 --- a/services/admin/src/admin/routes/force_refresh.py +++ b/services/admin/src/admin/routes/force_refresh.py @@ -28,0 +29 @@ def create_force_refresh_endpoint( + difficulty: int, @@ -64,0 +66 @@ def create_force_refresh_endpoint( + difficulty=difficulty, diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index eb5f4886..7ce3d60a 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -164 +164 @@ def test_get_cache_entry_from_steps() -> None: - queue.add_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config) + queue.add_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config, difficulty=50) diff --git a/services/worker/README.md b/services/worker/README.md index 9f794786..778a5ffb 100644 --- a/services/worker/README.md +++ b/services/worker/README.md @@ -13,0 +14,2 @@ Set environment variables to configure the worker. +- `WORKER_DIFFICULTY_MAX`: the maximum difficulty of the jobs to process. Defaults to None. +- `WORKER_DIFFICULTY_MIN`: the minimum difficulty of the jobs to process. Defaults to None. diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index d39a3c03..25ba41fa 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -18,0 +19,2 @@ WORKER_CONTENT_MAX_BYTES = 10_000_000 +WORKER_DIFFICULTY_MAX = None +WORKER_DIFFICULTY_MIN = None @@ -38,0 +41,2 @@ class WorkerConfig: + difficulty_max: Optional[int] = WORKER_DIFFICULTY_MAX + difficulty_min: Optional[int] = WORKER_DIFFICULTY_MIN @@ -58,0 +63,2 @@ class WorkerConfig: + difficulty_max=env.int(name="DIFFICULTY_MAX", default=WORKER_DIFFICULTY_MAX), + difficulty_min=env.int(name="DIFFICULTY_MIN", default=WORKER_DIFFICULTY_MIN), diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py index c0303b21..59e3f51b 100644 --- a/services/worker/src/worker/loop.py +++ b/services/worker/src/worker/loop.py @@ -129,0 +130,2 @@ class Loop: + difficulty_min=self.app_config.worker.difficulty_min, + difficulty_max=self.app_config.worker.difficulty_max, diff --git a/services/worker/tests/job_runners/config/test_config_job_runner.py b/services/worker/tests/job_runners/config/test_config_job_runner.py index 01a5ce8e..77a9ebec 100644 --- a/services/worker/tests/job_runners/config/test_config_job_runner.py +++ b/services/worker/tests/job_runners/config/test_config_job_runner.py @@ -41,0 +42 @@ def test_failed_creation(test_processing_step: ProcessingStep, app_config: AppCo + "difficulty": 50, @@ -62,0 +64 @@ def test_success_creation(test_processing_step: ProcessingStep, app_config: AppC + "difficulty": 50, diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py index 49be6df2..2b9b1983 100644 --- a/services/worker/tests/job_runners/config/test_info.py +++ b/services/worker/tests/job_runners/config/test_info.py @@ -165,0 +166 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py index bde0b947..8a06c6fb 100644 --- a/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py +++ b/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py @@ -59,0 +60 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index 1189453f..223f8f69 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -59,0 +60 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index a78ba301..2054d393 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -106,0 +107 @@ def get_job_runner( + "difficulty": 50, @@ -693,0 +695 @@ def get_dataset_config_names_job_runner( + "difficulty": 50, @@ -722,0 +725 @@ def launch_job_runner(job_runner_args: JobRunnerArgs) -> CompleteJobResult: + difficulty=50, @@ -728,0 +732 @@ def launch_job_runner(job_runner_args: JobRunnerArgs) -> CompleteJobResult: + difficulty=50, @@ -761,0 +766 @@ def test_concurrency( + difficulty=50, diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py index 571a8f8e..9e0079d6 100644 --- a/services/worker/tests/job_runners/config/test_parquet_metadata.py +++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py @@ -80,0 +81 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py index df89b2e2..33692159 100644 --- a/services/worker/tests/job_runners/config/test_size.py +++ b/services/worker/tests/job_runners/config/test_size.py @@ -58,0 +59 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/config/test_split_names_from_info.py b/services/worker/tests/job_runners/config/test_split_names_from_info.py index ee4b7ccb..5a06b413 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_info.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_info.py @@ -53,0 +54 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index 2e5f3215..c978f96a 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -57,0 +58 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py index dc6aafe2..3f2fcfd2 100644 --- a/services/worker/tests/job_runners/dataset/test_config_names.py +++ b/services/worker/tests/job_runners/dataset/test_config_names.py @@ -52,0 +53 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py index 49f7fa06..4fe427c9 100644 --- a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py +++ b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py @@ -42,0 +43 @@ def test_failed_creation(test_processing_step: ProcessingStep, app_config: AppCo + "difficulty": 50, @@ -63,0 +65 @@ def test_success_creation(test_processing_step: ProcessingStep, app_config: AppC + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py index 0d6b8158..385b5fbc 100644 --- a/services/worker/tests/job_runners/dataset/test_info.py +++ b/services/worker/tests/job_runners/dataset/test_info.py @@ -137,0 +138 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_is_valid.py b/services/worker/tests/job_runners/dataset/test_is_valid.py index a6c64001..fc9a9d66 100644 --- a/services/worker/tests/job_runners/dataset/test_is_valid.py +++ b/services/worker/tests/job_runners/dataset/test_is_valid.py @@ -106,0 +107 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py index 37f959de..71d37579 100644 --- a/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py +++ b/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py @@ -56,0 +57 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index 28c04d3f..8d53cad6 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -58,0 +59 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_size.py b/services/worker/tests/job_runners/dataset/test_size.py index 76326d1d..3d2e38d7 100644 --- a/services/worker/tests/job_runners/dataset/test_size.py +++ b/services/worker/tests/job_runners/dataset/test_size.py @@ -57,0 +58 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/dataset/test_split_names.py b/services/worker/tests/job_runners/dataset/test_split_names.py index 1afcf32b..6832f6e9 100644 --- a/services/worker/tests/job_runners/dataset/test_split_names.py +++ b/services/worker/tests/job_runners/dataset/test_split_names.py @@ -48,0 +49 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/split/test_duckdb_index.py b/services/worker/tests/job_runners/split/test_duckdb_index.py index a8fe40cf..ead4aed4 100644 --- a/services/worker/tests/job_runners/split/test_duckdb_index.py +++ b/services/worker/tests/job_runners/split/test_duckdb_index.py @@ -72,0 +73 @@ def get_job_runner( + "difficulty": 50, @@ -114,0 +116 @@ def get_parquet_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index 3dcb37c1..980eb479 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -67,0 +68 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index 9f255de8..9fdb0bad 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -64,0 +65 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/split/test_image_url_columns.py b/services/worker/tests/job_runners/split/test_image_url_columns.py index 68fb6f01..d49d4bd7 100644 --- a/services/worker/tests/job_runners/split/test_image_url_columns.py +++ b/services/worker/tests/job_runners/split/test_image_url_columns.py @@ -59,0 +60 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py index e3898d3c..399b4d8a 100644 --- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py +++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py @@ -61,0 +62 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py index 2de50d1d..72678d0e 100644 --- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py @@ -77,0 +78 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/split/test_split_job_runner.py b/services/worker/tests/job_runners/split/test_split_job_runner.py index 80d88ba4..2afde623 100644 --- a/services/worker/tests/job_runners/split/test_split_job_runner.py +++ b/services/worker/tests/job_runners/split/test_split_job_runner.py @@ -42,0 +43 @@ def test_failed_creation(test_processing_step: ProcessingStep, app_config: AppCo + "difficulty": 50, @@ -63,0 +65 @@ def test_success_creation(test_processing_step: ProcessingStep, app_config: AppC + "difficulty": 50, diff --git a/services/worker/tests/job_runners/test__job_runner_with_cache.py b/services/worker/tests/job_runners/test__job_runner_with_cache.py index 11aa9fd1..2ff25f6c 100644 --- a/services/worker/tests/job_runners/test__job_runner_with_cache.py +++ b/services/worker/tests/job_runners/test__job_runner_with_cache.py @@ -68,0 +69 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py index 2667da40..afb25020 100644 --- a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py +++ b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py @@ -72,0 +73 @@ def get_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py index 3e30ca26..165c8038 100644 --- a/services/worker/tests/test_executor.py +++ b/services/worker/tests/test_executor.py @@ -44,0 +45 @@ def get_job_info(prefix: str = "base") -> JobInfo: + difficulty=50, @@ -128,0 +130 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I + difficulty=job_info["difficulty"], @@ -161,0 +164 @@ def set_long_running_job_in_queue( + difficulty=job_info["difficulty"], @@ -191,0 +195 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato + difficulty=job_info["difficulty"], diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py index a6e54257..a01f4249 100644 --- a/services/worker/tests/test_job_manager.py +++ b/services/worker/tests/test_job_manager.py @@ -73,0 +74 @@ def test_check_type( + difficulty=50, @@ -95,0 +97 @@ def test_check_type( + difficulty=50, @@ -134,0 +137 @@ def test_backfill(priority: Priority, app_config: AppConfig) -> None: + difficulty=50, @@ -204,0 +208 @@ def test_job_runner_set_crashed( + difficulty=50, @@ -262,0 +267 @@ def test_raise_if_parallel_response_exists( + difficulty=50, @@ -293,0 +299 @@ def test_doesnotexist(app_config: AppConfig) -> None: + difficulty=50, diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py index e10bc8c0..84765d48 100644 --- a/services/worker/tests/test_job_runner_factory.py +++ b/services/worker/tests/test_job_runner_factory.py @@ -63,0 +64 @@ def test_create_job_runner( + "difficulty": 50, diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py index 2a766b92..3097e354 100644 --- a/services/worker/tests/test_loop.py +++ b/services/worker/tests/test_loop.py @@ -72 +72,3 @@ def test_process_next_job( - loop.queue.add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split) + loop.queue.add_job( + job_type=job_type, dataset=dataset, revision=revision, config=config, split=split, difficulty=50 + )
abf206283f7242efebc6d5fc601e75a2b758a5b0
Quentin Lhoest
2023-07-05T14:51:01
Block config-parquet-metadata job for light workers (#1490)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index d50a3bcd..61bdd330 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -311 +311 @@ workers: - workerJobTypesOnly: "config-parquet,config-parquet-metadata,dataset-parquet,config-info,dataset-info,config-split-names-from-info,config-size,dataset-size,dataset-split-names,dataset-is-valid,split-image-url-columns,split-opt-in-out-urls-count,config-opt-in-out-urls-count,dataset-opt-in-out-urls-count" + workerJobTypesOnly: "config-parquet,dataset-parquet,config-info,dataset-info,config-split-names-from-info,config-size,dataset-size,dataset-split-names,dataset-is-valid,split-image-url-columns,split-opt-in-out-urls-count,config-opt-in-out-urls-count,dataset-opt-in-out-urls-count"
ad0047367c77c8035e2bcc108dc51190f6562bd7
Quentin Lhoest
2023-07-05T13:00:47
More logging for /rows (#1485)
diff --git a/libs/libapi/src/libapi/utils.py b/libs/libapi/src/libapi/utils.py index 2f2480de..65e7f50f 100644 --- a/libs/libapi/src/libapi/utils.py +++ b/libs/libapi/src/libapi/utils.py @@ -3,0 +4 @@ +import logging @@ -100,0 +102 @@ def try_backfill_dataset( + logging.info(f"Set orchestrator revision for dataset={dataset}, revision={revision}") diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py index 351cb65e..b3bfcbbd 100644 --- a/libs/libcommon/src/libcommon/parquet_utils.py +++ b/libs/libcommon/src/libcommon/parquet_utils.py @@ -399,0 +400,4 @@ class RowsIndex: + logging.info( + f"Create ParquetIndexWithoutMetadata for dataset={self.dataset}, config={self.config}," + f" split={self.split}" + ) @@ -412,0 +417,4 @@ class RowsIndex: + logging.info( + f"Create ParquetIndexWithMetadata for dataset={self.dataset}, config={self.config}," + f" split={self.split}" + ) @@ -439,0 +448,4 @@ class RowsIndex: + logging.info( + f"Query {type(self.parquet_index).__name__} for dataset={self.dataset}, config={self.config}," + f" split={self.split}, offset={offset}, length={length}" + ) diff --git a/services/rows/src/rows/routes/rows.py b/services/rows/src/rows/routes/rows.py index 2529260f..efa0a503 100644 --- a/services/rows/src/rows/routes/rows.py +++ b/services/rows/src/rows/routes/rows.py @@ -313,2 +313,2 @@ def create_rows_endpoint( - with StepProfiler(method="rows_endpoint", step="get row groups index"): - try: + try: + with StepProfiler(method="rows_endpoint", step="get row groups index"): @@ -321,5 +321,6 @@ def create_rows_endpoint( - except CachedArtifactError: - config_parquet_processing_steps = processing_graph.get_config_parquet_processing_steps() - config_parquet_metadata_processing_steps = ( - processing_graph.get_config_parquet_metadata_processing_steps() - ) + except CachedArtifactError: + config_parquet_processing_steps = processing_graph.get_config_parquet_processing_steps() + config_parquet_metadata_processing_steps = ( + processing_graph.get_config_parquet_metadata_processing_steps() + ) + with StepProfiler(method="rows_endpoint", step="try backfill dataset"):
dee3002c03c7aa78720513bd960568065f5644a7
Polina Kazakova
2023-07-05T12:45:38
Delete `/parquet-and-dataset-info` endpoint (#1488)
diff --git a/e2e/tests/test_11_api.py b/e2e/tests/test_11_api.py index 48e60d68..8dd3e822 100644 --- a/e2e/tests/test_11_api.py +++ b/e2e/tests/test_11_api.py @@ -55 +54,0 @@ def test_auth_e2e( - ("/parquet-and-dataset-info", "config"), diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 28a62fc7..a7a95078 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -71,3 +70,0 @@ class EndpointConfig: - "/parquet-and-dataset-info": { - "config": ["config-parquet-and-info"], - }, diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index 37f8f820..eb5f4886 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -51,6 +50,0 @@ def test_endpoints_definition() -> None: - parquet_and_info = definition["/parquet-and-dataset-info"] - assert parquet_and_info is not None - assert sorted(list(parquet_and_info)) == ["config"] - assert parquet_and_info["config"] is not None - assert len(parquet_and_info["config"]) == 1 # Only has one processing step - @@ -91 +85 @@ def test_endpoints_definition() -> None: - # assert old endpoints don't exist + # assert old deleted endpoints don't exist @@ -93,0 +88,2 @@ def test_endpoints_definition() -> None: + with raises(KeyError): + _ = definition["/parquet-and-dataset-info"] diff --git a/services/worker/README.md b/services/worker/README.md index b6299326..9f794786 100644 --- a/services/worker/README.md +++ b/services/worker/README.md @@ -3 +3 @@ -> Worker that pre-computes and caches the response to /splits, /first-rows or /parquet-and-dataset-info. +> Workers that pre-compute and cache the response to /splits, /first-rows, /parquet, /info and /size. @@ -7 +7 @@ -Use environment variables to configure the worker. The prefix of each environment variable gives its scope. +Use environment variables to configure the workers. The prefix of each environment variable gives its scope. @@ -55 +55 @@ If the Hub is not https://huggingface.co (i.e., if you set the `COMMON_HF_ENDPOI -Set environment variables to configure the first rows worker (`FIRST_ROWS_` prefix): +Set environment variables to configure the `first-rows` worker (`FIRST_ROWS_` prefix): @@ -65 +65 @@ Also, set the assets-related configuration for the first-rows worker. See [../.. -### Parquet and dataset info worker +### Parquet and info worker @@ -67 +67 @@ Also, set the assets-related configuration for the first-rows worker. See [../.. -Set environment variables to configure the parquet worker (`PARQUET_AND_INFO_` prefix): +Set environment variables to configure the `parquet-and-info` worker (`PARQUET_AND_INFO_` prefix): @@ -80 +80 @@ Set environment variables to configure the parquet worker (`PARQUET_AND_INFO_` p -Set environment variables to configure the duckdb index worker (`DUCKDB_INDEX_` prefix): +Set environment variables to configure the `duckdb-index` worker (`DUCKDB_INDEX_` prefix): @@ -92 +92 @@ Set environment variables to configure the duckdb index worker (`DUCKDB_INDEX_` -The splits worker does not need any additional configuration. +The `splits` worker does not need any additional configuration.
4a8a64c9b19ae649214341f79c2c8cae3d637793
Sylvain Lesage
2023-07-05T11:23:32
feat: 🎸 query operation $in is faster then $nin (#1486)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 63eef88a..d50a3bcd 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -310,2 +310,2 @@ workers: - workerJobTypesBlocked: "dataset-config-names,config-split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan,split-duckdb-index" - workerJobTypesOnly: "" + workerJobTypesBlocked: "" + workerJobTypesOnly: "config-parquet,config-parquet-metadata,dataset-parquet,config-info,dataset-info,config-split-names-from-info,config-size,dataset-size,dataset-split-names,dataset-is-valid,split-image-url-columns,split-opt-in-out-urls-count,config-opt-in-out-urls-count,dataset-opt-in-out-urls-count"
1528ec93f4b61ec30cc6b7cc950050cfa608edfd
Quentin Lhoest
2023-07-04T20:31:29
minor (#1480)
diff --git a/libs/libcommon/src/libcommon/viewer_utils/asset.py b/libs/libcommon/src/libcommon/viewer_utils/asset.py index 28a8322c..f98eff12 100644 --- a/libs/libcommon/src/libcommon/viewer_utils/asset.py +++ b/libs/libcommon/src/libcommon/viewer_utils/asset.py @@ -50 +50,4 @@ def update_last_modified_date_of_rows_in_assets_dir( - (row_dirs_path / str(row_idx) / DATASETS_SERVER_MDATE_FILENAME).unlink() + try: + (row_dirs_path / str(row_idx) / DATASETS_SERVER_MDATE_FILENAME).unlink() + except FileNotFoundError: + pass
bd372054f7bf14f2430309f964d372cbbe8f6f50
Sylvain Lesage
2023-07-04T17:27:25
feat: 🎸 increase RAM for /rows service (#1484)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index f4f22785..63eef88a 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -287 +287 @@ rows: - memory: "512Mi" + memory: "2Gi" @@ -290 +290 @@ rows: - memory: "4Gi" + memory: "8Gi"
2619bd836a0a008485fcb62a07bfde325b0cdd9e
Sylvain Lesage
2023-07-04T16:26:15
Move the /rows endpoint to its own service (#1479)
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml index 0126945c..c4272b3e 100644 --- a/.github/workflows/_e2e_tests.yml +++ b/.github/workflows/_e2e_tests.yml @@ -34,0 +35,2 @@ jobs: + ROWS_UVICORN_NUM_WORKERS: "2" + ROWS_UVICORN_PORT: "8082" @@ -70,0 +73,2 @@ jobs: + ROWS_UVICORN_NUM_WORKERS: "2" + ROWS_UVICORN_PORT: "8082" diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml index d1394601..770cf6e9 100644 --- a/.github/workflows/cd.yml +++ b/.github/workflows/cd.yml @@ -31,0 +32,2 @@ jobs: + - directory: services + project: rows @@ -110,0 +113,2 @@ jobs: + rows: + tag: sha-${{ steps.vars.outputs.sha_short }} diff --git a/.github/workflows/s-api.yml b/.github/workflows/s-api.yml index 63c440cc..070b4c71 100644 --- a/.github/workflows/s-api.yml +++ b/.github/workflows/s-api.yml @@ -10,0 +11 @@ on: + - "libs/libapi/**" @@ -18,0 +20 @@ on: + - "libs/libapi/**" diff --git a/.github/workflows/s-rows.yml b/.github/workflows/s-rows.yml new file mode 100644 index 00000000..d49ac293 --- /dev/null +++ b/.github/workflows/s-rows.yml @@ -0,0 +1,35 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +name: services/rows +on: + workflow_dispatch: + push: + branches: + - main + paths: + - "libs/libapi/**" + - "libs/libcommon/**" + - "services/rows/**" + - ".github/workflows/s-rows.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" + pull_request: + paths: + - "libs/libapi/**" + - "libs/libcommon/**" + - "services/rows/**" + - ".github/workflows/s-rows.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" +jobs: + quality: + uses: ./.github/workflows/_quality-python.yml + with: + working-directory: services/rows + unit-tests: + uses: ./.github/workflows/_unit-tests-python.yml + with: + working-directory: services/rows diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace index 597c0f2f..b1be9dde 100644 --- a/.vscode/monorepo.code-workspace +++ b/.vscode/monorepo.code-workspace @@ -34,0 +35,4 @@ + { + "name": "services/rows", + "path": "../services/rows" + }, diff --git a/Makefile b/Makefile index bc174bb0..71571e13 100644 --- a/Makefile +++ b/Makefile @@ -4,0 +5 @@ export PORT_API := 8180 +export PORT_ROWS := 8182 @@ -23 +24 @@ start: - MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up + MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up @@ -27 +28 @@ stop: - MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down + MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down @@ -31 +32 @@ dev-start: - MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up + MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up @@ -35 +36 @@ dev-stop: - MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down + MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 732dcf41..24ea7c61 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.14.1 +version: 1.15.0 diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 953aeb10..f4f22785 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -13,0 +14 @@ global: + rows: 31371 @@ -44,0 +46,5 @@ images: + rows: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-services-rows + tag: sha-fb3399a @@ -250,0 +257,21 @@ api: + + nodeSelector: + role-datasets-server: "true" + replicas: 4 + service: + type: NodePort + resources: + requests: + cpu: 1 + memory: "512Mi" + limits: + cpu: 4 + memory: "4Gi" + +rows: + # the timeout in seconds for the requests to the Hugging Face Hub. + hfTimeoutSeconds: "1.5" + # Number of uvicorn workers for running the application + # (2 x $num_cores) + 1 + # https://docs.gunicorn.org/en/stable/design.html#how-many-workers + uvicornNumWorkers: "9" diff --git a/chart/env/staging.yaml b/chart/env/staging.yaml index 3014a6ef..5027dfd9 100644 --- a/chart/env/staging.yaml +++ b/chart/env/staging.yaml @@ -42,0 +43,5 @@ images: + rows: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-services-rows + tag: sha-fb3399a @@ -213,0 +219,14 @@ api: +rows: + uvicornNumWorkers: "1" + + replicas: 1 + service: + type: NodePort + resources: + requests: + cpu: 100m + memory: "512Mi" + limits: + cpu: 1 + memory: "4Gi" + diff --git a/chart/nginx-templates/default.conf.template b/chart/nginx-templates/default.conf.template index ac2edf9c..7efde8a0 100644 --- a/chart/nginx-templates/default.conf.template +++ b/chart/nginx-templates/default.conf.template @@ -49,0 +50,9 @@ server { + location /rows { + proxy_pass ${URL_ROWS}/rows; + proxy_set_header Host $proxy_host; + proxy_set_header X-Real-IP $remote_addr; + proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; + proxy_set_header X-Forwarded-Proto $scheme; + proxy_http_version 1.1; + } + diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index ac9370e6..cb73f80a 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -48,0 +49,4 @@ Docker image management +{{- define "services.rows.image" -}} +{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.services.rows "global" .Values.global.huggingface) }} +{{- end -}} + @@ -100,0 +105,5 @@ app.kubernetes.io/component: "{{ include "name" . }}-api" +{{- define "labels.rows" -}} +{{ include "hf.labels.commons" . }} +app.kubernetes.io/component: "{{ include "name" . }}-rows" +{{- end -}} + @@ -213,0 +223,8 @@ See https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#a-a +{{/* +The URL to access the rows service from another container +See https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#a-aaaa-records +*/}} +{{- define "rows.url" -}} +{{- printf "http://%s-rows.%s.svc.cluster.local:80" ( include "name" . ) ( .Release.Namespace ) }} +{{- end }} + diff --git a/chart/templates/reverse-proxy/_container.tpl b/chart/templates/reverse-proxy/_container.tpl index 5f1a0952..6ae0be54 100644 --- a/chart/templates/reverse-proxy/_container.tpl +++ b/chart/templates/reverse-proxy/_container.tpl @@ -20,0 +21,2 @@ + - name: URL_ROWS + value: {{ include "rows.url" . | quote }} diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl index 923532fc..4c83f00a 100644 --- a/chart/templates/services/api/_container.tpl +++ b/chart/templates/services/api/_container.tpl @@ -9 +8,0 @@ - {{ include "envCachedAssets" . | nindent 2 }} @@ -11 +9,0 @@ - {{ include "envParquetMetadata" . | nindent 2 }} @@ -49,3 +46,0 @@ - volumeMounts: - {{ include "volumeMountCachedAssetsRW" . | nindent 2 }} - {{ include "volumeMountParquetMetadataRO" . | nindent 2 }} diff --git a/chart/templates/services/api/deployment.yaml b/chart/templates/services/api/deployment.yaml index a34cef80..e77b45e3 100644 --- a/chart/templates/services/api/deployment.yaml +++ b/chart/templates/services/api/deployment.yaml @@ -26,3 +25,0 @@ spec: - initContainers: - {{ include "initContainerCachedAssets" . | nindent 8 }} - {{ include "initContainerParquetMetadata" . | nindent 8 }} diff --git a/chart/templates/services/rows/_container.tpl b/chart/templates/services/rows/_container.tpl new file mode 100644 index 00000000..2cb00e86 --- /dev/null +++ b/chart/templates/services/rows/_container.tpl @@ -0,0 +1,66 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +{{- define "containerRows" -}} +- name: "{{ include "name" . }}-rows" + image: {{ include "services.rows.image" . }} + imagePullPolicy: {{ .Values.images.pullPolicy }} + env: + {{ include "envCachedAssets" . | nindent 2 }} + {{ include "envCache" . | nindent 2 }} + {{ include "envParquetMetadata" . | nindent 2 }} + {{ include "envQueue" . | nindent 2 }} + {{ include "envCommon" . | nindent 2 }} + {{ include "envLog" . | nindent 2 }} + {{ include "envNumba" . | nindent 2 }} + # service + - name: API_HF_AUTH_PATH + value: {{ .Values.rows.hfAuthPath | quote }} + - name: API_HF_JWT_PUBLIC_KEY_URL + value: {{ .Values.rows.hfJwtPublicKeyUrl | quote }} + - name: API_HF_JWT_ALGORITHM + value: {{ .Values.rows.hfJwtAlgorithm | quote }} + - name: API_HF_TIMEOUT_SECONDS + value: {{ .Values.rows.hfTimeoutSeconds | quote }} + - name: API_HF_WEBHOOK_SECRET + {{- if .Values.secrets.hfWebhookSecret.fromSecret }} + valueFrom: + secretKeyRef: + name: {{ .Values.secrets.hfWebhookSecret.secretName | quote }} + key: WEBHOOK_SECRET + optional: false + {{- else }} + value: {{ .Values.secrets.hfWebhookSecret.value }} + {{- end }} + - name: API_MAX_AGE_LONG + value: {{ .Values.rows.maxAgeLong | quote }} + - name: API_MAX_AGE_SHORT + value: {{ .Values.rows.maxAgeShort | quote }} + # prometheus + - name: PROMETHEUS_MULTIPROC_DIR + value: {{ .Values.rows.prometheusMultiprocDirectory | quote }} + # uvicorn + - name: API_UVICORN_HOSTNAME + value: {{ .Values.rows.uvicornHostname | quote }} + - name: API_UVICORN_NUM_WORKERS + value: {{ .Values.rows.uvicornNumWorkers | quote }} + - name: API_UVICORN_PORT + value: {{ .Values.rows.uvicornPort | quote }} + volumeMounts: + {{ include "volumeMountCachedAssetsRW" . | nindent 2 }} + {{ include "volumeMountParquetMetadataRO" . | nindent 2 }} + securityContext: + allowPrivilegeEscalation: false + readinessProbe: + tcpSocket: + port: {{ .Values.rows.uvicornPort }} + livenessProbe: + tcpSocket: + port: {{ .Values.rows.uvicornPort }} + ports: + - containerPort: {{ .Values.rows.uvicornPort }} + name: http + protocol: TCP + resources: + {{ toYaml .Values.rows.resources | nindent 4 }} +{{- end -}} diff --git a/chart/templates/services/rows/deployment.yaml b/chart/templates/services/rows/deployment.yaml new file mode 100644 index 00000000..d5edea1c --- /dev/null +++ b/chart/templates/services/rows/deployment.yaml @@ -0,0 +1,33 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +apiVersion: apps/v1 +kind: Deployment +metadata: + labels: {{ include "labels.rows" . | nindent 4 }} + name: "{{ include "name" . }}-rows" + namespace: {{ .Release.Namespace }} +spec: + progressDeadlineSeconds: 600 + replicas: {{ .Values.rows.replicas }} + revisionHistoryLimit: 10 + selector: + matchLabels: {{ include "labels.rows" . | nindent 6 }} + strategy: + rollingUpdate: + maxSurge: 25% + maxUnavailable: 25% + type: RollingUpdate + template: + metadata: + labels: {{ include "labels.rows" . | nindent 8 }} + spec: + {{- include "image.imagePullSecrets" . | nindent 6 }} + initContainers: + {{ include "initContainerCachedAssets" . | nindent 8 }} + {{ include "initContainerParquetMetadata" . | nindent 8 }} + containers: {{ include "containerRows" . | nindent 8 }} + nodeSelector: {{ toYaml .Values.rows.nodeSelector | nindent 8 }} + tolerations: {{ toYaml .Values.rows.tolerations | nindent 8 }} + volumes: {{ include "volumeData" . | nindent 8 }} + securityContext: {{ include "securityContext" . | nindent 8 }} diff --git a/chart/templates/services/rows/pdb.yaml b/chart/templates/services/rows/pdb.yaml new file mode 100644 index 00000000..21c64d80 --- /dev/null +++ b/chart/templates/services/rows/pdb.yaml @@ -0,0 +1,13 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +apiVersion: policy/v1 +kind: PodDisruptionBudget +metadata: + labels: {{ include "labels.rows" . | nindent 4 }} + name: "{{ include "name" . }}-rows" + namespace: {{ .Release.Namespace }} +spec: + maxUnavailable: 1 + selector: + matchLabels: {{ include "labels.rows" . | nindent 6 }} diff --git a/chart/templates/services/rows/service.yaml b/chart/templates/services/rows/service.yaml new file mode 100644 index 00000000..3f5bdd23 --- /dev/null +++ b/chart/templates/services/rows/service.yaml @@ -0,0 +1,22 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +{{ $serviceType := .Values.reverseProxy.service.type | default .Values.global.huggingface.service.type }} +apiVersion: v1 +kind: Service +metadata: + name: "{{ include "name" . }}-rows" + annotations: {{ toYaml .Values.rows.service.annotations | nindent 4 }} + namespace: {{ .Release.Namespace }} + labels: {{ include "labels.rows" . | nindent 4 }} +spec: + ports: + - name: http + port: 80 + protocol: TCP + {{- if eq "NodePort" $serviceType }} + nodePort: {{ .Values.global.huggingface.service.ports.datasetsServer.rows }} + {{- end }} + targetPort: {{ .Values.rows.uvicornPort }} + selector: {{ include "labels.rows" . | nindent 4 }} + type: {{ $serviceType }} diff --git a/chart/templates/services/rows/servicemonitor.yaml b/chart/templates/services/rows/servicemonitor.yaml new file mode 100644 index 00000000..c53863ae --- /dev/null +++ b/chart/templates/services/rows/servicemonitor.yaml @@ -0,0 +1,20 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +{{- if .Values.monitoring.enabled }} +apiVersion: monitoring.coreos.com/v1 +kind: ServiceMonitor +metadata: + labels: {{ include "labels.rows" . | nindent 4 }} + name: "{{ include "name" . }}-rows" + namespace: {{ .Release.Namespace }} +spec: + endpoints: + - path: /metrics + port: http + namespaceSelector: + matchNames: + - {{ .Release.Namespace }} + selector: + matchLabels: {{ include "labels.rows" . | nindent 6 }} +{{- end }} diff --git a/chart/values.yaml b/chart/values.yaml index 212ccc51..e8f01d0b 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -19,0 +20 @@ global: + rows: 30023 @@ -50,0 +52,5 @@ images: + rows: + registry: huggingface + useGlobalRegistry: false + repository: datasets-server-services-rows + tag: sha-fb3399a @@ -409,0 +416,38 @@ api: +rows: + # the path of the external authentication service on the hub. + # The string must contain `%s` which will be replaced with the dataset name. + hfAuthPath: "/api/datasets/%s/auth-check" + # the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format. + # It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication + # check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored. + hfJwtPublicKeyUrl: "https://huggingface.co/api/keys/jwt" + # the algorithm used to encode the JWT. + hfJwtAlgorithm: "EdDSA" + # the timeout in seconds for the requests to the Hugging Face Hub. + hfTimeoutSeconds: "0.2" + # Number of seconds to set in the `max-age` header on data endpoints + maxAgeLong: "120" + # Number of seconds to set in the `max-age` header on technical endpoints + maxAgeShort: "10" + # Directory where the uvicorn workers will write the prometheus metrics + # see https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn + prometheusMultiprocDirectory: "/tmp" + # Hostname - it must not be set to localhost to work in Kube! + uvicornHostname: "0.0.0.0" + # Number of uvicorn workers for running the application + uvicornNumWorkers: "1" + # Application endpoint port + uvicornPort: 8082 + + nodeSelector: {} + replicas: 1 + resources: + requests: + cpu: 0 + limits: + cpu: 0 + service: + type: "" + annotations: {} + tolerations: [] + diff --git a/e2e/Makefile b/e2e/Makefile index dfe55a19..c1ac31e8 100644 --- a/e2e/Makefile +++ b/e2e/Makefile @@ -11,0 +12,2 @@ export API_UVICORN_PORT := 9080 +export ROWS_UVICORN_NUM_WORKERS := 2 +export ROWS_UVICORN_PORT := 9082 diff --git a/e2e/tests/test_40_rows_healthcheck.py b/e2e/tests/test_40_rows_healthcheck.py new file mode 100644 index 00000000..65ac68b1 --- /dev/null +++ b/e2e/tests/test_40_rows_healthcheck.py @@ -0,0 +1,11 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from .utils import ROWS_URL, poll + + +def test_healthcheck() -> None: + # this tests ensures the /healthcheck and the /metrics endpoints are hidden + response = poll("/healthcheck", expected_code=200, url=ROWS_URL) + assert response.status_code == 200, f"{response.status_code} - {response.text}" + assert "ok" in response.text, response.text diff --git a/e2e/tests/test_41_rows_metrics.py b/e2e/tests/test_41_rows_metrics.py new file mode 100644 index 00000000..5ef19930 --- /dev/null +++ b/e2e/tests/test_41_rows_metrics.py @@ -0,0 +1,49 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import os +import re +from typing import Mapping + +from .utils import ROWS_URL, get + + +def has_metric(name: str, labels: Mapping[str, str], metric_names: set[str]) -> bool: + label_str = ",".join([f'{k}="{v}"' for k, v in labels.items()]) + s = name + "{" + label_str + "}" + return any(re.match(s, metric_name) is not None for metric_name in metric_names) + + +def test_metrics() -> None: + assert "PROMETHEUS_MULTIPROC_DIR" in os.environ + response = get("/metrics", url=ROWS_URL) + assert response.status_code == 200, f"{response.status_code} - {response.text}" + content = response.text + lines = content.split("\n") + # examples: + # starlette_requests_total{method="GET",path_template="/metrics"} 1.0 + # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05 + metrics = { + parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1)) + } + # see https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn + assert "process_start_time_seconds" not in metrics + + # the middleware should have recorded the request + name = 'starlette_requests_total{method="GET",path_template="/metrics"}' + assert name in metrics, metrics + assert metrics[name] > 0, metrics + + metric_names = set(metrics.keys()) + for endpoint in ["/rows"]: + # these metrics are only available in the admin API + assert not has_metric( + name="queue_jobs_total", + labels={"pid": "[0-9]*", "queue": endpoint, "status": "started"}, + metric_names=metric_names, + ), f"queue_jobs_total - endpoint={endpoint} found in {metrics}" + assert not has_metric( + name="responses_in_cache_total", + labels={"error_code": "None", "http_status": "200", "path": endpoint, "pid": "[0-9]*"}, + metric_names=metric_names, + ), f"responses_in_cache_total - endpoint {endpoint} found in {metrics}" diff --git a/e2e/tests/utils.py b/e2e/tests/utils.py index 8e3466c7..681aa72a 100644 --- a/e2e/tests/utils.py +++ b/e2e/tests/utils.py @@ -15,0 +16 @@ ADMIN_UVICORN_PORT = os.environ.get("ADMIN_UVICORN_PORT", "8081") +ROWS_UVICORN_PORT = os.environ.get("ROWS_UVICORN_PORT", "8082") @@ -21,0 +23 @@ API_URL = f"http://localhost:{API_UVICORN_PORT}" +ROWS_URL = f"http://localhost:{ROWS_UVICORN_PORT}" @@ -127 +129,3 @@ def log(response: Response, url: str = URL, relative_url: Optional[str] = None, - return f"{response.status_code} - {response.headers} - {response.text} - {url}{extra}" + return ( + f"{dataset=} - {relative_url=} - {response.status_code} - {response.headers} - {response.text} - {url}{extra}" + ) diff --git a/services/admin/Makefile b/services/admin/Makefile index cd552232..23d6ba6e 100644 --- a/services/admin/Makefile +++ b/services/admin/Makefile @@ -30 +30 @@ test: - PROMETHEUS_MULTIPROC_DIR=/tmp/admin.prometheus poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH) + PROMETHEUS_MULTIPROC_DIR=/tmp/admin.prometheus poetry run python -m pytest -vv -x -k "test_metrics" ${ADDOPTS} $(TEST_PATH) diff --git a/services/api/Makefile b/services/api/Makefile index 44aa2fbe..439cb211 100644 --- a/services/api/Makefile +++ b/services/api/Makefile @@ -29 +29 @@ test: - PROMETHEUS_MULTIPROC_DIR=/tmp/api.prometheus poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH) + PROMETHEUS_MULTIPROC_DIR=/tmp/api.prometheus poetry run python -m pytest -vv -x -k "test_metrics" ${ADDOPTS} $(TEST_PATH) diff --git a/services/api/README.md b/services/api/README.md index 45239b60..5cf8431a 100644 --- a/services/api/README.md +++ b/services/api/README.md @@ -7 +7 @@ -The worker can be configured using environment variables. They are grouped by scope. +The service can be configured using environment variables. They are grouped by scope. diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index 8f6992b2..85625f38 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -12 +11,0 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource -from libcommon.storage import exists, init_cached_assets_dir, init_parquet_metadata_dir @@ -22 +20,0 @@ from api.routes.endpoint import EndpointsDefinition, create_endpoint -from api.routes.rows import create_rows_endpoint @@ -36,4 +33,0 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi - cached_assets_directory = init_cached_assets_dir(directory=app_config.cached_assets.storage_directory) - parquet_metadata_directory = init_parquet_metadata_dir(directory=app_config.parquet_metadata.storage_directory) - if not exists(cached_assets_directory): - raise RuntimeError("The assets storage directory could not be accessed. Exiting.") @@ -111,18 +104,0 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi - Route( - "/rows", - endpoint=create_rows_endpoint( - processing_graph=processing_graph, - cached_assets_base_url=app_config.cached_assets.base_url, - cached_assets_directory=cached_assets_directory, - parquet_metadata_directory=parquet_metadata_directory, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - hf_jwt_public_key=hf_jwt_public_key, - hf_jwt_algorithm=app_config.api.hf_jwt_algorithm, - external_auth_url=app_config.api.external_auth_url, - hf_timeout_seconds=app_config.api.hf_timeout_seconds, - max_age_long=app_config.api.max_age_long, - max_age_short=app_config.api.max_age_short, - cache_max_days=app_config.cache.max_days, - ), - ), diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py index cbc59c89..ea9ef490 100644 --- a/services/api/tests/conftest.py +++ b/services/api/tests/conftest.py @@ -4 +3,0 @@ -from pathlib import Path @@ -18,3 +16,0 @@ from api.routes.endpoint import EndpointsDefinition, StepsByInputTypeAndEndpoint -# Import fixture modules as plugins -pytest_plugins = ["tests.fixtures.fsspec"] - @@ -28 +23,0 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]: - monkeypatch_session.setenv("CACHED_ASSETS_BASE_URL", "http://localhost/cached-assets") @@ -142,7 +136,0 @@ def parquet_metadata_directory(app_config: AppConfig) -> StrPath: - - -@fixture -def image_path() -> str: - image_path = Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg" - assert image_path.is_file() - return str(image_path) diff --git a/services/reverse-proxy/README.md b/services/reverse-proxy/README.md index d322fda2..2fc6a99b 100644 --- a/services/reverse-proxy/README.md +++ b/services/reverse-proxy/README.md @@ -20 +20 @@ It takes various environment variables, all of them are mandatory: -- `URL_ADMIN`= URL of the admin, eg `http://admin:8080` +- `URL_ADMIN`= URL of the admin, eg `http://admin:8081` @@ -21,0 +22 @@ It takes various environment variables, all of them are mandatory: +- `URL_ROWS`= URL of the rows service, eg `http://rows:8082` diff --git a/services/rows/.flake8 b/services/rows/.flake8 new file mode 100644 index 00000000..f7d6157c --- /dev/null +++ b/services/rows/.flake8 @@ -0,0 +1,5 @@ +[flake8] +# Recommend matching the black line length (119), +# rather than using the flake8 default of 79: +max-line-length = 119 +extend-ignore = "E203" diff --git a/services/rows/.python-version b/services/rows/.python-version new file mode 100644 index 00000000..b326afbc --- /dev/null +++ b/services/rows/.python-version @@ -0,0 +1 @@ +3.9.15 diff --git a/services/rows/Dockerfile b/services/rows/Dockerfile new file mode 100644 index 00000000..6e0db885 --- /dev/null +++ b/services/rows/Dockerfile @@ -0,0 +1,35 @@ +# build with +# docker build -t some_tag_rows -f Dockerfile ../.. +FROM python:3.9.15-slim + +ENV PYTHONFAULTHANDLER=1 \ + PYTHONUNBUFFERED=1 \ + PYTHONHASHSEED=random \ + PIP_NO_CACHE_DIR=1 \ + PIP_DISABLE_PIP_VERSION_CHECK=on \ + PIP_DEFAULT_TIMEOUT=100 \ + POETRY_NO_INTERACTION=1 \ + # Versions: + POETRY_VERSION=1.4.0 \ + POETRY_VIRTUALENVS_IN_PROJECT=true \ + PATH="$PATH:/root/.local/bin" + +# System deps: +RUN apt-get update \ + && apt-get install -y build-essential unzip wget \ + libicu-dev ffmpeg libavcodec-extra libsndfile1 llvm pkg-config \ + && rm -rf /var/lib/apt/lists/* +RUN pip install -U pip +RUN pip install "poetry==$POETRY_VERSION" + +WORKDIR /src +COPY services/rows/poetry.lock ./services/rows/poetry.lock +COPY services/rows/pyproject.toml ./services/rows/pyproject.toml +COPY libs/libcommon ./libs/libcommon +COPY libs/libapi ./libs/libapi +WORKDIR /src/services/rows/ +RUN poetry install --no-cache +COPY services/rows/src ./src +RUN poetry install --no-cache + +ENTRYPOINT ["poetry", "run", "python", "src/rows/main.py"] diff --git a/services/rows/Makefile b/services/rows/Makefile new file mode 100644 index 00000000..02a11c6f --- /dev/null +++ b/services/rows/Makefile @@ -0,0 +1,37 @@ +# environment variables for the commands (docker compose, poetry) +export COMPOSE_PROJECT_NAME := rows +export MONGO_PORT := 27032 +export CACHE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT} +# makefile variables +DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml +TEST_PATH ?= tests + +include ../../tools/Python.mk +include ../../tools/Docker.mk + +.PHONY: run +run: + poetry run python src/rows/main.py + +.PHONY: watch +watch: + poetry run watchmedo auto-restart -d src/rows -p "*.py" -R python src/rows/main.py + +# override the default test target to test prometheus depending on the environment +# we cannot set the env var with pytest.MonkeyPatch, it's too late +.PHONY: test +test: + $(MAKE) up + poetry run python -m pytest -vv -x ${ADDOPTS} $(TEST_PATH) + rm -rf /tmp/rows.prometheus + mkdir /tmp/rows.prometheus + PROMETHEUS_MULTIPROC_DIR=/tmp/rows.prometheus poetry run python -m pytest -vv -x -k "test_metrics" ${ADDOPTS} $(TEST_PATH) + rm -rf /tmp/rows.prometheus + $(MAKE) down + +.PHONY: coverage +coverage: + $(MAKE) up + poetry run python -m pytest -s --cov --cov-report xml:coverage.xml --cov-report=term $(TEST_PATH) + $(MAKE) down diff --git a/services/rows/README.md b/services/rows/README.md new file mode 100644 index 00000000..ecf8a2ef --- /dev/null +++ b/services/rows/README.md @@ -0,0 +1,23 @@ +# Datasets server API - rows endpoint + +> /rows endpoint + +## Configuration + +The service can be configured using environment variables. They are grouped by scope. + +### API service + +See [../../libs/libapi/README.md](../../libs/libapi/README.md) for more information about the API configuration. + +### Common + +See [../../libs/libcommon/README.md](../../libs/libcommon/README.md) for more information about the common configuration. + +## Endpoints + +See https://huggingface.co/docs/datasets-server + +- /healthcheck: ensure the app is running +- /metrics: return a list of metrics in the Prometheus format +- /rows: get a slice of rows of a dataset split diff --git a/services/rows/dev.Dockerfile b/services/rows/dev.Dockerfile new file mode 100644 index 00000000..6c710bbf --- /dev/null +++ b/services/rows/dev.Dockerfile @@ -0,0 +1,50 @@ +# build with +# docker build -t some_tag_rows -f Dockerfile ../.. +FROM python:3.9.15-slim + +ENV PYTHONFAULTHANDLER=1 \ + PYTHONUNBUFFERED=1 \ + PYTHONHASHSEED=random \ + PIP_NO_CACHE_DIR=1 \ + PIP_DISABLE_PIP_VERSION_CHECK=on \ + PIP_DEFAULT_TIMEOUT=100 \ + POETRY_NO_INTERACTION=1 \ + # Versions: + POETRY_VERSION=1.4.0 \ + POETRY_VIRTUALENVS_IN_PROJECT=true \ + PATH="$PATH:/root/.local/bin" + +# System deps: +RUN apt-get update \ + && apt-get install -y build-essential unzip wget \ + libicu-dev ffmpeg libavcodec-extra libsndfile1 llvm pkg-config \ + && rm -rf /var/lib/apt/lists/* +RUN pip install -U pip +RUN pip install "poetry==$POETRY_VERSION" + +WORKDIR /src +COPY libs/libcommon/poetry.lock ./libs/libcommon/poetry.lock +COPY libs/libcommon/pyproject.toml ./libs/libcommon/pyproject.toml +COPY libs/libapi/poetry.lock ./libs/libapi/poetry.lock +COPY libs/libapi/pyproject.toml ./libs/libapi/pyproject.toml +COPY services/rows/poetry.lock ./services/rows/poetry.lock +COPY services/rows/pyproject.toml ./services/rows/pyproject.toml + +# FOR LOCAL DEVELOPMENT ENVIRONMENT +# Initialize an empty libcommon +# Mapping a volume to ./libs/libcommon/src is required when running this image. +RUN mkdir ./libs/libcommon/src && mkdir ./libs/libcommon/src/libcommon && touch ./libs/libcommon/src/libcommon/__init__.py +# Initialize an empty libapi +# Mapping a volume to ./libs/libapi/src is required when running this image. +RUN mkdir ./libs/libapi/src && mkdir ./libs/libapi/src/libapi && touch ./libs/libapi/src/libapi/__init__.py + +# Install dependencies +WORKDIR /src/services/rows/ +RUN --mount=type=cache,target=/home/.cache/pypoetry/cache \ + --mount=type=cache,target=/home/.cache/pypoetry/artifacts \ + poetry install --no-root + +# FOR LOCAL DEVELOPMENT ENVIRONMENT +# Install the rows package. +# Mapping a volume to ./services/rows/src is required when running this image. +ENTRYPOINT ["/bin/sh", "-c" , "poetry install --only-root && poetry run python src/rows/main.py"] diff --git a/services/rows/poetry.lock b/services/rows/poetry.lock new file mode 100644 index 00000000..31d00257 --- /dev/null +++ b/services/rows/poetry.lock @@ -0,0 +1,3356 @@ +# This file is automatically @generated by Poetry and should not be changed by hand. + +[[package]] +name = "aiohttp" +version = "3.8.4" +description = "Async http client/server framework (asyncio)" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"}, + {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"}, + {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"}, + {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"}, + {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"}, + {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"}, + {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"}, + {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"}, + {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"}, + {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"}, + {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"}, + {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"}, +] + +[package.dependencies] +aiosignal = ">=1.1.2" +async-timeout = ">=4.0.0a3,<5.0" +attrs = ">=17.3.0" +charset-normalizer = ">=2.0,<4.0" +frozenlist = ">=1.1.1" +multidict = ">=4.5,<7.0" +yarl = ">=1.0,<2.0" + +[package.extras] +speedups = ["Brotli", "aiodns", "cchardet"] + +[[package]] +name = "aiosignal" +version = "1.3.1" +description = "aiosignal: a list of registered asynchronous callbacks" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, + {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, +] + +[package.dependencies] +frozenlist = ">=1.1.0" + +[[package]] +name = "anyio" +version = "3.7.0" +description = "High level compatibility layer for multiple asynchronous event loop implementations" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "anyio-3.7.0-py3-none-any.whl", hash = "sha256:eddca883c4175f14df8aedce21054bfca3adb70ffe76a9f607aef9d7fa2ea7f0"}, + {file = "anyio-3.7.0.tar.gz", hash = "sha256:275d9973793619a5374e1c89a4f4ad3f4b0a5510a2b5b939444bee8f4c4d37ce"}, +] + +[package.dependencies] +exceptiongroup = {version = "*", markers = "python_version < \"3.11\""} +idna = ">=2.8" +sniffio = ">=1.1" + +[package.extras] +doc = ["Sphinx (>=6.1.0)", "packaging", "sphinx-autodoc-typehints (>=1.2.0)", "sphinx-rtd-theme", "sphinxcontrib-jquery"] +test = ["anyio[trio]", "coverage[toml] (>=4.5)", "hypothesis (>=4.0)", "mock (>=4)", "psutil (>=5.9)", "pytest (>=7.0)", "pytest-mock (>=3.6.1)", "trustme", "uvloop (>=0.17)"] +trio = ["trio (<0.22)"] + +[[package]] +name = "appdirs" +version = "1.4.4" +description = "A small Python module for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "appdirs-1.4.4-py2.py3-none-any.whl", hash = "sha256:a841dacd6b99318a741b166adb07e19ee71a274450e68237b4650ca1055ab128"}, + {file = "appdirs-1.4.4.tar.gz", hash = "sha256:7d5d0167b2b1ba821647616af46a749d1c653740dd0d2415100fe26e27afdf41"}, +] + +[[package]] +name = "async-timeout" +version = "4.0.2" +description = "Timeout context manager for asyncio programs" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"}, + {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"}, +] + +[[package]] +name = "attrs" +version = "23.1.0" +description = "Classes Without Boilerplate" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "attrs-23.1.0-py3-none-any.whl", hash = "sha256:1f28b4522cdc2fb4256ac1a020c78acf9cba2c6b461ccd2c126f3aa8e8335d04"}, + {file = "attrs-23.1.0.tar.gz", hash = "sha256:6279836d581513a26f1bf235f9acd333bc9115683f14f7e8fae46c98fc50e015"}, +] + +[package.extras] +cov = ["attrs[tests]", "coverage[toml] (>=5.3)"] +dev = ["attrs[docs,tests]", "pre-commit"] +docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope-interface"] +tests = ["attrs[tests-no-zope]", "zope-interface"] +tests-no-zope = ["cloudpickle", "hypothesis", "mypy (>=1.1.1)", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"] + +[[package]] +name = "audioread" +version = "3.0.0" +description = "multi-library, cross-platform audio decoding" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"}, +] + +[[package]] +name = "bandit" +version = "1.7.5" +description = "Security oriented static analyser for python code." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "bandit-1.7.5-py3-none-any.whl", hash = "sha256:75665181dc1e0096369112541a056c59d1c5f66f9bb74a8d686c3c362b83f549"}, + {file = "bandit-1.7.5.tar.gz", hash = "sha256:bdfc739baa03b880c2d15d0431b31c658ffc348e907fe197e54e0389dd59e11e"}, +] + +[package.dependencies] +colorama = {version = ">=0.3.9", markers = "platform_system == \"Windows\""} +GitPython = ">=1.0.1" +PyYAML = ">=5.3.1" +rich = "*" +stevedore = ">=1.20.0" + +[package.extras] +test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "tomli (>=1.1.0)"] +toml = ["tomli (>=1.1.0)"] +yaml = ["PyYAML"] + +[[package]] +name = "black" +version = "22.12.0" +description = "The uncompromising code formatter." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "black-22.12.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eedd20838bd5d75b80c9f5487dbcb06836a43833a37846cf1d8c1cc01cef59d"}, + {file = "black-22.12.0-cp310-cp310-win_amd64.whl", hash = "sha256:159a46a4947f73387b4d83e87ea006dbb2337eab6c879620a3ba52699b1f4351"}, + {file = "black-22.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d30b212bffeb1e252b31dd269dfae69dd17e06d92b87ad26e23890f3efea366f"}, + {file = "black-22.12.0-cp311-cp311-win_amd64.whl", hash = "sha256:7412e75863aa5c5411886804678b7d083c7c28421210180d67dfd8cf1221e1f4"}, + {file = "black-22.12.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c116eed0efb9ff870ded8b62fe9f28dd61ef6e9ddd28d83d7d264a38417dcee2"}, + {file = "black-22.12.0-cp37-cp37m-win_amd64.whl", hash = "sha256:1f58cbe16dfe8c12b7434e50ff889fa479072096d79f0a7f25e4ab8e94cd8350"}, + {file = "black-22.12.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77d86c9f3db9b1bf6761244bc0b3572a546f5fe37917a044e02f3166d5aafa7d"}, + {file = "black-22.12.0-cp38-cp38-win_amd64.whl", hash = "sha256:82d9fe8fee3401e02e79767016b4907820a7dc28d70d137eb397b92ef3cc5bfc"}, + {file = "black-22.12.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:101c69b23df9b44247bd88e1d7e90154336ac4992502d4197bdac35dd7ee3320"}, + {file = "black-22.12.0-cp39-cp39-win_amd64.whl", hash = "sha256:559c7a1ba9a006226f09e4916060982fd27334ae1998e7a38b3f33a37f7a2148"}, + {file = "black-22.12.0-py3-none-any.whl", hash = "sha256:436cc9167dd28040ad90d3b404aec22cedf24a6e4d7de221bec2730ec0c97bcf"}, + {file = "black-22.12.0.tar.gz", hash = "sha256:229351e5a18ca30f447bf724d007f890f97e13af070bb6ad4c0a441cd7596a2f"}, +] + +[package.dependencies] +click = ">=8.0.0" +mypy-extensions = ">=0.4.3" +pathspec = ">=0.9.0" +platformdirs = ">=2" +tomli = {version = ">=1.1.0", markers = "python_full_version < \"3.11.0a7\""} +typing-extensions = {version = ">=3.10.0.0", markers = "python_version < \"3.10\""} + +[package.extras] +colorama = ["colorama (>=0.4.3)"] +d = ["aiohttp (>=3.7.4)"] +jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"] +uvloop = ["uvloop (>=0.15.2)"] + +[[package]] +name = "cachecontrol" +version = "0.13.1" +description = "httplib2 caching for requests" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, +] + +[package.dependencies] +filelock = {version = ">=3.8.0", optional = true, markers = "extra == \"filecache\""} +msgpack = ">=0.5.2" +requests = ">=2.16.0" + +[package.extras] +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] +filecache = ["filelock (>=3.8.0)"] +redis = ["redis (>=2.10.5)"] + +[[package]] +name = "certifi" +version = "2023.5.7" +description = "Python package for providing Mozilla's CA Bundle." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "certifi-2023.5.7-py3-none-any.whl", hash = "sha256:c6c2e98f5c7869efca1f8916fed228dd91539f9f1b444c314c06eef02980c716"}, + {file = "certifi-2023.5.7.tar.gz", hash = "sha256:0f0d56dc5a6ad56fd4ba36484d6cc34451e1c6548c61daad8c320169f91eddc7"}, +] + +[[package]] +name = "cffi" +version = "1.15.1" +description = "Foreign Function Interface for Python calling C code." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"}, + {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"}, + {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"}, + {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"}, + {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"}, + {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"}, + {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"}, + {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"}, + {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"}, + {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"}, + {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"}, + {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"}, + {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"}, + {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"}, + {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"}, + {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"}, + {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"}, + {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"}, + {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"}, +] + +[package.dependencies] +pycparser = "*" + +[[package]] +name = "charset-normalizer" +version = "3.1.0" +description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"}, + {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"}, +] + +[[package]] +name = "click" +version = "8.1.3" +description = "Composable command line interface toolkit" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "click-8.1.3-py3-none-any.whl", hash = "sha256:bb4d8133cb15a609f44e8213d9b391b0809795062913b383c62be0ee95b1db48"}, + {file = "click-8.1.3.tar.gz", hash = "sha256:7682dc8afb30297001674575ea00d1814d808d6a36af415a82bd481d37ba7b8e"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[[package]] +name = "colorama" +version = "0.4.6" +description = "Cross-platform colored terminal text." +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7" +files = [ + {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"}, + {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"}, +] + +[[package]] +name = "coverage" +version = "7.2.7" +description = "Code coverage measurement for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "coverage-7.2.7-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d39b5b4f2a66ccae8b7263ac3c8170994b65266797fb96cbbfd3fb5b23921db8"}, + {file = "coverage-7.2.7-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:6d040ef7c9859bb11dfeb056ff5b3872436e3b5e401817d87a31e1750b9ae2fb"}, + {file = "coverage-7.2.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ba90a9563ba44a72fda2e85302c3abc71c5589cea608ca16c22b9804262aaeb6"}, + {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e7d9405291c6928619403db1d10bd07888888ec1abcbd9748fdaa971d7d661b2"}, + {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:31563e97dae5598556600466ad9beea39fb04e0229e61c12eaa206e0aa202063"}, + {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ebba1cd308ef115925421d3e6a586e655ca5a77b5bf41e02eb0e4562a111f2d1"}, + {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:cb017fd1b2603ef59e374ba2063f593abe0fc45f2ad9abdde5b4d83bd922a353"}, + {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d62a5c7dad11015c66fbb9d881bc4caa5b12f16292f857842d9d1871595f4495"}, + {file = "coverage-7.2.7-cp310-cp310-win32.whl", hash = "sha256:ee57190f24fba796e36bb6d3aa8a8783c643d8fa9760c89f7a98ab5455fbf818"}, + {file = "coverage-7.2.7-cp310-cp310-win_amd64.whl", hash = "sha256:f75f7168ab25dd93110c8a8117a22450c19976afbc44234cbf71481094c1b850"}, + {file = "coverage-7.2.7-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:06a9a2be0b5b576c3f18f1a241f0473575c4a26021b52b2a85263a00f034d51f"}, + {file = "coverage-7.2.7-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5baa06420f837184130752b7c5ea0808762083bf3487b5038d68b012e5937dbe"}, + {file = "coverage-7.2.7-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fdec9e8cbf13a5bf63290fc6013d216a4c7232efb51548594ca3631a7f13c3a3"}, + {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:52edc1a60c0d34afa421c9c37078817b2e67a392cab17d97283b64c5833f427f"}, + {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:63426706118b7f5cf6bb6c895dc215d8a418d5952544042c8a2d9fe87fcf09cb"}, + {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:afb17f84d56068a7c29f5fa37bfd38d5aba69e3304af08ee94da8ed5b0865833"}, + {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:48c19d2159d433ccc99e729ceae7d5293fbffa0bdb94952d3579983d1c8c9d97"}, + {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:0e1f928eaf5469c11e886fe0885ad2bf1ec606434e79842a879277895a50942a"}, + {file = "coverage-7.2.7-cp311-cp311-win32.whl", hash = "sha256:33d6d3ea29d5b3a1a632b3c4e4f4ecae24ef170b0b9ee493883f2df10039959a"}, + {file = "coverage-7.2.7-cp311-cp311-win_amd64.whl", hash = "sha256:5b7540161790b2f28143191f5f8ec02fb132660ff175b7747b95dcb77ac26562"}, + {file = "coverage-7.2.7-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:f2f67fe12b22cd130d34d0ef79206061bfb5eda52feb6ce0dba0644e20a03cf4"}, + {file = "coverage-7.2.7-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a342242fe22407f3c17f4b499276a02b01e80f861f1682ad1d95b04018e0c0d4"}, + {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:171717c7cb6b453aebac9a2ef603699da237f341b38eebfee9be75d27dc38e01"}, + {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:49969a9f7ffa086d973d91cec8d2e31080436ef0fb4a359cae927e742abfaaa6"}, + {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:b46517c02ccd08092f4fa99f24c3b83d8f92f739b4657b0f146246a0ca6a831d"}, + {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:a3d33a6b3eae87ceaefa91ffdc130b5e8536182cd6dfdbfc1aa56b46ff8c86de"}, + {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:976b9c42fb2a43ebf304fa7d4a310e5f16cc99992f33eced91ef6f908bd8f33d"}, + {file = "coverage-7.2.7-cp312-cp312-win32.whl", hash = "sha256:8de8bb0e5ad103888d65abef8bca41ab93721647590a3f740100cd65c3b00511"}, + {file = "coverage-7.2.7-cp312-cp312-win_amd64.whl", hash = "sha256:9e31cb64d7de6b6f09702bb27c02d1904b3aebfca610c12772452c4e6c21a0d3"}, + {file = "coverage-7.2.7-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:58c2ccc2f00ecb51253cbe5d8d7122a34590fac9646a960d1430d5b15321d95f"}, + {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d22656368f0e6189e24722214ed8d66b8022db19d182927b9a248a2a8a2f67eb"}, + {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a895fcc7b15c3fc72beb43cdcbdf0ddb7d2ebc959edac9cef390b0d14f39f8a9"}, + {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e84606b74eb7de6ff581a7915e2dab7a28a0517fbe1c9239eb227e1354064dcd"}, + {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0a5f9e1dbd7fbe30196578ca36f3fba75376fb99888c395c5880b355e2875f8a"}, + {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:419bfd2caae268623dd469eff96d510a920c90928b60f2073d79f8fe2bbc5959"}, + {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:2aee274c46590717f38ae5e4650988d1af340fe06167546cc32fe2f58ed05b02"}, + {file = "coverage-7.2.7-cp37-cp37m-win32.whl", hash = "sha256:61b9a528fb348373c433e8966535074b802c7a5d7f23c4f421e6c6e2f1697a6f"}, + {file = "coverage-7.2.7-cp37-cp37m-win_amd64.whl", hash = "sha256:b1c546aca0ca4d028901d825015dc8e4d56aac4b541877690eb76490f1dc8ed0"}, + {file = "coverage-7.2.7-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:54b896376ab563bd38453cecb813c295cf347cf5906e8b41d340b0321a5433e5"}, + {file = "coverage-7.2.7-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3d376df58cc111dc8e21e3b6e24606b5bb5dee6024f46a5abca99124b2229ef5"}, + {file = "coverage-7.2.7-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e330fc79bd7207e46c7d7fd2bb4af2963f5f635703925543a70b99574b0fea9"}, + {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e9d683426464e4a252bf70c3498756055016f99ddaec3774bf368e76bbe02b6"}, + {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8d13c64ee2d33eccf7437961b6ea7ad8673e2be040b4f7fd4fd4d4d28d9ccb1e"}, + {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:b7aa5f8a41217360e600da646004f878250a0d6738bcdc11a0a39928d7dc2050"}, + {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8fa03bce9bfbeeef9f3b160a8bed39a221d82308b4152b27d82d8daa7041fee5"}, + {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:245167dd26180ab4c91d5e1496a30be4cd721a5cf2abf52974f965f10f11419f"}, + {file = "coverage-7.2.7-cp38-cp38-win32.whl", hash = "sha256:d2c2db7fd82e9b72937969bceac4d6ca89660db0a0967614ce2481e81a0b771e"}, + {file = "coverage-7.2.7-cp38-cp38-win_amd64.whl", hash = "sha256:2e07b54284e381531c87f785f613b833569c14ecacdcb85d56b25c4622c16c3c"}, + {file = "coverage-7.2.7-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:537891ae8ce59ef63d0123f7ac9e2ae0fc8b72c7ccbe5296fec45fd68967b6c9"}, + {file = "coverage-7.2.7-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:06fb182e69f33f6cd1d39a6c597294cff3143554b64b9825d1dc69d18cc2fff2"}, + {file = "coverage-7.2.7-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:201e7389591af40950a6480bd9edfa8ed04346ff80002cec1a66cac4549c1ad7"}, + {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f6951407391b639504e3b3be51b7ba5f3528adbf1a8ac3302b687ecababf929e"}, + {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6f48351d66575f535669306aa7d6d6f71bc43372473b54a832222803eb956fd1"}, + {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b29019c76039dc3c0fd815c41392a044ce555d9bcdd38b0fb60fb4cd8e475ba9"}, + {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:81c13a1fc7468c40f13420732805a4c38a105d89848b7c10af65a90beff25250"}, + {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:975d70ab7e3c80a3fe86001d8751f6778905ec723f5b110aed1e450da9d4b7f2"}, + {file = "coverage-7.2.7-cp39-cp39-win32.whl", hash = "sha256:7ee7d9d4822c8acc74a5e26c50604dff824710bc8de424904c0982e25c39c6cb"}, + {file = "coverage-7.2.7-cp39-cp39-win_amd64.whl", hash = "sha256:eb393e5ebc85245347950143969b241d08b52b88a3dc39479822e073a1a8eb27"}, + {file = "coverage-7.2.7-pp37.pp38.pp39-none-any.whl", hash = "sha256:b7b4c971f05e6ae490fef852c218b0e79d4e52f79ef0c8475566584a8fb3e01d"}, + {file = "coverage-7.2.7.tar.gz", hash = "sha256:924d94291ca674905fe9481f12294eb11f2d3d3fd1adb20314ba89e94f44ed59"}, +] + +[package.extras] +toml = ["tomli"] + +[[package]] +name = "cryptography" +version = "41.0.1" +description = "cryptography is a package which provides cryptographic recipes and primitives to Python developers." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_universal2.whl", hash = "sha256:f73bff05db2a3e5974a6fd248af2566134d8981fd7ab012e5dd4ddb1d9a70699"}, + {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_x86_64.whl", hash = "sha256:1a5472d40c8f8e91ff7a3d8ac6dfa363d8e3138b961529c996f3e2df0c7a411a"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fa01527046ca5facdf973eef2535a27fec4cb651e4daec4d043ef63f6ecd4ca"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b46e37db3cc267b4dea1f56da7346c9727e1209aa98487179ee8ebed09d21e43"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:d198820aba55660b4d74f7b5fd1f17db3aa5eb3e6893b0a41b75e84e4f9e0e4b"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:948224d76c4b6457349d47c0c98657557f429b4e93057cf5a2f71d603e2fc3a3"}, + {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:059e348f9a3c1950937e1b5d7ba1f8e968508ab181e75fc32b879452f08356db"}, + {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:b4ceb5324b998ce2003bc17d519080b4ec8d5b7b70794cbd2836101406a9be31"}, + {file = "cryptography-41.0.1-cp37-abi3-win32.whl", hash = "sha256:8f4ab7021127a9b4323537300a2acfb450124b2def3756f64dc3a3d2160ee4b5"}, + {file = "cryptography-41.0.1-cp37-abi3-win_amd64.whl", hash = "sha256:1fee5aacc7367487b4e22484d3c7e547992ed726d14864ee33c0176ae43b0d7c"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-macosx_10_12_x86_64.whl", hash = "sha256:9a6c7a3c87d595608a39980ebaa04d5a37f94024c9f24eb7d10262b92f739ddb"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:5d092fdfedaec4cbbffbf98cddc915ba145313a6fdaab83c6e67f4e6c218e6f3"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:1a8e6c2de6fbbcc5e14fd27fb24414507cb3333198ea9ab1258d916f00bc3039"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cb33ccf15e89f7ed89b235cff9d49e2e62c6c981a6061c9c8bb47ed7951190bc"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:5f0ff6e18d13a3de56f609dd1fd11470918f770c6bd5d00d632076c727d35485"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:7bfc55a5eae8b86a287747053140ba221afc65eb06207bedf6e019b8934b477c"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:eb8163f5e549a22888c18b0d53d6bb62a20510060a22fd5a995ec8a05268df8a"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:8dde71c4169ec5ccc1087bb7521d54251c016f126f922ab2dfe6649170a3b8c5"}, + {file = "cryptography-41.0.1.tar.gz", hash = "sha256:d34579085401d3f49762d2f7d6634d6b6c2ae1242202e860f4d26b046e3a1006"}, +] + +[package.dependencies] +cffi = ">=1.12" + +[package.extras] +docs = ["sphinx (>=5.3.0)", "sphinx-rtd-theme (>=1.1.1)"] +docstest = ["pyenchant (>=1.6.11)", "sphinxcontrib-spelling (>=4.0.1)", "twine (>=1.12.0)"] +nox = ["nox"] +pep8test = ["black", "check-sdist", "mypy", "ruff"] +sdist = ["build"] +ssh = ["bcrypt (>=3.1.5)"] +test = ["pretend", "pytest (>=6.2.0)", "pytest-benchmark", "pytest-cov", "pytest-xdist"] +test-randomorder = ["pytest-randomly"] + +[[package]] +name = "cyclonedx-python-lib" +version = "4.0.1" +description = "A library for producing CycloneDX SBOM (Software Bill of Materials) files." +category = "dev" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "cyclonedx_python_lib-4.0.1-py3-none-any.whl", hash = "sha256:907b64f00df85d727a425de86604768b248cf19285993729e04f17bec767f692"}, + {file = "cyclonedx_python_lib-4.0.1.tar.gz", hash = "sha256:878e33b8e0080c786f6cbd4c6f87ad610db65d6a3a686a5698415d9cfcd8925d"}, +] + +[package.dependencies] +packageurl-python = ">=0.11" +py-serializable = ">=0.11.1,<0.12.0" +sortedcontainers = ">=2.4.0,<3.0.0" + +[[package]] +name = "datasets" +version = "2.13.1" +description = "HuggingFace community-driven open-source library of datasets" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, +] + +[package.dependencies] +aiohttp = "*" +dill = ">=0.3.0,<0.3.7" +fsspec = {version = ">=2021.11.1", extras = ["http"]} +huggingface-hub = ">=0.11.0,<1.0.0" +librosa = {version = "*", optional = true, markers = "extra == \"audio\""} +multiprocess = "*" +numpy = ">=1.17" +packaging = "*" +pandas = "*" +Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} +pyarrow = ">=8.0.0" +pyyaml = ">=5.1" +requests = ">=2.19.0" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} +tqdm = ">=4.62.1" +xxhash = "*" + +[package.extras] +apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] +audio = ["librosa", "soundfile (>=0.12.1)"] +benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +docs = ["s3fs"] +jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"] +s3 = ["s3fs"] +tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"] +tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +torch = ["torch"] +vision = ["Pillow (>=6.2.1)"] + +[[package]] +name = "decorator" +version = "5.1.1" +description = "Decorators for Humans" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"}, + {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"}, +] + +[[package]] +name = "defusedxml" +version = "0.7.1" +description = "XML bomb protection for Python stdlib modules" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "defusedxml-0.7.1-py2.py3-none-any.whl", hash = "sha256:a352e7e428770286cc899e2542b6cdaedb2b4953ff269a210103ec58f6198a61"}, + {file = "defusedxml-0.7.1.tar.gz", hash = "sha256:1bb3032db185915b62d7c6209c5a8792be6a32ab2fedacc84e01b52c51aa3e69"}, +] + +[[package]] +name = "dill" +version = "0.3.6" +description = "serialize all of python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"}, + {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"}, +] + +[package.extras] +graph = ["objgraph (>=1.7.2)"] + +[[package]] +name = "dnspython" +version = "1.16.0" +description = "DNS toolkit" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "dnspython-1.16.0-py2.py3-none-any.whl", hash = "sha256:f69c21288a962f4da86e56c4905b49d11aba7938d3d740e80d9e366ee4f1632d"}, + {file = "dnspython-1.16.0.zip", hash = "sha256:36c5e8e38d4369a08b6780b7f27d790a292b2b08eea01607865bf0936c558e01"}, +] + +[package.extras] +dnssec = ["ecdsa (>=0.13)", "pycryptodome"] +idna = ["idna (>=2.1)"] + +[[package]] +name = "environs" +version = "9.5.0" +description = "simplified environment variable parsing" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "environs-9.5.0-py2.py3-none-any.whl", hash = "sha256:1e549569a3de49c05f856f40bce86979e7d5ffbbc4398e7f338574c220189124"}, + {file = "environs-9.5.0.tar.gz", hash = "sha256:a76307b36fbe856bdca7ee9161e6c466fd7fcffc297109a118c59b54e27e30c9"}, +] + +[package.dependencies] +marshmallow = ">=3.0.0" +python-dotenv = "*" + +[package.extras] +dev = ["dj-database-url", "dj-email-url", "django-cache-url", "flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)", "pytest", "tox"] +django = ["dj-database-url", "dj-email-url", "django-cache-url"] +lint = ["flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)"] +tests = ["dj-database-url", "dj-email-url", "django-cache-url", "pytest"] + +[[package]] +name = "exceptiongroup" +version = "1.1.2" +description = "Backport of PEP 654 (exception groups)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "exceptiongroup-1.1.2-py3-none-any.whl", hash = "sha256:e346e69d186172ca7cf029c8c1d16235aa0e04035e5750b4b95039e65204328f"}, + {file = "exceptiongroup-1.1.2.tar.gz", hash = "sha256:12c3e887d6485d16943a309616de20ae5582633e0a2eda17f4e10fd61c1e8af5"}, +] + +[package.extras] +test = ["pytest (>=6)"] + +[[package]] +name = "filelock" +version = "3.12.2" +description = "A platform independent file lock." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "filelock-3.12.2-py3-none-any.whl", hash = "sha256:cbb791cdea2a72f23da6ac5b5269ab0a0d161e9ef0100e653b69049a7706d1ec"}, + {file = "filelock-3.12.2.tar.gz", hash = "sha256:002740518d8aa59a26b0c76e10fb8c6e15eae825d34b6fdf670333fd7b938d81"}, +] + +[package.extras] +docs = ["furo (>=2023.5.20)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"] +testing = ["covdefaults (>=2.3)", "coverage (>=7.2.7)", "diff-cover (>=7.5)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)", "pytest-timeout (>=2.1)"] + +[[package]] +name = "flake8" +version = "3.9.2" +description = "the modular source code checker: pep8 pyflakes and co" +category = "dev" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7" +files = [ + {file = "flake8-3.9.2-py2.py3-none-any.whl", hash = "sha256:bf8fd333346d844f616e8d47905ef3a3384edae6b4e9beb0c5101e25e3110907"}, + {file = "flake8-3.9.2.tar.gz", hash = "sha256:07528381786f2a6237b061f6e96610a4167b226cb926e2aa2b6b1d78057c576b"}, +] + +[package.dependencies] +mccabe = ">=0.6.0,<0.7.0" +pycodestyle = ">=2.7.0,<2.8.0" +pyflakes = ">=2.3.0,<2.4.0" + +[[package]] +name = "frozenlist" +version = "1.3.3" +description = "A list-like structure which implements collections.abc.MutableSequence" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"}, + {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"}, + {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"}, + {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"}, + {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"}, + {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"}, + {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"}, + {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"}, + {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"}, + {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"}, + {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"}, +] + +[[package]] +name = "fsspec" +version = "2023.6.0" +description = "File-system specification" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "fsspec-2023.6.0-py3-none-any.whl", hash = "sha256:1cbad1faef3e391fba6dc005ae9b5bdcbf43005c9167ce78c915549c352c869a"}, + {file = "fsspec-2023.6.0.tar.gz", hash = "sha256:d0b2f935446169753e7a5c5c55681c54ea91996cc67be93c39a154fb3a2742af"}, +] + +[package.dependencies] +aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""} +requests = {version = "*", optional = true, markers = "extra == \"http\""} + +[package.extras] +abfs = ["adlfs"] +adl = ["adlfs"] +arrow = ["pyarrow (>=1)"] +dask = ["dask", "distributed"] +devel = ["pytest", "pytest-cov"] +dropbox = ["dropbox", "dropboxdrivefs", "requests"] +full = ["adlfs", "aiohttp (!=4.0.0a0,!=4.0.0a1)", "dask", "distributed", "dropbox", "dropboxdrivefs", "fusepy", "gcsfs", "libarchive-c", "ocifs", "panel", "paramiko", "pyarrow (>=1)", "pygit2", "requests", "s3fs", "smbprotocol", "tqdm"] +fuse = ["fusepy"] +gcs = ["gcsfs"] +git = ["pygit2"] +github = ["requests"] +gs = ["gcsfs"] +gui = ["panel"] +hdfs = ["pyarrow (>=1)"] +http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"] +libarchive = ["libarchive-c"] +oci = ["ocifs"] +s3 = ["s3fs"] +sftp = ["paramiko"] +smb = ["smbprotocol"] +ssh = ["paramiko"] +tqdm = ["tqdm"] + +[[package]] +name = "gitdb" +version = "4.0.10" +description = "Git Object Database" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "gitdb-4.0.10-py3-none-any.whl", hash = "sha256:c286cf298426064079ed96a9e4a9d39e7f3e9bf15ba60701e95f5492f28415c7"}, + {file = "gitdb-4.0.10.tar.gz", hash = "sha256:6eb990b69df4e15bad899ea868dc46572c3f75339735663b81de79b06f17eb9a"}, +] + +[package.dependencies] +smmap = ">=3.0.1,<6" + +[[package]] +name = "gitpython" +version = "3.1.31" +description = "GitPython is a Python library used to interact with Git repositories" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"}, + {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"}, +] + +[package.dependencies] +gitdb = ">=4.0.1,<5" + +[[package]] +name = "h11" +version = "0.14.0" +description = "A pure-Python, bring-your-own-I/O implementation of HTTP/1.1" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "h11-0.14.0-py3-none-any.whl", hash = "sha256:e3fe4ac4b851c468cc8363d500db52c2ead036020723024a109d37346efaa761"}, + {file = "h11-0.14.0.tar.gz", hash = "sha256:8f19fbbe99e72420ff35c00b27a34cb9937e902a8b810e2c88300c6f0a3b699d"}, +] + +[[package]] +name = "html5lib" +version = "1.1" +description = "HTML parser based on the WHATWG HTML specification" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "html5lib-1.1-py2.py3-none-any.whl", hash = "sha256:0d78f8fde1c230e99fe37986a60526d7049ed4bf8a9fadbad5f00e22e58e041d"}, + {file = "html5lib-1.1.tar.gz", hash = "sha256:b2e5b40261e20f354d198eae92afc10d750afb487ed5e50f9c4eaf07c184146f"}, +] + +[package.dependencies] +six = ">=1.9" +webencodings = "*" + +[package.extras] +all = ["chardet (>=2.2)", "genshi", "lxml"] +chardet = ["chardet (>=2.2)"] +genshi = ["genshi"] +lxml = ["lxml"] + +[[package]] +name = "httpcore" +version = "0.16.3" +description = "A minimal low-level HTTP client." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "httpcore-0.16.3-py3-none-any.whl", hash = "sha256:da1fb708784a938aa084bde4feb8317056c55037247c787bd7e19eb2c2949dc0"}, + {file = "httpcore-0.16.3.tar.gz", hash = "sha256:c5d6f04e2fc530f39e0c077e6a30caa53f1451096120f1f38b954afd0b17c0cb"}, +] + +[package.dependencies] +anyio = ">=3.0,<5.0" +certifi = "*" +h11 = ">=0.13,<0.15" +sniffio = ">=1.0.0,<2.0.0" + +[package.extras] +http2 = ["h2 (>=3,<5)"] +socks = ["socksio (>=1.0.0,<2.0.0)"] + +[[package]] +name = "httpx" +version = "0.23.3" +description = "The next generation HTTP client." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "httpx-0.23.3-py3-none-any.whl", hash = "sha256:a211fcce9b1254ea24f0cd6af9869b3d29aba40154e947d2a07bb499b3e310d6"}, + {file = "httpx-0.23.3.tar.gz", hash = "sha256:9818458eb565bb54898ccb9b8b251a28785dd4a55afbc23d0eb410754fe7d0f9"}, +] + +[package.dependencies] +certifi = "*" +httpcore = ">=0.15.0,<0.17.0" +rfc3986 = {version = ">=1.3,<2", extras = ["idna2008"]} +sniffio = "*" + +[package.extras] +brotli = ["brotli", "brotlicffi"] +cli = ["click (>=8.0.0,<9.0.0)", "pygments (>=2.0.0,<3.0.0)", "rich (>=10,<13)"] +http2 = ["h2 (>=3,<5)"] +socks = ["socksio (>=1.0.0,<2.0.0)"] + +[[package]] +name = "huggingface-hub" +version = "0.15.1" +description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "huggingface_hub-0.15.1-py3-none-any.whl", hash = "sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445"}, + {file = "huggingface_hub-0.15.1.tar.gz", hash = "sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081"}, +] + +[package.dependencies] +filelock = "*" +fsspec = "*" +packaging = ">=20.9" +pyyaml = ">=5.1" +requests = "*" +tqdm = ">=4.42.1" +typing-extensions = ">=3.7.4.3" + +[package.extras] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"] +cli = ["InquirerPy (==0.3.4)"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"] +fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] +quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"] +tensorflow = ["graphviz", "pydot", "tensorflow"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "gradio", "jedi", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "soundfile", "urllib3 (<2.0)"] +torch = ["torch"] +typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] + +[[package]] +name = "idna" +version = "3.4" +description = "Internationalized Domain Names in Applications (IDNA)" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"}, + {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"}, +] + +[[package]] +name = "iniconfig" +version = "2.0.0" +description = "brain-dead simple config-ini parsing" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "iniconfig-2.0.0-py3-none-any.whl", hash = "sha256:b6a85871a79d2e3b22d2d1b94ac2824226a63c6b741c88f7ae975f18b6778374"}, + {file = "iniconfig-2.0.0.tar.gz", hash = "sha256:2d91e135bf72d31a410b17c16da610a82cb55f6b0477d1a902134b24a455b8b3"}, +] + +[[package]] +name = "isort" +version = "5.12.0" +description = "A Python utility / library to sort Python imports." +category = "dev" +optional = false +python-versions = ">=3.8.0" +files = [ + {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"}, + {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"}, +] + +[package.extras] +colors = ["colorama (>=0.4.3)"] +pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"] +plugins = ["setuptools"] +requirements-deprecated-finder = ["pip-api", "pipreqs"] + +[[package]] +name = "joblib" +version = "1.3.1" +description = "Lightweight pipelining with Python functions" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "joblib-1.3.1-py3-none-any.whl", hash = "sha256:89cf0529520e01b3de7ac7b74a8102c90d16d54c64b5dd98cafcd14307fdf915"}, + {file = "joblib-1.3.1.tar.gz", hash = "sha256:1f937906df65329ba98013dc9692fe22a4c5e4a648112de500508b18a21b41e3"}, +] + +[[package]] +name = "jsonschema" +version = "4.17.3" +description = "An implementation of JSON Schema validation for Python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "jsonschema-4.17.3-py3-none-any.whl", hash = "sha256:a870ad254da1a8ca84b6a2905cac29d265f805acc57af304784962a2aa6508f6"}, + {file = "jsonschema-4.17.3.tar.gz", hash = "sha256:0f864437ab8b6076ba6707453ef8f98a6a0d512a80e93f8abdb676f737ecb60d"}, +] + +[package.dependencies] +attrs = ">=17.4.0" +pyrsistent = ">=0.14.0,<0.17.0 || >0.17.0,<0.17.1 || >0.17.1,<0.17.2 || >0.17.2" + +[package.extras] +format = ["fqdn", "idna", "isoduration", "jsonpointer (>1.13)", "rfc3339-validator", "rfc3987", "uri-template", "webcolors (>=1.11)"] +format-nongpl = ["fqdn", "idna", "isoduration", "jsonpointer (>1.13)", "rfc3339-validator", "rfc3986-validator (>0.1.0)", "uri-template", "webcolors (>=1.11)"] + +[[package]] +name = "lazy-loader" +version = "0.3" +description = "lazy_loader" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "lazy_loader-0.3-py3-none-any.whl", hash = "sha256:1e9e76ee8631e264c62ce10006718e80b2cfc74340d17d1031e0f84af7478554"}, + {file = "lazy_loader-0.3.tar.gz", hash = "sha256:3b68898e34f5b2a29daaaac172c6555512d0f32074f147e2254e4a6d9d838f37"}, +] + +[package.extras] +lint = ["pre-commit (>=3.3)"] +test = ["pytest (>=7.4)", "pytest-cov (>=4.1)"] + +[[package]] +name = "libapi" +version = "0.1.0" +description = "Library for the API services" +category = "main" +optional = false +python-versions = "3.9.15" +files = [] +develop = true + +[package.dependencies] +cryptography = "^41.0.1" +environs = "^9.5.0" +libcommon = {path = "../../libs/libcommon", develop = true} +orjson = "^3.8.6" +pyjwt = {version = "^2.6.0", extras = ["crypto"]} +requests = "^2.28.2" +starlette = "^0.28.0" +starlette-prometheus = "^0.9.0" + +[package.source] +type = "directory" +url = "../../libs/libapi" + +[[package]] +name = "libcommon" +version = "0.6.8" +description = "Library for utils common to all the services" +category = "main" +optional = false +python-versions = "3.9.15" +files = [] +develop = true + +[package.dependencies] +appdirs = "^1.4.4" +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} +environs = "^9.5.0" +huggingface-hub = "^0.15.1" +mongo-types = "0.15.1" +mongoengine = "^0.24.2" +networkx = "^3.0" +numba = "0.56.4" +orjson = "^3.8.6" +pandas = "^2.0.1" +psutil = "^5.9.4" +pydub = "^0.25.1" +pymongo = {version = "^3.13.0", extras = ["srv"]} +pytz = "^2020.1" +requests = "^2.31.0" +soundfile = ">=0.12.1" +starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" + +[package.source] +type = "directory" +url = "../../libs/libcommon" + +[[package]] +name = "librosa" +version = "0.10.0.post2" +description = "Python module for audio and music processing" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"}, + {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"}, +] + +[package.dependencies] +audioread = ">=2.1.9" +decorator = ">=4.3.0" +joblib = ">=0.14" +lazy-loader = ">=0.1" +msgpack = ">=1.0" +numba = ">=0.51.0" +numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2" +pooch = ">=1.0,<1.7" +scikit-learn = ">=0.20.0" +scipy = ">=1.2.0" +soundfile = ">=0.12.1" +soxr = ">=0.3.2" +typing-extensions = ">=4.1.1" + +[package.extras] +display = ["matplotlib (>=3.3.0)"] +docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"] +tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"] + +[[package]] +name = "llvmlite" +version = "0.39.1" +description = "lightweight wrapper around basic LLVM functionality" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"}, + {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"}, + {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"}, + {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"}, + {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"}, + {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"}, + {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"}, + {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"}, + {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"}, + {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"}, +] + +[[package]] +name = "markdown-it-py" +version = "3.0.0" +description = "Python port of markdown-it. Markdown parsing, done right!" +category = "dev" +optional = false +python-versions = ">=3.8" +files = [ + {file = "markdown-it-py-3.0.0.tar.gz", hash = "sha256:e3f60a94fa066dc52ec76661e37c851cb232d92f9886b15cb560aaada2df8feb"}, + {file = "markdown_it_py-3.0.0-py3-none-any.whl", hash = "sha256:355216845c60bd96232cd8d8c40e8f9765cc86f46880e43a8fd22dc1a1a8cab1"}, +] + +[package.dependencies] +mdurl = ">=0.1,<1.0" + +[package.extras] +benchmarking = ["psutil", "pytest", "pytest-benchmark"] +code-style = ["pre-commit (>=3.0,<4.0)"] +compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"] +linkify = ["linkify-it-py (>=1,<3)"] +plugins = ["mdit-py-plugins"] +profiling = ["gprof2dot"] +rtd = ["jupyter_sphinx", "mdit-py-plugins", "myst-parser", "pyyaml", "sphinx", "sphinx-copybutton", "sphinx-design", "sphinx_book_theme"] +testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"] + +[[package]] +name = "marshmallow" +version = "3.19.0" +description = "A lightweight library for converting complex datatypes to and from native Python datatypes." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "marshmallow-3.19.0-py3-none-any.whl", hash = "sha256:93f0958568da045b0021ec6aeb7ac37c81bfcccbb9a0e7ed8559885070b3a19b"}, + {file = "marshmallow-3.19.0.tar.gz", hash = "sha256:90032c0fd650ce94b6ec6dc8dfeb0e3ff50c144586462c389b81a07205bedb78"}, +] + +[package.dependencies] +packaging = ">=17.0" + +[package.extras] +dev = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)", "pytest", "pytz", "simplejson", "tox"] +docs = ["alabaster (==0.7.12)", "autodocsumm (==0.2.9)", "sphinx (==5.3.0)", "sphinx-issues (==3.0.1)", "sphinx-version-warning (==1.1.2)"] +lint = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)"] +tests = ["pytest", "pytz", "simplejson"] + +[[package]] +name = "mccabe" +version = "0.6.1" +description = "McCabe checker, plugin for flake8" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "mccabe-0.6.1-py2.py3-none-any.whl", hash = "sha256:ab8a6258860da4b6677da4bd2fe5dc2c659cff31b3ee4f7f5d64e79735b80d42"}, + {file = "mccabe-0.6.1.tar.gz", hash = "sha256:dd8d182285a0fe56bace7f45b5e7d1a6ebcbf524e8f3bd87eb0f125271b8831f"}, +] + +[[package]] +name = "mdurl" +version = "0.1.2" +description = "Markdown URL utilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8"}, + {file = "mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba"}, +] + +[[package]] +name = "mongo-types" +version = "0.15.1" +description = "Type stubs for mongoengine w/ basic support for bson and pymongo" +category = "main" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "mongo-types-0.15.1.tar.gz", hash = "sha256:0a9deeb7733ea7da5db3711d92e22d93556b522f860bbff82e5df44c53bd06a9"}, + {file = "mongo_types-0.15.1-py3-none-any.whl", hash = "sha256:9417ae5b9a759c09630b5ec7d66904cc333c2d2fcfe75e2760a332ed5e267309"}, +] + +[[package]] +name = "mongoengine" +version = "0.24.2" +description = "MongoEngine is a Python Object-Document Mapper for working with MongoDB." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "mongoengine-0.24.2-py3-none-any.whl", hash = "sha256:f5c4e1b206b2ccffe4adc7a6283ed26dd799bd115a5fb1d2e885a075132cdb88"}, + {file = "mongoengine-0.24.2.tar.gz", hash = "sha256:c76d49658575bb995682e2e77c8ef7cda63faf939415b32ee923745d120f8b02"}, +] + +[package.dependencies] +pymongo = ">=3.4,<5.0" + +[[package]] +name = "msgpack" +version = "1.0.5" +description = "MessagePack serializer" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:525228efd79bb831cf6830a732e2e80bc1b05436b086d4264814b4b2955b2fa9"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4f8d8b3bf1ff2672567d6b5c725a1b347fe838b912772aa8ae2bf70338d5a198"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cdc793c50be3f01106245a61b739328f7dccc2c648b501e237f0699fe1395b81"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb47c21a8a65b165ce29f2bec852790cbc04936f502966768e4aae9fa763cb7"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e42b9594cc3bf4d838d67d6ed62b9e59e201862a25e9a157019e171fbe672dd3"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:55b56a24893105dc52c1253649b60f475f36b3aa0fc66115bffafb624d7cb30b"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:1967f6129fc50a43bfe0951c35acbb729be89a55d849fab7686004da85103f1c"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:20a97bf595a232c3ee6d57ddaadd5453d174a52594bf9c21d10407e2a2d9b3bd"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d25dd59bbbbb996eacf7be6b4ad082ed7eacc4e8f3d2df1ba43822da9bfa122a"}, + {file = "msgpack-1.0.5-cp310-cp310-win32.whl", hash = "sha256:382b2c77589331f2cb80b67cc058c00f225e19827dbc818d700f61513ab47bea"}, + {file = "msgpack-1.0.5-cp310-cp310-win_amd64.whl", hash = "sha256:4867aa2df9e2a5fa5f76d7d5565d25ec76e84c106b55509e78c1ede0f152659a"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9f5ae84c5c8a857ec44dc180a8b0cc08238e021f57abdf51a8182e915e6299f0"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:9e6ca5d5699bcd89ae605c150aee83b5321f2115695e741b99618f4856c50898"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5494ea30d517a3576749cad32fa27f7585c65f5f38309c88c6d137877fa28a5a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ab2f3331cb1b54165976a9d976cb251a83183631c88076613c6c780f0d6e45a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:28592e20bbb1620848256ebc105fc420436af59515793ed27d5c77a217477705"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe5c63197c55bce6385d9aee16c4d0641684628f63ace85f73571e65ad1c1e8d"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:ed40e926fa2f297e8a653c954b732f125ef97bdd4c889f243182299de27e2aa9"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b2de4c1c0538dcb7010902a2b97f4e00fc4ddf2c8cda9749af0e594d3b7fa3d7"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf22a83f973b50f9d38e55c6aade04c41ddda19b00c4ebc558930d78eecc64ed"}, + {file = "msgpack-1.0.5-cp311-cp311-win32.whl", hash = "sha256:c396e2cc213d12ce017b686e0f53497f94f8ba2b24799c25d913d46c08ec422c"}, + {file = "msgpack-1.0.5-cp311-cp311-win_amd64.whl", hash = "sha256:6c4c68d87497f66f96d50142a2b73b97972130d93677ce930718f68828b382e2"}, + {file = "msgpack-1.0.5-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:a2b031c2e9b9af485d5e3c4520f4220d74f4d222a5b8dc8c1a3ab9448ca79c57"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f837b93669ce4336e24d08286c38761132bc7ab29782727f8557e1eb21b2080"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1d46dfe3832660f53b13b925d4e0fa1432b00f5f7210eb3ad3bb9a13c6204a6"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:366c9a7b9057e1547f4ad51d8facad8b406bab69c7d72c0eb6f529cf76d4b85f"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4c075728a1095efd0634a7dccb06204919a2f67d1893b6aa8e00497258bf926c"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:f933bbda5a3ee63b8834179096923b094b76f0c7a73c1cfe8f07ad608c58844b"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:36961b0568c36027c76e2ae3ca1132e35123dcec0706c4b7992683cc26c1320c"}, + {file = "msgpack-1.0.5-cp36-cp36m-win32.whl", hash = "sha256:b5ef2f015b95f912c2fcab19c36814963b5463f1fb9049846994b007962743e9"}, + {file = "msgpack-1.0.5-cp36-cp36m-win_amd64.whl", hash = "sha256:288e32b47e67f7b171f86b030e527e302c91bd3f40fd9033483f2cacc37f327a"}, + {file = "msgpack-1.0.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:137850656634abddfb88236008339fdaba3178f4751b28f270d2ebe77a563b6c"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c05a4a96585525916b109bb85f8cb6511db1c6f5b9d9cbcbc940dc6b4be944b"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:56a62ec00b636583e5cb6ad313bbed36bb7ead5fa3a3e38938503142c72cba4f"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef8108f8dedf204bb7b42994abf93882da1159728a2d4c5e82012edd92c9da9f"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1835c84d65f46900920b3708f5ba829fb19b1096c1800ad60bae8418652a951d"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:e57916ef1bd0fee4f21c4600e9d1da352d8816b52a599c46460e93a6e9f17086"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:17358523b85973e5f242ad74aa4712b7ee560715562554aa2134d96e7aa4cbbf"}, + {file = "msgpack-1.0.5-cp37-cp37m-win32.whl", hash = "sha256:cb5aaa8c17760909ec6cb15e744c3ebc2ca8918e727216e79607b7bbce9c8f77"}, + {file = "msgpack-1.0.5-cp37-cp37m-win_amd64.whl", hash = "sha256:ab31e908d8424d55601ad7075e471b7d0140d4d3dd3272daf39c5c19d936bd82"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:b72d0698f86e8d9ddf9442bdedec15b71df3598199ba33322d9711a19f08145c"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:379026812e49258016dd84ad79ac8446922234d498058ae1d415f04b522d5b2d"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:332360ff25469c346a1c5e47cbe2a725517919892eda5cfaffe6046656f0b7bb"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:476a8fe8fae289fdf273d6d2a6cb6e35b5a58541693e8f9f019bfe990a51e4ba"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a9985b214f33311df47e274eb788a5893a761d025e2b92c723ba4c63936b69b1"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:48296af57cdb1d885843afd73c4656be5c76c0c6328db3440c9601a98f303d87"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:addab7e2e1fcc04bd08e4eb631c2a90960c340e40dfc4a5e24d2ff0d5a3b3edb"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:916723458c25dfb77ff07f4c66aed34e47503b2eb3188b3adbec8d8aa6e00f48"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:821c7e677cc6acf0fd3f7ac664c98803827ae6de594a9f99563e48c5a2f27eb0"}, + {file = "msgpack-1.0.5-cp38-cp38-win32.whl", hash = "sha256:1c0f7c47f0087ffda62961d425e4407961a7ffd2aa004c81b9c07d9269512f6e"}, + {file = "msgpack-1.0.5-cp38-cp38-win_amd64.whl", hash = "sha256:bae7de2026cbfe3782c8b78b0db9cbfc5455e079f1937cb0ab8d133496ac55e1"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:20c784e66b613c7f16f632e7b5e8a1651aa5702463d61394671ba07b2fc9e025"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:266fa4202c0eb94d26822d9bfd7af25d1e2c088927fe8de9033d929dd5ba24c5"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:18334484eafc2b1aa47a6d42427da7fa8f2ab3d60b674120bce7a895a0a85bdd"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57e1f3528bd95cc44684beda696f74d3aaa8a5e58c816214b9046512240ef437"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:586d0d636f9a628ddc6a17bfd45aa5b5efaf1606d2b60fa5d87b8986326e933f"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a740fa0e4087a734455f0fc3abf5e746004c9da72fbd541e9b113013c8dc3282"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3055b0455e45810820db1f29d900bf39466df96ddca11dfa6d074fa47054376d"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:a61215eac016f391129a013c9e46f3ab308db5f5ec9f25811e811f96962599a8"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:362d9655cd369b08fda06b6657a303eb7172d5279997abe094512e919cf74b11"}, + {file = "msgpack-1.0.5-cp39-cp39-win32.whl", hash = "sha256:ac9dd47af78cae935901a9a500104e2dea2e253207c924cc95de149606dc43cc"}, + {file = "msgpack-1.0.5-cp39-cp39-win_amd64.whl", hash = "sha256:06f5174b5f8ed0ed919da0e62cbd4ffde676a374aba4020034da05fab67b9164"}, + {file = "msgpack-1.0.5.tar.gz", hash = "sha256:c075544284eadc5cddc70f4757331d99dcbc16b2bbd4849d15f8aae4cf36d31c"}, +] + +[[package]] +name = "multidict" +version = "6.0.4" +description = "multidict implementation" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"}, + {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"}, + {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"}, + {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"}, + {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"}, + {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"}, + {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"}, + {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"}, + {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"}, + {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"}, + {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"}, + {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"}, + {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"}, +] + +[[package]] +name = "multiprocess" +version = "0.70.14" +description = "better multiprocessing and multithreading in python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"}, + {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"}, + {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"}, + {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"}, + {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"}, + {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"}, +] + +[package.dependencies] +dill = ">=0.3.6" + +[[package]] +name = "mypy" +version = "1.4.1" +description = "Optional static typing for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mypy-1.4.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:566e72b0cd6598503e48ea610e0052d1b8168e60a46e0bfd34b3acf2d57f96a8"}, + {file = "mypy-1.4.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ca637024ca67ab24a7fd6f65d280572c3794665eaf5edcc7e90a866544076878"}, + {file = "mypy-1.4.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0dde1d180cd84f0624c5dcaaa89c89775550a675aff96b5848de78fb11adabcd"}, + {file = "mypy-1.4.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:8c4d8e89aa7de683e2056a581ce63c46a0c41e31bd2b6d34144e2c80f5ea53dc"}, + {file = "mypy-1.4.1-cp310-cp310-win_amd64.whl", hash = "sha256:bfdca17c36ae01a21274a3c387a63aa1aafe72bff976522886869ef131b937f1"}, + {file = "mypy-1.4.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7549fbf655e5825d787bbc9ecf6028731973f78088fbca3a1f4145c39ef09462"}, + {file = "mypy-1.4.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:98324ec3ecf12296e6422939e54763faedbfcc502ea4a4c38502082711867258"}, + {file = "mypy-1.4.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:141dedfdbfe8a04142881ff30ce6e6653c9685b354876b12e4fe6c78598b45e2"}, + {file = "mypy-1.4.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8207b7105829eca6f3d774f64a904190bb2231de91b8b186d21ffd98005f14a7"}, + {file = "mypy-1.4.1-cp311-cp311-win_amd64.whl", hash = "sha256:16f0db5b641ba159eff72cff08edc3875f2b62b2fa2bc24f68c1e7a4e8232d01"}, + {file = "mypy-1.4.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:470c969bb3f9a9efcedbadcd19a74ffb34a25f8e6b0e02dae7c0e71f8372f97b"}, + {file = "mypy-1.4.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5952d2d18b79f7dc25e62e014fe5a23eb1a3d2bc66318df8988a01b1a037c5b"}, + {file = "mypy-1.4.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:190b6bab0302cec4e9e6767d3eb66085aef2a1cc98fe04936d8a42ed2ba77bb7"}, + {file = "mypy-1.4.1-cp37-cp37m-win_amd64.whl", hash = "sha256:9d40652cc4fe33871ad3338581dca3297ff5f2213d0df345bcfbde5162abf0c9"}, + {file = "mypy-1.4.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:01fd2e9f85622d981fd9063bfaef1aed6e336eaacca00892cd2d82801ab7c042"}, + {file = "mypy-1.4.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2460a58faeea905aeb1b9b36f5065f2dc9a9c6e4c992a6499a2360c6c74ceca3"}, + {file = "mypy-1.4.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a2746d69a8196698146a3dbe29104f9eb6a2a4d8a27878d92169a6c0b74435b6"}, + {file = "mypy-1.4.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:ae704dcfaa180ff7c4cfbad23e74321a2b774f92ca77fd94ce1049175a21c97f"}, + {file = "mypy-1.4.1-cp38-cp38-win_amd64.whl", hash = "sha256:43d24f6437925ce50139a310a64b2ab048cb2d3694c84c71c3f2a1626d8101dc"}, + {file = "mypy-1.4.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c482e1246726616088532b5e964e39765b6d1520791348e6c9dc3af25b233828"}, + {file = "mypy-1.4.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:43b592511672017f5b1a483527fd2684347fdffc041c9ef53428c8dc530f79a3"}, + {file = "mypy-1.4.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:34a9239d5b3502c17f07fd7c0b2ae6b7dd7d7f6af35fbb5072c6208e76295816"}, + {file = "mypy-1.4.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5703097c4936bbb9e9bce41478c8d08edd2865e177dc4c52be759f81ee4dd26c"}, + {file = "mypy-1.4.1-cp39-cp39-win_amd64.whl", hash = "sha256:e02d700ec8d9b1859790c0475df4e4092c7bf3272a4fd2c9f33d87fac4427b8f"}, + {file = "mypy-1.4.1-py3-none-any.whl", hash = "sha256:45d32cec14e7b97af848bddd97d85ea4f0db4d5a149ed9676caa4eb2f7402bb4"}, + {file = "mypy-1.4.1.tar.gz", hash = "sha256:9bbcd9ab8ea1f2e1c8031c21445b511442cc45c89951e49bbf852cbb70755b1b"}, +] + +[package.dependencies] +mypy-extensions = ">=1.0.0" +tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""} +typing-extensions = ">=4.1.0" + +[package.extras] +dmypy = ["psutil (>=4.0)"] +install-types = ["pip"] +python2 = ["typed-ast (>=1.4.0,<2)"] +reports = ["lxml"] + +[[package]] +name = "mypy-extensions" +version = "1.0.0" +description = "Type system extensions for programs checked with the mypy type checker." +category = "dev" +optional = false +python-versions = ">=3.5" +files = [ + {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"}, + {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"}, +] + +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + +[[package]] +name = "numba" +version = "0.56.4" +description = "compiling Python code using LLVM" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"}, + {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"}, + {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"}, + {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"}, + {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"}, + {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"}, + {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"}, + {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"}, + {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"}, + {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"}, + {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"}, + {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"}, + {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"}, + {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"}, + {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"}, + {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"}, +] + +[package.dependencies] +llvmlite = ">=0.39.0dev0,<0.40" +numpy = ">=1.18,<1.24" +setuptools = "*" + +[[package]] +name = "numpy" +version = "1.23.5" +description = "NumPy is the fundamental package for array computing with Python." +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"}, + {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"}, + {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"}, + {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"}, + {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"}, + {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"}, + {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"}, + {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"}, + {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"}, + {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"}, +] + +[[package]] +name = "orjson" +version = "3.9.1" +description = "Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "orjson-3.9.1-cp310-cp310-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:c4434b7b786fdc394b95d029fb99949d7c2b05bbd4bf5cb5e3906be96ffeee3b"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09faf14f74ed47e773fa56833be118e04aa534956f661eb491522970b7478e3b"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:503eb86a8d53a187fe66aa80c69295a3ca35475804da89a9547e4fce5f803822"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:20f2804b5a1dbd3609c086041bd243519224d47716efd7429db6c03ed28b7cc3"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0fd828e0656615a711c4cc4da70f3cac142e66a6703ba876c20156a14e28e3fa"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ec53d648176f873203b9c700a0abacab33ca1ab595066e9d616f98cdc56f4434"}, + {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e186ae76b0d97c505500664193ddf508c13c1e675d9b25f1f4414a7606100da6"}, + {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d4edee78503016f4df30aeede0d999b3cb11fb56f47e9db0e487bce0aaca9285"}, + {file = "orjson-3.9.1-cp310-none-win_amd64.whl", hash = "sha256:a4cc5d21e68af982d9a2528ac61e604f092c60eed27aef3324969c68f182ec7e"}, + {file = "orjson-3.9.1-cp311-cp311-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:761b6efd33c49de20dd73ce64cc59da62c0dab10aa6015f582680e0663cc792c"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:31229f9d0b8dc2ef7ee7e4393f2e4433a28e16582d4b25afbfccc9d68dc768f8"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b7ab18d55ecb1de543d452f0a5f8094b52282b916aa4097ac11a4c79f317b86"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:db774344c39041f4801c7dfe03483df9203cbd6c84e601a65908e5552228dd25"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ae47ef8c0fe89c4677db7e9e1fb2093ca6e66c3acbee5442d84d74e727edad5e"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:103952c21575b9805803c98add2eaecd005580a1e746292ed2ec0d76dd3b9746"}, + {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:2cb0121e6f2c9da3eddf049b99b95fef0adf8480ea7cb544ce858706cdf916eb"}, + {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:24d4ddaa2876e657c0fd32902b5c451fd2afc35159d66a58da7837357044b8c2"}, + {file = "orjson-3.9.1-cp311-none-win_amd64.whl", hash = "sha256:0b53b5f72cf536dd8aa4fc4c95e7e09a7adb119f8ff8ee6cc60f735d7740ad6a"}, + {file = "orjson-3.9.1-cp37-cp37m-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d4b68d01a506242316a07f1d2f29fb0a8b36cee30a7c35076f1ef59dce0890c1"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d9dd4abe6c6fd352f00f4246d85228f6a9847d0cc14f4d54ee553718c225388f"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9e20bca5e13041e31ceba7a09bf142e6d63c8a7467f5a9c974f8c13377c75af2"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d8ae0467d01eb1e4bcffef4486d964bfd1c2e608103e75f7074ed34be5df48cc"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:06f6ab4697fab090517f295915318763a97a12ee8186054adf21c1e6f6abbd3d"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8515867713301fa065c58ec4c9053ba1a22c35113ab4acad555317b8fd802e50"}, + {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:393d0697d1dfa18d27d193e980c04fdfb672c87f7765b87952f550521e21b627"}, + {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d96747662d3666f79119e5d28c124e7d356c7dc195cd4b09faea4031c9079dc9"}, + {file = "orjson-3.9.1-cp37-none-win_amd64.whl", hash = "sha256:6d173d3921dd58a068c88ec22baea7dbc87a137411501618b1292a9d6252318e"}, + {file = "orjson-3.9.1-cp38-cp38-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d1c2b0b4246c992ce2529fc610a446b945f1429445ece1c1f826a234c829a918"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:19f70ba1f441e1c4bb1a581f0baa092e8b3e3ce5b2aac2e1e090f0ac097966da"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:375d65f002e686212aac42680aed044872c45ee4bc656cf63d4a215137a6124a"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4751cee4a7b1daeacb90a7f5adf2170ccab893c3ab7c5cea58b45a13f89b30b3"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78d9a2a4b2302d5ebc3695498ebc305c3568e5ad4f3501eb30a6405a32d8af22"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:46b4facc32643b2689dfc292c0c463985dac4b6ab504799cf51fc3c6959ed668"}, + {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:ec7c8a0f1bf35da0d5fd14f8956f3b82a9a6918a3c6963d718dfd414d6d3b604"}, + {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:d3a40b0fbe06ccd4d6a99e523d20b47985655bcada8d1eba485b1b32a43e4904"}, + {file = "orjson-3.9.1-cp38-none-win_amd64.whl", hash = "sha256:402f9d3edfec4560a98880224ec10eba4c5f7b4791e4bc0d4f4d8df5faf2a006"}, + {file = "orjson-3.9.1-cp39-cp39-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:49c0d78dcd34626e2e934f1192d7c052b94e0ecadc5f386fd2bda6d2e03dadf5"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:125f63e56d38393daa0a1a6dc6fedefca16c538614b66ea5997c3bd3af35ef26"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:08927970365d2e1f3ce4894f9ff928a7b865d53f26768f1bbdd85dd4fee3e966"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f9a744e212d4780ecd67f4b6b128b2e727bee1df03e7059cddb2dfe1083e7dc4"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5d1dbf36db7240c61eec98c8d21545d671bce70be0730deb2c0d772e06b71af3"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:80a1e384626f76b66df615f7bb622a79a25c166d08c5d2151ffd41f24c4cc104"}, + {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:15d28872fb055bf17ffca913826e618af61b2f689d2b170f72ecae1a86f80d52"}, + {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:1e4d905338f9ef32c67566929dfbfbb23cc80287af8a2c38930fb0eda3d40b76"}, + {file = "orjson-3.9.1-cp39-none-win_amd64.whl", hash = "sha256:48a27da6c7306965846565cc385611d03382bbd84120008653aa2f6741e2105d"}, + {file = "orjson-3.9.1.tar.gz", hash = "sha256:db373a25ec4a4fccf8186f9a72a1b3442837e40807a736a815ab42481e83b7d0"}, +] + +[[package]] +name = "packageurl-python" +version = "0.11.1" +description = "A purl aka. Package URL parser and builder" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "packageurl-python-0.11.1.tar.gz", hash = "sha256:bbcc53d2cb5920c815c1626c75992f319bfc450b73893fa7bd8aac5869aa49fe"}, + {file = "packageurl_python-0.11.1-py3-none-any.whl", hash = "sha256:4bad1d3ea4feb5e7a1db5ca8fb690ac9c82ab18e08d500755947b853df68817d"}, +] + +[package.extras] +build = ["wheel"] +lint = ["black", "isort", "mypy"] +test = ["pytest"] + +[[package]] +name = "packaging" +version = "23.1" +description = "Core utilities for Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "packaging-23.1-py3-none-any.whl", hash = "sha256:994793af429502c4ea2ebf6bf664629d07c1a9fe974af92966e4b8d2df7edc61"}, + {file = "packaging-23.1.tar.gz", hash = "sha256:a392980d2b6cffa644431898be54b0045151319d1e7ec34f0cfed48767dd334f"}, +] + +[[package]] +name = "pandas" +version = "2.0.3" +description = "Powerful data structures for data analysis, time series, and statistics" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "pandas-2.0.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e4c7c9f27a4185304c7caf96dc7d91bc60bc162221152de697c98eb0b2648dd8"}, + {file = "pandas-2.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f167beed68918d62bffb6ec64f2e1d8a7d297a038f86d4aed056b9493fca407f"}, + {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ce0c6f76a0f1ba361551f3e6dceaff06bde7514a374aa43e33b588ec10420183"}, + {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ba619e410a21d8c387a1ea6e8a0e49bb42216474436245718d7f2e88a2f8d7c0"}, + {file = "pandas-2.0.3-cp310-cp310-win32.whl", hash = "sha256:3ef285093b4fe5058eefd756100a367f27029913760773c8bf1d2d8bebe5d210"}, + {file = "pandas-2.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:9ee1a69328d5c36c98d8e74db06f4ad518a1840e8ccb94a4ba86920986bb617e"}, + {file = "pandas-2.0.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b084b91d8d66ab19f5bb3256cbd5ea661848338301940e17f4492b2ce0801fe8"}, + {file = "pandas-2.0.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:37673e3bdf1551b95bf5d4ce372b37770f9529743d2498032439371fc7b7eb26"}, + {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b9cb1e14fdb546396b7e1b923ffaeeac24e4cedd14266c3497216dd4448e4f2d"}, + {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9cd88488cceb7635aebb84809d087468eb33551097d600c6dad13602029c2df"}, + {file = "pandas-2.0.3-cp311-cp311-win32.whl", hash = "sha256:694888a81198786f0e164ee3a581df7d505024fbb1f15202fc7db88a71d84ebd"}, + {file = "pandas-2.0.3-cp311-cp311-win_amd64.whl", hash = "sha256:6a21ab5c89dcbd57f78d0ae16630b090eec626360085a4148693def5452d8a6b"}, + {file = "pandas-2.0.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9e4da0d45e7f34c069fe4d522359df7d23badf83abc1d1cef398895822d11061"}, + {file = "pandas-2.0.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:32fca2ee1b0d93dd71d979726b12b61faa06aeb93cf77468776287f41ff8fdc5"}, + {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:258d3624b3ae734490e4d63c430256e716f488c4fcb7c8e9bde2d3aa46c29089"}, + {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eae3dc34fa1aa7772dd3fc60270d13ced7346fcbcfee017d3132ec625e23bb0"}, + {file = "pandas-2.0.3-cp38-cp38-win32.whl", hash = "sha256:f3421a7afb1a43f7e38e82e844e2bca9a6d793d66c1a7f9f0ff39a795bbc5e02"}, + {file = "pandas-2.0.3-cp38-cp38-win_amd64.whl", hash = "sha256:69d7f3884c95da3a31ef82b7618af5710dba95bb885ffab339aad925c3e8ce78"}, + {file = "pandas-2.0.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5247fb1ba347c1261cbbf0fcfba4a3121fbb4029d95d9ef4dc45406620b25c8b"}, + {file = "pandas-2.0.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:81af086f4543c9d8bb128328b5d32e9986e0c84d3ee673a2ac6fb57fd14f755e"}, + {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1994c789bf12a7c5098277fb43836ce090f1073858c10f9220998ac74f37c69b"}, + {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5ec591c48e29226bcbb316e0c1e9423622bc7a4eaf1ef7c3c9fa1a3981f89641"}, + {file = "pandas-2.0.3-cp39-cp39-win32.whl", hash = "sha256:04dbdbaf2e4d46ca8da896e1805bc04eb85caa9a82e259e8eed00254d5e0c682"}, + {file = "pandas-2.0.3-cp39-cp39-win_amd64.whl", hash = "sha256:1168574b036cd8b93abc746171c9b4f1b83467438a5e45909fed645cf8692dbc"}, + {file = "pandas-2.0.3.tar.gz", hash = "sha256:c02f372a88e0d17f36d3093a644c73cfc1788e876a7c4bcb4020a77512e2043c"}, +] + +[package.dependencies] +numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""} +python-dateutil = ">=2.8.2" +pytz = ">=2020.1" +tzdata = ">=2022.1" + +[package.extras] +all = ["PyQt5 (>=5.15.1)", "SQLAlchemy (>=1.4.16)", "beautifulsoup4 (>=4.9.3)", "bottleneck (>=1.3.2)", "brotlipy (>=0.7.0)", "fastparquet (>=0.6.3)", "fsspec (>=2021.07.0)", "gcsfs (>=2021.07.0)", "html5lib (>=1.1)", "hypothesis (>=6.34.2)", "jinja2 (>=3.0.0)", "lxml (>=4.6.3)", "matplotlib (>=3.6.1)", "numba (>=0.53.1)", "numexpr (>=2.7.3)", "odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pandas-gbq (>=0.15.0)", "psycopg2 (>=2.8.6)", "pyarrow (>=7.0.0)", "pymysql (>=1.0.2)", "pyreadstat (>=1.1.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)", "python-snappy (>=0.6.0)", "pyxlsb (>=1.0.8)", "qtpy (>=2.2.0)", "s3fs (>=2021.08.0)", "scipy (>=1.7.1)", "tables (>=3.6.1)", "tabulate (>=0.8.9)", "xarray (>=0.21.0)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)", "zstandard (>=0.15.2)"] +aws = ["s3fs (>=2021.08.0)"] +clipboard = ["PyQt5 (>=5.15.1)", "qtpy (>=2.2.0)"] +compression = ["brotlipy (>=0.7.0)", "python-snappy (>=0.6.0)", "zstandard (>=0.15.2)"] +computation = ["scipy (>=1.7.1)", "xarray (>=0.21.0)"] +excel = ["odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pyxlsb (>=1.0.8)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)"] +feather = ["pyarrow (>=7.0.0)"] +fss = ["fsspec (>=2021.07.0)"] +gcp = ["gcsfs (>=2021.07.0)", "pandas-gbq (>=0.15.0)"] +hdf5 = ["tables (>=3.6.1)"] +html = ["beautifulsoup4 (>=4.9.3)", "html5lib (>=1.1)", "lxml (>=4.6.3)"] +mysql = ["SQLAlchemy (>=1.4.16)", "pymysql (>=1.0.2)"] +output-formatting = ["jinja2 (>=3.0.0)", "tabulate (>=0.8.9)"] +parquet = ["pyarrow (>=7.0.0)"] +performance = ["bottleneck (>=1.3.2)", "numba (>=0.53.1)", "numexpr (>=2.7.1)"] +plot = ["matplotlib (>=3.6.1)"] +postgresql = ["SQLAlchemy (>=1.4.16)", "psycopg2 (>=2.8.6)"] +spss = ["pyreadstat (>=1.1.2)"] +sql-other = ["SQLAlchemy (>=1.4.16)"] +test = ["hypothesis (>=6.34.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)"] +xml = ["lxml (>=4.6.3)"] + +[[package]] +name = "pandas-stubs" +version = "1.5.3.230321" +description = "Type annotations for pandas" +category = "dev" +optional = false +python-versions = ">=3.8,<3.12" +files = [ + {file = "pandas_stubs-1.5.3.230321-py3-none-any.whl", hash = "sha256:4bf36b3071dd55f0e558ac8efe07676a120f2ed89e7a3df0fb78ddf2733bf247"}, + {file = "pandas_stubs-1.5.3.230321.tar.gz", hash = "sha256:2fa860df9e6058e9f0d2c09bc711c09abb8f0516eee7f0b9f9950d29b835fc6f"}, +] + +[package.dependencies] +types-pytz = ">=2022.1.1" + +[[package]] +name = "pathspec" +version = "0.11.1" +description = "Utility library for gitignore style pattern matching of file paths." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pathspec-0.11.1-py3-none-any.whl", hash = "sha256:d8af70af76652554bd134c22b3e8a1cc46ed7d91edcdd721ef1a0c51a84a5293"}, + {file = "pathspec-0.11.1.tar.gz", hash = "sha256:2798de800fa92780e33acca925945e9a19a133b715067cf165b8866c15a31687"}, +] + +[[package]] +name = "pbr" +version = "5.11.1" +description = "Python Build Reasonableness" +category = "dev" +optional = false +python-versions = ">=2.6" +files = [ + {file = "pbr-5.11.1-py2.py3-none-any.whl", hash = "sha256:567f09558bae2b3ab53cb3c1e2e33e726ff3338e7bae3db5dc954b3a44eef12b"}, + {file = "pbr-5.11.1.tar.gz", hash = "sha256:aefc51675b0b533d56bb5fd1c8c6c0522fe31896679882e1c4c63d5e4a0fccb3"}, +] + +[[package]] +name = "pillow" +version = "10.0.0" +description = "Python Imaging Library (Fork)" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "Pillow-10.0.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:1f62406a884ae75fb2f818694469519fb685cc7eaff05d3451a9ebe55c646891"}, + {file = "Pillow-10.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d5db32e2a6ccbb3d34d87c87b432959e0db29755727afb37290e10f6e8e62614"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:edf4392b77bdc81f36e92d3a07a5cd072f90253197f4a52a55a8cec48a12483b"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:520f2a520dc040512699f20fa1c363eed506e94248d71f85412b625026f6142c"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:8c11160913e3dd06c8ffdb5f233a4f254cb449f4dfc0f8f4549eda9e542c93d1"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a74ba0c356aaa3bb8e3eb79606a87669e7ec6444be352870623025d75a14a2bf"}, + {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d5d0dae4cfd56969d23d94dc8e89fb6a217be461c69090768227beb8ed28c0a3"}, + {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22c10cc517668d44b211717fd9775799ccec4124b9a7f7b3635fc5386e584992"}, + {file = "Pillow-10.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:dffe31a7f47b603318c609f378ebcd57f1554a3a6a8effbc59c3c69f804296de"}, + {file = "Pillow-10.0.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:9fb218c8a12e51d7ead2a7c9e101a04982237d4855716af2e9499306728fb485"}, + {file = "Pillow-10.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d35e3c8d9b1268cbf5d3670285feb3528f6680420eafe35cccc686b73c1e330f"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3ed64f9ca2f0a95411e88a4efbd7a29e5ce2cea36072c53dd9d26d9c76f753b3"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0b6eb5502f45a60a3f411c63187db83a3d3107887ad0d036c13ce836f8a36f1d"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:c1fbe7621c167ecaa38ad29643d77a9ce7311583761abf7836e1510c580bf3dd"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:cd25d2a9d2b36fcb318882481367956d2cf91329f6892fe5d385c346c0649629"}, + {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:3b08d4cc24f471b2c8ca24ec060abf4bebc6b144cb89cba638c720546b1cf538"}, + {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d737a602fbd82afd892ca746392401b634e278cb65d55c4b7a8f48e9ef8d008d"}, + {file = "Pillow-10.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:3a82c40d706d9aa9734289740ce26460a11aeec2d9c79b7af87bb35f0073c12f"}, + {file = "Pillow-10.0.0-cp312-cp312-macosx_10_10_x86_64.whl", hash = "sha256:d80cf684b541685fccdd84c485b31ce73fc5c9b5d7523bf1394ce134a60c6883"}, + {file = "Pillow-10.0.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:76de421f9c326da8f43d690110f0e79fe3ad1e54be811545d7d91898b4c8493e"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81ff539a12457809666fef6624684c008e00ff6bf455b4b89fd00a140eecd640"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce543ed15570eedbb85df19b0a1a7314a9c8141a36ce089c0a894adbfccb4568"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:685ac03cc4ed5ebc15ad5c23bc555d68a87777586d970c2c3e216619a5476223"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:d72e2ecc68a942e8cf9739619b7f408cc7b272b279b56b2c83c6123fcfa5cdff"}, + {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:d50b6aec14bc737742ca96e85d6d0a5f9bfbded018264b3b70ff9d8c33485551"}, + {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:00e65f5e822decd501e374b0650146063fbb30a7264b4d2744bdd7b913e0cab5"}, + {file = "Pillow-10.0.0-cp312-cp312-win_amd64.whl", hash = "sha256:f31f9fdbfecb042d046f9d91270a0ba28368a723302786c0009ee9b9f1f60199"}, + {file = "Pillow-10.0.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:349930d6e9c685c089284b013478d6f76e3a534e36ddfa912cde493f235372f3"}, + {file = "Pillow-10.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3a684105f7c32488f7153905a4e3015a3b6c7182e106fe3c37fbb5ef3e6994c3"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b4f69b3700201b80bb82c3a97d5e9254084f6dd5fb5b16fc1a7b974260f89f43"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3f07ea8d2f827d7d2a49ecf1639ec02d75ffd1b88dcc5b3a61bbb37a8759ad8d"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:040586f7d37b34547153fa383f7f9aed68b738992380ac911447bb78f2abe530"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:f88a0b92277de8e3ca715a0d79d68dc82807457dae3ab8699c758f07c20b3c51"}, + {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:c7cf14a27b0d6adfaebb3ae4153f1e516df54e47e42dcc073d7b3d76111a8d86"}, + {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:3400aae60685b06bb96f99a21e1ada7bc7a413d5f49bce739828ecd9391bb8f7"}, + {file = "Pillow-10.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:dbc02381779d412145331789b40cc7b11fdf449e5d94f6bc0b080db0a56ea3f0"}, + {file = "Pillow-10.0.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9211e7ad69d7c9401cfc0e23d49b69ca65ddd898976d660a2fa5904e3d7a9baa"}, + {file = "Pillow-10.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:faaf07ea35355b01a35cb442dd950d8f1bb5b040a7787791a535de13db15ed90"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c9f72a021fbb792ce98306ffb0c348b3c9cb967dce0f12a49aa4c3d3fdefa967"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f7c16705f44e0504a3a2a14197c1f0b32a95731d251777dcb060aa83022cb2d"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:76edb0a1fa2b4745fb0c99fb9fb98f8b180a1bbceb8be49b087e0b21867e77d3"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:368ab3dfb5f49e312231b6f27b8820c823652b7cd29cfbd34090565a015e99ba"}, + {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:608bfdee0d57cf297d32bcbb3c728dc1da0907519d1784962c5f0c68bb93e5a3"}, + {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5c6e3df6bdd396749bafd45314871b3d0af81ff935b2d188385e970052091017"}, + {file = "Pillow-10.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:7be600823e4c8631b74e4a0d38384c73f680e6105a7d3c6824fcf226c178c7e6"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-macosx_10_10_x86_64.whl", hash = "sha256:92be919bbc9f7d09f7ae343c38f5bb21c973d2576c1d45600fce4b74bafa7ac0"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f8182b523b2289f7c415f589118228d30ac8c355baa2f3194ced084dac2dbba"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:38250a349b6b390ee6047a62c086d3817ac69022c127f8a5dc058c31ccef17f3"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:88af2003543cc40c80f6fca01411892ec52b11021b3dc22ec3bc9d5afd1c5334"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:c189af0545965fa8d3b9613cfdb0cd37f9d71349e0f7750e1fd704648d475ed2"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce7b031a6fc11365970e6a5686d7ba8c63e4c1cf1ea143811acbb524295eabed"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:db24668940f82321e746773a4bc617bfac06ec831e5c88b643f91f122a785684"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:efe8c0681042536e0d06c11f48cebe759707c9e9abf880ee213541c5b46c5bf3"}, + {file = "Pillow-10.0.0.tar.gz", hash = "sha256:9c82b5b3e043c7af0d95792d0d20ccf68f61a1fec6b3530e718b688422727396"}, +] + +[package.extras] +docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-removed-in", "sphinxext-opengraph"] +tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"] + +[[package]] +name = "pip" +version = "23.1.2" +description = "The PyPA recommended tool for installing Python packages." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-23.1.2-py3-none-any.whl", hash = "sha256:3ef6ac33239e4027d9a5598a381b9d30880a1477e50039db2eac6e8a8f6d1b18"}, + {file = "pip-23.1.2.tar.gz", hash = "sha256:0e7c86f486935893c708287b30bd050a36ac827ec7fe5e43fe7cb198dd835fba"}, +] + +[[package]] +name = "pip-api" +version = "0.0.30" +description = "An unofficial, importable pip API" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-api-0.0.30.tar.gz", hash = "sha256:a05df2c7aa9b7157374bcf4273544201a0c7bae60a9c65bcf84f3959ef3896f3"}, + {file = "pip_api-0.0.30-py3-none-any.whl", hash = "sha256:2a0314bd31522eb9ffe8a99668b0d07fee34ebc537931e7b6483001dbedcbdc9"}, +] + +[package.dependencies] +pip = "*" + +[[package]] +name = "pip-audit" +version = "2.6.0" +description = "A tool for scanning Python environments for known vulnerabilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip_audit-2.6.0-py3-none-any.whl", hash = "sha256:49e97e3d6663d2ed0c00b7a7c468afcb816beb3988f32f8496d3fe3927cfd627"}, + {file = "pip_audit-2.6.0.tar.gz", hash = "sha256:6431c363efa80ef52c2599197c5b8a39ff8708ce316624b97fa35b5cdf493118"}, +] + +[package.dependencies] +CacheControl = {version = ">=0.13.0", extras = ["filecache"]} +cyclonedx-python-lib = ">=4.0,<5.0" +html5lib = ">=1.1" +packaging = ">=23.0.0" +pip-api = ">=0.0.28" +pip-requirements-parser = ">=32.0.0" +requests = ">=2.31.0" +rich = ">=12.4" +toml = ">=0.10" + +[package.extras] +dev = ["build", "bump (>=1.3.2)", "pip-audit[doc,lint,test]"] +doc = ["pdoc"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.276)", "types-html5lib", "types-requests", "types-toml"] +test = ["coverage[toml]", "pretend", "pytest", "pytest-cov"] + +[[package]] +name = "pip-requirements-parser" +version = "32.0.1" +description = "pip requirements parser - a mostly correct pip requirements parsing library because it uses pip's own code." +category = "dev" +optional = false +python-versions = ">=3.6.0" +files = [ + {file = "pip-requirements-parser-32.0.1.tar.gz", hash = "sha256:b4fa3a7a0be38243123cf9d1f3518da10c51bdb165a2b2985566247f9155a7d3"}, + {file = "pip_requirements_parser-32.0.1-py3-none-any.whl", hash = "sha256:4659bc2a667783e7a15d190f6fccf8b2486685b6dba4c19c3876314769c57526"}, +] + +[package.dependencies] +packaging = "*" +pyparsing = "*" + +[package.extras] +docs = ["Sphinx (>=3.3.1)", "doc8 (>=0.8.1)", "sphinx-rtd-theme (>=0.5.0)"] +testing = ["aboutcode-toolkit (>=6.0.0)", "black", "pytest (>=6,!=7.0.0)", "pytest-xdist (>=2)"] + +[[package]] +name = "platformdirs" +version = "3.8.0" +description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "platformdirs-3.8.0-py3-none-any.whl", hash = "sha256:ca9ed98ce73076ba72e092b23d3c93ea6c4e186b3f1c3dad6edd98ff6ffcca2e"}, + {file = "platformdirs-3.8.0.tar.gz", hash = "sha256:b0cabcb11063d21a0b261d557acb0a9d2126350e63b70cdf7db6347baea456dc"}, +] + +[package.extras] +docs = ["furo (>=2023.5.20)", "proselint (>=0.13)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"] +test = ["appdirs (==1.4.4)", "covdefaults (>=2.3)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)"] + +[[package]] +name = "pluggy" +version = "1.2.0" +description = "plugin and hook calling mechanisms for python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pluggy-1.2.0-py3-none-any.whl", hash = "sha256:c2fd55a7d7a3863cba1a013e4e2414658b1d07b6bc57b3919e0c63c9abb99849"}, + {file = "pluggy-1.2.0.tar.gz", hash = "sha256:d12f0c4b579b15f5e054301bb226ee85eeeba08ffec228092f8defbaa3a4c4b3"}, +] + +[package.extras] +dev = ["pre-commit", "tox"] +testing = ["pytest", "pytest-benchmark"] + +[[package]] +name = "pooch" +version = "1.6.0" +description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\"" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"}, + {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"}, +] + +[package.dependencies] +appdirs = ">=1.3.0" +packaging = ">=20.0" +requests = ">=2.19.0" + +[package.extras] +progress = ["tqdm (>=4.41.0,<5.0.0)"] +sftp = ["paramiko (>=2.7.0)"] +xxhash = ["xxhash (>=1.4.3)"] + +[[package]] +name = "prometheus-client" +version = "0.12.0" +description = "Python client for the Prometheus monitoring system." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "prometheus_client-0.12.0-py2.py3-none-any.whl", hash = "sha256:317453ebabff0a1b02df7f708efbab21e3489e7072b61cb6957230dd004a0af0"}, + {file = "prometheus_client-0.12.0.tar.gz", hash = "sha256:1b12ba48cee33b9b0b9de64a1047cbd3c5f2d0ab6ebcead7ddda613a750ec3c5"}, +] + +[package.extras] +twisted = ["twisted"] + +[[package]] +name = "psutil" +version = "5.9.5" +description = "Cross-platform lib for process and system monitoring in Python." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "psutil-5.9.5-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:be8929ce4313f9f8146caad4272f6abb8bf99fc6cf59344a3167ecd74f4f203f"}, + {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:ab8ed1a1d77c95453db1ae00a3f9c50227ebd955437bcf2a574ba8adbf6a74d5"}, + {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:4aef137f3345082a3d3232187aeb4ac4ef959ba3d7c10c33dd73763fbc063da4"}, + {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:ea8518d152174e1249c4f2a1c89e3e6065941df2fa13a1ab45327716a23c2b48"}, + {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:acf2aef9391710afded549ff602b5887d7a2349831ae4c26be7c807c0a39fac4"}, + {file = "psutil-5.9.5-cp27-none-win32.whl", hash = "sha256:5b9b8cb93f507e8dbaf22af6a2fd0ccbe8244bf30b1baad6b3954e935157ae3f"}, + {file = "psutil-5.9.5-cp27-none-win_amd64.whl", hash = "sha256:8c5f7c5a052d1d567db4ddd231a9d27a74e8e4a9c3f44b1032762bd7b9fdcd42"}, + {file = "psutil-5.9.5-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:3c6f686f4225553615612f6d9bc21f1c0e305f75d7d8454f9b46e901778e7217"}, + {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7a7dd9997128a0d928ed4fb2c2d57e5102bb6089027939f3b722f3a210f9a8da"}, + {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:89518112647f1276b03ca97b65cc7f64ca587b1eb0278383017c2a0dcc26cbe4"}, + {file = "psutil-5.9.5-cp36-abi3-win32.whl", hash = "sha256:104a5cc0e31baa2bcf67900be36acde157756b9c44017b86b2c049f11957887d"}, + {file = "psutil-5.9.5-cp36-abi3-win_amd64.whl", hash = "sha256:b258c0c1c9d145a1d5ceffab1134441c4c5113b2417fafff7315a917a026c3c9"}, + {file = "psutil-5.9.5-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:c607bb3b57dc779d55e1554846352b4e358c10fff3abf3514a7a6601beebdb30"}, + {file = "psutil-5.9.5.tar.gz", hash = "sha256:5410638e4df39c54d957fc51ce03048acd8e6d60abc0f5107af51e5fb566eb3c"}, +] + +[package.extras] +test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] + +[[package]] +name = "py-serializable" +version = "0.11.1" +description = "Library for serializing and deserializing Python Objects to and from JSON and XML." +category = "dev" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "py-serializable-0.11.1.tar.gz", hash = "sha256:ba0e1287b9e4f645a5334f1913abd8e647e7250209f84f55dce3909498a6f586"}, + {file = "py_serializable-0.11.1-py3-none-any.whl", hash = "sha256:79e21f0672822e6200b15f45ce9f636e8126466f62dbd7d488c67313c72b5c3e"}, +] + +[package.dependencies] +defusedxml = ">=0.7.1,<0.8.0" + +[[package]] +name = "pyarrow" +version = "11.0.0" +description = "Python library for Apache Arrow" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"}, + {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"}, + {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"}, + {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"}, + {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"}, + {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"}, + {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"}, + {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"}, + {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"}, + {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"}, + {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"}, + {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"}, + {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"}, + {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"}, + {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"}, + {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"}, + {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"}, +] + +[package.dependencies] +numpy = ">=1.16.6" + +[[package]] +name = "pycodestyle" +version = "2.7.0" +description = "Python style guide checker" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycodestyle-2.7.0-py2.py3-none-any.whl", hash = "sha256:514f76d918fcc0b55c6680472f0a37970994e07bbb80725808c17089be302068"}, + {file = "pycodestyle-2.7.0.tar.gz", hash = "sha256:c389c1d06bf7904078ca03399a4816f974a1d590090fecea0c63ec26ebaf1cef"}, +] + +[[package]] +name = "pycparser" +version = "2.21" +description = "C parser in Python" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, + {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, +] + +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + +[[package]] +name = "pyflakes" +version = "2.3.1" +description = "passive checker of Python programs" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pyflakes-2.3.1-py2.py3-none-any.whl", hash = "sha256:7893783d01b8a89811dd72d7dfd4d84ff098e5eed95cfa8905b22bbffe52efc3"}, + {file = "pyflakes-2.3.1.tar.gz", hash = "sha256:f5bc8ecabc05bb9d291eb5203d6810b49040f6ff446a756326104746cc00c1db"}, +] + +[[package]] +name = "pygments" +version = "2.15.1" +description = "Pygments is a syntax highlighting package written in Python." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pygments-2.15.1-py3-none-any.whl", hash = "sha256:db2db3deb4b4179f399a09054b023b6a586b76499d36965813c71aa8ed7b5fd1"}, + {file = "Pygments-2.15.1.tar.gz", hash = "sha256:8ace4d3c1dd481894b2005f560ead0f9f19ee64fe983366be1a21e171d12775c"}, +] + +[package.extras] +plugins = ["importlib-metadata"] + +[[package]] +name = "pyjwt" +version = "2.7.0" +description = "JSON Web Token implementation in Python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "PyJWT-2.7.0-py3-none-any.whl", hash = "sha256:ba2b425b15ad5ef12f200dc67dd56af4e26de2331f965c5439994dad075876e1"}, + {file = "PyJWT-2.7.0.tar.gz", hash = "sha256:bd6ca4a3c4285c1a2d4349e5a035fdf8fb94e04ccd0fcbe6ba289dae9cc3e074"}, +] + +[package.dependencies] +cryptography = {version = ">=3.4.0", optional = true, markers = "extra == \"crypto\""} + +[package.extras] +crypto = ["cryptography (>=3.4.0)"] +dev = ["coverage[toml] (==5.0.4)", "cryptography (>=3.4.0)", "pre-commit", "pytest (>=6.0.0,<7.0.0)", "sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"] +docs = ["sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"] +tests = ["coverage[toml] (==5.0.4)", "pytest (>=6.0.0,<7.0.0)"] + +[[package]] +name = "pymongo" +version = "3.13.0" +description = "Python driver for MongoDB <http://www.mongodb.org>" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl", hash = "sha256:3ad3a3df830f7df7e0856c2bdb54d19f5bf188bd7420985e18643b8e4d2a075f"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:b96e0e9d2d48948240b510bac81614458fc10adcd3a93240c2fd96448b4efd35"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9f592b202d77923498b32ddc5b376e5fa9ba280d3e16ed56cb8c932fe6d6a478"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:851f2bb52b5cb2f4711171ca925e0e05344a8452972a748a8a8ffdda1e1d72a7"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:1c9d23f62a3fa7523d849c4942acc0d9ff7081ebc00c808ee7cfdc070df0687f"}, + {file = "pymongo-3.13.0-cp27-cp27m-win32.whl", hash = "sha256:a17b81f22398e3e0f72bdf938e98c810286994b2bcc0a125cd5ad8fd4ea54ad7"}, + {file = "pymongo-3.13.0-cp27-cp27m-win_amd64.whl", hash = "sha256:4f6dd55dab77adf60b445c11f426ee5cdfa1b86f6d54cb937bfcbf09572333ab"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:776f90bf2252f90a4ae838e7917638894c6356bef7265f424592e2fd1f577d05"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:50b99f4d3eee6f03778fe841d6f470e6c18e744dc665156da6da3bc6e65b398d"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50a81b2d9f188c7909e0a1084fa969bb92a788076809c437ac1ae80393f46df9"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c7c45a8a1a752002b0a7c81ab3a4c5e3b6f67f9826b16fbe3943f5329f565f24"}, + {file = "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1037097708498bdc85f23c8798a5c46c7bce432d77d23608ff14e0d831f1a971"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl", hash = "sha256:b5b733694e7df22d5c049581acfc487695a6ff813322318bed8dd66f79978636"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:d7c91747ec8dde51440dd594603158cc98abb3f7df84b2ed8a836f138285e4fb"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl", hash = "sha256:f4175fcdddf764d371ee52ec4505a40facee2533e84abf2953cda86d050cfa1f"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl", hash = "sha256:93d4e9a02c17813b34e4bd9f6fbf07310c140c8f74341537c24d07c1cdeb24d1"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl", hash = "sha256:3b261d593f2563299062733ae003a925420a86ff4ddda68a69097d67204e43f3"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl", hash = "sha256:172db03182a22e9002157b262c1ea3b0045c73d4ff465adc152ce5b4b0e7b8d4"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09de3bfc995ae8cb955abb0c9ae963c134dba1b5622be3bcc527b89b0fd4091c"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c0379447587ee4b8f983ba183202496e86c0358f47c45612619d634d1fcd82bd"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30245a8747dc90019a3c9ad9df987e0280a3ea632ad36227cde7d1d8dcba0830"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65b6fddf6a7b91da044f202771a38e71bbb9bf42720a406b26b25fe2256e7102"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5831a377d15a626fbec10890ffebc4c6abcd37e4126737932cd780a171eabdc1"}, + {file = "pymongo-3.13.0-cp310-cp310-win32.whl", hash = "sha256:944249aa83dee314420c37d0f40c30a8f6dc4a3877566017b87062e53af449f4"}, + {file = "pymongo-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea8824ebc9a1a5c8269e8f1e3989b5a6bec876726e2f3c33ebd036cb488277f0"}, + {file = "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:bdd34c57b4da51a7961beb33645646d197e41f8517801dc76b37c1441e7a4e10"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26f9cc42a162faa241c82e117ac85734ae9f14343dc2df1c90c6b2181f791b22"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a82a1c10f5608e6494913faa169e213d703194bfca0aa710901f303be212414"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8927f22ef6a16229da7f18944deac8605bdc2c0858be5184259f2f7ce7fd4459"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e6f8191a282ef77e526f8f8f63753a437e4aa4bc78f5edd8b6b6ed0eaebd5363"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4d9ed67c987bf9ac2ac684590ba3d2599cdfb0f331ee3db607f9684469b3b59d"}, + {file = "pymongo-3.13.0-cp311-cp311-win32.whl", hash = "sha256:e8f6979664ff477cd61b06bf8aba206df7b2334209815ab3b1019931dab643d6"}, + {file = "pymongo-3.13.0-cp311-cp311-win_amd64.whl", hash = "sha256:174fd1000e896d0dfbc7f6d7e6a1992a4868796c7dec31679e38218c78d6a942"}, + {file = "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl", hash = "sha256:d1ee773fb72ba024e7e3bb6ea8907fe52bccafcb5184aaced6bad995bd30ea20"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:28565e3dbd69fe5fe35a210067064dbb6ed5abe997079f653c19c873c3896fe6"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:5c1db7d366004d6c699eb08c716a63ae0a3e946d061cbebea65d7ce361950265"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e1956f3338c10308e2f99c2c9ff46ae412035cbcd7aaa76c39ccdb806854a247"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:10f0fddc1d63ba3d4a4bffcc7720184c1b7efd570726ad5e2f55818da320239f"}, + {file = "pymongo-3.13.0-cp35-cp35m-win32.whl", hash = "sha256:570ae3365b23d4fd8c669cb57613b1a90b2757e993588d3370ef90945dbeec4b"}, + {file = "pymongo-3.13.0-cp35-cp35m-win_amd64.whl", hash = "sha256:79f777eaf3f5b2c6d81f9ef00d87837001d7063302503bbcbfdbf3e9bc27c96f"}, + {file = "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl", hash = "sha256:d42eb29ba314adfd9c11234b4b646f61b0448bf9b00f14db4b317e6e4b947e77"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:e5e87c0eb774561c546f979342a8ff36ebee153c60a0b6c6b03ba989ceb9538c"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:0f2c5a5984599a88d087a15859860579b825098b473d8c843f1979a83d159f2e"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:59c98e86c5e861032b71e6e5b65f23e6afaacea6e82483b66f1191a5021a7b4f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:70b67390e27e58876853efbb87e43c85252de2515e2887f7dd901b4fa3d21973"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:42ba8606492d76e6f9e4c7a458ed4bc712603be393259a52450345f0945da2cf"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:0e5536994cf2d8488c6fd9dea71df3c4dbb3e0d2ba5e695da06d9142a29a0969"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:fe8194f107f0fa3cabd14e9e809f174eca335993c1db72d1e74e0f496e7afe1f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d593d50815771f517d3ac4367ff716e3f3c78edae51d98e1e25791459f8848ff"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5136ebe8da6a1604998a8eb96be55935aa5f7129c41cc7bddc400d48e8df43be"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a424bdedfd84454d2905a861e0d4bb947cc5bd024fdeb3600c1a97d2be0f4255"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5161167b3840e9c84c80f2534ea6a099f51749d5673b662a3dd248be17c3208"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:644470442beaf969df99c4e00367a817eee05f0bba5d888f1ba6fe97b5e1c102"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2406df90b2335371706c59b7d79e9633b81ed2a7ecd48c1faf8584552bdf2d90"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:222591b828de10ac90064047b5d4916953f38c38b155009c4b8b5e0d33117c2b"}, + {file = "pymongo-3.13.0-cp36-cp36m-win32.whl", hash = "sha256:7cb987b199fa223ad78eebaa9fbc183d5a5944bfe568a9d6f617316ca1c1f32f"}, + {file = "pymongo-3.13.0-cp36-cp36m-win_amd64.whl", hash = "sha256:a6cbb73d9fc2282677e2b7a137d13da987bd0b13abd88ed27bba5534c226db06"}, + {file = "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl", hash = "sha256:b1223b826acbef07a7f5eb9bf37247b0b580119916dca9eae19d92b1290f5855"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:398fb86d374dc351a4abc2e24cd15e5e14b2127f6d90ce0df3fdf2adcc55ac1b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:9c3d07ea19cd2856d9943dce37e75d69ecbb5baf93c3e4c82f73b6075c481292"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:2943d739715f265a2983ac43747595b6af3312d0a370614040959fd293763adf"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:c3b70ed82f20d18d22eafc9bda0ea656605071762f7d31f3c5afc35c59d3393b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:7ec2bb598847569ae34292f580842d37619eea3e546005042f485e15710180d5"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:8cc37b437cba909bef06499dadd91a39c15c14225e8d8c7870020049f8a549fe"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:65a063970e15a4f338f14b820561cf6cdaf2839691ac0adb2474ddff9d0b8b0b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02f0e1a75d3bc0e16c7e15daf9c56185642be055e425f3b34888fc6eb1b22401"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16e74b9c2aca2734c7f49f00fe68d6830a30d26df60e2ace7fe40ccb92087b94"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:24e954be35ad4537840f20bbc8d75320ae647d3cb4fab12cb8fcd2d55f408e76"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a149377d1ff766fd618500798d0d94637f66d0ae222bb6d28f41f3e15c626297"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61660710b054ae52c8fc10368e91d74719eb05554b631d7f8ca93d21d2bff2e6"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4bbc0d27dfef7689285e54f2e0a224f0c7cd9d5c46d2638fabad5500b951c92f"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:9b2ed9c3b30f11cd4a3fbfc22167af7987b01b444215c2463265153fe7cf66d6"}, + {file = "pymongo-3.13.0-cp37-cp37m-win32.whl", hash = "sha256:1c2c5e2b00e2fadcd590c0b2e293d71215e98ed1cb635cfca2be4998d197e534"}, + {file = "pymongo-3.13.0-cp37-cp37m-win_amd64.whl", hash = "sha256:32eac95bbb030b2376ffd897376c6f870222a3457f01a9ce466b9057876132f8"}, + {file = "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a796ef39dadf9d73af05d24937644d386495e43a7d13617aa3651d836da542c8"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:b6793baf4639c72a500698a49e9250b293e17ae1faf11ac1699d8141194786fe"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:80d8576b04d0824f63bf803190359c0d3bcb6e7fa63fefbd4bc0ceaa7faae38c"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:db2e11507fe9cc2a722be21ccc62c1b1295398fe9724c1f14900cdc7166fc0d7"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:b01ce58eec5edeededf1992d2dce63fb8565e437be12d6f139d75b15614c4d08"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:d1a19d6c5098f1f4e11430cd74621699453cbc534dd7ade9167e582f50814b19"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:7219b1a726ced3bacecabef9bd114529bbb69477901373e800d7d0140baadc95"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:2dae3b353a10c3767e0aa1c1492f2af388f1012b08117695ab3fd1f219e5814e"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:12721d926d43d33dd3318e58dce9b0250e8a9c6e1093fa8e09f4805193ff4b43"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6af0a4b17faf26779d5caee8542a4f2cba040cea27d3bffc476cbc6ccbd4c8ee"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:09b9d0f5a445c7e0ddcc021b09835aa6556f0166afc498f57dfdd72cdf6f02ad"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db5b4f8ad8607a3d612da1d4c89a84e4cf5c88f98b46365820d9babe5884ba45"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:34dbf5fecf653c152edb75a35a8b15dfdc4549473484ee768aeb12c97983cead"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:34cd48df7e1fc69222f296d8f69e3957eb7c6b5aa0709d3467184880ed7538c0"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c8f755ff1f4ab4ca790d1d6d3229006100b301475948021b6b2757822e0d6c97"}, + {file = "pymongo-3.13.0-cp38-cp38-win32.whl", hash = "sha256:b0746d0d4535f56bbaa63a8f6da362f330804d578e66e126b226eebe76c2bf00"}, + {file = "pymongo-3.13.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ad0515abb132f52ce9d8abd1a29681a1e65dba7b7fe13ea01e1a8db5715bf80"}, + {file = "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:3c5cb6c93c94df76a879bad4b89db0104b01806d17c2b803c1316ba50962b6d6"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2e0854170813238f0c3131050c67cb1fb1ade75c93bf6cd156c1bd9a16095528"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:1410faa51ce835cc1234c99ec42e98ab4f3c6f50d92d86a2d4f6e11c97ee7a4e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:d7910135f5de1c5c3578e61d6f4b087715b15e365f11d4fa51a9cee92988b2bd"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:028175dd8d2979a889153a2308e8e500b3df7d9e3fd1c33ca7fdeadf61cc87a2"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:2bfc39276c0e6d07c95bd1088b5003f049e986e089509f7dbd68bb7a4b1e65ac"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:4092b660ec720d44d3ca81074280dc25c7a3718df1b6c0fe9fe36ac6ed2833e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:5bdeb71a610a7b801416268e500e716d0fe693fb10d809e17f0fb3dac5be5a34"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aa3bca8e76f5c00ed2bb4325e0e383a547d71595926d5275d7c88175aaf7435e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7c7cab8155f430ca460a6fc7ae8a705b34f3e279a57adb5f900eb81943ec777c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4a32f3dfcca4a4816373bdb6256c18c78974ebb3430e7da988516cd95b2bd6e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30ed2788a6ec68743e2040ab1d16573d7d9f6e7333e45070ce9268cbc93d148c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21e61a536ffed84d10376c21c13a6ed1ebefb61989a844952547c229d6aeedf3"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0665412dce26b2318092a33bd2d2327d487c4490cfcde158d6946d39b1e28d78"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:64ed1a5ce5e5926727eb0f87c698c4d9a7a9f7b0953683a65e9ce2b7cc5f8e91"}, + {file = "pymongo-3.13.0-cp39-cp39-win32.whl", hash = "sha256:7593cb1214185a0c5b43b96effc51ce82ddc933298ee36db7dc2bd45d61b4adc"}, + {file = "pymongo-3.13.0-cp39-cp39-win_amd64.whl", hash = "sha256:3cfc9bc1e8b5667bc1f3dbe46d2f85b3f24ff7533893bdc1203058012db2c046"}, + {file = "pymongo-3.13.0.tar.gz", hash = "sha256:e22d6cf5802cd09b674c307cc9e03870b8c37c503ebec3d25b86f2ce8c535dc7"}, +] + +[package.dependencies] +dnspython = {version = ">=1.16.0,<1.17.0", optional = true, markers = "extra == \"srv\""} + +[package.extras] +aws = ["pymongo-auth-aws (<2.0.0)"] +encryption = ["pymongocrypt (>=1.1.0,<2.0.0)"] +gssapi = ["pykerberos"] +ocsp = ["certifi", "pyopenssl (>=17.2.0)", "requests (<3.0.0)", "service-identity (>=18.1.0)"] +snappy = ["python-snappy"] +srv = ["dnspython (>=1.16.0,<1.17.0)"] +tls = ["ipaddress"] +zstd = ["zstandard"] + +[[package]] +name = "pyparsing" +version = "3.1.0" +description = "pyparsing module - Classes and methods to define and execute parsing grammars" +category = "dev" +optional = false +python-versions = ">=3.6.8" +files = [ + {file = "pyparsing-3.1.0-py3-none-any.whl", hash = "sha256:d554a96d1a7d3ddaf7183104485bc19fd80543ad6ac5bdb6426719d766fb06c1"}, + {file = "pyparsing-3.1.0.tar.gz", hash = "sha256:edb662d6fe322d6e990b1594b5feaeadf806803359e3d4d42f11e295e588f0ea"}, +] + +[package.extras] +diagrams = ["jinja2", "railroad-diagrams"] + +[[package]] +name = "pyrsistent" +version = "0.19.3" +description = "Persistent/Functional/Immutable data structures" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyrsistent-0.19.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:20460ac0ea439a3e79caa1dbd560344b64ed75e85d8703943e0b66c2a6150e4a"}, + {file = "pyrsistent-0.19.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4c18264cb84b5e68e7085a43723f9e4c1fd1d935ab240ce02c0324a8e01ccb64"}, + {file = "pyrsistent-0.19.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4b774f9288dda8d425adb6544e5903f1fb6c273ab3128a355c6b972b7df39dcf"}, + {file = "pyrsistent-0.19.3-cp310-cp310-win32.whl", hash = "sha256:5a474fb80f5e0d6c9394d8db0fc19e90fa540b82ee52dba7d246a7791712f74a"}, + {file = "pyrsistent-0.19.3-cp310-cp310-win_amd64.whl", hash = "sha256:49c32f216c17148695ca0e02a5c521e28a4ee6c5089f97e34fe24163113722da"}, + {file = "pyrsistent-0.19.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:f0774bf48631f3a20471dd7c5989657b639fd2d285b861237ea9e82c36a415a9"}, + {file = "pyrsistent-0.19.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3ab2204234c0ecd8b9368dbd6a53e83c3d4f3cab10ecaf6d0e772f456c442393"}, + {file = "pyrsistent-0.19.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e42296a09e83028b3476f7073fcb69ffebac0e66dbbfd1bd847d61f74db30f19"}, + {file = "pyrsistent-0.19.3-cp311-cp311-win32.whl", hash = "sha256:64220c429e42a7150f4bfd280f6f4bb2850f95956bde93c6fda1b70507af6ef3"}, + {file = "pyrsistent-0.19.3-cp311-cp311-win_amd64.whl", hash = "sha256:016ad1afadf318eb7911baa24b049909f7f3bb2c5b1ed7b6a8f21db21ea3faa8"}, + {file = "pyrsistent-0.19.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c4db1bd596fefd66b296a3d5d943c94f4fac5bcd13e99bffe2ba6a759d959a28"}, + {file = "pyrsistent-0.19.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aeda827381f5e5d65cced3024126529ddc4289d944f75e090572c77ceb19adbf"}, + {file = "pyrsistent-0.19.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:42ac0b2f44607eb92ae88609eda931a4f0dfa03038c44c772e07f43e738bcac9"}, + {file = "pyrsistent-0.19.3-cp37-cp37m-win32.whl", hash = "sha256:e8f2b814a3dc6225964fa03d8582c6e0b6650d68a232df41e3cc1b66a5d2f8d1"}, + {file = "pyrsistent-0.19.3-cp37-cp37m-win_amd64.whl", hash = "sha256:c9bb60a40a0ab9aba40a59f68214eed5a29c6274c83b2cc206a359c4a89fa41b"}, + {file = "pyrsistent-0.19.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:a2471f3f8693101975b1ff85ffd19bb7ca7dd7c38f8a81701f67d6b4f97b87d8"}, + {file = "pyrsistent-0.19.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cc5d149f31706762c1f8bda2e8c4f8fead6e80312e3692619a75301d3dbb819a"}, + {file = "pyrsistent-0.19.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3311cb4237a341aa52ab8448c27e3a9931e2ee09561ad150ba94e4cfd3fc888c"}, + {file = "pyrsistent-0.19.3-cp38-cp38-win32.whl", hash = "sha256:f0e7c4b2f77593871e918be000b96c8107da48444d57005b6a6bc61fb4331b2c"}, + {file = "pyrsistent-0.19.3-cp38-cp38-win_amd64.whl", hash = "sha256:c147257a92374fde8498491f53ffa8f4822cd70c0d85037e09028e478cababb7"}, + {file = "pyrsistent-0.19.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:b735e538f74ec31378f5a1e3886a26d2ca6351106b4dfde376a26fc32a044edc"}, + {file = "pyrsistent-0.19.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:99abb85579e2165bd8522f0c0138864da97847875ecbd45f3e7e2af569bfc6f2"}, + {file = "pyrsistent-0.19.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a8cb235fa6d3fd7aae6a4f1429bbb1fec1577d978098da1252f0489937786f3"}, + {file = "pyrsistent-0.19.3-cp39-cp39-win32.whl", hash = "sha256:c74bed51f9b41c48366a286395c67f4e894374306b197e62810e0fdaf2364da2"}, + {file = "pyrsistent-0.19.3-cp39-cp39-win_amd64.whl", hash = "sha256:878433581fc23e906d947a6814336eee031a00e6defba224234169ae3d3d6a98"}, + {file = "pyrsistent-0.19.3-py3-none-any.whl", hash = "sha256:ccf0d6bd208f8111179f0c26fdf84ed7c3891982f2edaeae7422575f47e66b64"}, + {file = "pyrsistent-0.19.3.tar.gz", hash = "sha256:1a2994773706bbb4995c31a97bc94f1418314923bd1048c6d964837040376440"}, +] + +[[package]] +name = "pytest" +version = "7.4.0" +description = "pytest: simple powerful testing with Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pytest-7.4.0-py3-none-any.whl", hash = "sha256:78bf16451a2eb8c7a2ea98e32dc119fd2aa758f1d5d66dbf0a59d69a3969df32"}, + {file = "pytest-7.4.0.tar.gz", hash = "sha256:b4bf8c45bd59934ed84001ad51e11b4ee40d40a1229d2c79f9c592b0a3f6bd8a"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "sys_platform == \"win32\""} +exceptiongroup = {version = ">=1.0.0rc8", markers = "python_version < \"3.11\""} +iniconfig = "*" +packaging = "*" +pluggy = ">=0.12,<2.0" +tomli = {version = ">=1.0.0", markers = "python_version < \"3.11\""} + +[package.extras] +testing = ["argcomplete", "attrs (>=19.2.0)", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.7.2)", "requests", "setuptools", "xmlschema"] + +[[package]] +name = "pytest-cov" +version = "2.12.1" +description = "Pytest plugin for measuring coverage." +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"}, + {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"}, +] + +[package.dependencies] +coverage = ">=5.2.1" +pytest = ">=4.6" +toml = "*" + +[package.extras] +testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"] + +[[package]] +name = "python-dateutil" +version = "2.8.2" +description = "Extensions to the standard Python datetime module" +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" +files = [ + {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"}, + {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"}, +] + +[package.dependencies] +six = ">=1.5" + +[[package]] +name = "python-dotenv" +version = "1.0.0" +description = "Read key-value pairs from a .env file and set them as environment variables" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"}, + {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"}, +] + +[package.extras] +cli = ["click (>=5.0)"] + +[[package]] +name = "pytz" +version = "2020.5" +description = "World timezone definitions, modern and historical" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"}, + {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"}, +] + +[[package]] +name = "pyyaml" +version = "6.0" +description = "YAML parser and emitter for Python" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d4db7c7aef085872ef65a8fd7d6d09a14ae91f691dec3e87ee5ee0539d516f53"}, + {file = "PyYAML-6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9df7ed3b3d2e0ecfe09e14741b857df43adb5a3ddadc919a2d94fbdf78fea53c"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77f396e6ef4c73fdc33a9157446466f1cff553d979bd00ecb64385760c6babdc"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a80a78046a72361de73f8f395f1f1e49f956c6be882eed58505a15f3e430962b"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f84fbc98b019fef2ee9a1cb3ce93e3187a6df0b2538a651bfb890254ba9f90b5"}, + {file = "PyYAML-6.0-cp310-cp310-win32.whl", hash = "sha256:2cd5df3de48857ed0544b34e2d40e9fac445930039f3cfe4bcc592a1f836d513"}, + {file = "PyYAML-6.0-cp310-cp310-win_amd64.whl", hash = "sha256:daf496c58a8c52083df09b80c860005194014c3698698d1a57cbcfa182142a3a"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d4b0ba9512519522b118090257be113b9468d804b19d63c71dbcf4a48fa32358"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:81957921f441d50af23654aa6c5e5eaf9b06aba7f0a19c18a538dc7ef291c5a1"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:afa17f5bc4d1b10afd4466fd3a44dc0e245382deca5b3c353d8b757f9e3ecb8d"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dbad0e9d368bb989f4515da330b88a057617d16b6a8245084f1b05400f24609f"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:432557aa2c09802be39460360ddffd48156e30721f5e8d917f01d31694216782"}, + {file = "PyYAML-6.0-cp311-cp311-win32.whl", hash = "sha256:bfaef573a63ba8923503d27530362590ff4f576c626d86a9fed95822a8255fd7"}, + {file = "PyYAML-6.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b45c0191e6d66c470b6cf1b9531a771a83c1c4208272ead47a3ae4f2f603bf"}, + {file = "PyYAML-6.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:897b80890765f037df3403d22bab41627ca8811ae55e9a722fd0392850ec4d86"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50602afada6d6cbfad699b0c7bb50d5ccffa7e46a3d738092afddc1f9758427f"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:48c346915c114f5fdb3ead70312bd042a953a8ce5c7106d5bfb1a5254e47da92"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:98c4d36e99714e55cfbaaee6dd5badbc9a1ec339ebfc3b1f52e293aee6bb71a4"}, + {file = "PyYAML-6.0-cp36-cp36m-win32.whl", hash = "sha256:0283c35a6a9fbf047493e3a0ce8d79ef5030852c51e9d911a27badfde0605293"}, + {file = "PyYAML-6.0-cp36-cp36m-win_amd64.whl", hash = "sha256:07751360502caac1c067a8132d150cf3d61339af5691fe9e87803040dbc5db57"}, + {file = "PyYAML-6.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:819b3830a1543db06c4d4b865e70ded25be52a2e0631ccd2f6a47a2822f2fd7c"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:473f9edb243cb1935ab5a084eb238d842fb8f404ed2193a915d1784b5a6b5fc0"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0ce82d761c532fe4ec3f87fc45688bdd3a4c1dc5e0b4a19814b9009a29baefd4"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:231710d57adfd809ef5d34183b8ed1eeae3f76459c18fb4a0b373ad56bedcdd9"}, + {file = "PyYAML-6.0-cp37-cp37m-win32.whl", hash = "sha256:c5687b8d43cf58545ade1fe3e055f70eac7a5a1a0bf42824308d868289a95737"}, + {file = "PyYAML-6.0-cp37-cp37m-win_amd64.whl", hash = "sha256:d15a181d1ecd0d4270dc32edb46f7cb7733c7c508857278d3d378d14d606db2d"}, + {file = "PyYAML-6.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b4624f379dab24d3725ffde76559cff63d9ec94e1736b556dacdfebe5ab6d4b"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:213c60cd50106436cc818accf5baa1aba61c0189ff610f64f4a3e8c6726218ba"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9fa600030013c4de8165339db93d182b9431076eb98eb40ee068700c9c813e34"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:277a0ef2981ca40581a47093e9e2d13b3f1fbbeffae064c1d21bfceba2030287"}, + {file = "PyYAML-6.0-cp38-cp38-win32.whl", hash = "sha256:d4eccecf9adf6fbcc6861a38015c2a64f38b9d94838ac1810a9023a0609e1b78"}, + {file = "PyYAML-6.0-cp38-cp38-win_amd64.whl", hash = "sha256:1e4747bc279b4f613a09eb64bba2ba602d8a6664c6ce6396a4d0cd413a50ce07"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:055d937d65826939cb044fc8c9b08889e8c743fdc6a32b33e2390f66013e449b"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e61ceaab6f49fb8bdfaa0f92c4b57bcfbea54c09277b1b4f7ac376bfb7a7c174"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d67d839ede4ed1b28a4e8909735fc992a923cdb84e618544973d7dfc71540803"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba8c411ef271aa037d7357a2bc8f9ee8b58b9965831d9e51baf703280dc73d3"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:40527857252b61eacd1d9af500c3337ba8deb8fc298940291486c465c8b46ec0"}, + {file = "PyYAML-6.0-cp39-cp39-win32.whl", hash = "sha256:b5b9eccad747aabaaffbc6064800670f0c297e52c12754eb1d976c57e4f74dcb"}, + {file = "PyYAML-6.0-cp39-cp39-win_amd64.whl", hash = "sha256:b3d267842bf12586ba6c734f89d1f5b871df0273157918b0ccefa29deb05c21c"}, + {file = "PyYAML-6.0.tar.gz", hash = "sha256:68fb519c14306fec9720a2a5b45bc9f0c8d1b9c72adf45c37baedfcd949c35a2"}, +] + +[[package]] +name = "requests" +version = "2.31.0" +description = "Python HTTP for Humans." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "requests-2.31.0-py3-none-any.whl", hash = "sha256:58cd2187c01e70e6e26505bca751777aa9f2ee0b7f4300988b709f44e013003f"}, + {file = "requests-2.31.0.tar.gz", hash = "sha256:942c5a758f98d790eaed1a29cb6eefc7ffb0d1cf7af05c3d2791656dbd6ad1e1"}, +] + +[package.dependencies] +certifi = ">=2017.4.17" +charset-normalizer = ">=2,<4" +idna = ">=2.5,<4" +urllib3 = ">=1.21.1,<3" + +[package.extras] +socks = ["PySocks (>=1.5.6,!=1.5.7)"] +use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] + +[[package]] +name = "rfc3986" +version = "1.5.0" +description = "Validating URI References per RFC 3986" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "rfc3986-1.5.0-py2.py3-none-any.whl", hash = "sha256:a86d6e1f5b1dc238b218b012df0aa79409667bb209e58da56d0b94704e712a97"}, + {file = "rfc3986-1.5.0.tar.gz", hash = "sha256:270aaf10d87d0d4e095063c65bf3ddbc6ee3d0b226328ce21e036f946e421835"}, +] + +[package.dependencies] +idna = {version = "*", optional = true, markers = "extra == \"idna2008\""} + +[package.extras] +idna2008 = ["idna"] + +[[package]] +name = "rich" +version = "13.4.2" +description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal" +category = "dev" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "rich-13.4.2-py3-none-any.whl", hash = "sha256:8f87bc7ee54675732fa66a05ebfe489e27264caeeff3728c945d25971b6485ec"}, + {file = "rich-13.4.2.tar.gz", hash = "sha256:d653d6bccede5844304c605d5aac802c7cf9621efd700b46c7ec2b51ea914898"}, +] + +[package.dependencies] +markdown-it-py = ">=2.2.0" +pygments = ">=2.13.0,<3.0.0" + +[package.extras] +jupyter = ["ipywidgets (>=7.5.1,<9)"] + +[[package]] +name = "scikit-learn" +version = "1.3.0" +description = "A set of python modules for machine learning and data mining" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "scikit-learn-1.3.0.tar.gz", hash = "sha256:8be549886f5eda46436b6e555b0e4873b4f10aa21c07df45c4bc1735afbccd7a"}, + {file = "scikit_learn-1.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:981287869e576d42c682cf7ca96af0c6ac544ed9316328fd0d9292795c742cf5"}, + {file = "scikit_learn-1.3.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:436aaaae2c916ad16631142488e4c82f4296af2404f480e031d866863425d2a2"}, + {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c7e28d8fa47a0b30ae1bd7a079519dd852764e31708a7804da6cb6f8b36e3630"}, + {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ae80c08834a473d08a204d966982a62e11c976228d306a2648c575e3ead12111"}, + {file = "scikit_learn-1.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:552fd1b6ee22900cf1780d7386a554bb96949e9a359999177cf30211e6b20df6"}, + {file = "scikit_learn-1.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:79970a6d759eb00a62266a31e2637d07d2d28446fca8079cf9afa7c07b0427f8"}, + {file = "scikit_learn-1.3.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:850a00b559e636b23901aabbe79b73dc604b4e4248ba9e2d6e72f95063765603"}, + {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee04835fb016e8062ee9fe9074aef9b82e430504e420bff51e3e5fffe72750ca"}, + {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9d953531f5d9f00c90c34fa3b7d7cfb43ecff4c605dac9e4255a20b114a27369"}, + {file = "scikit_learn-1.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:151ac2bf65ccf363664a689b8beafc9e6aae36263db114b4ca06fbbbf827444a"}, + {file = "scikit_learn-1.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:6a885a9edc9c0a341cab27ec4f8a6c58b35f3d449c9d2503a6fd23e06bbd4f6a"}, + {file = "scikit_learn-1.3.0-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:9877af9c6d1b15486e18a94101b742e9d0d2f343d35a634e337411ddb57783f3"}, + {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c470f53cea065ff3d588050955c492793bb50c19a92923490d18fcb637f6383a"}, + {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fd6e2d7389542eae01077a1ee0318c4fec20c66c957f45c7aac0c6eb0fe3c612"}, + {file = "scikit_learn-1.3.0-cp38-cp38-win_amd64.whl", hash = "sha256:3a11936adbc379a6061ea32fa03338d4ca7248d86dd507c81e13af428a5bc1db"}, + {file = "scikit_learn-1.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:998d38fcec96584deee1e79cd127469b3ad6fefd1ea6c2dfc54e8db367eb396b"}, + {file = "scikit_learn-1.3.0-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:ded35e810438a527e17623ac6deae3b360134345b7c598175ab7741720d7ffa7"}, + {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0e8102d5036e28d08ab47166b48c8d5e5810704daecf3a476a4282d562be9a28"}, + {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7617164951c422747e7c32be4afa15d75ad8044f42e7d70d3e2e0429a50e6718"}, + {file = "scikit_learn-1.3.0-cp39-cp39-win_amd64.whl", hash = "sha256:1d54fb9e6038284548072df22fd34777e434153f7ffac72c8596f2d6987110dd"}, +] + +[package.dependencies] +joblib = ">=1.1.1" +numpy = ">=1.17.3" +scipy = ">=1.5.0" +threadpoolctl = ">=2.0.0" + +[package.extras] +benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"] +docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=6.0.0)", "sphinx-copybutton (>=0.5.2)", "sphinx-gallery (>=0.10.1)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"] +examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"] +tests = ["black (>=23.3.0)", "matplotlib (>=3.1.3)", "mypy (>=1.3)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=7.1.2)", "pytest-cov (>=2.9.0)", "ruff (>=0.0.272)", "scikit-image (>=0.16.2)"] + +[[package]] +name = "scipy" +version = "1.11.1" +description = "Fundamental algorithms for scientific computing in Python" +category = "main" +optional = false +python-versions = "<3.13,>=3.9" +files = [ + {file = "scipy-1.11.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:aec8c62fbe52914f9cf28d846cf0401dd80ab80788bbab909434eb336ed07c04"}, + {file = "scipy-1.11.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:3b9963798df1d8a52db41a6fc0e6fa65b1c60e85d73da27ae8bb754de4792481"}, + {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3e8eb42db36526b130dfbc417609498a6192381abc1975b91e3eb238e0b41c1a"}, + {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:366a6a937110d80dca4f63b3f5b00cc89d36f678b2d124a01067b154e692bab1"}, + {file = "scipy-1.11.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:08d957ca82d3535b3b9ba6c8ff355d78fe975271874e2af267cb5add5bd78625"}, + {file = "scipy-1.11.1-cp310-cp310-win_amd64.whl", hash = "sha256:e866514bc2d660608447b6ba95c8900d591f2865c07cca0aa4f7ff3c4ca70f30"}, + {file = "scipy-1.11.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ba94eeef3c9caa4cea7b402a35bb02a5714ee1ee77eb98aca1eed4543beb0f4c"}, + {file = "scipy-1.11.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:512fdc18c65f76dadaca139348e525646d440220d8d05f6d21965b8d4466bccd"}, + {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cce154372f0ebe88556ed06d7b196e9c2e0c13080ecb58d0f35062dc7cc28b47"}, + {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b4bb943010203465ac81efa392e4645265077b4d9e99b66cf3ed33ae12254173"}, + {file = "scipy-1.11.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:249cfa465c379c9bb2c20123001e151ff5e29b351cbb7f9c91587260602c58d0"}, + {file = "scipy-1.11.1-cp311-cp311-win_amd64.whl", hash = "sha256:ffb28e3fa31b9c376d0fb1f74c1f13911c8c154a760312fbee87a21eb21efe31"}, + {file = "scipy-1.11.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:39154437654260a52871dfde852adf1b93b1d1bc5dc0ffa70068f16ec0be2624"}, + {file = "scipy-1.11.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:b588311875c58d1acd4ef17c983b9f1ab5391755a47c3d70b6bd503a45bfaf71"}, + {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d51565560565a0307ed06fa0ec4c6f21ff094947d4844d6068ed04400c72d0c3"}, + {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b41a0f322b4eb51b078cb3441e950ad661ede490c3aca66edef66f4b37ab1877"}, + {file = "scipy-1.11.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:396fae3f8c12ad14c5f3eb40499fd06a6fef8393a6baa352a652ecd51e74e029"}, + {file = "scipy-1.11.1-cp39-cp39-win_amd64.whl", hash = "sha256:be8c962a821957fdde8c4044efdab7a140c13294997a407eaee777acf63cbf0c"}, + {file = "scipy-1.11.1.tar.gz", hash = "sha256:fb5b492fa035334fd249f0973cc79ecad8b09c604b42a127a677b45a9a3d4289"}, +] + +[package.dependencies] +numpy = ">=1.21.6,<1.28.0" + +[package.extras] +dev = ["click", "cython-lint (>=0.12.2)", "doit (>=0.36.0)", "mypy", "pycodestyle", "pydevtool", "rich-click", "ruff", "types-psutil", "typing_extensions"] +doc = ["jupytext", "matplotlib (>2)", "myst-nb", "numpydoc", "pooch", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"] +test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"] + +[[package]] +name = "setuptools" +version = "68.0.0" +description = "Easily download, build, install, upgrade, and uninstall Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "setuptools-68.0.0-py3-none-any.whl", hash = "sha256:11e52c67415a381d10d6b462ced9cfb97066179f0e871399e006c4ab101fc85f"}, + {file = "setuptools-68.0.0.tar.gz", hash = "sha256:baf1fdb41c6da4cd2eae722e135500da913332ab3f2f5c7d33af9b492acb5235"}, +] + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-hoverxref (<2)", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (==0.8.3)", "sphinx-reredirects", "sphinxcontrib-towncrier"] +testing = ["build[virtualenv]", "filelock (>=3.4.0)", "flake8-2020", "ini2toml[lite] (>=0.9)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pip (>=19.1)", "pip-run (>=8.8)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-mypy (>=0.9.1)", "pytest-perf", "pytest-ruff", "pytest-timeout", "pytest-xdist", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel"] +testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pytest", "pytest-enabler", "pytest-xdist", "tomli", "virtualenv (>=13.0.0)", "wheel"] + +[[package]] +name = "six" +version = "1.16.0" +description = "Python 2 and 3 compatibility utilities" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "six-1.16.0-py2.py3-none-any.whl", hash = "sha256:8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254"}, + {file = "six-1.16.0.tar.gz", hash = "sha256:1e61c37477a1626458e36f7b1d82aa5c9b094fa4802892072e49de9c60c4c926"}, +] + +[[package]] +name = "smmap" +version = "5.0.0" +description = "A pure Python implementation of a sliding window memory map manager" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "smmap-5.0.0-py3-none-any.whl", hash = "sha256:2aba19d6a040e78d8b09de5c57e96207b09ed71d8e55ce0959eeee6c8e190d94"}, + {file = "smmap-5.0.0.tar.gz", hash = "sha256:c840e62059cd3be204b0c9c9f74be2c09d5648eddd4580d9314c3ecde0b30936"}, +] + +[[package]] +name = "sniffio" +version = "1.3.0" +description = "Sniff out which async library your code is running under" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "sniffio-1.3.0-py3-none-any.whl", hash = "sha256:eecefdce1e5bbfb7ad2eeaabf7c1eeb404d7757c379bd1f7e5cce9d8bf425384"}, + {file = "sniffio-1.3.0.tar.gz", hash = "sha256:e60305c5e5d314f5389259b7f22aaa33d8f7dee49763119234af3755c55b9101"}, +] + +[[package]] +name = "sortedcontainers" +version = "2.4.0" +description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"}, + {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"}, +] + +[[package]] +name = "soundfile" +version = "0.12.1" +description = "An audio library based on libsndfile, CFFI and NumPy" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"}, + {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"}, + {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"}, + {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"}, +] + +[package.dependencies] +cffi = ">=1.0" + +[package.extras] +numpy = ["numpy"] + +[[package]] +name = "soxr" +version = "0.3.5" +description = "High quality, one-dimensional sample-rate conversion library" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "soxr-0.3.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:21c3aa3b2e12351b4310eea9d56cf52ec0769e6832f911ee6ba32f85b7c92baa"}, + {file = "soxr-0.3.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ac3d7abc96082ff18a31fb1d678ddc0562f0c5e6d91f1cf0024b044989f63e93"}, + {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:145e1e9d1b873a59ce0b5aa463ccacc40cf4bb74d9d8e6cef23433c752bfecea"}, + {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4a376b3678801ffc1d0b9ae918b958be29d5884ca1b4bbeab32e29c567723bb3"}, + {file = "soxr-0.3.5-cp310-cp310-win32.whl", hash = "sha256:907e2eb176bdefec40cc8f6015b7cef7f3d525a34219b3580b603ee696cb25c6"}, + {file = "soxr-0.3.5-cp310-cp310-win_amd64.whl", hash = "sha256:0a6dbf9c7b7a3642916aba264c1d0b872b2e173be56204ed1895dbe381a32077"}, + {file = "soxr-0.3.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:22c08a41e8eee99241fc0e9afb510f9bc7ada4a149d469b8891b596281a27db3"}, + {file = "soxr-0.3.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bdacbe4ce4a1001043f1f8f0744480e294f5c5106e7861fd7033a83a869ba371"}, + {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4b9acd5c42159eac4a90807524d9aa450d6ea0c750df94455c151165896d922e"}, + {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:44b5d30f4e0d98b6d0034c00b04d5571ad070ce5cf3772f93193095b01b373de"}, + {file = "soxr-0.3.5-cp311-cp311-win32.whl", hash = "sha256:677d5f44e85fdf0fdef33cd0e6087470732dd2e08fa73286c3659814110d1183"}, + {file = "soxr-0.3.5-cp311-cp311-win_amd64.whl", hash = "sha256:a479984dd17bf0b50fb9fd659eba54a2dc59bf6eba9c29bb3a4a79ecec7dc9a4"}, + {file = "soxr-0.3.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a2eb4f273ca14d7cfa882b234a03497d0e5dfd6f769a488a0962fe500450838c"}, + {file = "soxr-0.3.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:a254c5e1adddb1204d8f327158b6c11a854908a10b5782103f38a67156108334"}, + {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5766727dfee4d3616edd2a866a9a0d2f272c01545bed165c5a2676fbfd278723"}, + {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2578664c6f94329685d864cdcae59794121bcbd808441572b2ffd01e7adc45dd"}, + {file = "soxr-0.3.5-cp38-cp38-win32.whl", hash = "sha256:8a6f03804f48d986610eab8ca2b52e50b495f40ec13507741cd95f00ef7c2cb6"}, + {file = "soxr-0.3.5-cp38-cp38-win_amd64.whl", hash = "sha256:592e9393e433501769a7e36b10460f4578c8e4ec3cddeec1aaaea4688e3558ef"}, + {file = "soxr-0.3.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:93adbf04f51c7a5113059395633c2647f73bf195fa820256e1dd4da78af59275"}, + {file = "soxr-0.3.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:37c4ec7ce275f284b0bf9741e5e6844a211ba1a850b2bf1c6a47769cdd3d109e"}, + {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:18d5f3151fe4a88dfc37447bc6c397072aedcf36aeffb325cc817350ac5ad78e"}, + {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:549a8358ba3b99a75588453c96aaa802e0c84d40957bdbe1f820f14f83a052ca"}, + {file = "soxr-0.3.5-cp39-cp39-win32.whl", hash = "sha256:799df1875803dc9c4a4d3a7c285b8c1cb34b40dc39dba7ac7bac85d072f936a5"}, + {file = "soxr-0.3.5-cp39-cp39-win_amd64.whl", hash = "sha256:4dd3f61929eb304c109f1f3b6cc8243e3a1a46d636d5bd86b5a7f50609ecd7d6"}, + {file = "soxr-0.3.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:028af32bd4ce4b4c8183bb36da99e23ae954a114034d74538b4cae1bf40a0555"}, + {file = "soxr-0.3.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1299e2aae4d659e222bcbbaca69a51ee99571486070ed49a393725ea6010a8e9"}, + {file = "soxr-0.3.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:162f4e8b9a014c6819b4db6def2d43f7f4d97432ae33f2edfc8e5d0c97cf1cb3"}, + {file = "soxr-0.3.5.tar.gz", hash = "sha256:b6b60f6381c98249a2f2a594e9234b647b78856c76c060597d53ed27b6efd249"}, +] + +[package.dependencies] +numpy = "*" + +[package.extras] +docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"] +test = ["pytest"] + +[[package]] +name = "starlette" +version = "0.28.0" +description = "The little ASGI library that shines." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "starlette-0.28.0-py3-none-any.whl", hash = "sha256:e58b9fc402c579950260fbb6d57173395c4e62804c40d3ede7e9ef1074f0c579"}, + {file = "starlette-0.28.0.tar.gz", hash = "sha256:7bf3da5e997e796cc202cef2bd3f96a7d9b1e1943203c2fe2b42e020bc658482"}, +] + +[package.dependencies] +anyio = ">=3.4.0,<5" +typing-extensions = {version = ">=3.10.0", markers = "python_version < \"3.10\""} + +[package.extras] +full = ["httpx (>=0.22.0)", "itsdangerous", "jinja2", "python-multipart", "pyyaml"] + +[[package]] +name = "starlette-prometheus" +version = "0.9.0" +description = "Prometheus integration for Starlette" +category = "main" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "starlette-prometheus-0.9.0.tar.gz", hash = "sha256:a52fb0f1df52b44a7a677a792759337ef0ce0d59ddf3e684a7d6459a93a90e99"}, + {file = "starlette_prometheus-0.9.0-py3-none-any.whl", hash = "sha256:b4702e4ec67dce508d28551db0e45f12f58411afdb5d1078c92ff74331915381"}, +] + +[package.dependencies] +prometheus_client = ">=0.12,<0.13" +starlette = ">=0.12.2" + +[[package]] +name = "stevedore" +version = "5.1.0" +description = "Manage dynamic plugins for Python applications" +category = "dev" +optional = false +python-versions = ">=3.8" +files = [ + {file = "stevedore-5.1.0-py3-none-any.whl", hash = "sha256:8cc040628f3cea5d7128f2e76cf486b2251a4e543c7b938f58d9a377f6694a2d"}, + {file = "stevedore-5.1.0.tar.gz", hash = "sha256:a54534acf9b89bc7ed264807013b505bf07f74dbe4bcfa37d32bd063870b087c"}, +] + +[package.dependencies] +pbr = ">=2.0.0,<2.1.0 || >2.1.0" + +[[package]] +name = "threadpoolctl" +version = "3.1.0" +description = "threadpoolctl" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, + {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"}, +] + +[[package]] +name = "toml" +version = "0.10.2" +description = "Python Library for Tom's Obvious, Minimal Language" +category = "dev" +optional = false +python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b"}, + {file = "toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f"}, +] + +[[package]] +name = "tomli" +version = "2.0.1" +description = "A lil' TOML parser" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tomli-2.0.1-py3-none-any.whl", hash = "sha256:939de3e7a6161af0c887ef91b7d41a53e7c5a1ca976325f429cb46ea9bc30ecc"}, + {file = "tomli-2.0.1.tar.gz", hash = "sha256:de526c12914f0c550d15924c62d72abc48d6fe7364aa87328337a31007fe8a4f"}, +] + +[[package]] +name = "tqdm" +version = "4.65.0" +description = "Fast, Extensible Progress Meter" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"}, + {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[package.extras] +dev = ["py-make (>=0.1.0)", "twine", "wheel"] +notebook = ["ipywidgets (>=6)"] +slack = ["slack-sdk"] +telegram = ["requests"] + +[[package]] +name = "types-jsonschema" +version = "4.17.0.8" +description = "Typing stubs for jsonschema" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-jsonschema-4.17.0.8.tar.gz", hash = "sha256:96a56990910f405e62de58862c0bbb3ac29ee6dba6d3d99aa0ba7f874cc547de"}, + {file = "types_jsonschema-4.17.0.8-py3-none-any.whl", hash = "sha256:f5958eb7b53217dfb5125f0412aeaef226a8a9013eac95816c95b5b523f6796b"}, +] + +[[package]] +name = "types-psutil" +version = "5.9.5.15" +description = "Typing stubs for psutil" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-psutil-5.9.5.15.tar.gz", hash = "sha256:943a5f9556c1995097c89f50274330dd8d71d17bd73cbc6b09c61dbf4e114cc6"}, + {file = "types_psutil-5.9.5.15-py3-none-any.whl", hash = "sha256:7f183072d1fdb4e092fd6dd88ea017b957056bf10e58eafcdca26bf4372c0742"}, +] + +[[package]] +name = "types-pytz" +version = "2023.3.0.0" +description = "Typing stubs for pytz" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-pytz-2023.3.0.0.tar.gz", hash = "sha256:ecdc70d543aaf3616a7e48631543a884f74205f284cefd6649ddf44c6a820aac"}, + {file = "types_pytz-2023.3.0.0-py3-none-any.whl", hash = "sha256:4fc2a7fbbc315f0b6630e0b899fd6c743705abe1094d007b0e612d10da15e0f3"}, +] + +[[package]] +name = "types-requests" +version = "2.31.0.1" +description = "Typing stubs for requests" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-requests-2.31.0.1.tar.gz", hash = "sha256:3de667cffa123ce698591de0ad7db034a5317457a596eb0b4944e5a9d9e8d1ac"}, + {file = "types_requests-2.31.0.1-py3-none-any.whl", hash = "sha256:afb06ef8f25ba83d59a1d424bd7a5a939082f94b94e90ab5e6116bd2559deaa3"}, +] + +[package.dependencies] +types-urllib3 = "*" + +[[package]] +name = "types-urllib3" +version = "1.26.25.13" +description = "Typing stubs for urllib3" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-urllib3-1.26.25.13.tar.gz", hash = "sha256:3300538c9dc11dad32eae4827ac313f5d986b8b21494801f1bf97a1ac6c03ae5"}, + {file = "types_urllib3-1.26.25.13-py3-none-any.whl", hash = "sha256:5dbd1d2bef14efee43f5318b5d36d805a489f6600252bb53626d4bfafd95e27c"}, +] + +[[package]] +name = "typing-extensions" +version = "4.7.1" +description = "Backported and Experimental Type Hints for Python 3.7+" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "typing_extensions-4.7.1-py3-none-any.whl", hash = "sha256:440d5dd3af93b060174bf433bccd69b0babc3b15b1a8dca43789fd7f61514b36"}, + {file = "typing_extensions-4.7.1.tar.gz", hash = "sha256:b75ddc264f0ba5615db7ba217daeb99701ad295353c45f9e95963337ceeeffb2"}, +] + +[[package]] +name = "tzdata" +version = "2023.3" +description = "Provider of IANA time zone data" +category = "main" +optional = false +python-versions = ">=2" +files = [ + {file = "tzdata-2023.3-py2.py3-none-any.whl", hash = "sha256:7e65763eef3120314099b6939b5546db7adce1e7d6f2e179e3df563c70511eda"}, + {file = "tzdata-2023.3.tar.gz", hash = "sha256:11ef1e08e54acb0d4f95bdb1be05da659673de4acbd21bf9c69e94cc5e907a3a"}, +] + +[[package]] +name = "urllib3" +version = "2.0.3" +description = "HTTP library with thread-safe connection pooling, file post, and more." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "urllib3-2.0.3-py3-none-any.whl", hash = "sha256:48e7fafa40319d358848e1bc6809b208340fafe2096f1725d05d67443d0483d1"}, + {file = "urllib3-2.0.3.tar.gz", hash = "sha256:bee28b5e56addb8226c96f7f13ac28cb4c301dd5ea8a6ca179c0b9835e032825"}, +] + +[package.extras] +brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)"] +secure = ["certifi", "cryptography (>=1.9)", "idna (>=2.0.0)", "pyopenssl (>=17.1.0)", "urllib3-secure-extra"] +socks = ["pysocks (>=1.5.6,!=1.5.7,<2.0)"] +zstd = ["zstandard (>=0.18.0)"] + +[[package]] +name = "uvicorn" +version = "0.20.0" +description = "The lightning-fast ASGI server." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "uvicorn-0.20.0-py3-none-any.whl", hash = "sha256:c3ed1598a5668208723f2bb49336f4509424ad198d6ab2615b7783db58d919fd"}, + {file = "uvicorn-0.20.0.tar.gz", hash = "sha256:a4e12017b940247f836bc90b72e725d7dfd0c8ed1c51eb365f5ba30d9f5127d8"}, +] + +[package.dependencies] +click = ">=7.0" +h11 = ">=0.8" + +[package.extras] +standard = ["colorama (>=0.4)", "httptools (>=0.5.0)", "python-dotenv (>=0.13)", "pyyaml (>=5.1)", "uvloop (>=0.14.0,!=0.15.0,!=0.15.1)", "watchfiles (>=0.13)", "websockets (>=10.4)"] + +[[package]] +name = "watchdog" +version = "2.3.1" +description = "Filesystem events monitoring" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:d1f1200d4ec53b88bf04ab636f9133cb703eb19768a39351cee649de21a33697"}, + {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:564e7739abd4bd348aeafbf71cc006b6c0ccda3160c7053c4a53b67d14091d42"}, + {file = "watchdog-2.3.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:95ad708a9454050a46f741ba5e2f3468655ea22da1114e4c40b8cbdaca572565"}, + {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:a073c91a6ef0dda488087669586768195c3080c66866144880f03445ca23ef16"}, + {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aa8b028750b43e80eea9946d01925168eeadb488dfdef1d82be4b1e28067f375"}, + {file = "watchdog-2.3.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:964fd236cd443933268ae49b59706569c8b741073dbfd7ca705492bae9d39aab"}, + {file = "watchdog-2.3.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:91fd146d723392b3e6eb1ac21f122fcce149a194a2ba0a82c5e4d0ee29cd954c"}, + {file = "watchdog-2.3.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:efe3252137392a471a2174d721e1037a0e6a5da7beb72a021e662b7000a9903f"}, + {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:85bf2263290591b7c5fa01140601b64c831be88084de41efbcba6ea289874f44"}, + {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8f2df370cd8e4e18499dd0bfdef476431bcc396108b97195d9448d90924e3131"}, + {file = "watchdog-2.3.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ea5d86d1bcf4a9d24610aa2f6f25492f441960cf04aed2bd9a97db439b643a7b"}, + {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:6f5d0f7eac86807275eba40b577c671b306f6f335ba63a5c5a348da151aba0fc"}, + {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5b848c71ef2b15d0ef02f69da8cc120d335cec0ed82a3fa7779e27a5a8527225"}, + {file = "watchdog-2.3.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0d9878be36d2b9271e3abaa6f4f051b363ff54dbbe7e7df1af3c920e4311ee43"}, + {file = "watchdog-2.3.1-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4cd61f98cb37143206818cb1786d2438626aa78d682a8f2ecee239055a9771d5"}, + {file = "watchdog-2.3.1-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3d2dbcf1acd96e7a9c9aefed201c47c8e311075105d94ce5e899f118155709fd"}, + {file = "watchdog-2.3.1-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:03f342a9432fe08107defbe8e405a2cb922c5d00c4c6c168c68b633c64ce6190"}, + {file = "watchdog-2.3.1-py3-none-manylinux2014_aarch64.whl", hash = "sha256:7a596f9415a378d0339681efc08d2249e48975daae391d58f2e22a3673b977cf"}, + {file = "watchdog-2.3.1-py3-none-manylinux2014_armv7l.whl", hash = "sha256:0e1dd6d449267cc7d6935d7fe27ee0426af6ee16578eed93bacb1be9ff824d2d"}, + {file = "watchdog-2.3.1-py3-none-manylinux2014_i686.whl", hash = "sha256:7a1876f660e32027a1a46f8a0fa5747ad4fcf86cb451860eae61a26e102c8c79"}, + {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64.whl", hash = "sha256:2caf77ae137935c1466f8cefd4a3aec7017b6969f425d086e6a528241cba7256"}, + {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:53f3e95081280898d9e4fc51c5c69017715929e4eea1ab45801d5e903dd518ad"}, + {file = "watchdog-2.3.1-py3-none-manylinux2014_s390x.whl", hash = "sha256:9da7acb9af7e4a272089bd2af0171d23e0d6271385c51d4d9bde91fe918c53ed"}, + {file = "watchdog-2.3.1-py3-none-manylinux2014_x86_64.whl", hash = "sha256:8a4d484e846dcd75e96b96d80d80445302621be40e293bfdf34a631cab3b33dc"}, + {file = "watchdog-2.3.1-py3-none-win32.whl", hash = "sha256:a74155398434937ac2780fd257c045954de5b11b5c52fc844e2199ce3eecf4cf"}, + {file = "watchdog-2.3.1-py3-none-win_amd64.whl", hash = "sha256:5defe4f0918a2a1a4afbe4dbb967f743ac3a93d546ea4674567806375b024adb"}, + {file = "watchdog-2.3.1-py3-none-win_ia64.whl", hash = "sha256:4109cccf214b7e3462e8403ab1e5b17b302ecce6c103eb2fc3afa534a7f27b96"}, + {file = "watchdog-2.3.1.tar.gz", hash = "sha256:d9f9ed26ed22a9d331820a8432c3680707ea8b54121ddcc9dc7d9f2ceeb36906"}, +] + +[package.dependencies] +PyYAML = {version = ">=3.10", optional = true, markers = "extra == \"watchmedo\""} + +[package.extras] +watchmedo = ["PyYAML (>=3.10)"] + +[[package]] +name = "webencodings" +version = "0.5.1" +description = "Character encoding aliases for legacy web content" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78"}, + {file = "webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923"}, +] + +[[package]] +name = "xxhash" +version = "3.2.0" +description = "Python binding for xxHash" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"}, + {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"}, + {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"}, + {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"}, + {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"}, + {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"}, + {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"}, + {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"}, + {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"}, + {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"}, + {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"}, + {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"}, + {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"}, + {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"}, + {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"}, + {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"}, + {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"}, +] + +[[package]] +name = "yarl" +version = "1.9.2" +description = "Yet another URL library" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:8c2ad583743d16ddbdf6bb14b5cd76bf43b0d0006e918809d5d4ddf7bde8dd82"}, + {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:82aa6264b36c50acfb2424ad5ca537a2060ab6de158a5bd2a72a032cc75b9eb8"}, + {file = "yarl-1.9.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c0c77533b5ed4bcc38e943178ccae29b9bcf48ffd1063f5821192f23a1bd27b9"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee4afac41415d52d53a9833ebae7e32b344be72835bbb589018c9e938045a560"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9bf345c3a4f5ba7f766430f97f9cc1320786f19584acc7086491f45524a551ac"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2a96c19c52ff442a808c105901d0bdfd2e28575b3d5f82e2f5fd67e20dc5f4ea"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:891c0e3ec5ec881541f6c5113d8df0315ce5440e244a716b95f2525b7b9f3608"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c3a53ba34a636a256d767c086ceb111358876e1fb6b50dfc4d3f4951d40133d5"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:566185e8ebc0898b11f8026447eacd02e46226716229cea8db37496c8cdd26e0"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:2b0738fb871812722a0ac2154be1f049c6223b9f6f22eec352996b69775b36d4"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:32f1d071b3f362c80f1a7d322bfd7b2d11e33d2adf395cc1dd4df36c9c243095"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e9fdc7ac0d42bc3ea78818557fab03af6181e076a2944f43c38684b4b6bed8e3"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:56ff08ab5df8429901ebdc5d15941b59f6253393cb5da07b4170beefcf1b2528"}, + {file = "yarl-1.9.2-cp310-cp310-win32.whl", hash = "sha256:8ea48e0a2f931064469bdabca50c2f578b565fc446f302a79ba6cc0ee7f384d3"}, + {file = "yarl-1.9.2-cp310-cp310-win_amd64.whl", hash = "sha256:50f33040f3836e912ed16d212f6cc1efb3231a8a60526a407aeb66c1c1956dde"}, + {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:646d663eb2232d7909e6601f1a9107e66f9791f290a1b3dc7057818fe44fc2b6"}, + {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aff634b15beff8902d1f918012fc2a42e0dbae6f469fce134c8a0dc51ca423bb"}, + {file = "yarl-1.9.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:a83503934c6273806aed765035716216cc9ab4e0364f7f066227e1aaea90b8d0"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b25322201585c69abc7b0e89e72790469f7dad90d26754717f3310bfe30331c2"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:22a94666751778629f1ec4280b08eb11815783c63f52092a5953faf73be24191"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8ec53a0ea2a80c5cd1ab397925f94bff59222aa3cf9c6da938ce05c9ec20428d"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:159d81f22d7a43e6eabc36d7194cb53f2f15f498dbbfa8edc8a3239350f59fe7"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:832b7e711027c114d79dffb92576acd1bd2decc467dec60e1cac96912602d0e6"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:95d2ecefbcf4e744ea952d073c6922e72ee650ffc79028eb1e320e732898d7e8"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:d4e2c6d555e77b37288eaf45b8f60f0737c9efa3452c6c44626a5455aeb250b9"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:783185c75c12a017cc345015ea359cc801c3b29a2966c2655cd12b233bf5a2be"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:b8cc1863402472f16c600e3e93d542b7e7542a540f95c30afd472e8e549fc3f7"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:822b30a0f22e588b32d3120f6d41e4ed021806418b4c9f0bc3048b8c8cb3f92a"}, + {file = "yarl-1.9.2-cp311-cp311-win32.whl", hash = "sha256:a60347f234c2212a9f0361955007fcf4033a75bf600a33c88a0a8e91af77c0e8"}, + {file = "yarl-1.9.2-cp311-cp311-win_amd64.whl", hash = "sha256:be6b3fdec5c62f2a67cb3f8c6dbf56bbf3f61c0f046f84645cd1ca73532ea051"}, + {file = "yarl-1.9.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:38a3928ae37558bc1b559f67410df446d1fbfa87318b124bf5032c31e3447b74"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ac9bb4c5ce3975aeac288cfcb5061ce60e0d14d92209e780c93954076c7c4367"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3da8a678ca8b96c8606bbb8bfacd99a12ad5dd288bc6f7979baddd62f71c63ef"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:13414591ff516e04fcdee8dc051c13fd3db13b673c7a4cb1350e6b2ad9639ad3"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf74d08542c3a9ea97bb8f343d4fcbd4d8f91bba5ec9d5d7f792dbe727f88938"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6e7221580dc1db478464cfeef9b03b95c5852cc22894e418562997df0d074ccc"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:494053246b119b041960ddcd20fd76224149cfea8ed8777b687358727911dd33"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:52a25809fcbecfc63ac9ba0c0fb586f90837f5425edfd1ec9f3372b119585e45"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:e65610c5792870d45d7b68c677681376fcf9cc1c289f23e8e8b39c1485384185"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1b1bba902cba32cdec51fca038fd53f8beee88b77efc373968d1ed021024cc04"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:662e6016409828ee910f5d9602a2729a8a57d74b163c89a837de3fea050c7582"}, + {file = "yarl-1.9.2-cp37-cp37m-win32.whl", hash = "sha256:f364d3480bffd3aa566e886587eaca7c8c04d74f6e8933f3f2c996b7f09bee1b"}, + {file = "yarl-1.9.2-cp37-cp37m-win_amd64.whl", hash = "sha256:6a5883464143ab3ae9ba68daae8e7c5c95b969462bbe42e2464d60e7e2698368"}, + {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5610f80cf43b6202e2c33ba3ec2ee0a2884f8f423c8f4f62906731d876ef4fac"}, + {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b9a4e67ad7b646cd6f0938c7ebfd60e481b7410f574c560e455e938d2da8e0f4"}, + {file = "yarl-1.9.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:83fcc480d7549ccebe9415d96d9263e2d4226798c37ebd18c930fce43dfb9574"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5fcd436ea16fee7d4207c045b1e340020e58a2597301cfbcfdbe5abd2356c2fb"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:84e0b1599334b1e1478db01b756e55937d4614f8654311eb26012091be109d59"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:3458a24e4ea3fd8930e934c129b676c27452e4ebda80fbe47b56d8c6c7a63a9e"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:838162460b3a08987546e881a2bfa573960bb559dfa739e7800ceeec92e64417"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f4e2d08f07a3d7d3e12549052eb5ad3eab1c349c53ac51c209a0e5991bbada78"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:de119f56f3c5f0e2fb4dee508531a32b069a5f2c6e827b272d1e0ff5ac040333"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:149ddea5abf329752ea5051b61bd6c1d979e13fbf122d3a1f9f0c8be6cb6f63c"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:674ca19cbee4a82c9f54e0d1eee28116e63bc6fd1e96c43031d11cbab8b2afd5"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:9b3152f2f5677b997ae6c804b73da05a39daa6a9e85a512e0e6823d81cdad7cc"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:5415d5a4b080dc9612b1b63cba008db84e908b95848369aa1da3686ae27b6d2b"}, + {file = "yarl-1.9.2-cp38-cp38-win32.whl", hash = "sha256:f7a3d8146575e08c29ed1cd287068e6d02f1c7bdff8970db96683b9591b86ee7"}, + {file = "yarl-1.9.2-cp38-cp38-win_amd64.whl", hash = "sha256:63c48f6cef34e6319a74c727376e95626f84ea091f92c0250a98e53e62c77c72"}, + {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:75df5ef94c3fdc393c6b19d80e6ef1ecc9ae2f4263c09cacb178d871c02a5ba9"}, + {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c027a6e96ef77d401d8d5a5c8d6bc478e8042f1e448272e8d9752cb0aff8b5c8"}, + {file = "yarl-1.9.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f3b078dbe227f79be488ffcfc7a9edb3409d018e0952cf13f15fd6512847f3f7"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:59723a029760079b7d991a401386390c4be5bfec1e7dd83e25a6a0881859e716"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b03917871bf859a81ccb180c9a2e6c1e04d2f6a51d953e6a5cdd70c93d4e5a2a"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c1012fa63eb6c032f3ce5d2171c267992ae0c00b9e164efe4d73db818465fac3"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a74dcbfe780e62f4b5a062714576f16c2f3493a0394e555ab141bf0d746bb955"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8c56986609b057b4839968ba901944af91b8e92f1725d1a2d77cbac6972b9ed1"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:2c315df3293cd521033533d242d15eab26583360b58f7ee5d9565f15fee1bef4"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:b7232f8dfbd225d57340e441d8caf8652a6acd06b389ea2d3222b8bc89cbfca6"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:53338749febd28935d55b41bf0bcc79d634881195a39f6b2f767870b72514caf"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:066c163aec9d3d073dc9ffe5dd3ad05069bcb03fcaab8d221290ba99f9f69ee3"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:8288d7cd28f8119b07dd49b7230d6b4562f9b61ee9a4ab02221060d21136be80"}, + {file = "yarl-1.9.2-cp39-cp39-win32.whl", hash = "sha256:b124e2a6d223b65ba8768d5706d103280914d61f5cae3afbc50fc3dfcc016623"}, + {file = "yarl-1.9.2-cp39-cp39-win_amd64.whl", hash = "sha256:61016e7d582bc46a5378ffdd02cd0314fb8ba52f40f9cf4d9a5e7dbef88dee18"}, + {file = "yarl-1.9.2.tar.gz", hash = "sha256:04ab9d4b9f587c06d801c2abfe9317b77cdf996c65a90d5e84ecc45010823571"}, +] + +[package.dependencies] +idna = ">=2.0" +multidict = ">=4.0" + +[metadata] +lock-version = "2.0" +python-versions = "3.9.15" +content-hash = "c46efd608bb9fde12d1d6aec4385e7b26da6d08011d84e287d5a1663f3c93217" diff --git a/services/rows/poetry.toml b/services/rows/poetry.toml new file mode 100644 index 00000000..5fcef8cd --- /dev/null +++ b/services/rows/poetry.toml @@ -0,0 +1,3 @@ +[virtualenvs] +in-project = true +prefer-active-python = true diff --git a/services/rows/pyproject.toml b/services/rows/pyproject.toml new file mode 100644 index 00000000..7072a9cf --- /dev/null +++ b/services/rows/pyproject.toml @@ -0,0 +1,69 @@ +[tool.poetry] +authors = ["The HuggingFace Authors."] +description = "REST API app - /rows endpoint" +name = "rows" +version = "0.1.0" +license = "Apache-2.0" + +[tool.poetry.dependencies] +environs = "^9.5.0" +jsonschema = "^4.17.0" +libapi = {path = "../../libs/libapi", develop = true} +pyarrow = "^11.0.0" +python = "3.9.15" +soundfile = ">=0.12.1" +uvicorn = "^0.20.0" +watchdog = { extras = ["watchmedo"], version = "^2.2.1" } + +[tool.poetry.group.dev.dependencies] +bandit = "^1.7.4" +black = "^22.12.0" +flake8 = "^3.9.2" +httpx = "^0.23.3" +isort = "^5.12.0" +mypy = "^1.0.0" +pandas-stubs = "^1.5.3" +pip-audit = "^2.5.4" +pytest = "^7.2.1" +pytest-cov = "^2.12.1" +types-psutil = "^5.9.5" +types-jsonschema = "^4.17.0.4" +types-requests = "^2.28.11" + +[build-system] +build-backend = "poetry.core.masonry.api" +requires = ["poetry-core>=1.0.0"] + +[tool.pytest.ini_options] +filterwarnings = ["ignore::DeprecationWarning"] +markers = [ + "real_dataset: tests on the Hub", + "wip: tests being developed" +] + +[tool.coverage.run] +source = ["rows"] + +[tool.isort] +profile = "black" + +[tool.black] +line-length = 119 +preview = true + +[tool.mypy] +strict = true +disallow_untyped_calls = false +# ^ call to expected_algorithm.from_jwk forces to set this to false + +[[tool.mypy.overrides]] +module = [ + "datasets.*", + "huggingface_hub.*", + "prometheus_client.*", + "pyarrow.*", + "tqdm.*", + "fsspec.*" +] +# ^ prometheus_client is now typed, but starlette-prometheus requires an old version +ignore_missing_imports = true diff --git a/services/rows/src/rows/__init__.py b/services/rows/src/rows/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/services/rows/src/rows/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/services/rows/src/rows/app.py b/services/rows/src/rows/app.py new file mode 100644 index 00000000..d792fc80 --- /dev/null +++ b/services/rows/src/rows/app.py @@ -0,0 +1,101 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +import uvicorn +from libapi.config import UvicornConfig +from libapi.jwt_token import fetch_jwt_public_key +from libapi.routes.healthcheck import healthcheck_endpoint +from libapi.routes.metrics import create_metrics_endpoint +from libcommon.log import init_logging +from libcommon.processing_graph import ProcessingGraph +from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource +from libcommon.storage import exists, init_cached_assets_dir, init_parquet_metadata_dir +from starlette.applications import Starlette +from starlette.middleware import Middleware +from starlette.middleware.cors import CORSMiddleware +from starlette.middleware.gzip import GZipMiddleware +from starlette.routing import Route +from starlette_prometheus import PrometheusMiddleware + +from rows.config import AppConfig +from rows.routes.rows import create_rows_endpoint + + +def create_app() -> Starlette: + app_config = AppConfig.from_env() + return create_app_with_config(app_config=app_config) + + +def create_app_with_config(app_config: AppConfig) -> Starlette: + init_logging(level=app_config.log.level) + # ^ set first to have logs as soon as possible + cached_assets_directory = init_cached_assets_dir(directory=app_config.cached_assets.storage_directory) + if not exists(cached_assets_directory): + raise RuntimeError("The assets storage directory could not be accessed. Exiting.") + parquet_metadata_directory = init_parquet_metadata_dir(directory=app_config.parquet_metadata.storage_directory) + if not exists(parquet_metadata_directory): + raise RuntimeError("The parquet metadata storage directory could not be accessed. Exiting.") + + processing_graph = ProcessingGraph(app_config.processing_graph.specification) + hf_jwt_public_key = ( + fetch_jwt_public_key( + url=app_config.api.hf_jwt_public_key_url, + hf_jwt_algorithm=app_config.api.hf_jwt_algorithm, + hf_timeout_seconds=app_config.api.hf_timeout_seconds, + ) + if app_config.api.hf_jwt_public_key_url and app_config.api.hf_jwt_algorithm + else None + ) + + middleware = [ + Middleware( + CORSMiddleware, allow_origins=["*"], allow_methods=["*"], allow_headers=["*"], allow_credentials=True + ), + Middleware(GZipMiddleware), + Middleware(PrometheusMiddleware, filter_unhandled_paths=True), + ] + + cache_resource = CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url) + queue_resource = QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url) + resources: list[Resource] = [cache_resource, queue_resource] + if not cache_resource.is_available(): + raise RuntimeError("The connection to the cache database could not be established. Exiting.") + if not queue_resource.is_available(): + raise RuntimeError("The connection to the queue database could not be established. Exiting.") + + routes = [ + Route("/healthcheck", endpoint=healthcheck_endpoint), + Route("/metrics", endpoint=create_metrics_endpoint()), + # ^ called by Prometheus + Route( + "/rows", + endpoint=create_rows_endpoint( + processing_graph=processing_graph, + cached_assets_base_url=app_config.cached_assets.base_url, + cached_assets_directory=cached_assets_directory, + parquet_metadata_directory=parquet_metadata_directory, + hf_endpoint=app_config.common.hf_endpoint, + hf_token=app_config.common.hf_token, + hf_jwt_public_key=hf_jwt_public_key, + hf_jwt_algorithm=app_config.api.hf_jwt_algorithm, + external_auth_url=app_config.api.external_auth_url, + hf_timeout_seconds=app_config.api.hf_timeout_seconds, + max_age_long=app_config.api.max_age_long, + max_age_short=app_config.api.max_age_short, + cache_max_days=app_config.cache.max_days, + ), + ), + ] + + return Starlette(routes=routes, middleware=middleware, on_shutdown=[resource.release for resource in resources]) + + +def start() -> None: + uvicorn_config = UvicornConfig.from_env() + uvicorn.run( + "app:create_app", + host=uvicorn_config.hostname, + port=uvicorn_config.port, + factory=True, + workers=uvicorn_config.num_workers, + ) diff --git a/services/rows/src/rows/config.py b/services/rows/src/rows/config.py new file mode 100644 index 00000000..722f4b93 --- /dev/null +++ b/services/rows/src/rows/config.py @@ -0,0 +1,41 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from dataclasses import dataclass, field + +from libapi.config import ApiConfig +from libcommon.config import ( + CacheConfig, + CachedAssetsConfig, + CommonConfig, + LogConfig, + ParquetMetadataConfig, + ProcessingGraphConfig, + QueueConfig, +) + + +@dataclass(frozen=True) +class AppConfig: + api: ApiConfig = field(default_factory=ApiConfig) + cached_assets: CachedAssetsConfig = field(default_factory=CachedAssetsConfig) + cache: CacheConfig = field(default_factory=CacheConfig) + common: CommonConfig = field(default_factory=CommonConfig) + log: LogConfig = field(default_factory=LogConfig) + queue: QueueConfig = field(default_factory=QueueConfig) + processing_graph: ProcessingGraphConfig = field(default_factory=ProcessingGraphConfig) + parquet_metadata: ParquetMetadataConfig = field(default_factory=ParquetMetadataConfig) + + @classmethod + def from_env(cls) -> "AppConfig": + common_config = CommonConfig.from_env() + return cls( + common=common_config, + cached_assets=CachedAssetsConfig.from_env(), + cache=CacheConfig.from_env(), + log=LogConfig.from_env(), + processing_graph=ProcessingGraphConfig.from_env(), + queue=QueueConfig.from_env(), + api=ApiConfig.from_env(hf_endpoint=common_config.hf_endpoint), + parquet_metadata=ParquetMetadataConfig.from_env(), + ) diff --git a/services/rows/src/rows/main.py b/services/rows/src/rows/main.py new file mode 100644 index 00000000..b4790c09 --- /dev/null +++ b/services/rows/src/rows/main.py @@ -0,0 +1,7 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from rows.app import start + +if __name__ == "__main__": + start() diff --git a/services/rows/src/rows/py.typed b/services/rows/src/rows/py.typed new file mode 100644 index 00000000..e69de29b diff --git a/services/rows/src/rows/routes/__init__.py b/services/rows/src/rows/routes/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/services/rows/src/rows/routes/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/services/api/src/api/routes/rows.py b/services/rows/src/rows/routes/rows.py similarity index 100% rename from services/api/src/api/routes/rows.py rename to services/rows/src/rows/routes/rows.py diff --git a/services/rows/tests/__init__.py b/services/rows/tests/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/services/rows/tests/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/services/rows/tests/conftest.py b/services/rows/tests/conftest.py new file mode 100644 index 00000000..e6f5323b --- /dev/null +++ b/services/rows/tests/conftest.py @@ -0,0 +1,104 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from pathlib import Path +from typing import Iterator + +from libapi.config import UvicornConfig +from libcommon.processing_graph import ProcessingGraph +from libcommon.queue import _clean_queue_database +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import _clean_cache_database +from libcommon.storage import StrPath, init_cached_assets_dir, init_parquet_metadata_dir +from pytest import MonkeyPatch, fixture + +from rows.config import AppConfig + +# Import fixture modules as plugins +pytest_plugins = ["tests.fixtures.fsspec"] + + +# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200 +@fixture(scope="session") +def monkeypatch_session() -> Iterator[MonkeyPatch]: + monkeypatch_session = MonkeyPatch() + monkeypatch_session.setenv("CACHE_MONGO_DATABASE", "datasets_server_cache_test") + monkeypatch_session.setenv("QUEUE_MONGO_DATABASE", "datasets_server_queue_test") + monkeypatch_session.setenv("CACHED_ASSETS_BASE_URL", "http://localhost/cached-assets") + hostname = "localhost" + port = "8888" + monkeypatch_session.setenv("API_HF_TIMEOUT_SECONDS", "10") + monkeypatch_session.setenv("API_UVICORN_HOSTNAME", hostname) + monkeypatch_session.setenv("API_UVICORN_PORT", port) + monkeypatch_session.setenv("COMMON_HF_ENDPOINT", f"http://{hostname}:{port}") + yield monkeypatch_session + monkeypatch_session.undo() + + +@fixture(scope="session") +def app_config(monkeypatch_session: MonkeyPatch) -> AppConfig: + app_config = AppConfig.from_env() + if "test" not in app_config.cache.mongo_database or "test" not in app_config.queue.mongo_database: + raise ValueError("Test must be launched on a test mongo database") + return app_config + + +@fixture(scope="session") +def processing_graph(app_config: AppConfig) -> ProcessingGraph: + return ProcessingGraph(app_config.processing_graph.specification) + + +@fixture(scope="session") +def rows_endpoint() -> str: + return "/rows" + + +@fixture(autouse=True) +def cache_mongo_resource(app_config: AppConfig) -> Iterator[CacheMongoResource]: + with CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url) as resource: + yield resource + _clean_cache_database() + + +@fixture(autouse=True) +def queue_mongo_resource(app_config: AppConfig) -> Iterator[QueueMongoResource]: + with QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url) as resource: + yield resource + _clean_queue_database() + + +@fixture(scope="session") +def uvicorn_config(monkeypatch_session: MonkeyPatch) -> UvicornConfig: + return UvicornConfig.from_env() + + +@fixture(scope="session") +def httpserver_listen_address(uvicorn_config: UvicornConfig) -> tuple[str, int]: + return (uvicorn_config.hostname, uvicorn_config.port) + + +@fixture(scope="session") +def hf_endpoint(app_config: AppConfig) -> str: + return app_config.common.hf_endpoint + + +@fixture(scope="session") +def hf_auth_path(app_config: AppConfig) -> str: + return app_config.api.hf_auth_path + + +@fixture +def cached_assets_directory(app_config: AppConfig) -> StrPath: + return init_cached_assets_dir(app_config.cached_assets.storage_directory) + + +@fixture +def parquet_metadata_directory(app_config: AppConfig) -> StrPath: + return init_parquet_metadata_dir(app_config.parquet_metadata.storage_directory) + + +@fixture +def image_path() -> str: + image_path = Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg" + assert image_path.is_file() + return str(image_path) diff --git a/services/api/tests/data/test_image_rgb.jpg b/services/rows/tests/data/test_image_rgb.jpg similarity index 100% rename from services/api/tests/data/test_image_rgb.jpg rename to services/rows/tests/data/test_image_rgb.jpg diff --git a/services/api/tests/fixtures/fsspec.py b/services/rows/tests/fixtures/fsspec.py similarity index 100% rename from services/api/tests/fixtures/fsspec.py rename to services/rows/tests/fixtures/fsspec.py diff --git a/services/rows/tests/routes/__init__.py b/services/rows/tests/routes/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/services/rows/tests/routes/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/services/api/tests/routes/test_rows.py b/services/rows/tests/routes/test_rows.py similarity index 98% rename from services/api/tests/routes/test_rows.py rename to services/rows/tests/routes/test_rows.py index e8c944db..fdb55d06 100644 --- a/services/api/tests/routes/test_rows.py +++ b/services/rows/tests/routes/test_rows.py @@ -30,2 +30,2 @@ from libcommon.viewer_utils.asset import update_last_modified_date_of_rows_in_as -from api.config import AppConfig -from api.routes.rows import clean_cached_assets, create_response +from rows.config import AppConfig +from rows.routes.rows import clean_cached_assets, create_response @@ -41 +41 @@ def enable_parquet_metadata_on_all_datasets() -> Generator[None, None, None]: - with patch("api.routes.rows.ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST", "all"): + with patch("rows.routes.rows.ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST", "all"): @@ -470 +470 @@ def test_clean_cached_assets( - with patch("api.routes.rows.glob_rows_in_assets_dir", deterministic_glob_rows_in_assets_dir): + with patch("rows.routes.rows.glob_rows_in_assets_dir", deterministic_glob_rows_in_assets_dir): diff --git a/services/rows/tests/test_app.py b/services/rows/tests/test_app.py new file mode 100644 index 00000000..cfde7342 --- /dev/null +++ b/services/rows/tests/test_app.py @@ -0,0 +1,96 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from typing import Optional + +import pytest +from starlette.testclient import TestClient + +from rows.app import create_app_with_config +from rows.config import AppConfig + + [email protected](scope="module") +def client(monkeypatch_session: pytest.MonkeyPatch, app_config: AppConfig) -> TestClient: + return TestClient(create_app_with_config(app_config=app_config)) + + +def test_cors(client: TestClient) -> None: + origin = "http://localhost:3000" + method = "GET" + header = "X-Requested-With" + response = client.options( + "/rows?dataset=dataset1&config=config1&split=train", + headers={ + "Origin": origin, + "Access-Control-Request-Method": method, + "Access-Control-Request-Headers": header, + }, + ) + assert response.status_code == 200 + assert ( + origin in [o.strip() for o in response.headers["Access-Control-Allow-Origin"].split(",")] + or response.headers["Access-Control-Allow-Origin"] == "*" + ) + assert ( + header in [o.strip() for o in response.headers["Access-Control-Allow-Headers"].split(",")] + or response.headers["Access-Control-Expose-Headers"] == "*" + ) + assert ( + method in [o.strip() for o in response.headers["Access-Control-Allow-Methods"].split(",")] + or response.headers["Access-Control-Expose-Headers"] == "*" + ) + assert response.headers["Access-Control-Allow-Credentials"] == "true" + + +def test_get_healthcheck(client: TestClient) -> None: + response = client.get("/healthcheck") + assert response.status_code == 200 + assert response.text == "ok" + + +def test_get_rows(client: TestClient) -> None: + # missing parameter + response = client.get("/rows") + assert response.status_code == 422 + + [email protected]( + "dataset,config,split", + [ + (None, None, None), + ("a", None, None), + ("a", "b", None), + ("a", "b", ""), + ], +) +def test_get_split_missing_parameter( + client: TestClient, + dataset: Optional[str], + config: Optional[str], + split: Optional[str], +) -> None: + response = client.get("/rows", params={"dataset": dataset, "config": config, "split": split}) + assert response.status_code == 422 + + +def test_metrics(client: TestClient) -> None: + response = client.get("/healthcheck") + response = client.get("/metrics") + assert response.status_code == 200 + text = response.text + lines = text.split("\n") + # examples: + # starlette_requests_total{method="GET",path_template="/metrics"} 1.0 + # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05 + metrics = { + parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1)) + } + + # the metrics should contain at least the following + for name in [ + 'starlette_requests_total{method="GET",path_template="/metrics"}', + 'method_steps_processing_time_seconds_sum{context="None",method="healthcheck_endpoint",step="all"}', + ]: + assert name in metrics, metrics + assert metrics[name] > 0, metrics diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml index 37b1c87d..542101d5 100644 --- a/tools/docker-compose-datasets-server.yml +++ b/tools/docker-compose-datasets-server.yml @@ -19,0 +20 @@ services: + URL_ROWS: http://rows:${ROWS_UVICORN_PORT-8082} @@ -21 +21,0 @@ services: - - api @@ -22,0 +23,2 @@ services: + - api + - rows @@ -56,0 +59,27 @@ services: + environment: + # service + API_HF_AUTH_PATH: ${API_HF_AUTH_PATH-/api/datasets/%s/auth-check} + API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://huggingface.co/api/keys/jwt} + API_HF_JWT_ALGORITHM: ${API_HF_JWT_ALGORITHM-EdDSA} + API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-0.2} + API_MAX_AGE_LONG: ${API_MAX_AGE_LONG-120} + API_MAX_AGE_SHORT: ${API_MAX_AGE_SHORT-10} + # prometheus + PROMETHEUS_MULTIPROC_DIR: ${PROMETHEUS_MULTIPROC_DIR-} + # uvicorn + API_UVICORN_HOSTNAME: 0.0.0.0 # required for docker compose + API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2} + API_UVICORN_PORT: ${API_UVICORN_PORT-8080} + ports: + # for debug + - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080} + depends_on: + - mongodb + restart: unless-stopped + rows: + build: + context: .. + dockerfile: services/rows/Dockerfile + extends: + file: docker-compose-base.yml + service: api @@ -79,2 +108,2 @@ services: - API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2} - API_UVICORN_PORT: ${API_UVICORN_PORT-8080} + API_UVICORN_NUM_WORKERS: ${ROWS_UVICORN_NUM_WORKERS-2} + API_UVICORN_PORT: ${ROWS_UVICORN_PORT-8082} @@ -83 +112 @@ services: - - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080} + - ${ROWS_UVICORN_PORT-8082}:${ROWS_UVICORN_PORT-8082} diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index 233e90f2..3a5cb0f4 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -19,0 +20 @@ services: + URL_ROWS: http://host.docker.internal:${ROWS_UVICORN_PORT-8082} @@ -23 +23,0 @@ services: - - api @@ -24,0 +25,2 @@ services: + - api + - rows @@ -58,0 +61,27 @@ services: + environment: + # service + API_HF_AUTH_PATH: ${API_HF_AUTH_PATH-/api/datasets/%s/auth-check} + API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://hub-ci.huggingface.co/api/keys/jwt} + API_HF_JWT_ALGORITHM: ${API_HF_JWT_ALGORITHM-EdDSA} + API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-1.0} + API_MAX_AGE_LONG: ${API_MAX_AGE_LONG-120} + API_MAX_AGE_SHORT: ${API_MAX_AGE_SHORT-10} + # prometheus + PROMETHEUS_MULTIPROC_DIR: ${PROMETHEUS_MULTIPROC_DIR-} + # uvicorn + API_UVICORN_HOSTNAME: 0.0.0.0 # required for docker compose + API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2} + API_UVICORN_PORT: ${API_UVICORN_PORT-8080} + network_mode: ${DEV_NETWORK_MODE} + ports: + - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080} + depends_on: + - mongodb + restart: unless-stopped + rows: + build: + context: .. + dockerfile: services/rows/dev.Dockerfile + extends: + file: docker-compose-dev-base.yml + service: api @@ -81,2 +110,2 @@ services: - API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2} - API_UVICORN_PORT: ${API_UVICORN_PORT-8080} + API_UVICORN_NUM_WORKERS: ${ROWS_UVICORN_NUM_WORKERS-2} + API_UVICORN_PORT: ${ROWS_UVICORN_PORT-8082} @@ -85 +114,2 @@ services: - - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080} + # for debug + - ${ROWS_UVICORN_PORT-8082}:${ROWS_UVICORN_PORT-8082}
6343f2cdd06772efbc44a1e184f9cab2761745b0
Andrea Francis Soria Jimenez
2023-07-04T13:32:37
Moving configs and adding doc in readme (#1474)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 3a713a1f..59006344 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -106,36 +105,0 @@ class ParquetMetadataConfig: - - -DUCKDB_INDEX_STORAGE_DIRECTORY = None -DUCKDB_INDEX_COMMIT_MESSAGE = "Update duckdb index file" -DUCKDB_INDEX_COMMITTER_HF_TOKEN = None -DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES = 100_000_000 -DUCKDB_INDEX_TARGET_REVISION = "refs/convert/parquet" -DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s" -DUCKDB_INDEX_EXTENSIONS_DIRECTORY: Optional[str] = None - - -@dataclass(frozen=True) -class DuckDbIndexConfig: - storage_directory: Optional[str] = DUCKDB_INDEX_STORAGE_DIRECTORY - commit_message: str = DUCKDB_INDEX_COMMIT_MESSAGE - committer_hf_token: Optional[str] = DUCKDB_INDEX_COMMITTER_HF_TOKEN - target_revision: str = DUCKDB_INDEX_TARGET_REVISION - url_template: str = DUCKDB_INDEX_URL_TEMPLATE - max_parquet_size_bytes: int = DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES - extensions_directory: Optional[str] = DUCKDB_INDEX_EXTENSIONS_DIRECTORY - - @classmethod - def from_env(cls) -> "DuckDbIndexConfig": - env = Env(expand_vars=True) - with env.prefixed("DUCKDB_INDEX_"): - return cls( - storage_directory=env.str(name="STORAGE_DIRECTORY", default=DUCKDB_INDEX_STORAGE_DIRECTORY), - commit_message=env.str(name="COMMIT_MESSAGE", default=DUCKDB_INDEX_COMMIT_MESSAGE), - committer_hf_token=env.str(name="COMMITTER_HF_TOKEN", default=DUCKDB_INDEX_COMMITTER_HF_TOKEN), - target_revision=env.str(name="TARGET_REVISION", default=DUCKDB_INDEX_TARGET_REVISION), - url_template=env.str(name="URL_TEMPLATE", default=DUCKDB_INDEX_URL_TEMPLATE), - max_parquet_size_bytes=env.int( - name="MAX_PARQUET_SIZE_BYTES", default=DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES - ), - extensions_directory=env.str(name="EXTENSIONS_DIRECTORY", default=DUCKDB_INDEX_EXTENSIONS_DIRECTORY), - ) diff --git a/services/worker/README.md b/services/worker/README.md index 00bd203e..b6299326 100644 --- a/services/worker/README.md +++ b/services/worker/README.md @@ -77,0 +78,12 @@ Set environment variables to configure the parquet worker (`PARQUET_AND_INFO_` p +### Duckdb Index worker + +Set environment variables to configure the duckdb index worker (`DUCKDB_INDEX_` prefix): + +- `DUCKDB_INDEX_STORAGE_DIRECTORY`: directory where the temporal duckdb index files are stored. Defaults to empty. +- `DUCKDB_INDEX_COMMIT_MESSAGE`: the git commit message when the worker uploads the duckdb index file to the Hub. Defaults to `Update duckdb index file`. +- `DUCKDB_INDEX_COMMITTER_HF_TOKEN`: the HuggingFace token to commit the duckdb index file to the Hub. The token must be an app token associated with a user that has the right to 1. create the `refs/convert/parquet` branch (see `DUCKDB_INDEX_TARGET_REVISION`) and 2. push commits to it on any dataset. [Datasets maintainers](https://huggingface.co/datasets-maintainers) members have these rights. The token must have permission to write. If not set, the worker will fail. Defaults to None. +- `DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES`: the maximum size in bytes of the dataset's parquet files to index. Datasets with bigger size are ignored. Defaults to `100_000_000`. +- `DUCKDB_INDEX_TARGET_REVISION`: the git revision of the dataset where to store the duckdb index file. Make sure the committer token (`DUCKDB_INDEX_COMMITTER_HF_TOKEN`) has the permission to write there. Defaults to `refs/convert/parquet`. +- `DUCKDB_INDEX_URL_TEMPLATE`: the URL template to build the duckdb index file URL. Defaults to `/datasets/%s/resolve/%s/%s`. +- `DUCKDB_INDEX_EXTENSIONS_DIRECTORY`: directory where the duckdb extensions will be downloaded. Defaults to empty. + diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index 0a57557e..d39a3c03 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -12 +11,0 @@ from libcommon.config import ( - DuckDbIndexConfig, @@ -236,0 +236,36 @@ class ConfigNamesConfig: +DUCKDB_INDEX_STORAGE_DIRECTORY = None +DUCKDB_INDEX_COMMIT_MESSAGE = "Update duckdb index file" +DUCKDB_INDEX_COMMITTER_HF_TOKEN = None +DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES = 100_000_000 +DUCKDB_INDEX_TARGET_REVISION = "refs/convert/parquet" +DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s" +DUCKDB_INDEX_EXTENSIONS_DIRECTORY: Optional[str] = None + + +@dataclass(frozen=True) +class DuckDbIndexConfig: + storage_directory: Optional[str] = DUCKDB_INDEX_STORAGE_DIRECTORY + commit_message: str = DUCKDB_INDEX_COMMIT_MESSAGE + committer_hf_token: Optional[str] = DUCKDB_INDEX_COMMITTER_HF_TOKEN + target_revision: str = DUCKDB_INDEX_TARGET_REVISION + url_template: str = DUCKDB_INDEX_URL_TEMPLATE + max_parquet_size_bytes: int = DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES + extensions_directory: Optional[str] = DUCKDB_INDEX_EXTENSIONS_DIRECTORY + + @classmethod + def from_env(cls) -> "DuckDbIndexConfig": + env = Env(expand_vars=True) + with env.prefixed("DUCKDB_INDEX_"): + return cls( + storage_directory=env.str(name="STORAGE_DIRECTORY", default=DUCKDB_INDEX_STORAGE_DIRECTORY), + commit_message=env.str(name="COMMIT_MESSAGE", default=DUCKDB_INDEX_COMMIT_MESSAGE), + committer_hf_token=env.str(name="COMMITTER_HF_TOKEN", default=DUCKDB_INDEX_COMMITTER_HF_TOKEN), + target_revision=env.str(name="TARGET_REVISION", default=DUCKDB_INDEX_TARGET_REVISION), + url_template=env.str(name="URL_TEMPLATE", default=DUCKDB_INDEX_URL_TEMPLATE), + max_parquet_size_bytes=env.int( + name="MAX_PARQUET_SIZE_BYTES", default=DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES + ), + extensions_directory=env.str(name="EXTENSIONS_DIRECTORY", default=DUCKDB_INDEX_EXTENSIONS_DIRECTORY), + ) + + diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py index 97fbb673..26ff821b 100644 --- a/services/worker/src/worker/job_runners/split/duckdb_index.py +++ b/services/worker/src/worker/job_runners/split/duckdb_index.py @@ -16 +15,0 @@ from huggingface_hub.utils._errors import HfHubHTTPError, RepositoryNotFoundErro -from libcommon.config import DuckDbIndexConfig @@ -36 +35 @@ from libcommon.utils import JobInfo, SplitHubFile -from worker.config import AppConfig +from worker.config import AppConfig, DuckDbIndexConfig
4c81b9616c0f570985a7589eb7a15b0e35b445f9
Sylvain Lesage
2023-07-04T13:32:18
Create libapi (#1475)
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml index c285df42..b282f8e3 100644 --- a/.github/workflows/_quality-python.yml +++ b/.github/workflows/_quality-python.yml @@ -52 +52 @@ jobs: - run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed 's/^huggingface-hub @ git.*/huggingface-hub==0.15.1 ; python_full_version == \"3.9.15\" --hash=sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445 --hash=sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081/')" + run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libapi @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed 's/^huggingface-hub @ git.*/huggingface-hub==0.15.1 ; python_full_version == \"3.9.15\" --hash=sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445 --hash=sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081/')" diff --git a/.github/workflows/l-libapi.yml b/.github/workflows/l-libapi.yml new file mode 100644 index 00000000..e7f75a41 --- /dev/null +++ b/.github/workflows/l-libapi.yml @@ -0,0 +1,31 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +name: libs/libapi +on: + workflow_dispatch: + push: + branches: + - main + paths: + - "libs/libapi/**" + - ".github/workflows/l-libapi.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" + pull_request: + paths: + - "libs/libapi/**" + - ".github/workflows/l-libapi.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" +jobs: + quality: + uses: ./.github/workflows/_quality-python.yml + with: + working-directory: libs/libapi + unit-tests: + uses: ./.github/workflows/_unit-tests-python.yml + with: + working-directory: libs/libapi diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace index d73dd79a..597c0f2f 100644 --- a/.vscode/monorepo.code-workspace +++ b/.vscode/monorepo.code-workspace @@ -22,0 +23,4 @@ + { + "name": "libs/libapi", + "path": "../libs/libapi" + }, diff --git a/libs/libapi/.flake8 b/libs/libapi/.flake8 new file mode 100644 index 00000000..f7d6157c --- /dev/null +++ b/libs/libapi/.flake8 @@ -0,0 +1,5 @@ +[flake8] +# Recommend matching the black line length (119), +# rather than using the flake8 default of 79: +max-line-length = 119 +extend-ignore = "E203" diff --git a/libs/libapi/.python-version b/libs/libapi/.python-version new file mode 100644 index 00000000..b326afbc --- /dev/null +++ b/libs/libapi/.python-version @@ -0,0 +1 @@ +3.9.15 diff --git a/libs/libapi/Makefile b/libs/libapi/Makefile new file mode 100644 index 00000000..ae7d6714 --- /dev/null +++ b/libs/libapi/Makefile @@ -0,0 +1,10 @@ +# environment variables for the commands (docker compose, poetry) +export MONGO_PORT := 27021 +export METRICS_MONGO_URL := mongodb://localhost:${MONGO_PORT} +export COMPOSE_PROJECT_NAME := libapi +# makefile variables +DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml + +include ../../tools/Python.mk +include ../../tools/PythonTest.mk +include ../../tools/Docker.mk diff --git a/libs/libapi/README.md b/libs/libapi/README.md new file mode 100644 index 00000000..8ab47fe8 --- /dev/null +++ b/libs/libapi/README.md @@ -0,0 +1,31 @@ +# libapi + +A Python library for the API services + +## Configuration + +The APIs can be configured using environment variables. They are grouped by scope. + +### API service + +Set environment variables to configure the application (`API_` prefix): + +- `API_HF_AUTH_PATH`: the path of the external authentication service, on the hub (see `HF_ENDPOINT`). The string must contain `%s` which will be replaced with the dataset name. The external authentication service must return 200, 401, 403 or 404. Defaults to "/api/datasets/%s/auth-check". +- `API_HF_JWT_PUBLIC_KEY_URL`: the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format. It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored. Defaults to empty. +- `API_HF_JWT_ALGORITHM`: the algorithm used to encode the JWT. Defaults to `"EdDSA"`. +- `API_HF_TIMEOUT_SECONDS`: the timeout in seconds for the requests to the Hugging Face Hub. Defaults to `0.2` (200 ms). +- `API_HF_WEBHOOK_SECRET`: a shared secret sent by the Hub in the "X-Webhook-Secret" header of POST requests sent to /webhook, to authenticate the originator and bypass some validation of the content (avoiding roundtrip to the Hub). If not set, all the validations are done. Defaults to empty. +- `API_MAX_AGE_LONG`: number of seconds to set in the `max-age` header on data endpoints. Defaults to `120` (2 minutes). +- `API_MAX_AGE_SHORT`: number of seconds to set in the `max-age` header on technical endpoints. Defaults to `10` (10 seconds). + +### Uvicorn + +The following environment variables are used to configure the Uvicorn server (`API_UVICORN_` prefix): + +- `API_UVICORN_HOSTNAME`: the hostname. Defaults to `"localhost"`. +- `API_UVICORN_NUM_WORKERS`: the number of uvicorn workers. Defaults to `2`. +- `API_UVICORN_PORT`: the port. Defaults to `8000`. + +### Prometheus + +- `PROMETHEUS_MULTIPROC_DIR`: the directory where the uvicorn workers share their prometheus metrics. See https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn. Defaults to empty, in which case every worker manages its own metrics, and the /metrics endpoint returns the metrics of a random worker. diff --git a/libs/libapi/poetry.lock b/libs/libapi/poetry.lock new file mode 100644 index 00000000..c970667c --- /dev/null +++ b/libs/libapi/poetry.lock @@ -0,0 +1,3178 @@ +# This file is automatically @generated by Poetry and should not be changed by hand. + +[[package]] +name = "aiohttp" +version = "3.8.4" +description = "Async http client/server framework (asyncio)" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"}, + {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"}, + {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"}, + {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"}, + {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"}, + {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"}, + {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"}, + {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"}, + {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"}, + {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"}, + {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"}, + {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"}, + {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"}, + {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"}, + {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"}, + {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"}, + {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"}, + {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"}, + {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"}, + {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"}, + {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"}, + {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"}, + {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"}, + {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"}, + {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"}, + {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"}, + {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"}, + {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"}, + {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"}, + {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"}, +] + +[package.dependencies] +aiosignal = ">=1.1.2" +async-timeout = ">=4.0.0a3,<5.0" +attrs = ">=17.3.0" +charset-normalizer = ">=2.0,<4.0" +frozenlist = ">=1.1.1" +multidict = ">=4.5,<7.0" +yarl = ">=1.0,<2.0" + +[package.extras] +speedups = ["Brotli", "aiodns", "cchardet"] + +[[package]] +name = "aiosignal" +version = "1.3.1" +description = "aiosignal: a list of registered asynchronous callbacks" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, + {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, +] + +[package.dependencies] +frozenlist = ">=1.1.0" + +[[package]] +name = "anyio" +version = "3.7.0" +description = "High level compatibility layer for multiple asynchronous event loop implementations" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "anyio-3.7.0-py3-none-any.whl", hash = "sha256:eddca883c4175f14df8aedce21054bfca3adb70ffe76a9f607aef9d7fa2ea7f0"}, + {file = "anyio-3.7.0.tar.gz", hash = "sha256:275d9973793619a5374e1c89a4f4ad3f4b0a5510a2b5b939444bee8f4c4d37ce"}, +] + +[package.dependencies] +exceptiongroup = {version = "*", markers = "python_version < \"3.11\""} +idna = ">=2.8" +sniffio = ">=1.1" + +[package.extras] +doc = ["Sphinx (>=6.1.0)", "packaging", "sphinx-autodoc-typehints (>=1.2.0)", "sphinx-rtd-theme", "sphinxcontrib-jquery"] +test = ["anyio[trio]", "coverage[toml] (>=4.5)", "hypothesis (>=4.0)", "mock (>=4)", "psutil (>=5.9)", "pytest (>=7.0)", "pytest-mock (>=3.6.1)", "trustme", "uvloop (>=0.17)"] +trio = ["trio (<0.22)"] + +[[package]] +name = "appdirs" +version = "1.4.4" +description = "A small Python module for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "appdirs-1.4.4-py2.py3-none-any.whl", hash = "sha256:a841dacd6b99318a741b166adb07e19ee71a274450e68237b4650ca1055ab128"}, + {file = "appdirs-1.4.4.tar.gz", hash = "sha256:7d5d0167b2b1ba821647616af46a749d1c653740dd0d2415100fe26e27afdf41"}, +] + +[[package]] +name = "async-timeout" +version = "4.0.2" +description = "Timeout context manager for asyncio programs" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"}, + {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"}, +] + +[[package]] +name = "attrs" +version = "23.1.0" +description = "Classes Without Boilerplate" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "attrs-23.1.0-py3-none-any.whl", hash = "sha256:1f28b4522cdc2fb4256ac1a020c78acf9cba2c6b461ccd2c126f3aa8e8335d04"}, + {file = "attrs-23.1.0.tar.gz", hash = "sha256:6279836d581513a26f1bf235f9acd333bc9115683f14f7e8fae46c98fc50e015"}, +] + +[package.extras] +cov = ["attrs[tests]", "coverage[toml] (>=5.3)"] +dev = ["attrs[docs,tests]", "pre-commit"] +docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope-interface"] +tests = ["attrs[tests-no-zope]", "zope-interface"] +tests-no-zope = ["cloudpickle", "hypothesis", "mypy (>=1.1.1)", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"] + +[[package]] +name = "audioread" +version = "3.0.0" +description = "multi-library, cross-platform audio decoding" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"}, +] + +[[package]] +name = "bandit" +version = "1.7.5" +description = "Security oriented static analyser for python code." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "bandit-1.7.5-py3-none-any.whl", hash = "sha256:75665181dc1e0096369112541a056c59d1c5f66f9bb74a8d686c3c362b83f549"}, + {file = "bandit-1.7.5.tar.gz", hash = "sha256:bdfc739baa03b880c2d15d0431b31c658ffc348e907fe197e54e0389dd59e11e"}, +] + +[package.dependencies] +colorama = {version = ">=0.3.9", markers = "platform_system == \"Windows\""} +GitPython = ">=1.0.1" +PyYAML = ">=5.3.1" +rich = "*" +stevedore = ">=1.20.0" + +[package.extras] +test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "tomli (>=1.1.0)"] +toml = ["tomli (>=1.1.0)"] +yaml = ["PyYAML"] + +[[package]] +name = "black" +version = "22.12.0" +description = "The uncompromising code formatter." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "black-22.12.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eedd20838bd5d75b80c9f5487dbcb06836a43833a37846cf1d8c1cc01cef59d"}, + {file = "black-22.12.0-cp310-cp310-win_amd64.whl", hash = "sha256:159a46a4947f73387b4d83e87ea006dbb2337eab6c879620a3ba52699b1f4351"}, + {file = "black-22.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d30b212bffeb1e252b31dd269dfae69dd17e06d92b87ad26e23890f3efea366f"}, + {file = "black-22.12.0-cp311-cp311-win_amd64.whl", hash = "sha256:7412e75863aa5c5411886804678b7d083c7c28421210180d67dfd8cf1221e1f4"}, + {file = "black-22.12.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c116eed0efb9ff870ded8b62fe9f28dd61ef6e9ddd28d83d7d264a38417dcee2"}, + {file = "black-22.12.0-cp37-cp37m-win_amd64.whl", hash = "sha256:1f58cbe16dfe8c12b7434e50ff889fa479072096d79f0a7f25e4ab8e94cd8350"}, + {file = "black-22.12.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77d86c9f3db9b1bf6761244bc0b3572a546f5fe37917a044e02f3166d5aafa7d"}, + {file = "black-22.12.0-cp38-cp38-win_amd64.whl", hash = "sha256:82d9fe8fee3401e02e79767016b4907820a7dc28d70d137eb397b92ef3cc5bfc"}, + {file = "black-22.12.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:101c69b23df9b44247bd88e1d7e90154336ac4992502d4197bdac35dd7ee3320"}, + {file = "black-22.12.0-cp39-cp39-win_amd64.whl", hash = "sha256:559c7a1ba9a006226f09e4916060982fd27334ae1998e7a38b3f33a37f7a2148"}, + {file = "black-22.12.0-py3-none-any.whl", hash = "sha256:436cc9167dd28040ad90d3b404aec22cedf24a6e4d7de221bec2730ec0c97bcf"}, + {file = "black-22.12.0.tar.gz", hash = "sha256:229351e5a18ca30f447bf724d007f890f97e13af070bb6ad4c0a441cd7596a2f"}, +] + +[package.dependencies] +click = ">=8.0.0" +mypy-extensions = ">=0.4.3" +pathspec = ">=0.9.0" +platformdirs = ">=2" +tomli = {version = ">=1.1.0", markers = "python_full_version < \"3.11.0a7\""} +typing-extensions = {version = ">=3.10.0.0", markers = "python_version < \"3.10\""} + +[package.extras] +colorama = ["colorama (>=0.4.3)"] +d = ["aiohttp (>=3.7.4)"] +jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"] +uvloop = ["uvloop (>=0.15.2)"] + +[[package]] +name = "cachecontrol" +version = "0.13.1" +description = "httplib2 caching for requests" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, +] + +[package.dependencies] +filelock = {version = ">=3.8.0", optional = true, markers = "extra == \"filecache\""} +msgpack = ">=0.5.2" +requests = ">=2.16.0" + +[package.extras] +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] +filecache = ["filelock (>=3.8.0)"] +redis = ["redis (>=2.10.5)"] + +[[package]] +name = "certifi" +version = "2023.5.7" +description = "Python package for providing Mozilla's CA Bundle." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "certifi-2023.5.7-py3-none-any.whl", hash = "sha256:c6c2e98f5c7869efca1f8916fed228dd91539f9f1b444c314c06eef02980c716"}, + {file = "certifi-2023.5.7.tar.gz", hash = "sha256:0f0d56dc5a6ad56fd4ba36484d6cc34451e1c6548c61daad8c320169f91eddc7"}, +] + +[[package]] +name = "cffi" +version = "1.15.1" +description = "Foreign Function Interface for Python calling C code." +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"}, + {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"}, + {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"}, + {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"}, + {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"}, + {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"}, + {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"}, + {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"}, + {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"}, + {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"}, + {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"}, + {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"}, + {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"}, + {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"}, + {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"}, + {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"}, + {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"}, + {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"}, + {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"}, + {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"}, + {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"}, + {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"}, + {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"}, + {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"}, + {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"}, + {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"}, + {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"}, + {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"}, + {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"}, + {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"}, + {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"}, + {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"}, + {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"}, +] + +[package.dependencies] +pycparser = "*" + +[[package]] +name = "charset-normalizer" +version = "3.1.0" +description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"}, + {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"}, + {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"}, + {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"}, + {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"}, + {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"}, + {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"}, +] + +[[package]] +name = "click" +version = "8.1.3" +description = "Composable command line interface toolkit" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "click-8.1.3-py3-none-any.whl", hash = "sha256:bb4d8133cb15a609f44e8213d9b391b0809795062913b383c62be0ee95b1db48"}, + {file = "click-8.1.3.tar.gz", hash = "sha256:7682dc8afb30297001674575ea00d1814d808d6a36af415a82bd481d37ba7b8e"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[[package]] +name = "colorama" +version = "0.4.6" +description = "Cross-platform colored terminal text." +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7" +files = [ + {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"}, + {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"}, +] + +[[package]] +name = "coverage" +version = "7.2.7" +description = "Code coverage measurement for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "coverage-7.2.7-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d39b5b4f2a66ccae8b7263ac3c8170994b65266797fb96cbbfd3fb5b23921db8"}, + {file = "coverage-7.2.7-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:6d040ef7c9859bb11dfeb056ff5b3872436e3b5e401817d87a31e1750b9ae2fb"}, + {file = "coverage-7.2.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ba90a9563ba44a72fda2e85302c3abc71c5589cea608ca16c22b9804262aaeb6"}, + {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e7d9405291c6928619403db1d10bd07888888ec1abcbd9748fdaa971d7d661b2"}, + {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:31563e97dae5598556600466ad9beea39fb04e0229e61c12eaa206e0aa202063"}, + {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ebba1cd308ef115925421d3e6a586e655ca5a77b5bf41e02eb0e4562a111f2d1"}, + {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:cb017fd1b2603ef59e374ba2063f593abe0fc45f2ad9abdde5b4d83bd922a353"}, + {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d62a5c7dad11015c66fbb9d881bc4caa5b12f16292f857842d9d1871595f4495"}, + {file = "coverage-7.2.7-cp310-cp310-win32.whl", hash = "sha256:ee57190f24fba796e36bb6d3aa8a8783c643d8fa9760c89f7a98ab5455fbf818"}, + {file = "coverage-7.2.7-cp310-cp310-win_amd64.whl", hash = "sha256:f75f7168ab25dd93110c8a8117a22450c19976afbc44234cbf71481094c1b850"}, + {file = "coverage-7.2.7-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:06a9a2be0b5b576c3f18f1a241f0473575c4a26021b52b2a85263a00f034d51f"}, + {file = "coverage-7.2.7-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5baa06420f837184130752b7c5ea0808762083bf3487b5038d68b012e5937dbe"}, + {file = "coverage-7.2.7-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fdec9e8cbf13a5bf63290fc6013d216a4c7232efb51548594ca3631a7f13c3a3"}, + {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:52edc1a60c0d34afa421c9c37078817b2e67a392cab17d97283b64c5833f427f"}, + {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:63426706118b7f5cf6bb6c895dc215d8a418d5952544042c8a2d9fe87fcf09cb"}, + {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:afb17f84d56068a7c29f5fa37bfd38d5aba69e3304af08ee94da8ed5b0865833"}, + {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:48c19d2159d433ccc99e729ceae7d5293fbffa0bdb94952d3579983d1c8c9d97"}, + {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:0e1f928eaf5469c11e886fe0885ad2bf1ec606434e79842a879277895a50942a"}, + {file = "coverage-7.2.7-cp311-cp311-win32.whl", hash = "sha256:33d6d3ea29d5b3a1a632b3c4e4f4ecae24ef170b0b9ee493883f2df10039959a"}, + {file = "coverage-7.2.7-cp311-cp311-win_amd64.whl", hash = "sha256:5b7540161790b2f28143191f5f8ec02fb132660ff175b7747b95dcb77ac26562"}, + {file = "coverage-7.2.7-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:f2f67fe12b22cd130d34d0ef79206061bfb5eda52feb6ce0dba0644e20a03cf4"}, + {file = "coverage-7.2.7-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a342242fe22407f3c17f4b499276a02b01e80f861f1682ad1d95b04018e0c0d4"}, + {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:171717c7cb6b453aebac9a2ef603699da237f341b38eebfee9be75d27dc38e01"}, + {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:49969a9f7ffa086d973d91cec8d2e31080436ef0fb4a359cae927e742abfaaa6"}, + {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:b46517c02ccd08092f4fa99f24c3b83d8f92f739b4657b0f146246a0ca6a831d"}, + {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:a3d33a6b3eae87ceaefa91ffdc130b5e8536182cd6dfdbfc1aa56b46ff8c86de"}, + {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:976b9c42fb2a43ebf304fa7d4a310e5f16cc99992f33eced91ef6f908bd8f33d"}, + {file = "coverage-7.2.7-cp312-cp312-win32.whl", hash = "sha256:8de8bb0e5ad103888d65abef8bca41ab93721647590a3f740100cd65c3b00511"}, + {file = "coverage-7.2.7-cp312-cp312-win_amd64.whl", hash = "sha256:9e31cb64d7de6b6f09702bb27c02d1904b3aebfca610c12772452c4e6c21a0d3"}, + {file = "coverage-7.2.7-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:58c2ccc2f00ecb51253cbe5d8d7122a34590fac9646a960d1430d5b15321d95f"}, + {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d22656368f0e6189e24722214ed8d66b8022db19d182927b9a248a2a8a2f67eb"}, + {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a895fcc7b15c3fc72beb43cdcbdf0ddb7d2ebc959edac9cef390b0d14f39f8a9"}, + {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e84606b74eb7de6ff581a7915e2dab7a28a0517fbe1c9239eb227e1354064dcd"}, + {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0a5f9e1dbd7fbe30196578ca36f3fba75376fb99888c395c5880b355e2875f8a"}, + {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:419bfd2caae268623dd469eff96d510a920c90928b60f2073d79f8fe2bbc5959"}, + {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:2aee274c46590717f38ae5e4650988d1af340fe06167546cc32fe2f58ed05b02"}, + {file = "coverage-7.2.7-cp37-cp37m-win32.whl", hash = "sha256:61b9a528fb348373c433e8966535074b802c7a5d7f23c4f421e6c6e2f1697a6f"}, + {file = "coverage-7.2.7-cp37-cp37m-win_amd64.whl", hash = "sha256:b1c546aca0ca4d028901d825015dc8e4d56aac4b541877690eb76490f1dc8ed0"}, + {file = "coverage-7.2.7-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:54b896376ab563bd38453cecb813c295cf347cf5906e8b41d340b0321a5433e5"}, + {file = "coverage-7.2.7-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3d376df58cc111dc8e21e3b6e24606b5bb5dee6024f46a5abca99124b2229ef5"}, + {file = "coverage-7.2.7-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e330fc79bd7207e46c7d7fd2bb4af2963f5f635703925543a70b99574b0fea9"}, + {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e9d683426464e4a252bf70c3498756055016f99ddaec3774bf368e76bbe02b6"}, + {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8d13c64ee2d33eccf7437961b6ea7ad8673e2be040b4f7fd4fd4d4d28d9ccb1e"}, + {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:b7aa5f8a41217360e600da646004f878250a0d6738bcdc11a0a39928d7dc2050"}, + {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8fa03bce9bfbeeef9f3b160a8bed39a221d82308b4152b27d82d8daa7041fee5"}, + {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:245167dd26180ab4c91d5e1496a30be4cd721a5cf2abf52974f965f10f11419f"}, + {file = "coverage-7.2.7-cp38-cp38-win32.whl", hash = "sha256:d2c2db7fd82e9b72937969bceac4d6ca89660db0a0967614ce2481e81a0b771e"}, + {file = "coverage-7.2.7-cp38-cp38-win_amd64.whl", hash = "sha256:2e07b54284e381531c87f785f613b833569c14ecacdcb85d56b25c4622c16c3c"}, + {file = "coverage-7.2.7-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:537891ae8ce59ef63d0123f7ac9e2ae0fc8b72c7ccbe5296fec45fd68967b6c9"}, + {file = "coverage-7.2.7-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:06fb182e69f33f6cd1d39a6c597294cff3143554b64b9825d1dc69d18cc2fff2"}, + {file = "coverage-7.2.7-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:201e7389591af40950a6480bd9edfa8ed04346ff80002cec1a66cac4549c1ad7"}, + {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f6951407391b639504e3b3be51b7ba5f3528adbf1a8ac3302b687ecababf929e"}, + {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6f48351d66575f535669306aa7d6d6f71bc43372473b54a832222803eb956fd1"}, + {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b29019c76039dc3c0fd815c41392a044ce555d9bcdd38b0fb60fb4cd8e475ba9"}, + {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:81c13a1fc7468c40f13420732805a4c38a105d89848b7c10af65a90beff25250"}, + {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:975d70ab7e3c80a3fe86001d8751f6778905ec723f5b110aed1e450da9d4b7f2"}, + {file = "coverage-7.2.7-cp39-cp39-win32.whl", hash = "sha256:7ee7d9d4822c8acc74a5e26c50604dff824710bc8de424904c0982e25c39c6cb"}, + {file = "coverage-7.2.7-cp39-cp39-win_amd64.whl", hash = "sha256:eb393e5ebc85245347950143969b241d08b52b88a3dc39479822e073a1a8eb27"}, + {file = "coverage-7.2.7-pp37.pp38.pp39-none-any.whl", hash = "sha256:b7b4c971f05e6ae490fef852c218b0e79d4e52f79ef0c8475566584a8fb3e01d"}, + {file = "coverage-7.2.7.tar.gz", hash = "sha256:924d94291ca674905fe9481f12294eb11f2d3d3fd1adb20314ba89e94f44ed59"}, +] + +[package.extras] +toml = ["tomli"] + +[[package]] +name = "cryptography" +version = "41.0.1" +description = "cryptography is a package which provides cryptographic recipes and primitives to Python developers." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_universal2.whl", hash = "sha256:f73bff05db2a3e5974a6fd248af2566134d8981fd7ab012e5dd4ddb1d9a70699"}, + {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_x86_64.whl", hash = "sha256:1a5472d40c8f8e91ff7a3d8ac6dfa363d8e3138b961529c996f3e2df0c7a411a"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fa01527046ca5facdf973eef2535a27fec4cb651e4daec4d043ef63f6ecd4ca"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b46e37db3cc267b4dea1f56da7346c9727e1209aa98487179ee8ebed09d21e43"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:d198820aba55660b4d74f7b5fd1f17db3aa5eb3e6893b0a41b75e84e4f9e0e4b"}, + {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:948224d76c4b6457349d47c0c98657557f429b4e93057cf5a2f71d603e2fc3a3"}, + {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:059e348f9a3c1950937e1b5d7ba1f8e968508ab181e75fc32b879452f08356db"}, + {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:b4ceb5324b998ce2003bc17d519080b4ec8d5b7b70794cbd2836101406a9be31"}, + {file = "cryptography-41.0.1-cp37-abi3-win32.whl", hash = "sha256:8f4ab7021127a9b4323537300a2acfb450124b2def3756f64dc3a3d2160ee4b5"}, + {file = "cryptography-41.0.1-cp37-abi3-win_amd64.whl", hash = "sha256:1fee5aacc7367487b4e22484d3c7e547992ed726d14864ee33c0176ae43b0d7c"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-macosx_10_12_x86_64.whl", hash = "sha256:9a6c7a3c87d595608a39980ebaa04d5a37f94024c9f24eb7d10262b92f739ddb"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:5d092fdfedaec4cbbffbf98cddc915ba145313a6fdaab83c6e67f4e6c218e6f3"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:1a8e6c2de6fbbcc5e14fd27fb24414507cb3333198ea9ab1258d916f00bc3039"}, + {file = "cryptography-41.0.1-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cb33ccf15e89f7ed89b235cff9d49e2e62c6c981a6061c9c8bb47ed7951190bc"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:5f0ff6e18d13a3de56f609dd1fd11470918f770c6bd5d00d632076c727d35485"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:7bfc55a5eae8b86a287747053140ba221afc65eb06207bedf6e019b8934b477c"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:eb8163f5e549a22888c18b0d53d6bb62a20510060a22fd5a995ec8a05268df8a"}, + {file = "cryptography-41.0.1-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:8dde71c4169ec5ccc1087bb7521d54251c016f126f922ab2dfe6649170a3b8c5"}, + {file = "cryptography-41.0.1.tar.gz", hash = "sha256:d34579085401d3f49762d2f7d6634d6b6c2ae1242202e860f4d26b046e3a1006"}, +] + +[package.dependencies] +cffi = ">=1.12" + +[package.extras] +docs = ["sphinx (>=5.3.0)", "sphinx-rtd-theme (>=1.1.1)"] +docstest = ["pyenchant (>=1.6.11)", "sphinxcontrib-spelling (>=4.0.1)", "twine (>=1.12.0)"] +nox = ["nox"] +pep8test = ["black", "check-sdist", "mypy", "ruff"] +sdist = ["build"] +ssh = ["bcrypt (>=3.1.5)"] +test = ["pretend", "pytest (>=6.2.0)", "pytest-benchmark", "pytest-cov", "pytest-xdist"] +test-randomorder = ["pytest-randomly"] + +[[package]] +name = "cyclonedx-python-lib" +version = "4.0.1" +description = "A library for producing CycloneDX SBOM (Software Bill of Materials) files." +category = "dev" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "cyclonedx_python_lib-4.0.1-py3-none-any.whl", hash = "sha256:907b64f00df85d727a425de86604768b248cf19285993729e04f17bec767f692"}, + {file = "cyclonedx_python_lib-4.0.1.tar.gz", hash = "sha256:878e33b8e0080c786f6cbd4c6f87ad610db65d6a3a686a5698415d9cfcd8925d"}, +] + +[package.dependencies] +packageurl-python = ">=0.11" +py-serializable = ">=0.11.1,<0.12.0" +sortedcontainers = ">=2.4.0,<3.0.0" + +[[package]] +name = "datasets" +version = "2.13.1" +description = "HuggingFace community-driven open-source library of datasets" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, +] + +[package.dependencies] +aiohttp = "*" +dill = ">=0.3.0,<0.3.7" +fsspec = {version = ">=2021.11.1", extras = ["http"]} +huggingface-hub = ">=0.11.0,<1.0.0" +librosa = {version = "*", optional = true, markers = "extra == \"audio\""} +multiprocess = "*" +numpy = ">=1.17" +packaging = "*" +pandas = "*" +Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""} +pyarrow = ">=8.0.0" +pyyaml = ">=5.1" +requests = ">=2.19.0" +soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""} +tqdm = ">=4.62.1" +xxhash = "*" + +[package.extras] +apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"] +audio = ["librosa", "soundfile (>=0.12.1)"] +benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +docs = ["s3fs"] +jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"] +s3 = ["s3fs"] +tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"] +tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +torch = ["torch"] +vision = ["Pillow (>=6.2.1)"] + +[[package]] +name = "decorator" +version = "5.1.1" +description = "Decorators for Humans" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"}, + {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"}, +] + +[[package]] +name = "defusedxml" +version = "0.7.1" +description = "XML bomb protection for Python stdlib modules" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "defusedxml-0.7.1-py2.py3-none-any.whl", hash = "sha256:a352e7e428770286cc899e2542b6cdaedb2b4953ff269a210103ec58f6198a61"}, + {file = "defusedxml-0.7.1.tar.gz", hash = "sha256:1bb3032db185915b62d7c6209c5a8792be6a32ab2fedacc84e01b52c51aa3e69"}, +] + +[[package]] +name = "dill" +version = "0.3.6" +description = "serialize all of python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"}, + {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"}, +] + +[package.extras] +graph = ["objgraph (>=1.7.2)"] + +[[package]] +name = "dnspython" +version = "1.16.0" +description = "DNS toolkit" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "dnspython-1.16.0-py2.py3-none-any.whl", hash = "sha256:f69c21288a962f4da86e56c4905b49d11aba7938d3d740e80d9e366ee4f1632d"}, + {file = "dnspython-1.16.0.zip", hash = "sha256:36c5e8e38d4369a08b6780b7f27d790a292b2b08eea01607865bf0936c558e01"}, +] + +[package.extras] +dnssec = ["ecdsa (>=0.13)", "pycryptodome"] +idna = ["idna (>=2.1)"] + +[[package]] +name = "environs" +version = "9.5.0" +description = "simplified environment variable parsing" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "environs-9.5.0-py2.py3-none-any.whl", hash = "sha256:1e549569a3de49c05f856f40bce86979e7d5ffbbc4398e7f338574c220189124"}, + {file = "environs-9.5.0.tar.gz", hash = "sha256:a76307b36fbe856bdca7ee9161e6c466fd7fcffc297109a118c59b54e27e30c9"}, +] + +[package.dependencies] +marshmallow = ">=3.0.0" +python-dotenv = "*" + +[package.extras] +dev = ["dj-database-url", "dj-email-url", "django-cache-url", "flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)", "pytest", "tox"] +django = ["dj-database-url", "dj-email-url", "django-cache-url"] +lint = ["flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)"] +tests = ["dj-database-url", "dj-email-url", "django-cache-url", "pytest"] + +[[package]] +name = "exceptiongroup" +version = "1.1.2" +description = "Backport of PEP 654 (exception groups)" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "exceptiongroup-1.1.2-py3-none-any.whl", hash = "sha256:e346e69d186172ca7cf029c8c1d16235aa0e04035e5750b4b95039e65204328f"}, + {file = "exceptiongroup-1.1.2.tar.gz", hash = "sha256:12c3e887d6485d16943a309616de20ae5582633e0a2eda17f4e10fd61c1e8af5"}, +] + +[package.extras] +test = ["pytest (>=6)"] + +[[package]] +name = "filelock" +version = "3.12.2" +description = "A platform independent file lock." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "filelock-3.12.2-py3-none-any.whl", hash = "sha256:cbb791cdea2a72f23da6ac5b5269ab0a0d161e9ef0100e653b69049a7706d1ec"}, + {file = "filelock-3.12.2.tar.gz", hash = "sha256:002740518d8aa59a26b0c76e10fb8c6e15eae825d34b6fdf670333fd7b938d81"}, +] + +[package.extras] +docs = ["furo (>=2023.5.20)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"] +testing = ["covdefaults (>=2.3)", "coverage (>=7.2.7)", "diff-cover (>=7.5)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)", "pytest-timeout (>=2.1)"] + +[[package]] +name = "flake8" +version = "3.9.2" +description = "the modular source code checker: pep8 pyflakes and co" +category = "dev" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7" +files = [ + {file = "flake8-3.9.2-py2.py3-none-any.whl", hash = "sha256:bf8fd333346d844f616e8d47905ef3a3384edae6b4e9beb0c5101e25e3110907"}, + {file = "flake8-3.9.2.tar.gz", hash = "sha256:07528381786f2a6237b061f6e96610a4167b226cb926e2aa2b6b1d78057c576b"}, +] + +[package.dependencies] +mccabe = ">=0.6.0,<0.7.0" +pycodestyle = ">=2.7.0,<2.8.0" +pyflakes = ">=2.3.0,<2.4.0" + +[[package]] +name = "frozenlist" +version = "1.3.3" +description = "A list-like structure which implements collections.abc.MutableSequence" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"}, + {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"}, + {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"}, + {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"}, + {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"}, + {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"}, + {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"}, + {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"}, + {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"}, + {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"}, + {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"}, + {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"}, + {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"}, + {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"}, + {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"}, + {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"}, + {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"}, + {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"}, + {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"}, + {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"}, + {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"}, + {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"}, + {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"}, + {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"}, + {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"}, + {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"}, +] + +[[package]] +name = "fsspec" +version = "2023.6.0" +description = "File-system specification" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "fsspec-2023.6.0-py3-none-any.whl", hash = "sha256:1cbad1faef3e391fba6dc005ae9b5bdcbf43005c9167ce78c915549c352c869a"}, + {file = "fsspec-2023.6.0.tar.gz", hash = "sha256:d0b2f935446169753e7a5c5c55681c54ea91996cc67be93c39a154fb3a2742af"}, +] + +[package.dependencies] +aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""} +requests = {version = "*", optional = true, markers = "extra == \"http\""} + +[package.extras] +abfs = ["adlfs"] +adl = ["adlfs"] +arrow = ["pyarrow (>=1)"] +dask = ["dask", "distributed"] +devel = ["pytest", "pytest-cov"] +dropbox = ["dropbox", "dropboxdrivefs", "requests"] +full = ["adlfs", "aiohttp (!=4.0.0a0,!=4.0.0a1)", "dask", "distributed", "dropbox", "dropboxdrivefs", "fusepy", "gcsfs", "libarchive-c", "ocifs", "panel", "paramiko", "pyarrow (>=1)", "pygit2", "requests", "s3fs", "smbprotocol", "tqdm"] +fuse = ["fusepy"] +gcs = ["gcsfs"] +git = ["pygit2"] +github = ["requests"] +gs = ["gcsfs"] +gui = ["panel"] +hdfs = ["pyarrow (>=1)"] +http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"] +libarchive = ["libarchive-c"] +oci = ["ocifs"] +s3 = ["s3fs"] +sftp = ["paramiko"] +smb = ["smbprotocol"] +ssh = ["paramiko"] +tqdm = ["tqdm"] + +[[package]] +name = "gitdb" +version = "4.0.10" +description = "Git Object Database" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "gitdb-4.0.10-py3-none-any.whl", hash = "sha256:c286cf298426064079ed96a9e4a9d39e7f3e9bf15ba60701e95f5492f28415c7"}, + {file = "gitdb-4.0.10.tar.gz", hash = "sha256:6eb990b69df4e15bad899ea868dc46572c3f75339735663b81de79b06f17eb9a"}, +] + +[package.dependencies] +smmap = ">=3.0.1,<6" + +[[package]] +name = "gitpython" +version = "3.1.31" +description = "GitPython is a Python library used to interact with Git repositories" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"}, + {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"}, +] + +[package.dependencies] +gitdb = ">=4.0.1,<5" + +[[package]] +name = "html5lib" +version = "1.1" +description = "HTML parser based on the WHATWG HTML specification" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "html5lib-1.1-py2.py3-none-any.whl", hash = "sha256:0d78f8fde1c230e99fe37986a60526d7049ed4bf8a9fadbad5f00e22e58e041d"}, + {file = "html5lib-1.1.tar.gz", hash = "sha256:b2e5b40261e20f354d198eae92afc10d750afb487ed5e50f9c4eaf07c184146f"}, +] + +[package.dependencies] +six = ">=1.9" +webencodings = "*" + +[package.extras] +all = ["chardet (>=2.2)", "genshi", "lxml"] +chardet = ["chardet (>=2.2)"] +genshi = ["genshi"] +lxml = ["lxml"] + +[[package]] +name = "huggingface-hub" +version = "0.15.1" +description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub" +category = "main" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "huggingface_hub-0.15.1-py3-none-any.whl", hash = "sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445"}, + {file = "huggingface_hub-0.15.1.tar.gz", hash = "sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081"}, +] + +[package.dependencies] +filelock = "*" +fsspec = "*" +packaging = ">=20.9" +pyyaml = ">=5.1" +requests = "*" +tqdm = ">=4.42.1" +typing-extensions = ">=3.7.4.3" + +[package.extras] +all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"] +cli = ["InquirerPy (==0.3.4)"] +dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"] +fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"] +quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"] +tensorflow = ["graphviz", "pydot", "tensorflow"] +testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "gradio", "jedi", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "soundfile", "urllib3 (<2.0)"] +torch = ["torch"] +typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"] + +[[package]] +name = "idna" +version = "3.4" +description = "Internationalized Domain Names in Applications (IDNA)" +category = "main" +optional = false +python-versions = ">=3.5" +files = [ + {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"}, + {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"}, +] + +[[package]] +name = "iniconfig" +version = "2.0.0" +description = "brain-dead simple config-ini parsing" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "iniconfig-2.0.0-py3-none-any.whl", hash = "sha256:b6a85871a79d2e3b22d2d1b94ac2824226a63c6b741c88f7ae975f18b6778374"}, + {file = "iniconfig-2.0.0.tar.gz", hash = "sha256:2d91e135bf72d31a410b17c16da610a82cb55f6b0477d1a902134b24a455b8b3"}, +] + +[[package]] +name = "isort" +version = "5.12.0" +description = "A Python utility / library to sort Python imports." +category = "dev" +optional = false +python-versions = ">=3.8.0" +files = [ + {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"}, + {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"}, +] + +[package.extras] +colors = ["colorama (>=0.4.3)"] +pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"] +plugins = ["setuptools"] +requirements-deprecated-finder = ["pip-api", "pipreqs"] + +[[package]] +name = "joblib" +version = "1.3.1" +description = "Lightweight pipelining with Python functions" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "joblib-1.3.1-py3-none-any.whl", hash = "sha256:89cf0529520e01b3de7ac7b74a8102c90d16d54c64b5dd98cafcd14307fdf915"}, + {file = "joblib-1.3.1.tar.gz", hash = "sha256:1f937906df65329ba98013dc9692fe22a4c5e4a648112de500508b18a21b41e3"}, +] + +[[package]] +name = "lazy-loader" +version = "0.3" +description = "lazy_loader" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "lazy_loader-0.3-py3-none-any.whl", hash = "sha256:1e9e76ee8631e264c62ce10006718e80b2cfc74340d17d1031e0f84af7478554"}, + {file = "lazy_loader-0.3.tar.gz", hash = "sha256:3b68898e34f5b2a29daaaac172c6555512d0f32074f147e2254e4a6d9d838f37"}, +] + +[package.extras] +lint = ["pre-commit (>=3.3)"] +test = ["pytest (>=7.4)", "pytest-cov (>=4.1)"] + +[[package]] +name = "libcommon" +version = "0.6.8" +description = "Library for utils common to all the services" +category = "main" +optional = false +python-versions = "3.9.15" +files = [] +develop = true + +[package.dependencies] +appdirs = "^1.4.4" +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} +environs = "^9.5.0" +huggingface-hub = "^0.15.1" +mongo-types = "0.15.1" +mongoengine = "^0.24.2" +networkx = "^3.0" +numba = "0.56.4" +orjson = "^3.8.6" +pandas = "^2.0.1" +psutil = "^5.9.4" +pydub = "^0.25.1" +pymongo = {version = "^3.13.0", extras = ["srv"]} +pytz = "^2020.1" +requests = "^2.31.0" +soundfile = ">=0.12.1" +starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" + +[package.source] +type = "directory" +url = "../libcommon" + +[[package]] +name = "librosa" +version = "0.10.0.post2" +description = "Python module for audio and music processing" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"}, + {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"}, +] + +[package.dependencies] +audioread = ">=2.1.9" +decorator = ">=4.3.0" +joblib = ">=0.14" +lazy-loader = ">=0.1" +msgpack = ">=1.0" +numba = ">=0.51.0" +numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2" +pooch = ">=1.0,<1.7" +scikit-learn = ">=0.20.0" +scipy = ">=1.2.0" +soundfile = ">=0.12.1" +soxr = ">=0.3.2" +typing-extensions = ">=4.1.1" + +[package.extras] +display = ["matplotlib (>=3.3.0)"] +docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"] +tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"] + +[[package]] +name = "llvmlite" +version = "0.39.1" +description = "lightweight wrapper around basic LLVM functionality" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"}, + {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"}, + {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"}, + {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"}, + {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"}, + {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"}, + {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"}, + {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"}, + {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"}, + {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"}, + {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"}, + {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"}, + {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"}, + {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"}, + {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"}, + {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"}, + {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"}, +] + +[[package]] +name = "markdown-it-py" +version = "3.0.0" +description = "Python port of markdown-it. Markdown parsing, done right!" +category = "dev" +optional = false +python-versions = ">=3.8" +files = [ + {file = "markdown-it-py-3.0.0.tar.gz", hash = "sha256:e3f60a94fa066dc52ec76661e37c851cb232d92f9886b15cb560aaada2df8feb"}, + {file = "markdown_it_py-3.0.0-py3-none-any.whl", hash = "sha256:355216845c60bd96232cd8d8c40e8f9765cc86f46880e43a8fd22dc1a1a8cab1"}, +] + +[package.dependencies] +mdurl = ">=0.1,<1.0" + +[package.extras] +benchmarking = ["psutil", "pytest", "pytest-benchmark"] +code-style = ["pre-commit (>=3.0,<4.0)"] +compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"] +linkify = ["linkify-it-py (>=1,<3)"] +plugins = ["mdit-py-plugins"] +profiling = ["gprof2dot"] +rtd = ["jupyter_sphinx", "mdit-py-plugins", "myst-parser", "pyyaml", "sphinx", "sphinx-copybutton", "sphinx-design", "sphinx_book_theme"] +testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"] + +[[package]] +name = "markupsafe" +version = "2.1.3" +description = "Safely add untrusted strings to HTML/XML markup." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cd0f502fe016460680cd20aaa5a76d241d6f35a1c3350c474bac1273803893fa"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e09031c87a1e51556fdcb46e5bd4f59dfb743061cf93c4d6831bf894f125eb57"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:68e78619a61ecf91e76aa3e6e8e33fc4894a2bebe93410754bd28fce0a8a4f9f"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65c1a9bcdadc6c28eecee2c119465aebff8f7a584dd719facdd9e825ec61ab52"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:525808b8019e36eb524b8c68acdd63a37e75714eac50e988180b169d64480a00"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:962f82a3086483f5e5f64dbad880d31038b698494799b097bc59c2edf392fce6"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:aa7bd130efab1c280bed0f45501b7c8795f9fdbeb02e965371bbef3523627779"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c9c804664ebe8f83a211cace637506669e7890fec1b4195b505c214e50dd4eb7"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-win32.whl", hash = "sha256:10bbfe99883db80bdbaff2dcf681dfc6533a614f700da1287707e8a5d78a8431"}, + {file = "MarkupSafe-2.1.3-cp310-cp310-win_amd64.whl", hash = "sha256:1577735524cdad32f9f694208aa75e422adba74f1baee7551620e43a3141f559"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:ad9e82fb8f09ade1c3e1b996a6337afac2b8b9e365f926f5a61aacc71adc5b3c"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3c0fae6c3be832a0a0473ac912810b2877c8cb9d76ca48de1ed31e1c68386575"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b076b6226fb84157e3f7c971a47ff3a679d837cf338547532ab866c57930dbee"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bfce63a9e7834b12b87c64d6b155fdd9b3b96191b6bd334bf37db7ff1fe457f2"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:338ae27d6b8745585f87218a3f23f1512dbf52c26c28e322dbe54bcede54ccb9"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:e4dd52d80b8c83fdce44e12478ad2e85c64ea965e75d66dbeafb0a3e77308fcc"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:df0be2b576a7abbf737b1575f048c23fb1d769f267ec4358296f31c2479db8f9"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:5bbe06f8eeafd38e5d0a4894ffec89378b6c6a625ff57e3028921f8ff59318ac"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-win32.whl", hash = "sha256:dd15ff04ffd7e05ffcb7fe79f1b98041b8ea30ae9234aed2a9168b5797c3effb"}, + {file = "MarkupSafe-2.1.3-cp311-cp311-win_amd64.whl", hash = "sha256:134da1eca9ec0ae528110ccc9e48041e0828d79f24121a1a146161103c76e686"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8e254ae696c88d98da6555f5ace2279cf7cd5b3f52be2b5cf97feafe883b58d2"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cb0932dc158471523c9637e807d9bfb93e06a95cbf010f1a38b98623b929ef2b"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9402b03f1a1b4dc4c19845e5c749e3ab82d5078d16a2a4c2cd2df62d57bb0707"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca379055a47383d02a5400cb0d110cef0a776fc644cda797db0c5696cfd7e18e"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7ff0f54cb4ff66dd38bebd335a38e2c22c41a8ee45aa608efc890ac3e3931bc"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c011a4149cfbcf9f03994ec2edffcb8b1dc2d2aede7ca243746df97a5d41ce48"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:56d9f2ecac662ca1611d183feb03a3fa4406469dafe241673d521dd5ae92a155"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-win32.whl", hash = "sha256:8758846a7e80910096950b67071243da3e5a20ed2546e6392603c096778d48e0"}, + {file = "MarkupSafe-2.1.3-cp37-cp37m-win_amd64.whl", hash = "sha256:787003c0ddb00500e49a10f2844fac87aa6ce977b90b0feaaf9de23c22508b24"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:2ef12179d3a291be237280175b542c07a36e7f60718296278d8593d21ca937d4"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:2c1b19b3aaacc6e57b7e25710ff571c24d6c3613a45e905b1fde04d691b98ee0"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8afafd99945ead6e075b973fefa56379c5b5c53fd8937dad92c662da5d8fd5ee"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8c41976a29d078bb235fea9b2ecd3da465df42a562910f9022f1a03107bd02be"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d080e0a5eb2529460b30190fcfcc4199bd7f827663f858a226a81bc27beaa97e"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:69c0f17e9f5a7afdf2cc9fb2d1ce6aabdb3bafb7f38017c0b77862bcec2bbad8"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:504b320cd4b7eff6f968eddf81127112db685e81f7e36e75f9f84f0df46041c3"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:42de32b22b6b804f42c5d98be4f7e5e977ecdd9ee9b660fda1a3edf03b11792d"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-win32.whl", hash = "sha256:ceb01949af7121f9fc39f7d27f91be8546f3fb112c608bc4029aef0bab86a2a5"}, + {file = "MarkupSafe-2.1.3-cp38-cp38-win_amd64.whl", hash = "sha256:1b40069d487e7edb2676d3fbdb2b0829ffa2cd63a2ec26c4938b2d34391b4ecc"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:8023faf4e01efadfa183e863fefde0046de576c6f14659e8782065bcece22198"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6b2b56950d93e41f33b4223ead100ea0fe11f8e6ee5f641eb753ce4b77a7042b"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9dcdfd0eaf283af041973bff14a2e143b8bd64e069f4c383416ecd79a81aab58"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:05fb21170423db021895e1ea1e1f3ab3adb85d1c2333cbc2310f2a26bc77272e"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:282c2cb35b5b673bbcadb33a585408104df04f14b2d9b01d4c345a3b92861c2c"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ab4a0df41e7c16a1392727727e7998a467472d0ad65f3ad5e6e765015df08636"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:7ef3cb2ebbf91e330e3bb937efada0edd9003683db6b57bb108c4001f37a02ea"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:0a4e4a1aff6c7ac4cd55792abf96c915634c2b97e3cc1c7129578aa68ebd754e"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-win32.whl", hash = "sha256:fec21693218efe39aa7f8599346e90c705afa52c5b31ae019b2e57e8f6542bb2"}, + {file = "MarkupSafe-2.1.3-cp39-cp39-win_amd64.whl", hash = "sha256:3fd4abcb888d15a94f32b75d8fd18ee162ca0c064f35b11134be77050296d6ba"}, + {file = "MarkupSafe-2.1.3.tar.gz", hash = "sha256:af598ed32d6ae86f1b747b82783958b1a4ab8f617b06fe68795c7f026abbdcad"}, +] + +[[package]] +name = "marshmallow" +version = "3.19.0" +description = "A lightweight library for converting complex datatypes to and from native Python datatypes." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "marshmallow-3.19.0-py3-none-any.whl", hash = "sha256:93f0958568da045b0021ec6aeb7ac37c81bfcccbb9a0e7ed8559885070b3a19b"}, + {file = "marshmallow-3.19.0.tar.gz", hash = "sha256:90032c0fd650ce94b6ec6dc8dfeb0e3ff50c144586462c389b81a07205bedb78"}, +] + +[package.dependencies] +packaging = ">=17.0" + +[package.extras] +dev = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)", "pytest", "pytz", "simplejson", "tox"] +docs = ["alabaster (==0.7.12)", "autodocsumm (==0.2.9)", "sphinx (==5.3.0)", "sphinx-issues (==3.0.1)", "sphinx-version-warning (==1.1.2)"] +lint = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)"] +tests = ["pytest", "pytz", "simplejson"] + +[[package]] +name = "mccabe" +version = "0.6.1" +description = "McCabe checker, plugin for flake8" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "mccabe-0.6.1-py2.py3-none-any.whl", hash = "sha256:ab8a6258860da4b6677da4bd2fe5dc2c659cff31b3ee4f7f5d64e79735b80d42"}, + {file = "mccabe-0.6.1.tar.gz", hash = "sha256:dd8d182285a0fe56bace7f45b5e7d1a6ebcbf524e8f3bd87eb0f125271b8831f"}, +] + +[[package]] +name = "mdurl" +version = "0.1.2" +description = "Markdown URL utilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8"}, + {file = "mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba"}, +] + +[[package]] +name = "mongo-types" +version = "0.15.1" +description = "Type stubs for mongoengine w/ basic support for bson and pymongo" +category = "main" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "mongo-types-0.15.1.tar.gz", hash = "sha256:0a9deeb7733ea7da5db3711d92e22d93556b522f860bbff82e5df44c53bd06a9"}, + {file = "mongo_types-0.15.1-py3-none-any.whl", hash = "sha256:9417ae5b9a759c09630b5ec7d66904cc333c2d2fcfe75e2760a332ed5e267309"}, +] + +[[package]] +name = "mongoengine" +version = "0.24.2" +description = "MongoEngine is a Python Object-Document Mapper for working with MongoDB." +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "mongoengine-0.24.2-py3-none-any.whl", hash = "sha256:f5c4e1b206b2ccffe4adc7a6283ed26dd799bd115a5fb1d2e885a075132cdb88"}, + {file = "mongoengine-0.24.2.tar.gz", hash = "sha256:c76d49658575bb995682e2e77c8ef7cda63faf939415b32ee923745d120f8b02"}, +] + +[package.dependencies] +pymongo = ">=3.4,<5.0" + +[[package]] +name = "msgpack" +version = "1.0.5" +description = "MessagePack serializer" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:525228efd79bb831cf6830a732e2e80bc1b05436b086d4264814b4b2955b2fa9"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4f8d8b3bf1ff2672567d6b5c725a1b347fe838b912772aa8ae2bf70338d5a198"}, + {file = "msgpack-1.0.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cdc793c50be3f01106245a61b739328f7dccc2c648b501e237f0699fe1395b81"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb47c21a8a65b165ce29f2bec852790cbc04936f502966768e4aae9fa763cb7"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e42b9594cc3bf4d838d67d6ed62b9e59e201862a25e9a157019e171fbe672dd3"}, + {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:55b56a24893105dc52c1253649b60f475f36b3aa0fc66115bffafb624d7cb30b"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:1967f6129fc50a43bfe0951c35acbb729be89a55d849fab7686004da85103f1c"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:20a97bf595a232c3ee6d57ddaadd5453d174a52594bf9c21d10407e2a2d9b3bd"}, + {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d25dd59bbbbb996eacf7be6b4ad082ed7eacc4e8f3d2df1ba43822da9bfa122a"}, + {file = "msgpack-1.0.5-cp310-cp310-win32.whl", hash = "sha256:382b2c77589331f2cb80b67cc058c00f225e19827dbc818d700f61513ab47bea"}, + {file = "msgpack-1.0.5-cp310-cp310-win_amd64.whl", hash = "sha256:4867aa2df9e2a5fa5f76d7d5565d25ec76e84c106b55509e78c1ede0f152659a"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9f5ae84c5c8a857ec44dc180a8b0cc08238e021f57abdf51a8182e915e6299f0"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:9e6ca5d5699bcd89ae605c150aee83b5321f2115695e741b99618f4856c50898"}, + {file = "msgpack-1.0.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5494ea30d517a3576749cad32fa27f7585c65f5f38309c88c6d137877fa28a5a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ab2f3331cb1b54165976a9d976cb251a83183631c88076613c6c780f0d6e45a"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:28592e20bbb1620848256ebc105fc420436af59515793ed27d5c77a217477705"}, + {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe5c63197c55bce6385d9aee16c4d0641684628f63ace85f73571e65ad1c1e8d"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:ed40e926fa2f297e8a653c954b732f125ef97bdd4c889f243182299de27e2aa9"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b2de4c1c0538dcb7010902a2b97f4e00fc4ddf2c8cda9749af0e594d3b7fa3d7"}, + {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf22a83f973b50f9d38e55c6aade04c41ddda19b00c4ebc558930d78eecc64ed"}, + {file = "msgpack-1.0.5-cp311-cp311-win32.whl", hash = "sha256:c396e2cc213d12ce017b686e0f53497f94f8ba2b24799c25d913d46c08ec422c"}, + {file = "msgpack-1.0.5-cp311-cp311-win_amd64.whl", hash = "sha256:6c4c68d87497f66f96d50142a2b73b97972130d93677ce930718f68828b382e2"}, + {file = "msgpack-1.0.5-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:a2b031c2e9b9af485d5e3c4520f4220d74f4d222a5b8dc8c1a3ab9448ca79c57"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f837b93669ce4336e24d08286c38761132bc7ab29782727f8557e1eb21b2080"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1d46dfe3832660f53b13b925d4e0fa1432b00f5f7210eb3ad3bb9a13c6204a6"}, + {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:366c9a7b9057e1547f4ad51d8facad8b406bab69c7d72c0eb6f529cf76d4b85f"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4c075728a1095efd0634a7dccb06204919a2f67d1893b6aa8e00497258bf926c"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:f933bbda5a3ee63b8834179096923b094b76f0c7a73c1cfe8f07ad608c58844b"}, + {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:36961b0568c36027c76e2ae3ca1132e35123dcec0706c4b7992683cc26c1320c"}, + {file = "msgpack-1.0.5-cp36-cp36m-win32.whl", hash = "sha256:b5ef2f015b95f912c2fcab19c36814963b5463f1fb9049846994b007962743e9"}, + {file = "msgpack-1.0.5-cp36-cp36m-win_amd64.whl", hash = "sha256:288e32b47e67f7b171f86b030e527e302c91bd3f40fd9033483f2cacc37f327a"}, + {file = "msgpack-1.0.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:137850656634abddfb88236008339fdaba3178f4751b28f270d2ebe77a563b6c"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c05a4a96585525916b109bb85f8cb6511db1c6f5b9d9cbcbc940dc6b4be944b"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:56a62ec00b636583e5cb6ad313bbed36bb7ead5fa3a3e38938503142c72cba4f"}, + {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef8108f8dedf204bb7b42994abf93882da1159728a2d4c5e82012edd92c9da9f"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1835c84d65f46900920b3708f5ba829fb19b1096c1800ad60bae8418652a951d"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:e57916ef1bd0fee4f21c4600e9d1da352d8816b52a599c46460e93a6e9f17086"}, + {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:17358523b85973e5f242ad74aa4712b7ee560715562554aa2134d96e7aa4cbbf"}, + {file = "msgpack-1.0.5-cp37-cp37m-win32.whl", hash = "sha256:cb5aaa8c17760909ec6cb15e744c3ebc2ca8918e727216e79607b7bbce9c8f77"}, + {file = "msgpack-1.0.5-cp37-cp37m-win_amd64.whl", hash = "sha256:ab31e908d8424d55601ad7075e471b7d0140d4d3dd3272daf39c5c19d936bd82"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:b72d0698f86e8d9ddf9442bdedec15b71df3598199ba33322d9711a19f08145c"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:379026812e49258016dd84ad79ac8446922234d498058ae1d415f04b522d5b2d"}, + {file = "msgpack-1.0.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:332360ff25469c346a1c5e47cbe2a725517919892eda5cfaffe6046656f0b7bb"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:476a8fe8fae289fdf273d6d2a6cb6e35b5a58541693e8f9f019bfe990a51e4ba"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a9985b214f33311df47e274eb788a5893a761d025e2b92c723ba4c63936b69b1"}, + {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:48296af57cdb1d885843afd73c4656be5c76c0c6328db3440c9601a98f303d87"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:addab7e2e1fcc04bd08e4eb631c2a90960c340e40dfc4a5e24d2ff0d5a3b3edb"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:916723458c25dfb77ff07f4c66aed34e47503b2eb3188b3adbec8d8aa6e00f48"}, + {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:821c7e677cc6acf0fd3f7ac664c98803827ae6de594a9f99563e48c5a2f27eb0"}, + {file = "msgpack-1.0.5-cp38-cp38-win32.whl", hash = "sha256:1c0f7c47f0087ffda62961d425e4407961a7ffd2aa004c81b9c07d9269512f6e"}, + {file = "msgpack-1.0.5-cp38-cp38-win_amd64.whl", hash = "sha256:bae7de2026cbfe3782c8b78b0db9cbfc5455e079f1937cb0ab8d133496ac55e1"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:20c784e66b613c7f16f632e7b5e8a1651aa5702463d61394671ba07b2fc9e025"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:266fa4202c0eb94d26822d9bfd7af25d1e2c088927fe8de9033d929dd5ba24c5"}, + {file = "msgpack-1.0.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:18334484eafc2b1aa47a6d42427da7fa8f2ab3d60b674120bce7a895a0a85bdd"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57e1f3528bd95cc44684beda696f74d3aaa8a5e58c816214b9046512240ef437"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:586d0d636f9a628ddc6a17bfd45aa5b5efaf1606d2b60fa5d87b8986326e933f"}, + {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a740fa0e4087a734455f0fc3abf5e746004c9da72fbd541e9b113013c8dc3282"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3055b0455e45810820db1f29d900bf39466df96ddca11dfa6d074fa47054376d"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:a61215eac016f391129a013c9e46f3ab308db5f5ec9f25811e811f96962599a8"}, + {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:362d9655cd369b08fda06b6657a303eb7172d5279997abe094512e919cf74b11"}, + {file = "msgpack-1.0.5-cp39-cp39-win32.whl", hash = "sha256:ac9dd47af78cae935901a9a500104e2dea2e253207c924cc95de149606dc43cc"}, + {file = "msgpack-1.0.5-cp39-cp39-win_amd64.whl", hash = "sha256:06f5174b5f8ed0ed919da0e62cbd4ffde676a374aba4020034da05fab67b9164"}, + {file = "msgpack-1.0.5.tar.gz", hash = "sha256:c075544284eadc5cddc70f4757331d99dcbc16b2bbd4849d15f8aae4cf36d31c"}, +] + +[[package]] +name = "multidict" +version = "6.0.4" +description = "multidict implementation" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"}, + {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"}, + {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"}, + {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"}, + {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"}, + {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"}, + {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"}, + {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"}, + {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"}, + {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"}, + {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"}, + {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"}, + {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"}, + {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"}, + {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"}, + {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"}, + {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"}, + {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"}, + {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"}, + {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"}, + {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"}, + {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"}, + {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"}, + {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"}, + {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"}, + {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"}, + {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"}, +] + +[[package]] +name = "multiprocess" +version = "0.70.14" +description = "better multiprocessing and multithreading in python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"}, + {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"}, + {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"}, + {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"}, + {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"}, + {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"}, + {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"}, + {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"}, + {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"}, +] + +[package.dependencies] +dill = ">=0.3.6" + +[[package]] +name = "mypy" +version = "1.4.1" +description = "Optional static typing for Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "mypy-1.4.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:566e72b0cd6598503e48ea610e0052d1b8168e60a46e0bfd34b3acf2d57f96a8"}, + {file = "mypy-1.4.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ca637024ca67ab24a7fd6f65d280572c3794665eaf5edcc7e90a866544076878"}, + {file = "mypy-1.4.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0dde1d180cd84f0624c5dcaaa89c89775550a675aff96b5848de78fb11adabcd"}, + {file = "mypy-1.4.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:8c4d8e89aa7de683e2056a581ce63c46a0c41e31bd2b6d34144e2c80f5ea53dc"}, + {file = "mypy-1.4.1-cp310-cp310-win_amd64.whl", hash = "sha256:bfdca17c36ae01a21274a3c387a63aa1aafe72bff976522886869ef131b937f1"}, + {file = "mypy-1.4.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7549fbf655e5825d787bbc9ecf6028731973f78088fbca3a1f4145c39ef09462"}, + {file = "mypy-1.4.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:98324ec3ecf12296e6422939e54763faedbfcc502ea4a4c38502082711867258"}, + {file = "mypy-1.4.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:141dedfdbfe8a04142881ff30ce6e6653c9685b354876b12e4fe6c78598b45e2"}, + {file = "mypy-1.4.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8207b7105829eca6f3d774f64a904190bb2231de91b8b186d21ffd98005f14a7"}, + {file = "mypy-1.4.1-cp311-cp311-win_amd64.whl", hash = "sha256:16f0db5b641ba159eff72cff08edc3875f2b62b2fa2bc24f68c1e7a4e8232d01"}, + {file = "mypy-1.4.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:470c969bb3f9a9efcedbadcd19a74ffb34a25f8e6b0e02dae7c0e71f8372f97b"}, + {file = "mypy-1.4.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5952d2d18b79f7dc25e62e014fe5a23eb1a3d2bc66318df8988a01b1a037c5b"}, + {file = "mypy-1.4.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:190b6bab0302cec4e9e6767d3eb66085aef2a1cc98fe04936d8a42ed2ba77bb7"}, + {file = "mypy-1.4.1-cp37-cp37m-win_amd64.whl", hash = "sha256:9d40652cc4fe33871ad3338581dca3297ff5f2213d0df345bcfbde5162abf0c9"}, + {file = "mypy-1.4.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:01fd2e9f85622d981fd9063bfaef1aed6e336eaacca00892cd2d82801ab7c042"}, + {file = "mypy-1.4.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2460a58faeea905aeb1b9b36f5065f2dc9a9c6e4c992a6499a2360c6c74ceca3"}, + {file = "mypy-1.4.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a2746d69a8196698146a3dbe29104f9eb6a2a4d8a27878d92169a6c0b74435b6"}, + {file = "mypy-1.4.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:ae704dcfaa180ff7c4cfbad23e74321a2b774f92ca77fd94ce1049175a21c97f"}, + {file = "mypy-1.4.1-cp38-cp38-win_amd64.whl", hash = "sha256:43d24f6437925ce50139a310a64b2ab048cb2d3694c84c71c3f2a1626d8101dc"}, + {file = "mypy-1.4.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c482e1246726616088532b5e964e39765b6d1520791348e6c9dc3af25b233828"}, + {file = "mypy-1.4.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:43b592511672017f5b1a483527fd2684347fdffc041c9ef53428c8dc530f79a3"}, + {file = "mypy-1.4.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:34a9239d5b3502c17f07fd7c0b2ae6b7dd7d7f6af35fbb5072c6208e76295816"}, + {file = "mypy-1.4.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5703097c4936bbb9e9bce41478c8d08edd2865e177dc4c52be759f81ee4dd26c"}, + {file = "mypy-1.4.1-cp39-cp39-win_amd64.whl", hash = "sha256:e02d700ec8d9b1859790c0475df4e4092c7bf3272a4fd2c9f33d87fac4427b8f"}, + {file = "mypy-1.4.1-py3-none-any.whl", hash = "sha256:45d32cec14e7b97af848bddd97d85ea4f0db4d5a149ed9676caa4eb2f7402bb4"}, + {file = "mypy-1.4.1.tar.gz", hash = "sha256:9bbcd9ab8ea1f2e1c8031c21445b511442cc45c89951e49bbf852cbb70755b1b"}, +] + +[package.dependencies] +mypy-extensions = ">=1.0.0" +tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""} +typing-extensions = ">=4.1.0" + +[package.extras] +dmypy = ["psutil (>=4.0)"] +install-types = ["pip"] +python2 = ["typed-ast (>=1.4.0,<2)"] +reports = ["lxml"] + +[[package]] +name = "mypy-extensions" +version = "1.0.0" +description = "Type system extensions for programs checked with the mypy type checker." +category = "dev" +optional = false +python-versions = ">=3.5" +files = [ + {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"}, + {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"}, +] + +[[package]] +name = "networkx" +version = "3.1" +description = "Python package for creating and manipulating graphs and networks" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"}, + {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"}, +] + +[package.extras] +default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"] +developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"] +doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"] +extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"] +test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"] + +[[package]] +name = "numba" +version = "0.56.4" +description = "compiling Python code using LLVM" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"}, + {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"}, + {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"}, + {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"}, + {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"}, + {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"}, + {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"}, + {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"}, + {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"}, + {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"}, + {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"}, + {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"}, + {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"}, + {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"}, + {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"}, + {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"}, + {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"}, + {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"}, + {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"}, + {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"}, +] + +[package.dependencies] +llvmlite = ">=0.39.0dev0,<0.40" +numpy = ">=1.18,<1.24" +setuptools = "*" + +[[package]] +name = "numpy" +version = "1.23.5" +description = "NumPy is the fundamental package for array computing with Python." +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"}, + {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"}, + {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"}, + {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"}, + {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"}, + {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"}, + {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"}, + {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"}, + {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"}, + {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"}, + {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"}, + {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"}, + {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"}, + {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"}, + {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"}, + {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"}, + {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"}, + {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"}, + {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"}, +] + +[[package]] +name = "orjson" +version = "3.9.1" +description = "Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "orjson-3.9.1-cp310-cp310-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:c4434b7b786fdc394b95d029fb99949d7c2b05bbd4bf5cb5e3906be96ffeee3b"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09faf14f74ed47e773fa56833be118e04aa534956f661eb491522970b7478e3b"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:503eb86a8d53a187fe66aa80c69295a3ca35475804da89a9547e4fce5f803822"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:20f2804b5a1dbd3609c086041bd243519224d47716efd7429db6c03ed28b7cc3"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0fd828e0656615a711c4cc4da70f3cac142e66a6703ba876c20156a14e28e3fa"}, + {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ec53d648176f873203b9c700a0abacab33ca1ab595066e9d616f98cdc56f4434"}, + {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e186ae76b0d97c505500664193ddf508c13c1e675d9b25f1f4414a7606100da6"}, + {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d4edee78503016f4df30aeede0d999b3cb11fb56f47e9db0e487bce0aaca9285"}, + {file = "orjson-3.9.1-cp310-none-win_amd64.whl", hash = "sha256:a4cc5d21e68af982d9a2528ac61e604f092c60eed27aef3324969c68f182ec7e"}, + {file = "orjson-3.9.1-cp311-cp311-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:761b6efd33c49de20dd73ce64cc59da62c0dab10aa6015f582680e0663cc792c"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:31229f9d0b8dc2ef7ee7e4393f2e4433a28e16582d4b25afbfccc9d68dc768f8"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b7ab18d55ecb1de543d452f0a5f8094b52282b916aa4097ac11a4c79f317b86"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:db774344c39041f4801c7dfe03483df9203cbd6c84e601a65908e5552228dd25"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ae47ef8c0fe89c4677db7e9e1fb2093ca6e66c3acbee5442d84d74e727edad5e"}, + {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:103952c21575b9805803c98add2eaecd005580a1e746292ed2ec0d76dd3b9746"}, + {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:2cb0121e6f2c9da3eddf049b99b95fef0adf8480ea7cb544ce858706cdf916eb"}, + {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:24d4ddaa2876e657c0fd32902b5c451fd2afc35159d66a58da7837357044b8c2"}, + {file = "orjson-3.9.1-cp311-none-win_amd64.whl", hash = "sha256:0b53b5f72cf536dd8aa4fc4c95e7e09a7adb119f8ff8ee6cc60f735d7740ad6a"}, + {file = "orjson-3.9.1-cp37-cp37m-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d4b68d01a506242316a07f1d2f29fb0a8b36cee30a7c35076f1ef59dce0890c1"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d9dd4abe6c6fd352f00f4246d85228f6a9847d0cc14f4d54ee553718c225388f"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9e20bca5e13041e31ceba7a09bf142e6d63c8a7467f5a9c974f8c13377c75af2"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d8ae0467d01eb1e4bcffef4486d964bfd1c2e608103e75f7074ed34be5df48cc"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:06f6ab4697fab090517f295915318763a97a12ee8186054adf21c1e6f6abbd3d"}, + {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8515867713301fa065c58ec4c9053ba1a22c35113ab4acad555317b8fd802e50"}, + {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:393d0697d1dfa18d27d193e980c04fdfb672c87f7765b87952f550521e21b627"}, + {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d96747662d3666f79119e5d28c124e7d356c7dc195cd4b09faea4031c9079dc9"}, + {file = "orjson-3.9.1-cp37-none-win_amd64.whl", hash = "sha256:6d173d3921dd58a068c88ec22baea7dbc87a137411501618b1292a9d6252318e"}, + {file = "orjson-3.9.1-cp38-cp38-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d1c2b0b4246c992ce2529fc610a446b945f1429445ece1c1f826a234c829a918"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:19f70ba1f441e1c4bb1a581f0baa092e8b3e3ce5b2aac2e1e090f0ac097966da"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:375d65f002e686212aac42680aed044872c45ee4bc656cf63d4a215137a6124a"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4751cee4a7b1daeacb90a7f5adf2170ccab893c3ab7c5cea58b45a13f89b30b3"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78d9a2a4b2302d5ebc3695498ebc305c3568e5ad4f3501eb30a6405a32d8af22"}, + {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:46b4facc32643b2689dfc292c0c463985dac4b6ab504799cf51fc3c6959ed668"}, + {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:ec7c8a0f1bf35da0d5fd14f8956f3b82a9a6918a3c6963d718dfd414d6d3b604"}, + {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:d3a40b0fbe06ccd4d6a99e523d20b47985655bcada8d1eba485b1b32a43e4904"}, + {file = "orjson-3.9.1-cp38-none-win_amd64.whl", hash = "sha256:402f9d3edfec4560a98880224ec10eba4c5f7b4791e4bc0d4f4d8df5faf2a006"}, + {file = "orjson-3.9.1-cp39-cp39-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:49c0d78dcd34626e2e934f1192d7c052b94e0ecadc5f386fd2bda6d2e03dadf5"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:125f63e56d38393daa0a1a6dc6fedefca16c538614b66ea5997c3bd3af35ef26"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:08927970365d2e1f3ce4894f9ff928a7b865d53f26768f1bbdd85dd4fee3e966"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f9a744e212d4780ecd67f4b6b128b2e727bee1df03e7059cddb2dfe1083e7dc4"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5d1dbf36db7240c61eec98c8d21545d671bce70be0730deb2c0d772e06b71af3"}, + {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:80a1e384626f76b66df615f7bb622a79a25c166d08c5d2151ffd41f24c4cc104"}, + {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:15d28872fb055bf17ffca913826e618af61b2f689d2b170f72ecae1a86f80d52"}, + {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:1e4d905338f9ef32c67566929dfbfbb23cc80287af8a2c38930fb0eda3d40b76"}, + {file = "orjson-3.9.1-cp39-none-win_amd64.whl", hash = "sha256:48a27da6c7306965846565cc385611d03382bbd84120008653aa2f6741e2105d"}, + {file = "orjson-3.9.1.tar.gz", hash = "sha256:db373a25ec4a4fccf8186f9a72a1b3442837e40807a736a815ab42481e83b7d0"}, +] + +[[package]] +name = "packageurl-python" +version = "0.11.1" +description = "A purl aka. Package URL parser and builder" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "packageurl-python-0.11.1.tar.gz", hash = "sha256:bbcc53d2cb5920c815c1626c75992f319bfc450b73893fa7bd8aac5869aa49fe"}, + {file = "packageurl_python-0.11.1-py3-none-any.whl", hash = "sha256:4bad1d3ea4feb5e7a1db5ca8fb690ac9c82ab18e08d500755947b853df68817d"}, +] + +[package.extras] +build = ["wheel"] +lint = ["black", "isort", "mypy"] +test = ["pytest"] + +[[package]] +name = "packaging" +version = "23.1" +description = "Core utilities for Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "packaging-23.1-py3-none-any.whl", hash = "sha256:994793af429502c4ea2ebf6bf664629d07c1a9fe974af92966e4b8d2df7edc61"}, + {file = "packaging-23.1.tar.gz", hash = "sha256:a392980d2b6cffa644431898be54b0045151319d1e7ec34f0cfed48767dd334f"}, +] + +[[package]] +name = "pandas" +version = "2.0.3" +description = "Powerful data structures for data analysis, time series, and statistics" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "pandas-2.0.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e4c7c9f27a4185304c7caf96dc7d91bc60bc162221152de697c98eb0b2648dd8"}, + {file = "pandas-2.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f167beed68918d62bffb6ec64f2e1d8a7d297a038f86d4aed056b9493fca407f"}, + {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ce0c6f76a0f1ba361551f3e6dceaff06bde7514a374aa43e33b588ec10420183"}, + {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ba619e410a21d8c387a1ea6e8a0e49bb42216474436245718d7f2e88a2f8d7c0"}, + {file = "pandas-2.0.3-cp310-cp310-win32.whl", hash = "sha256:3ef285093b4fe5058eefd756100a367f27029913760773c8bf1d2d8bebe5d210"}, + {file = "pandas-2.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:9ee1a69328d5c36c98d8e74db06f4ad518a1840e8ccb94a4ba86920986bb617e"}, + {file = "pandas-2.0.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b084b91d8d66ab19f5bb3256cbd5ea661848338301940e17f4492b2ce0801fe8"}, + {file = "pandas-2.0.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:37673e3bdf1551b95bf5d4ce372b37770f9529743d2498032439371fc7b7eb26"}, + {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b9cb1e14fdb546396b7e1b923ffaeeac24e4cedd14266c3497216dd4448e4f2d"}, + {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9cd88488cceb7635aebb84809d087468eb33551097d600c6dad13602029c2df"}, + {file = "pandas-2.0.3-cp311-cp311-win32.whl", hash = "sha256:694888a81198786f0e164ee3a581df7d505024fbb1f15202fc7db88a71d84ebd"}, + {file = "pandas-2.0.3-cp311-cp311-win_amd64.whl", hash = "sha256:6a21ab5c89dcbd57f78d0ae16630b090eec626360085a4148693def5452d8a6b"}, + {file = "pandas-2.0.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9e4da0d45e7f34c069fe4d522359df7d23badf83abc1d1cef398895822d11061"}, + {file = "pandas-2.0.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:32fca2ee1b0d93dd71d979726b12b61faa06aeb93cf77468776287f41ff8fdc5"}, + {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:258d3624b3ae734490e4d63c430256e716f488c4fcb7c8e9bde2d3aa46c29089"}, + {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eae3dc34fa1aa7772dd3fc60270d13ced7346fcbcfee017d3132ec625e23bb0"}, + {file = "pandas-2.0.3-cp38-cp38-win32.whl", hash = "sha256:f3421a7afb1a43f7e38e82e844e2bca9a6d793d66c1a7f9f0ff39a795bbc5e02"}, + {file = "pandas-2.0.3-cp38-cp38-win_amd64.whl", hash = "sha256:69d7f3884c95da3a31ef82b7618af5710dba95bb885ffab339aad925c3e8ce78"}, + {file = "pandas-2.0.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5247fb1ba347c1261cbbf0fcfba4a3121fbb4029d95d9ef4dc45406620b25c8b"}, + {file = "pandas-2.0.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:81af086f4543c9d8bb128328b5d32e9986e0c84d3ee673a2ac6fb57fd14f755e"}, + {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1994c789bf12a7c5098277fb43836ce090f1073858c10f9220998ac74f37c69b"}, + {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5ec591c48e29226bcbb316e0c1e9423622bc7a4eaf1ef7c3c9fa1a3981f89641"}, + {file = "pandas-2.0.3-cp39-cp39-win32.whl", hash = "sha256:04dbdbaf2e4d46ca8da896e1805bc04eb85caa9a82e259e8eed00254d5e0c682"}, + {file = "pandas-2.0.3-cp39-cp39-win_amd64.whl", hash = "sha256:1168574b036cd8b93abc746171c9b4f1b83467438a5e45909fed645cf8692dbc"}, + {file = "pandas-2.0.3.tar.gz", hash = "sha256:c02f372a88e0d17f36d3093a644c73cfc1788e876a7c4bcb4020a77512e2043c"}, +] + +[package.dependencies] +numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""} +python-dateutil = ">=2.8.2" +pytz = ">=2020.1" +tzdata = ">=2022.1" + +[package.extras] +all = ["PyQt5 (>=5.15.1)", "SQLAlchemy (>=1.4.16)", "beautifulsoup4 (>=4.9.3)", "bottleneck (>=1.3.2)", "brotlipy (>=0.7.0)", "fastparquet (>=0.6.3)", "fsspec (>=2021.07.0)", "gcsfs (>=2021.07.0)", "html5lib (>=1.1)", "hypothesis (>=6.34.2)", "jinja2 (>=3.0.0)", "lxml (>=4.6.3)", "matplotlib (>=3.6.1)", "numba (>=0.53.1)", "numexpr (>=2.7.3)", "odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pandas-gbq (>=0.15.0)", "psycopg2 (>=2.8.6)", "pyarrow (>=7.0.0)", "pymysql (>=1.0.2)", "pyreadstat (>=1.1.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)", "python-snappy (>=0.6.0)", "pyxlsb (>=1.0.8)", "qtpy (>=2.2.0)", "s3fs (>=2021.08.0)", "scipy (>=1.7.1)", "tables (>=3.6.1)", "tabulate (>=0.8.9)", "xarray (>=0.21.0)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)", "zstandard (>=0.15.2)"] +aws = ["s3fs (>=2021.08.0)"] +clipboard = ["PyQt5 (>=5.15.1)", "qtpy (>=2.2.0)"] +compression = ["brotlipy (>=0.7.0)", "python-snappy (>=0.6.0)", "zstandard (>=0.15.2)"] +computation = ["scipy (>=1.7.1)", "xarray (>=0.21.0)"] +excel = ["odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pyxlsb (>=1.0.8)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)"] +feather = ["pyarrow (>=7.0.0)"] +fss = ["fsspec (>=2021.07.0)"] +gcp = ["gcsfs (>=2021.07.0)", "pandas-gbq (>=0.15.0)"] +hdf5 = ["tables (>=3.6.1)"] +html = ["beautifulsoup4 (>=4.9.3)", "html5lib (>=1.1)", "lxml (>=4.6.3)"] +mysql = ["SQLAlchemy (>=1.4.16)", "pymysql (>=1.0.2)"] +output-formatting = ["jinja2 (>=3.0.0)", "tabulate (>=0.8.9)"] +parquet = ["pyarrow (>=7.0.0)"] +performance = ["bottleneck (>=1.3.2)", "numba (>=0.53.1)", "numexpr (>=2.7.1)"] +plot = ["matplotlib (>=3.6.1)"] +postgresql = ["SQLAlchemy (>=1.4.16)", "psycopg2 (>=2.8.6)"] +spss = ["pyreadstat (>=1.1.2)"] +sql-other = ["SQLAlchemy (>=1.4.16)"] +test = ["hypothesis (>=6.34.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)"] +xml = ["lxml (>=4.6.3)"] + +[[package]] +name = "pathspec" +version = "0.11.1" +description = "Utility library for gitignore style pattern matching of file paths." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pathspec-0.11.1-py3-none-any.whl", hash = "sha256:d8af70af76652554bd134c22b3e8a1cc46ed7d91edcdd721ef1a0c51a84a5293"}, + {file = "pathspec-0.11.1.tar.gz", hash = "sha256:2798de800fa92780e33acca925945e9a19a133b715067cf165b8866c15a31687"}, +] + +[[package]] +name = "pbr" +version = "5.11.1" +description = "Python Build Reasonableness" +category = "dev" +optional = false +python-versions = ">=2.6" +files = [ + {file = "pbr-5.11.1-py2.py3-none-any.whl", hash = "sha256:567f09558bae2b3ab53cb3c1e2e33e726ff3338e7bae3db5dc954b3a44eef12b"}, + {file = "pbr-5.11.1.tar.gz", hash = "sha256:aefc51675b0b533d56bb5fd1c8c6c0522fe31896679882e1c4c63d5e4a0fccb3"}, +] + +[[package]] +name = "pillow" +version = "10.0.0" +description = "Python Imaging Library (Fork)" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "Pillow-10.0.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:1f62406a884ae75fb2f818694469519fb685cc7eaff05d3451a9ebe55c646891"}, + {file = "Pillow-10.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d5db32e2a6ccbb3d34d87c87b432959e0db29755727afb37290e10f6e8e62614"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:edf4392b77bdc81f36e92d3a07a5cd072f90253197f4a52a55a8cec48a12483b"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:520f2a520dc040512699f20fa1c363eed506e94248d71f85412b625026f6142c"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:8c11160913e3dd06c8ffdb5f233a4f254cb449f4dfc0f8f4549eda9e542c93d1"}, + {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a74ba0c356aaa3bb8e3eb79606a87669e7ec6444be352870623025d75a14a2bf"}, + {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d5d0dae4cfd56969d23d94dc8e89fb6a217be461c69090768227beb8ed28c0a3"}, + {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22c10cc517668d44b211717fd9775799ccec4124b9a7f7b3635fc5386e584992"}, + {file = "Pillow-10.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:dffe31a7f47b603318c609f378ebcd57f1554a3a6a8effbc59c3c69f804296de"}, + {file = "Pillow-10.0.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:9fb218c8a12e51d7ead2a7c9e101a04982237d4855716af2e9499306728fb485"}, + {file = "Pillow-10.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d35e3c8d9b1268cbf5d3670285feb3528f6680420eafe35cccc686b73c1e330f"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3ed64f9ca2f0a95411e88a4efbd7a29e5ce2cea36072c53dd9d26d9c76f753b3"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0b6eb5502f45a60a3f411c63187db83a3d3107887ad0d036c13ce836f8a36f1d"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:c1fbe7621c167ecaa38ad29643d77a9ce7311583761abf7836e1510c580bf3dd"}, + {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:cd25d2a9d2b36fcb318882481367956d2cf91329f6892fe5d385c346c0649629"}, + {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:3b08d4cc24f471b2c8ca24ec060abf4bebc6b144cb89cba638c720546b1cf538"}, + {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d737a602fbd82afd892ca746392401b634e278cb65d55c4b7a8f48e9ef8d008d"}, + {file = "Pillow-10.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:3a82c40d706d9aa9734289740ce26460a11aeec2d9c79b7af87bb35f0073c12f"}, + {file = "Pillow-10.0.0-cp312-cp312-macosx_10_10_x86_64.whl", hash = "sha256:d80cf684b541685fccdd84c485b31ce73fc5c9b5d7523bf1394ce134a60c6883"}, + {file = "Pillow-10.0.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:76de421f9c326da8f43d690110f0e79fe3ad1e54be811545d7d91898b4c8493e"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81ff539a12457809666fef6624684c008e00ff6bf455b4b89fd00a140eecd640"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce543ed15570eedbb85df19b0a1a7314a9c8141a36ce089c0a894adbfccb4568"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:685ac03cc4ed5ebc15ad5c23bc555d68a87777586d970c2c3e216619a5476223"}, + {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:d72e2ecc68a942e8cf9739619b7f408cc7b272b279b56b2c83c6123fcfa5cdff"}, + {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:d50b6aec14bc737742ca96e85d6d0a5f9bfbded018264b3b70ff9d8c33485551"}, + {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:00e65f5e822decd501e374b0650146063fbb30a7264b4d2744bdd7b913e0cab5"}, + {file = "Pillow-10.0.0-cp312-cp312-win_amd64.whl", hash = "sha256:f31f9fdbfecb042d046f9d91270a0ba28368a723302786c0009ee9b9f1f60199"}, + {file = "Pillow-10.0.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:349930d6e9c685c089284b013478d6f76e3a534e36ddfa912cde493f235372f3"}, + {file = "Pillow-10.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3a684105f7c32488f7153905a4e3015a3b6c7182e106fe3c37fbb5ef3e6994c3"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b4f69b3700201b80bb82c3a97d5e9254084f6dd5fb5b16fc1a7b974260f89f43"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3f07ea8d2f827d7d2a49ecf1639ec02d75ffd1b88dcc5b3a61bbb37a8759ad8d"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:040586f7d37b34547153fa383f7f9aed68b738992380ac911447bb78f2abe530"}, + {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:f88a0b92277de8e3ca715a0d79d68dc82807457dae3ab8699c758f07c20b3c51"}, + {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:c7cf14a27b0d6adfaebb3ae4153f1e516df54e47e42dcc073d7b3d76111a8d86"}, + {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:3400aae60685b06bb96f99a21e1ada7bc7a413d5f49bce739828ecd9391bb8f7"}, + {file = "Pillow-10.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:dbc02381779d412145331789b40cc7b11fdf449e5d94f6bc0b080db0a56ea3f0"}, + {file = "Pillow-10.0.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9211e7ad69d7c9401cfc0e23d49b69ca65ddd898976d660a2fa5904e3d7a9baa"}, + {file = "Pillow-10.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:faaf07ea35355b01a35cb442dd950d8f1bb5b040a7787791a535de13db15ed90"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c9f72a021fbb792ce98306ffb0c348b3c9cb967dce0f12a49aa4c3d3fdefa967"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f7c16705f44e0504a3a2a14197c1f0b32a95731d251777dcb060aa83022cb2d"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:76edb0a1fa2b4745fb0c99fb9fb98f8b180a1bbceb8be49b087e0b21867e77d3"}, + {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:368ab3dfb5f49e312231b6f27b8820c823652b7cd29cfbd34090565a015e99ba"}, + {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:608bfdee0d57cf297d32bcbb3c728dc1da0907519d1784962c5f0c68bb93e5a3"}, + {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5c6e3df6bdd396749bafd45314871b3d0af81ff935b2d188385e970052091017"}, + {file = "Pillow-10.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:7be600823e4c8631b74e4a0d38384c73f680e6105a7d3c6824fcf226c178c7e6"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-macosx_10_10_x86_64.whl", hash = "sha256:92be919bbc9f7d09f7ae343c38f5bb21c973d2576c1d45600fce4b74bafa7ac0"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f8182b523b2289f7c415f589118228d30ac8c355baa2f3194ced084dac2dbba"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:38250a349b6b390ee6047a62c086d3817ac69022c127f8a5dc058c31ccef17f3"}, + {file = "Pillow-10.0.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:88af2003543cc40c80f6fca01411892ec52b11021b3dc22ec3bc9d5afd1c5334"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:c189af0545965fa8d3b9613cfdb0cd37f9d71349e0f7750e1fd704648d475ed2"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce7b031a6fc11365970e6a5686d7ba8c63e4c1cf1ea143811acbb524295eabed"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:db24668940f82321e746773a4bc617bfac06ec831e5c88b643f91f122a785684"}, + {file = "Pillow-10.0.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:efe8c0681042536e0d06c11f48cebe759707c9e9abf880ee213541c5b46c5bf3"}, + {file = "Pillow-10.0.0.tar.gz", hash = "sha256:9c82b5b3e043c7af0d95792d0d20ccf68f61a1fec6b3530e718b688422727396"}, +] + +[package.extras] +docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-removed-in", "sphinxext-opengraph"] +tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"] + +[[package]] +name = "pip" +version = "23.1.2" +description = "The PyPA recommended tool for installing Python packages." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-23.1.2-py3-none-any.whl", hash = "sha256:3ef6ac33239e4027d9a5598a381b9d30880a1477e50039db2eac6e8a8f6d1b18"}, + {file = "pip-23.1.2.tar.gz", hash = "sha256:0e7c86f486935893c708287b30bd050a36ac827ec7fe5e43fe7cb198dd835fba"}, +] + +[[package]] +name = "pip-api" +version = "0.0.30" +description = "An unofficial, importable pip API" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip-api-0.0.30.tar.gz", hash = "sha256:a05df2c7aa9b7157374bcf4273544201a0c7bae60a9c65bcf84f3959ef3896f3"}, + {file = "pip_api-0.0.30-py3-none-any.whl", hash = "sha256:2a0314bd31522eb9ffe8a99668b0d07fee34ebc537931e7b6483001dbedcbdc9"}, +] + +[package.dependencies] +pip = "*" + +[[package]] +name = "pip-audit" +version = "2.6.0" +description = "A tool for scanning Python environments for known vulnerabilities" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pip_audit-2.6.0-py3-none-any.whl", hash = "sha256:49e97e3d6663d2ed0c00b7a7c468afcb816beb3988f32f8496d3fe3927cfd627"}, + {file = "pip_audit-2.6.0.tar.gz", hash = "sha256:6431c363efa80ef52c2599197c5b8a39ff8708ce316624b97fa35b5cdf493118"}, +] + +[package.dependencies] +CacheControl = {version = ">=0.13.0", extras = ["filecache"]} +cyclonedx-python-lib = ">=4.0,<5.0" +html5lib = ">=1.1" +packaging = ">=23.0.0" +pip-api = ">=0.0.28" +pip-requirements-parser = ">=32.0.0" +requests = ">=2.31.0" +rich = ">=12.4" +toml = ">=0.10" + +[package.extras] +dev = ["build", "bump (>=1.3.2)", "pip-audit[doc,lint,test]"] +doc = ["pdoc"] +lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.276)", "types-html5lib", "types-requests", "types-toml"] +test = ["coverage[toml]", "pretend", "pytest", "pytest-cov"] + +[[package]] +name = "pip-requirements-parser" +version = "32.0.1" +description = "pip requirements parser - a mostly correct pip requirements parsing library because it uses pip's own code." +category = "dev" +optional = false +python-versions = ">=3.6.0" +files = [ + {file = "pip-requirements-parser-32.0.1.tar.gz", hash = "sha256:b4fa3a7a0be38243123cf9d1f3518da10c51bdb165a2b2985566247f9155a7d3"}, + {file = "pip_requirements_parser-32.0.1-py3-none-any.whl", hash = "sha256:4659bc2a667783e7a15d190f6fccf8b2486685b6dba4c19c3876314769c57526"}, +] + +[package.dependencies] +packaging = "*" +pyparsing = "*" + +[package.extras] +docs = ["Sphinx (>=3.3.1)", "doc8 (>=0.8.1)", "sphinx-rtd-theme (>=0.5.0)"] +testing = ["aboutcode-toolkit (>=6.0.0)", "black", "pytest (>=6,!=7.0.0)", "pytest-xdist (>=2)"] + +[[package]] +name = "platformdirs" +version = "3.8.0" +description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "platformdirs-3.8.0-py3-none-any.whl", hash = "sha256:ca9ed98ce73076ba72e092b23d3c93ea6c4e186b3f1c3dad6edd98ff6ffcca2e"}, + {file = "platformdirs-3.8.0.tar.gz", hash = "sha256:b0cabcb11063d21a0b261d557acb0a9d2126350e63b70cdf7db6347baea456dc"}, +] + +[package.extras] +docs = ["furo (>=2023.5.20)", "proselint (>=0.13)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"] +test = ["appdirs (==1.4.4)", "covdefaults (>=2.3)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)"] + +[[package]] +name = "pluggy" +version = "1.2.0" +description = "plugin and hook calling mechanisms for python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pluggy-1.2.0-py3-none-any.whl", hash = "sha256:c2fd55a7d7a3863cba1a013e4e2414658b1d07b6bc57b3919e0c63c9abb99849"}, + {file = "pluggy-1.2.0.tar.gz", hash = "sha256:d12f0c4b579b15f5e054301bb226ee85eeeba08ffec228092f8defbaa3a4c4b3"}, +] + +[package.extras] +dev = ["pre-commit", "tox"] +testing = ["pytest", "pytest-benchmark"] + +[[package]] +name = "pooch" +version = "1.6.0" +description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\"" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"}, + {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"}, +] + +[package.dependencies] +appdirs = ">=1.3.0" +packaging = ">=20.0" +requests = ">=2.19.0" + +[package.extras] +progress = ["tqdm (>=4.41.0,<5.0.0)"] +sftp = ["paramiko (>=2.7.0)"] +xxhash = ["xxhash (>=1.4.3)"] + +[[package]] +name = "prometheus-client" +version = "0.12.0" +description = "Python client for the Prometheus monitoring system." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "prometheus_client-0.12.0-py2.py3-none-any.whl", hash = "sha256:317453ebabff0a1b02df7f708efbab21e3489e7072b61cb6957230dd004a0af0"}, + {file = "prometheus_client-0.12.0.tar.gz", hash = "sha256:1b12ba48cee33b9b0b9de64a1047cbd3c5f2d0ab6ebcead7ddda613a750ec3c5"}, +] + +[package.extras] +twisted = ["twisted"] + +[[package]] +name = "psutil" +version = "5.9.5" +description = "Cross-platform lib for process and system monitoring in Python." +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "psutil-5.9.5-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:be8929ce4313f9f8146caad4272f6abb8bf99fc6cf59344a3167ecd74f4f203f"}, + {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:ab8ed1a1d77c95453db1ae00a3f9c50227ebd955437bcf2a574ba8adbf6a74d5"}, + {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:4aef137f3345082a3d3232187aeb4ac4ef959ba3d7c10c33dd73763fbc063da4"}, + {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:ea8518d152174e1249c4f2a1c89e3e6065941df2fa13a1ab45327716a23c2b48"}, + {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:acf2aef9391710afded549ff602b5887d7a2349831ae4c26be7c807c0a39fac4"}, + {file = "psutil-5.9.5-cp27-none-win32.whl", hash = "sha256:5b9b8cb93f507e8dbaf22af6a2fd0ccbe8244bf30b1baad6b3954e935157ae3f"}, + {file = "psutil-5.9.5-cp27-none-win_amd64.whl", hash = "sha256:8c5f7c5a052d1d567db4ddd231a9d27a74e8e4a9c3f44b1032762bd7b9fdcd42"}, + {file = "psutil-5.9.5-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:3c6f686f4225553615612f6d9bc21f1c0e305f75d7d8454f9b46e901778e7217"}, + {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7a7dd9997128a0d928ed4fb2c2d57e5102bb6089027939f3b722f3a210f9a8da"}, + {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:89518112647f1276b03ca97b65cc7f64ca587b1eb0278383017c2a0dcc26cbe4"}, + {file = "psutil-5.9.5-cp36-abi3-win32.whl", hash = "sha256:104a5cc0e31baa2bcf67900be36acde157756b9c44017b86b2c049f11957887d"}, + {file = "psutil-5.9.5-cp36-abi3-win_amd64.whl", hash = "sha256:b258c0c1c9d145a1d5ceffab1134441c4c5113b2417fafff7315a917a026c3c9"}, + {file = "psutil-5.9.5-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:c607bb3b57dc779d55e1554846352b4e358c10fff3abf3514a7a6601beebdb30"}, + {file = "psutil-5.9.5.tar.gz", hash = "sha256:5410638e4df39c54d957fc51ce03048acd8e6d60abc0f5107af51e5fb566eb3c"}, +] + +[package.extras] +test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"] + +[[package]] +name = "py-serializable" +version = "0.11.1" +description = "Library for serializing and deserializing Python Objects to and from JSON and XML." +category = "dev" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "py-serializable-0.11.1.tar.gz", hash = "sha256:ba0e1287b9e4f645a5334f1913abd8e647e7250209f84f55dce3909498a6f586"}, + {file = "py_serializable-0.11.1-py3-none-any.whl", hash = "sha256:79e21f0672822e6200b15f45ce9f636e8126466f62dbd7d488c67313c72b5c3e"}, +] + +[package.dependencies] +defusedxml = ">=0.7.1,<0.8.0" + +[[package]] +name = "pyarrow" +version = "12.0.1" +description = "Python library for Apache Arrow" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pyarrow-12.0.1-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:6d288029a94a9bb5407ceebdd7110ba398a00412c5b0155ee9813a40d246c5df"}, + {file = "pyarrow-12.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:345e1828efdbd9aa4d4de7d5676778aba384a2c3add896d995b23d368e60e5af"}, + {file = "pyarrow-12.0.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d6009fdf8986332b2169314da482baed47ac053311c8934ac6651e614deacd6"}, + {file = "pyarrow-12.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2d3c4cbbf81e6dd23fe921bc91dc4619ea3b79bc58ef10bce0f49bdafb103daf"}, + {file = "pyarrow-12.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:cdacf515ec276709ac8042c7d9bd5be83b4f5f39c6c037a17a60d7ebfd92c890"}, + {file = "pyarrow-12.0.1-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:749be7fd2ff260683f9cc739cb862fb11be376de965a2a8ccbf2693b098db6c7"}, + {file = "pyarrow-12.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:6895b5fb74289d055c43db3af0de6e16b07586c45763cb5e558d38b86a91e3a7"}, + {file = "pyarrow-12.0.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1887bdae17ec3b4c046fcf19951e71b6a619f39fa674f9881216173566c8f718"}, + {file = "pyarrow-12.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e2c9cb8eeabbadf5fcfc3d1ddea616c7ce893db2ce4dcef0ac13b099ad7ca082"}, + {file = "pyarrow-12.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:ce4aebdf412bd0eeb800d8e47db854f9f9f7e2f5a0220440acf219ddfddd4f63"}, + {file = "pyarrow-12.0.1-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:e0d8730c7f6e893f6db5d5b86eda42c0a130842d101992b581e2138e4d5663d3"}, + {file = "pyarrow-12.0.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:43364daec02f69fec89d2315f7fbfbeec956e0d991cbbef471681bd77875c40f"}, + {file = "pyarrow-12.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:051f9f5ccf585f12d7de836e50965b3c235542cc896959320d9776ab93f3b33d"}, + {file = "pyarrow-12.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:be2757e9275875d2a9c6e6052ac7957fbbfc7bc7370e4a036a9b893e96fedaba"}, + {file = "pyarrow-12.0.1-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:cf812306d66f40f69e684300f7af5111c11f6e0d89d6b733e05a3de44961529d"}, + {file = "pyarrow-12.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:459a1c0ed2d68671188b2118c63bac91eaef6fc150c77ddd8a583e3c795737bf"}, + {file = "pyarrow-12.0.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:85e705e33eaf666bbe508a16fd5ba27ca061e177916b7a317ba5a51bee43384c"}, + {file = "pyarrow-12.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9120c3eb2b1f6f516a3b7a9714ed860882d9ef98c4b17edcdc91d95b7528db60"}, + {file = "pyarrow-12.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:c780f4dc40460015d80fcd6a6140de80b615349ed68ef9adb653fe351778c9b3"}, + {file = "pyarrow-12.0.1-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:a3c63124fc26bf5f95f508f5d04e1ece8cc23a8b0af2a1e6ab2b1ec3fdc91b24"}, + {file = "pyarrow-12.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:b13329f79fa4472324f8d32dc1b1216616d09bd1e77cfb13104dec5463632c36"}, + {file = "pyarrow-12.0.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb656150d3d12ec1396f6dde542db1675a95c0cc8366d507347b0beed96e87ca"}, + {file = "pyarrow-12.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6251e38470da97a5b2e00de5c6a049149f7b2bd62f12fa5dbb9ac674119ba71a"}, + {file = "pyarrow-12.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:3de26da901216149ce086920547dfff5cd22818c9eab67ebc41e863a5883bac7"}, + {file = "pyarrow-12.0.1.tar.gz", hash = "sha256:cce317fc96e5b71107bf1f9f184d5e54e2bd14bbf3f9a3d62819961f0af86fec"}, +] + +[package.dependencies] +numpy = ">=1.16.6" + +[[package]] +name = "pycodestyle" +version = "2.7.0" +description = "Python style guide checker" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycodestyle-2.7.0-py2.py3-none-any.whl", hash = "sha256:514f76d918fcc0b55c6680472f0a37970994e07bbb80725808c17089be302068"}, + {file = "pycodestyle-2.7.0.tar.gz", hash = "sha256:c389c1d06bf7904078ca03399a4816f974a1d590090fecea0c63ec26ebaf1cef"}, +] + +[[package]] +name = "pycparser" +version = "2.21" +description = "C parser in Python" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, + {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, +] + +[[package]] +name = "pydub" +version = "0.25.1" +description = "Manipulate audio with an simple and easy high level interface" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"}, + {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"}, +] + +[[package]] +name = "pyflakes" +version = "2.3.1" +description = "passive checker of Python programs" +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +files = [ + {file = "pyflakes-2.3.1-py2.py3-none-any.whl", hash = "sha256:7893783d01b8a89811dd72d7dfd4d84ff098e5eed95cfa8905b22bbffe52efc3"}, + {file = "pyflakes-2.3.1.tar.gz", hash = "sha256:f5bc8ecabc05bb9d291eb5203d6810b49040f6ff446a756326104746cc00c1db"}, +] + +[[package]] +name = "pygments" +version = "2.15.1" +description = "Pygments is a syntax highlighting package written in Python." +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pygments-2.15.1-py3-none-any.whl", hash = "sha256:db2db3deb4b4179f399a09054b023b6a586b76499d36965813c71aa8ed7b5fd1"}, + {file = "Pygments-2.15.1.tar.gz", hash = "sha256:8ace4d3c1dd481894b2005f560ead0f9f19ee64fe983366be1a21e171d12775c"}, +] + +[package.extras] +plugins = ["importlib-metadata"] + +[[package]] +name = "pyjwt" +version = "2.7.0" +description = "JSON Web Token implementation in Python" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "PyJWT-2.7.0-py3-none-any.whl", hash = "sha256:ba2b425b15ad5ef12f200dc67dd56af4e26de2331f965c5439994dad075876e1"}, + {file = "PyJWT-2.7.0.tar.gz", hash = "sha256:bd6ca4a3c4285c1a2d4349e5a035fdf8fb94e04ccd0fcbe6ba289dae9cc3e074"}, +] + +[package.dependencies] +cryptography = {version = ">=3.4.0", optional = true, markers = "extra == \"crypto\""} + +[package.extras] +crypto = ["cryptography (>=3.4.0)"] +dev = ["coverage[toml] (==5.0.4)", "cryptography (>=3.4.0)", "pre-commit", "pytest (>=6.0.0,<7.0.0)", "sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"] +docs = ["sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"] +tests = ["coverage[toml] (==5.0.4)", "pytest (>=6.0.0,<7.0.0)"] + +[[package]] +name = "pymongo" +version = "3.13.0" +description = "Python driver for MongoDB <http://www.mongodb.org>" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl", hash = "sha256:3ad3a3df830f7df7e0856c2bdb54d19f5bf188bd7420985e18643b8e4d2a075f"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:b96e0e9d2d48948240b510bac81614458fc10adcd3a93240c2fd96448b4efd35"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9f592b202d77923498b32ddc5b376e5fa9ba280d3e16ed56cb8c932fe6d6a478"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:851f2bb52b5cb2f4711171ca925e0e05344a8452972a748a8a8ffdda1e1d72a7"}, + {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:1c9d23f62a3fa7523d849c4942acc0d9ff7081ebc00c808ee7cfdc070df0687f"}, + {file = "pymongo-3.13.0-cp27-cp27m-win32.whl", hash = "sha256:a17b81f22398e3e0f72bdf938e98c810286994b2bcc0a125cd5ad8fd4ea54ad7"}, + {file = "pymongo-3.13.0-cp27-cp27m-win_amd64.whl", hash = "sha256:4f6dd55dab77adf60b445c11f426ee5cdfa1b86f6d54cb937bfcbf09572333ab"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:776f90bf2252f90a4ae838e7917638894c6356bef7265f424592e2fd1f577d05"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:50b99f4d3eee6f03778fe841d6f470e6c18e744dc665156da6da3bc6e65b398d"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50a81b2d9f188c7909e0a1084fa969bb92a788076809c437ac1ae80393f46df9"}, + {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c7c45a8a1a752002b0a7c81ab3a4c5e3b6f67f9826b16fbe3943f5329f565f24"}, + {file = "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1037097708498bdc85f23c8798a5c46c7bce432d77d23608ff14e0d831f1a971"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl", hash = "sha256:b5b733694e7df22d5c049581acfc487695a6ff813322318bed8dd66f79978636"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:d7c91747ec8dde51440dd594603158cc98abb3f7df84b2ed8a836f138285e4fb"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl", hash = "sha256:f4175fcdddf764d371ee52ec4505a40facee2533e84abf2953cda86d050cfa1f"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl", hash = "sha256:93d4e9a02c17813b34e4bd9f6fbf07310c140c8f74341537c24d07c1cdeb24d1"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl", hash = "sha256:3b261d593f2563299062733ae003a925420a86ff4ddda68a69097d67204e43f3"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl", hash = "sha256:172db03182a22e9002157b262c1ea3b0045c73d4ff465adc152ce5b4b0e7b8d4"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09de3bfc995ae8cb955abb0c9ae963c134dba1b5622be3bcc527b89b0fd4091c"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c0379447587ee4b8f983ba183202496e86c0358f47c45612619d634d1fcd82bd"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30245a8747dc90019a3c9ad9df987e0280a3ea632ad36227cde7d1d8dcba0830"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65b6fddf6a7b91da044f202771a38e71bbb9bf42720a406b26b25fe2256e7102"}, + {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5831a377d15a626fbec10890ffebc4c6abcd37e4126737932cd780a171eabdc1"}, + {file = "pymongo-3.13.0-cp310-cp310-win32.whl", hash = "sha256:944249aa83dee314420c37d0f40c30a8f6dc4a3877566017b87062e53af449f4"}, + {file = "pymongo-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea8824ebc9a1a5c8269e8f1e3989b5a6bec876726e2f3c33ebd036cb488277f0"}, + {file = "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:bdd34c57b4da51a7961beb33645646d197e41f8517801dc76b37c1441e7a4e10"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26f9cc42a162faa241c82e117ac85734ae9f14343dc2df1c90c6b2181f791b22"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a82a1c10f5608e6494913faa169e213d703194bfca0aa710901f303be212414"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8927f22ef6a16229da7f18944deac8605bdc2c0858be5184259f2f7ce7fd4459"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e6f8191a282ef77e526f8f8f63753a437e4aa4bc78f5edd8b6b6ed0eaebd5363"}, + {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4d9ed67c987bf9ac2ac684590ba3d2599cdfb0f331ee3db607f9684469b3b59d"}, + {file = "pymongo-3.13.0-cp311-cp311-win32.whl", hash = "sha256:e8f6979664ff477cd61b06bf8aba206df7b2334209815ab3b1019931dab643d6"}, + {file = "pymongo-3.13.0-cp311-cp311-win_amd64.whl", hash = "sha256:174fd1000e896d0dfbc7f6d7e6a1992a4868796c7dec31679e38218c78d6a942"}, + {file = "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl", hash = "sha256:d1ee773fb72ba024e7e3bb6ea8907fe52bccafcb5184aaced6bad995bd30ea20"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:28565e3dbd69fe5fe35a210067064dbb6ed5abe997079f653c19c873c3896fe6"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:5c1db7d366004d6c699eb08c716a63ae0a3e946d061cbebea65d7ce361950265"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e1956f3338c10308e2f99c2c9ff46ae412035cbcd7aaa76c39ccdb806854a247"}, + {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:10f0fddc1d63ba3d4a4bffcc7720184c1b7efd570726ad5e2f55818da320239f"}, + {file = "pymongo-3.13.0-cp35-cp35m-win32.whl", hash = "sha256:570ae3365b23d4fd8c669cb57613b1a90b2757e993588d3370ef90945dbeec4b"}, + {file = "pymongo-3.13.0-cp35-cp35m-win_amd64.whl", hash = "sha256:79f777eaf3f5b2c6d81f9ef00d87837001d7063302503bbcbfdbf3e9bc27c96f"}, + {file = "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl", hash = "sha256:d42eb29ba314adfd9c11234b4b646f61b0448bf9b00f14db4b317e6e4b947e77"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:e5e87c0eb774561c546f979342a8ff36ebee153c60a0b6c6b03ba989ceb9538c"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:0f2c5a5984599a88d087a15859860579b825098b473d8c843f1979a83d159f2e"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:59c98e86c5e861032b71e6e5b65f23e6afaacea6e82483b66f1191a5021a7b4f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:70b67390e27e58876853efbb87e43c85252de2515e2887f7dd901b4fa3d21973"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:42ba8606492d76e6f9e4c7a458ed4bc712603be393259a52450345f0945da2cf"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:0e5536994cf2d8488c6fd9dea71df3c4dbb3e0d2ba5e695da06d9142a29a0969"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:fe8194f107f0fa3cabd14e9e809f174eca335993c1db72d1e74e0f496e7afe1f"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d593d50815771f517d3ac4367ff716e3f3c78edae51d98e1e25791459f8848ff"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5136ebe8da6a1604998a8eb96be55935aa5f7129c41cc7bddc400d48e8df43be"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a424bdedfd84454d2905a861e0d4bb947cc5bd024fdeb3600c1a97d2be0f4255"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5161167b3840e9c84c80f2534ea6a099f51749d5673b662a3dd248be17c3208"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:644470442beaf969df99c4e00367a817eee05f0bba5d888f1ba6fe97b5e1c102"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2406df90b2335371706c59b7d79e9633b81ed2a7ecd48c1faf8584552bdf2d90"}, + {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:222591b828de10ac90064047b5d4916953f38c38b155009c4b8b5e0d33117c2b"}, + {file = "pymongo-3.13.0-cp36-cp36m-win32.whl", hash = "sha256:7cb987b199fa223ad78eebaa9fbc183d5a5944bfe568a9d6f617316ca1c1f32f"}, + {file = "pymongo-3.13.0-cp36-cp36m-win_amd64.whl", hash = "sha256:a6cbb73d9fc2282677e2b7a137d13da987bd0b13abd88ed27bba5534c226db06"}, + {file = "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl", hash = "sha256:b1223b826acbef07a7f5eb9bf37247b0b580119916dca9eae19d92b1290f5855"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:398fb86d374dc351a4abc2e24cd15e5e14b2127f6d90ce0df3fdf2adcc55ac1b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:9c3d07ea19cd2856d9943dce37e75d69ecbb5baf93c3e4c82f73b6075c481292"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:2943d739715f265a2983ac43747595b6af3312d0a370614040959fd293763adf"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:c3b70ed82f20d18d22eafc9bda0ea656605071762f7d31f3c5afc35c59d3393b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:7ec2bb598847569ae34292f580842d37619eea3e546005042f485e15710180d5"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:8cc37b437cba909bef06499dadd91a39c15c14225e8d8c7870020049f8a549fe"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:65a063970e15a4f338f14b820561cf6cdaf2839691ac0adb2474ddff9d0b8b0b"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02f0e1a75d3bc0e16c7e15daf9c56185642be055e425f3b34888fc6eb1b22401"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16e74b9c2aca2734c7f49f00fe68d6830a30d26df60e2ace7fe40ccb92087b94"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:24e954be35ad4537840f20bbc8d75320ae647d3cb4fab12cb8fcd2d55f408e76"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a149377d1ff766fd618500798d0d94637f66d0ae222bb6d28f41f3e15c626297"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61660710b054ae52c8fc10368e91d74719eb05554b631d7f8ca93d21d2bff2e6"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4bbc0d27dfef7689285e54f2e0a224f0c7cd9d5c46d2638fabad5500b951c92f"}, + {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:9b2ed9c3b30f11cd4a3fbfc22167af7987b01b444215c2463265153fe7cf66d6"}, + {file = "pymongo-3.13.0-cp37-cp37m-win32.whl", hash = "sha256:1c2c5e2b00e2fadcd590c0b2e293d71215e98ed1cb635cfca2be4998d197e534"}, + {file = "pymongo-3.13.0-cp37-cp37m-win_amd64.whl", hash = "sha256:32eac95bbb030b2376ffd897376c6f870222a3457f01a9ce466b9057876132f8"}, + {file = "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a796ef39dadf9d73af05d24937644d386495e43a7d13617aa3651d836da542c8"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:b6793baf4639c72a500698a49e9250b293e17ae1faf11ac1699d8141194786fe"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:80d8576b04d0824f63bf803190359c0d3bcb6e7fa63fefbd4bc0ceaa7faae38c"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:db2e11507fe9cc2a722be21ccc62c1b1295398fe9724c1f14900cdc7166fc0d7"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:b01ce58eec5edeededf1992d2dce63fb8565e437be12d6f139d75b15614c4d08"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:d1a19d6c5098f1f4e11430cd74621699453cbc534dd7ade9167e582f50814b19"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:7219b1a726ced3bacecabef9bd114529bbb69477901373e800d7d0140baadc95"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:2dae3b353a10c3767e0aa1c1492f2af388f1012b08117695ab3fd1f219e5814e"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:12721d926d43d33dd3318e58dce9b0250e8a9c6e1093fa8e09f4805193ff4b43"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6af0a4b17faf26779d5caee8542a4f2cba040cea27d3bffc476cbc6ccbd4c8ee"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:09b9d0f5a445c7e0ddcc021b09835aa6556f0166afc498f57dfdd72cdf6f02ad"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db5b4f8ad8607a3d612da1d4c89a84e4cf5c88f98b46365820d9babe5884ba45"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:34dbf5fecf653c152edb75a35a8b15dfdc4549473484ee768aeb12c97983cead"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:34cd48df7e1fc69222f296d8f69e3957eb7c6b5aa0709d3467184880ed7538c0"}, + {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c8f755ff1f4ab4ca790d1d6d3229006100b301475948021b6b2757822e0d6c97"}, + {file = "pymongo-3.13.0-cp38-cp38-win32.whl", hash = "sha256:b0746d0d4535f56bbaa63a8f6da362f330804d578e66e126b226eebe76c2bf00"}, + {file = "pymongo-3.13.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ad0515abb132f52ce9d8abd1a29681a1e65dba7b7fe13ea01e1a8db5715bf80"}, + {file = "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:3c5cb6c93c94df76a879bad4b89db0104b01806d17c2b803c1316ba50962b6d6"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2e0854170813238f0c3131050c67cb1fb1ade75c93bf6cd156c1bd9a16095528"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:1410faa51ce835cc1234c99ec42e98ab4f3c6f50d92d86a2d4f6e11c97ee7a4e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:d7910135f5de1c5c3578e61d6f4b087715b15e365f11d4fa51a9cee92988b2bd"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:028175dd8d2979a889153a2308e8e500b3df7d9e3fd1c33ca7fdeadf61cc87a2"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:2bfc39276c0e6d07c95bd1088b5003f049e986e089509f7dbd68bb7a4b1e65ac"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:4092b660ec720d44d3ca81074280dc25c7a3718df1b6c0fe9fe36ac6ed2833e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:5bdeb71a610a7b801416268e500e716d0fe693fb10d809e17f0fb3dac5be5a34"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aa3bca8e76f5c00ed2bb4325e0e383a547d71595926d5275d7c88175aaf7435e"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7c7cab8155f430ca460a6fc7ae8a705b34f3e279a57adb5f900eb81943ec777c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4a32f3dfcca4a4816373bdb6256c18c78974ebb3430e7da988516cd95b2bd6e4"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30ed2788a6ec68743e2040ab1d16573d7d9f6e7333e45070ce9268cbc93d148c"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21e61a536ffed84d10376c21c13a6ed1ebefb61989a844952547c229d6aeedf3"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0665412dce26b2318092a33bd2d2327d487c4490cfcde158d6946d39b1e28d78"}, + {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:64ed1a5ce5e5926727eb0f87c698c4d9a7a9f7b0953683a65e9ce2b7cc5f8e91"}, + {file = "pymongo-3.13.0-cp39-cp39-win32.whl", hash = "sha256:7593cb1214185a0c5b43b96effc51ce82ddc933298ee36db7dc2bd45d61b4adc"}, + {file = "pymongo-3.13.0-cp39-cp39-win_amd64.whl", hash = "sha256:3cfc9bc1e8b5667bc1f3dbe46d2f85b3f24ff7533893bdc1203058012db2c046"}, + {file = "pymongo-3.13.0.tar.gz", hash = "sha256:e22d6cf5802cd09b674c307cc9e03870b8c37c503ebec3d25b86f2ce8c535dc7"}, +] + +[package.dependencies] +dnspython = {version = ">=1.16.0,<1.17.0", optional = true, markers = "extra == \"srv\""} + +[package.extras] +aws = ["pymongo-auth-aws (<2.0.0)"] +encryption = ["pymongocrypt (>=1.1.0,<2.0.0)"] +gssapi = ["pykerberos"] +ocsp = ["certifi", "pyopenssl (>=17.2.0)", "requests (<3.0.0)", "service-identity (>=18.1.0)"] +snappy = ["python-snappy"] +srv = ["dnspython (>=1.16.0,<1.17.0)"] +tls = ["ipaddress"] +zstd = ["zstandard"] + +[[package]] +name = "pyparsing" +version = "3.1.0" +description = "pyparsing module - Classes and methods to define and execute parsing grammars" +category = "dev" +optional = false +python-versions = ">=3.6.8" +files = [ + {file = "pyparsing-3.1.0-py3-none-any.whl", hash = "sha256:d554a96d1a7d3ddaf7183104485bc19fd80543ad6ac5bdb6426719d766fb06c1"}, + {file = "pyparsing-3.1.0.tar.gz", hash = "sha256:edb662d6fe322d6e990b1594b5feaeadf806803359e3d4d42f11e295e588f0ea"}, +] + +[package.extras] +diagrams = ["jinja2", "railroad-diagrams"] + +[[package]] +name = "pytest" +version = "7.4.0" +description = "pytest: simple powerful testing with Python" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "pytest-7.4.0-py3-none-any.whl", hash = "sha256:78bf16451a2eb8c7a2ea98e32dc119fd2aa758f1d5d66dbf0a59d69a3969df32"}, + {file = "pytest-7.4.0.tar.gz", hash = "sha256:b4bf8c45bd59934ed84001ad51e11b4ee40d40a1229d2c79f9c592b0a3f6bd8a"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "sys_platform == \"win32\""} +exceptiongroup = {version = ">=1.0.0rc8", markers = "python_version < \"3.11\""} +iniconfig = "*" +packaging = "*" +pluggy = ">=0.12,<2.0" +tomli = {version = ">=1.0.0", markers = "python_version < \"3.11\""} + +[package.extras] +testing = ["argcomplete", "attrs (>=19.2.0)", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.7.2)", "requests", "setuptools", "xmlschema"] + +[[package]] +name = "pytest-cov" +version = "2.12.1" +description = "Pytest plugin for measuring coverage." +category = "dev" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*" +files = [ + {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"}, + {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"}, +] + +[package.dependencies] +coverage = ">=5.2.1" +pytest = ">=4.6" +toml = "*" + +[package.extras] +testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"] + +[[package]] +name = "pytest-httpserver" +version = "1.0.8" +description = "pytest-httpserver is a httpserver for pytest" +category = "dev" +optional = false +python-versions = ">=3.8,<4.0" +files = [ + {file = "pytest_httpserver-1.0.8-py3-none-any.whl", hash = "sha256:24cd3d9f6a0b927c7bfc400d0b3fda7442721b8267ce29942bf307b190f0bb09"}, + {file = "pytest_httpserver-1.0.8.tar.gz", hash = "sha256:e052f69bc8a9073db02484681e8e47004dd1fb3763b0ae833bd899e5895c559a"}, +] + +[package.dependencies] +Werkzeug = ">=2.0.0" + +[[package]] +name = "python-dateutil" +version = "2.8.2" +description = "Extensions to the standard Python datetime module" +category = "main" +optional = false +python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" +files = [ + {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"}, + {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"}, +] + +[package.dependencies] +six = ">=1.5" + +[[package]] +name = "python-dotenv" +version = "1.0.0" +description = "Read key-value pairs from a .env file and set them as environment variables" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"}, + {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"}, +] + +[package.extras] +cli = ["click (>=5.0)"] + +[[package]] +name = "pytz" +version = "2020.5" +description = "World timezone definitions, modern and historical" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"}, + {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"}, +] + +[[package]] +name = "pyyaml" +version = "6.0" +description = "YAML parser and emitter for Python" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d4db7c7aef085872ef65a8fd7d6d09a14ae91f691dec3e87ee5ee0539d516f53"}, + {file = "PyYAML-6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9df7ed3b3d2e0ecfe09e14741b857df43adb5a3ddadc919a2d94fbdf78fea53c"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77f396e6ef4c73fdc33a9157446466f1cff553d979bd00ecb64385760c6babdc"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a80a78046a72361de73f8f395f1f1e49f956c6be882eed58505a15f3e430962b"}, + {file = "PyYAML-6.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f84fbc98b019fef2ee9a1cb3ce93e3187a6df0b2538a651bfb890254ba9f90b5"}, + {file = "PyYAML-6.0-cp310-cp310-win32.whl", hash = "sha256:2cd5df3de48857ed0544b34e2d40e9fac445930039f3cfe4bcc592a1f836d513"}, + {file = "PyYAML-6.0-cp310-cp310-win_amd64.whl", hash = "sha256:daf496c58a8c52083df09b80c860005194014c3698698d1a57cbcfa182142a3a"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d4b0ba9512519522b118090257be113b9468d804b19d63c71dbcf4a48fa32358"}, + {file = "PyYAML-6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:81957921f441d50af23654aa6c5e5eaf9b06aba7f0a19c18a538dc7ef291c5a1"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:afa17f5bc4d1b10afd4466fd3a44dc0e245382deca5b3c353d8b757f9e3ecb8d"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dbad0e9d368bb989f4515da330b88a057617d16b6a8245084f1b05400f24609f"}, + {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:432557aa2c09802be39460360ddffd48156e30721f5e8d917f01d31694216782"}, + {file = "PyYAML-6.0-cp311-cp311-win32.whl", hash = "sha256:bfaef573a63ba8923503d27530362590ff4f576c626d86a9fed95822a8255fd7"}, + {file = "PyYAML-6.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b45c0191e6d66c470b6cf1b9531a771a83c1c4208272ead47a3ae4f2f603bf"}, + {file = "PyYAML-6.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:897b80890765f037df3403d22bab41627ca8811ae55e9a722fd0392850ec4d86"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50602afada6d6cbfad699b0c7bb50d5ccffa7e46a3d738092afddc1f9758427f"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:48c346915c114f5fdb3ead70312bd042a953a8ce5c7106d5bfb1a5254e47da92"}, + {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:98c4d36e99714e55cfbaaee6dd5badbc9a1ec339ebfc3b1f52e293aee6bb71a4"}, + {file = "PyYAML-6.0-cp36-cp36m-win32.whl", hash = "sha256:0283c35a6a9fbf047493e3a0ce8d79ef5030852c51e9d911a27badfde0605293"}, + {file = "PyYAML-6.0-cp36-cp36m-win_amd64.whl", hash = "sha256:07751360502caac1c067a8132d150cf3d61339af5691fe9e87803040dbc5db57"}, + {file = "PyYAML-6.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:819b3830a1543db06c4d4b865e70ded25be52a2e0631ccd2f6a47a2822f2fd7c"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:473f9edb243cb1935ab5a084eb238d842fb8f404ed2193a915d1784b5a6b5fc0"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0ce82d761c532fe4ec3f87fc45688bdd3a4c1dc5e0b4a19814b9009a29baefd4"}, + {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:231710d57adfd809ef5d34183b8ed1eeae3f76459c18fb4a0b373ad56bedcdd9"}, + {file = "PyYAML-6.0-cp37-cp37m-win32.whl", hash = "sha256:c5687b8d43cf58545ade1fe3e055f70eac7a5a1a0bf42824308d868289a95737"}, + {file = "PyYAML-6.0-cp37-cp37m-win_amd64.whl", hash = "sha256:d15a181d1ecd0d4270dc32edb46f7cb7733c7c508857278d3d378d14d606db2d"}, + {file = "PyYAML-6.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b4624f379dab24d3725ffde76559cff63d9ec94e1736b556dacdfebe5ab6d4b"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:213c60cd50106436cc818accf5baa1aba61c0189ff610f64f4a3e8c6726218ba"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9fa600030013c4de8165339db93d182b9431076eb98eb40ee068700c9c813e34"}, + {file = "PyYAML-6.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:277a0ef2981ca40581a47093e9e2d13b3f1fbbeffae064c1d21bfceba2030287"}, + {file = "PyYAML-6.0-cp38-cp38-win32.whl", hash = "sha256:d4eccecf9adf6fbcc6861a38015c2a64f38b9d94838ac1810a9023a0609e1b78"}, + {file = "PyYAML-6.0-cp38-cp38-win_amd64.whl", hash = "sha256:1e4747bc279b4f613a09eb64bba2ba602d8a6664c6ce6396a4d0cd413a50ce07"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:055d937d65826939cb044fc8c9b08889e8c743fdc6a32b33e2390f66013e449b"}, + {file = "PyYAML-6.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e61ceaab6f49fb8bdfaa0f92c4b57bcfbea54c09277b1b4f7ac376bfb7a7c174"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d67d839ede4ed1b28a4e8909735fc992a923cdb84e618544973d7dfc71540803"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba8c411ef271aa037d7357a2bc8f9ee8b58b9965831d9e51baf703280dc73d3"}, + {file = "PyYAML-6.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:40527857252b61eacd1d9af500c3337ba8deb8fc298940291486c465c8b46ec0"}, + {file = "PyYAML-6.0-cp39-cp39-win32.whl", hash = "sha256:b5b9eccad747aabaaffbc6064800670f0c297e52c12754eb1d976c57e4f74dcb"}, + {file = "PyYAML-6.0-cp39-cp39-win_amd64.whl", hash = "sha256:b3d267842bf12586ba6c734f89d1f5b871df0273157918b0ccefa29deb05c21c"}, + {file = "PyYAML-6.0.tar.gz", hash = "sha256:68fb519c14306fec9720a2a5b45bc9f0c8d1b9c72adf45c37baedfcd949c35a2"}, +] + +[[package]] +name = "requests" +version = "2.31.0" +description = "Python HTTP for Humans." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "requests-2.31.0-py3-none-any.whl", hash = "sha256:58cd2187c01e70e6e26505bca751777aa9f2ee0b7f4300988b709f44e013003f"}, + {file = "requests-2.31.0.tar.gz", hash = "sha256:942c5a758f98d790eaed1a29cb6eefc7ffb0d1cf7af05c3d2791656dbd6ad1e1"}, +] + +[package.dependencies] +certifi = ">=2017.4.17" +charset-normalizer = ">=2,<4" +idna = ">=2.5,<4" +urllib3 = ">=1.21.1,<3" + +[package.extras] +socks = ["PySocks (>=1.5.6,!=1.5.7)"] +use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] + +[[package]] +name = "rich" +version = "13.4.2" +description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal" +category = "dev" +optional = false +python-versions = ">=3.7.0" +files = [ + {file = "rich-13.4.2-py3-none-any.whl", hash = "sha256:8f87bc7ee54675732fa66a05ebfe489e27264caeeff3728c945d25971b6485ec"}, + {file = "rich-13.4.2.tar.gz", hash = "sha256:d653d6bccede5844304c605d5aac802c7cf9621efd700b46c7ec2b51ea914898"}, +] + +[package.dependencies] +markdown-it-py = ">=2.2.0" +pygments = ">=2.13.0,<3.0.0" + +[package.extras] +jupyter = ["ipywidgets (>=7.5.1,<9)"] + +[[package]] +name = "scikit-learn" +version = "1.3.0" +description = "A set of python modules for machine learning and data mining" +category = "main" +optional = false +python-versions = ">=3.8" +files = [ + {file = "scikit-learn-1.3.0.tar.gz", hash = "sha256:8be549886f5eda46436b6e555b0e4873b4f10aa21c07df45c4bc1735afbccd7a"}, + {file = "scikit_learn-1.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:981287869e576d42c682cf7ca96af0c6ac544ed9316328fd0d9292795c742cf5"}, + {file = "scikit_learn-1.3.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:436aaaae2c916ad16631142488e4c82f4296af2404f480e031d866863425d2a2"}, + {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c7e28d8fa47a0b30ae1bd7a079519dd852764e31708a7804da6cb6f8b36e3630"}, + {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ae80c08834a473d08a204d966982a62e11c976228d306a2648c575e3ead12111"}, + {file = "scikit_learn-1.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:552fd1b6ee22900cf1780d7386a554bb96949e9a359999177cf30211e6b20df6"}, + {file = "scikit_learn-1.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:79970a6d759eb00a62266a31e2637d07d2d28446fca8079cf9afa7c07b0427f8"}, + {file = "scikit_learn-1.3.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:850a00b559e636b23901aabbe79b73dc604b4e4248ba9e2d6e72f95063765603"}, + {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee04835fb016e8062ee9fe9074aef9b82e430504e420bff51e3e5fffe72750ca"}, + {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9d953531f5d9f00c90c34fa3b7d7cfb43ecff4c605dac9e4255a20b114a27369"}, + {file = "scikit_learn-1.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:151ac2bf65ccf363664a689b8beafc9e6aae36263db114b4ca06fbbbf827444a"}, + {file = "scikit_learn-1.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:6a885a9edc9c0a341cab27ec4f8a6c58b35f3d449c9d2503a6fd23e06bbd4f6a"}, + {file = "scikit_learn-1.3.0-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:9877af9c6d1b15486e18a94101b742e9d0d2f343d35a634e337411ddb57783f3"}, + {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c470f53cea065ff3d588050955c492793bb50c19a92923490d18fcb637f6383a"}, + {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fd6e2d7389542eae01077a1ee0318c4fec20c66c957f45c7aac0c6eb0fe3c612"}, + {file = "scikit_learn-1.3.0-cp38-cp38-win_amd64.whl", hash = "sha256:3a11936adbc379a6061ea32fa03338d4ca7248d86dd507c81e13af428a5bc1db"}, + {file = "scikit_learn-1.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:998d38fcec96584deee1e79cd127469b3ad6fefd1ea6c2dfc54e8db367eb396b"}, + {file = "scikit_learn-1.3.0-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:ded35e810438a527e17623ac6deae3b360134345b7c598175ab7741720d7ffa7"}, + {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0e8102d5036e28d08ab47166b48c8d5e5810704daecf3a476a4282d562be9a28"}, + {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7617164951c422747e7c32be4afa15d75ad8044f42e7d70d3e2e0429a50e6718"}, + {file = "scikit_learn-1.3.0-cp39-cp39-win_amd64.whl", hash = "sha256:1d54fb9e6038284548072df22fd34777e434153f7ffac72c8596f2d6987110dd"}, +] + +[package.dependencies] +joblib = ">=1.1.1" +numpy = ">=1.17.3" +scipy = ">=1.5.0" +threadpoolctl = ">=2.0.0" + +[package.extras] +benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"] +docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=6.0.0)", "sphinx-copybutton (>=0.5.2)", "sphinx-gallery (>=0.10.1)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"] +examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"] +tests = ["black (>=23.3.0)", "matplotlib (>=3.1.3)", "mypy (>=1.3)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=7.1.2)", "pytest-cov (>=2.9.0)", "ruff (>=0.0.272)", "scikit-image (>=0.16.2)"] + +[[package]] +name = "scipy" +version = "1.11.1" +description = "Fundamental algorithms for scientific computing in Python" +category = "main" +optional = false +python-versions = "<3.13,>=3.9" +files = [ + {file = "scipy-1.11.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:aec8c62fbe52914f9cf28d846cf0401dd80ab80788bbab909434eb336ed07c04"}, + {file = "scipy-1.11.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:3b9963798df1d8a52db41a6fc0e6fa65b1c60e85d73da27ae8bb754de4792481"}, + {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3e8eb42db36526b130dfbc417609498a6192381abc1975b91e3eb238e0b41c1a"}, + {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:366a6a937110d80dca4f63b3f5b00cc89d36f678b2d124a01067b154e692bab1"}, + {file = "scipy-1.11.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:08d957ca82d3535b3b9ba6c8ff355d78fe975271874e2af267cb5add5bd78625"}, + {file = "scipy-1.11.1-cp310-cp310-win_amd64.whl", hash = "sha256:e866514bc2d660608447b6ba95c8900d591f2865c07cca0aa4f7ff3c4ca70f30"}, + {file = "scipy-1.11.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ba94eeef3c9caa4cea7b402a35bb02a5714ee1ee77eb98aca1eed4543beb0f4c"}, + {file = "scipy-1.11.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:512fdc18c65f76dadaca139348e525646d440220d8d05f6d21965b8d4466bccd"}, + {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cce154372f0ebe88556ed06d7b196e9c2e0c13080ecb58d0f35062dc7cc28b47"}, + {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b4bb943010203465ac81efa392e4645265077b4d9e99b66cf3ed33ae12254173"}, + {file = "scipy-1.11.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:249cfa465c379c9bb2c20123001e151ff5e29b351cbb7f9c91587260602c58d0"}, + {file = "scipy-1.11.1-cp311-cp311-win_amd64.whl", hash = "sha256:ffb28e3fa31b9c376d0fb1f74c1f13911c8c154a760312fbee87a21eb21efe31"}, + {file = "scipy-1.11.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:39154437654260a52871dfde852adf1b93b1d1bc5dc0ffa70068f16ec0be2624"}, + {file = "scipy-1.11.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:b588311875c58d1acd4ef17c983b9f1ab5391755a47c3d70b6bd503a45bfaf71"}, + {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d51565560565a0307ed06fa0ec4c6f21ff094947d4844d6068ed04400c72d0c3"}, + {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b41a0f322b4eb51b078cb3441e950ad661ede490c3aca66edef66f4b37ab1877"}, + {file = "scipy-1.11.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:396fae3f8c12ad14c5f3eb40499fd06a6fef8393a6baa352a652ecd51e74e029"}, + {file = "scipy-1.11.1-cp39-cp39-win_amd64.whl", hash = "sha256:be8c962a821957fdde8c4044efdab7a140c13294997a407eaee777acf63cbf0c"}, + {file = "scipy-1.11.1.tar.gz", hash = "sha256:fb5b492fa035334fd249f0973cc79ecad8b09c604b42a127a677b45a9a3d4289"}, +] + +[package.dependencies] +numpy = ">=1.21.6,<1.28.0" + +[package.extras] +dev = ["click", "cython-lint (>=0.12.2)", "doit (>=0.36.0)", "mypy", "pycodestyle", "pydevtool", "rich-click", "ruff", "types-psutil", "typing_extensions"] +doc = ["jupytext", "matplotlib (>2)", "myst-nb", "numpydoc", "pooch", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"] +test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"] + +[[package]] +name = "setuptools" +version = "68.0.0" +description = "Easily download, build, install, upgrade, and uninstall Python packages" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "setuptools-68.0.0-py3-none-any.whl", hash = "sha256:11e52c67415a381d10d6b462ced9cfb97066179f0e871399e006c4ab101fc85f"}, + {file = "setuptools-68.0.0.tar.gz", hash = "sha256:baf1fdb41c6da4cd2eae722e135500da913332ab3f2f5c7d33af9b492acb5235"}, +] + +[package.extras] +docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-hoverxref (<2)", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (==0.8.3)", "sphinx-reredirects", "sphinxcontrib-towncrier"] +testing = ["build[virtualenv]", "filelock (>=3.4.0)", "flake8-2020", "ini2toml[lite] (>=0.9)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pip (>=19.1)", "pip-run (>=8.8)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-mypy (>=0.9.1)", "pytest-perf", "pytest-ruff", "pytest-timeout", "pytest-xdist", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel"] +testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pytest", "pytest-enabler", "pytest-xdist", "tomli", "virtualenv (>=13.0.0)", "wheel"] + +[[package]] +name = "six" +version = "1.16.0" +description = "Python 2 and 3 compatibility utilities" +category = "main" +optional = false +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "six-1.16.0-py2.py3-none-any.whl", hash = "sha256:8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254"}, + {file = "six-1.16.0.tar.gz", hash = "sha256:1e61c37477a1626458e36f7b1d82aa5c9b094fa4802892072e49de9c60c4c926"}, +] + +[[package]] +name = "smmap" +version = "5.0.0" +description = "A pure Python implementation of a sliding window memory map manager" +category = "dev" +optional = false +python-versions = ">=3.6" +files = [ + {file = "smmap-5.0.0-py3-none-any.whl", hash = "sha256:2aba19d6a040e78d8b09de5c57e96207b09ed71d8e55ce0959eeee6c8e190d94"}, + {file = "smmap-5.0.0.tar.gz", hash = "sha256:c840e62059cd3be204b0c9c9f74be2c09d5648eddd4580d9314c3ecde0b30936"}, +] + +[[package]] +name = "sniffio" +version = "1.3.0" +description = "Sniff out which async library your code is running under" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "sniffio-1.3.0-py3-none-any.whl", hash = "sha256:eecefdce1e5bbfb7ad2eeaabf7c1eeb404d7757c379bd1f7e5cce9d8bf425384"}, + {file = "sniffio-1.3.0.tar.gz", hash = "sha256:e60305c5e5d314f5389259b7f22aaa33d8f7dee49763119234af3755c55b9101"}, +] + +[[package]] +name = "sortedcontainers" +version = "2.4.0" +description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"}, + {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"}, +] + +[[package]] +name = "soundfile" +version = "0.12.1" +description = "An audio library based on libsndfile, CFFI and NumPy" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"}, + {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"}, + {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"}, + {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"}, + {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"}, +] + +[package.dependencies] +cffi = ">=1.0" + +[package.extras] +numpy = ["numpy"] + +[[package]] +name = "soxr" +version = "0.3.5" +description = "High quality, one-dimensional sample-rate conversion library" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "soxr-0.3.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:21c3aa3b2e12351b4310eea9d56cf52ec0769e6832f911ee6ba32f85b7c92baa"}, + {file = "soxr-0.3.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ac3d7abc96082ff18a31fb1d678ddc0562f0c5e6d91f1cf0024b044989f63e93"}, + {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:145e1e9d1b873a59ce0b5aa463ccacc40cf4bb74d9d8e6cef23433c752bfecea"}, + {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4a376b3678801ffc1d0b9ae918b958be29d5884ca1b4bbeab32e29c567723bb3"}, + {file = "soxr-0.3.5-cp310-cp310-win32.whl", hash = "sha256:907e2eb176bdefec40cc8f6015b7cef7f3d525a34219b3580b603ee696cb25c6"}, + {file = "soxr-0.3.5-cp310-cp310-win_amd64.whl", hash = "sha256:0a6dbf9c7b7a3642916aba264c1d0b872b2e173be56204ed1895dbe381a32077"}, + {file = "soxr-0.3.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:22c08a41e8eee99241fc0e9afb510f9bc7ada4a149d469b8891b596281a27db3"}, + {file = "soxr-0.3.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bdacbe4ce4a1001043f1f8f0744480e294f5c5106e7861fd7033a83a869ba371"}, + {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4b9acd5c42159eac4a90807524d9aa450d6ea0c750df94455c151165896d922e"}, + {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:44b5d30f4e0d98b6d0034c00b04d5571ad070ce5cf3772f93193095b01b373de"}, + {file = "soxr-0.3.5-cp311-cp311-win32.whl", hash = "sha256:677d5f44e85fdf0fdef33cd0e6087470732dd2e08fa73286c3659814110d1183"}, + {file = "soxr-0.3.5-cp311-cp311-win_amd64.whl", hash = "sha256:a479984dd17bf0b50fb9fd659eba54a2dc59bf6eba9c29bb3a4a79ecec7dc9a4"}, + {file = "soxr-0.3.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a2eb4f273ca14d7cfa882b234a03497d0e5dfd6f769a488a0962fe500450838c"}, + {file = "soxr-0.3.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:a254c5e1adddb1204d8f327158b6c11a854908a10b5782103f38a67156108334"}, + {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5766727dfee4d3616edd2a866a9a0d2f272c01545bed165c5a2676fbfd278723"}, + {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2578664c6f94329685d864cdcae59794121bcbd808441572b2ffd01e7adc45dd"}, + {file = "soxr-0.3.5-cp38-cp38-win32.whl", hash = "sha256:8a6f03804f48d986610eab8ca2b52e50b495f40ec13507741cd95f00ef7c2cb6"}, + {file = "soxr-0.3.5-cp38-cp38-win_amd64.whl", hash = "sha256:592e9393e433501769a7e36b10460f4578c8e4ec3cddeec1aaaea4688e3558ef"}, + {file = "soxr-0.3.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:93adbf04f51c7a5113059395633c2647f73bf195fa820256e1dd4da78af59275"}, + {file = "soxr-0.3.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:37c4ec7ce275f284b0bf9741e5e6844a211ba1a850b2bf1c6a47769cdd3d109e"}, + {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:18d5f3151fe4a88dfc37447bc6c397072aedcf36aeffb325cc817350ac5ad78e"}, + {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:549a8358ba3b99a75588453c96aaa802e0c84d40957bdbe1f820f14f83a052ca"}, + {file = "soxr-0.3.5-cp39-cp39-win32.whl", hash = "sha256:799df1875803dc9c4a4d3a7c285b8c1cb34b40dc39dba7ac7bac85d072f936a5"}, + {file = "soxr-0.3.5-cp39-cp39-win_amd64.whl", hash = "sha256:4dd3f61929eb304c109f1f3b6cc8243e3a1a46d636d5bd86b5a7f50609ecd7d6"}, + {file = "soxr-0.3.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:028af32bd4ce4b4c8183bb36da99e23ae954a114034d74538b4cae1bf40a0555"}, + {file = "soxr-0.3.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1299e2aae4d659e222bcbbaca69a51ee99571486070ed49a393725ea6010a8e9"}, + {file = "soxr-0.3.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:162f4e8b9a014c6819b4db6def2d43f7f4d97432ae33f2edfc8e5d0c97cf1cb3"}, + {file = "soxr-0.3.5.tar.gz", hash = "sha256:b6b60f6381c98249a2f2a594e9234b647b78856c76c060597d53ed27b6efd249"}, +] + +[package.dependencies] +numpy = "*" + +[package.extras] +docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"] +test = ["pytest"] + +[[package]] +name = "starlette" +version = "0.28.0" +description = "The little ASGI library that shines." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "starlette-0.28.0-py3-none-any.whl", hash = "sha256:e58b9fc402c579950260fbb6d57173395c4e62804c40d3ede7e9ef1074f0c579"}, + {file = "starlette-0.28.0.tar.gz", hash = "sha256:7bf3da5e997e796cc202cef2bd3f96a7d9b1e1943203c2fe2b42e020bc658482"}, +] + +[package.dependencies] +anyio = ">=3.4.0,<5" +typing-extensions = {version = ">=3.10.0", markers = "python_version < \"3.10\""} + +[package.extras] +full = ["httpx (>=0.22.0)", "itsdangerous", "jinja2", "python-multipart", "pyyaml"] + +[[package]] +name = "starlette-prometheus" +version = "0.9.0" +description = "Prometheus integration for Starlette" +category = "main" +optional = false +python-versions = ">=3.7,<4.0" +files = [ + {file = "starlette-prometheus-0.9.0.tar.gz", hash = "sha256:a52fb0f1df52b44a7a677a792759337ef0ce0d59ddf3e684a7d6459a93a90e99"}, + {file = "starlette_prometheus-0.9.0-py3-none-any.whl", hash = "sha256:b4702e4ec67dce508d28551db0e45f12f58411afdb5d1078c92ff74331915381"}, +] + +[package.dependencies] +prometheus_client = ">=0.12,<0.13" +starlette = ">=0.12.2" + +[[package]] +name = "stevedore" +version = "5.1.0" +description = "Manage dynamic plugins for Python applications" +category = "dev" +optional = false +python-versions = ">=3.8" +files = [ + {file = "stevedore-5.1.0-py3-none-any.whl", hash = "sha256:8cc040628f3cea5d7128f2e76cf486b2251a4e543c7b938f58d9a377f6694a2d"}, + {file = "stevedore-5.1.0.tar.gz", hash = "sha256:a54534acf9b89bc7ed264807013b505bf07f74dbe4bcfa37d32bd063870b087c"}, +] + +[package.dependencies] +pbr = ">=2.0.0,<2.1.0 || >2.1.0" + +[[package]] +name = "threadpoolctl" +version = "3.1.0" +description = "threadpoolctl" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, + {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"}, +] + +[[package]] +name = "toml" +version = "0.10.2" +description = "Python Library for Tom's Obvious, Minimal Language" +category = "dev" +optional = false +python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*" +files = [ + {file = "toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b"}, + {file = "toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f"}, +] + +[[package]] +name = "tomli" +version = "2.0.1" +description = "A lil' TOML parser" +category = "dev" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tomli-2.0.1-py3-none-any.whl", hash = "sha256:939de3e7a6161af0c887ef91b7d41a53e7c5a1ca976325f429cb46ea9bc30ecc"}, + {file = "tomli-2.0.1.tar.gz", hash = "sha256:de526c12914f0c550d15924c62d72abc48d6fe7364aa87328337a31007fe8a4f"}, +] + +[[package]] +name = "tqdm" +version = "4.65.0" +description = "Fast, Extensible Progress Meter" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"}, + {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"}, +] + +[package.dependencies] +colorama = {version = "*", markers = "platform_system == \"Windows\""} + +[package.extras] +dev = ["py-make (>=0.1.0)", "twine", "wheel"] +notebook = ["ipywidgets (>=6)"] +slack = ["slack-sdk"] +telegram = ["requests"] + +[[package]] +name = "types-requests" +version = "2.31.0.1" +description = "Typing stubs for requests" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-requests-2.31.0.1.tar.gz", hash = "sha256:3de667cffa123ce698591de0ad7db034a5317457a596eb0b4944e5a9d9e8d1ac"}, + {file = "types_requests-2.31.0.1-py3-none-any.whl", hash = "sha256:afb06ef8f25ba83d59a1d424bd7a5a939082f94b94e90ab5e6116bd2559deaa3"}, +] + +[package.dependencies] +types-urllib3 = "*" + +[[package]] +name = "types-urllib3" +version = "1.26.25.13" +description = "Typing stubs for urllib3" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-urllib3-1.26.25.13.tar.gz", hash = "sha256:3300538c9dc11dad32eae4827ac313f5d986b8b21494801f1bf97a1ac6c03ae5"}, + {file = "types_urllib3-1.26.25.13-py3-none-any.whl", hash = "sha256:5dbd1d2bef14efee43f5318b5d36d805a489f6600252bb53626d4bfafd95e27c"}, +] + +[[package]] +name = "typing-extensions" +version = "4.7.1" +description = "Backported and Experimental Type Hints for Python 3.7+" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "typing_extensions-4.7.1-py3-none-any.whl", hash = "sha256:440d5dd3af93b060174bf433bccd69b0babc3b15b1a8dca43789fd7f61514b36"}, + {file = "typing_extensions-4.7.1.tar.gz", hash = "sha256:b75ddc264f0ba5615db7ba217daeb99701ad295353c45f9e95963337ceeeffb2"}, +] + +[[package]] +name = "tzdata" +version = "2023.3" +description = "Provider of IANA time zone data" +category = "main" +optional = false +python-versions = ">=2" +files = [ + {file = "tzdata-2023.3-py2.py3-none-any.whl", hash = "sha256:7e65763eef3120314099b6939b5546db7adce1e7d6f2e179e3df563c70511eda"}, + {file = "tzdata-2023.3.tar.gz", hash = "sha256:11ef1e08e54acb0d4f95bdb1be05da659673de4acbd21bf9c69e94cc5e907a3a"}, +] + +[[package]] +name = "urllib3" +version = "2.0.3" +description = "HTTP library with thread-safe connection pooling, file post, and more." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "urllib3-2.0.3-py3-none-any.whl", hash = "sha256:48e7fafa40319d358848e1bc6809b208340fafe2096f1725d05d67443d0483d1"}, + {file = "urllib3-2.0.3.tar.gz", hash = "sha256:bee28b5e56addb8226c96f7f13ac28cb4c301dd5ea8a6ca179c0b9835e032825"}, +] + +[package.extras] +brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)"] +secure = ["certifi", "cryptography (>=1.9)", "idna (>=2.0.0)", "pyopenssl (>=17.1.0)", "urllib3-secure-extra"] +socks = ["pysocks (>=1.5.6,!=1.5.7,<2.0)"] +zstd = ["zstandard (>=0.18.0)"] + +[[package]] +name = "webencodings" +version = "0.5.1" +description = "Character encoding aliases for legacy web content" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78"}, + {file = "webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923"}, +] + +[[package]] +name = "werkzeug" +version = "2.3.6" +description = "The comprehensive WSGI web application library." +category = "dev" +optional = false +python-versions = ">=3.8" +files = [ + {file = "Werkzeug-2.3.6-py3-none-any.whl", hash = "sha256:935539fa1413afbb9195b24880778422ed620c0fc09670945185cce4d91a8890"}, + {file = "Werkzeug-2.3.6.tar.gz", hash = "sha256:98c774df2f91b05550078891dee5f0eb0cb797a522c757a2452b9cee5b202330"}, +] + +[package.dependencies] +MarkupSafe = ">=2.1.1" + +[package.extras] +watchdog = ["watchdog (>=2.3)"] + +[[package]] +name = "xxhash" +version = "3.2.0" +description = "Python binding for xxHash" +category = "main" +optional = false +python-versions = ">=3.6" +files = [ + {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"}, + {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"}, + {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"}, + {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"}, + {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"}, + {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"}, + {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"}, + {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"}, + {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"}, + {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"}, + {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"}, + {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"}, + {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"}, + {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"}, + {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"}, + {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"}, + {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"}, + {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"}, + {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"}, + {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"}, + {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"}, + {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"}, + {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"}, + {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"}, + {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"}, + {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"}, + {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"}, + {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"}, + {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"}, + {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"}, + {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"}, + {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"}, + {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"}, + {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"}, + {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"}, +] + +[[package]] +name = "yarl" +version = "1.9.2" +description = "Yet another URL library" +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:8c2ad583743d16ddbdf6bb14b5cd76bf43b0d0006e918809d5d4ddf7bde8dd82"}, + {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:82aa6264b36c50acfb2424ad5ca537a2060ab6de158a5bd2a72a032cc75b9eb8"}, + {file = "yarl-1.9.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c0c77533b5ed4bcc38e943178ccae29b9bcf48ffd1063f5821192f23a1bd27b9"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee4afac41415d52d53a9833ebae7e32b344be72835bbb589018c9e938045a560"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9bf345c3a4f5ba7f766430f97f9cc1320786f19584acc7086491f45524a551ac"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2a96c19c52ff442a808c105901d0bdfd2e28575b3d5f82e2f5fd67e20dc5f4ea"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:891c0e3ec5ec881541f6c5113d8df0315ce5440e244a716b95f2525b7b9f3608"}, + {file = "yarl-1.9.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c3a53ba34a636a256d767c086ceb111358876e1fb6b50dfc4d3f4951d40133d5"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:566185e8ebc0898b11f8026447eacd02e46226716229cea8db37496c8cdd26e0"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:2b0738fb871812722a0ac2154be1f049c6223b9f6f22eec352996b69775b36d4"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:32f1d071b3f362c80f1a7d322bfd7b2d11e33d2adf395cc1dd4df36c9c243095"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e9fdc7ac0d42bc3ea78818557fab03af6181e076a2944f43c38684b4b6bed8e3"}, + {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:56ff08ab5df8429901ebdc5d15941b59f6253393cb5da07b4170beefcf1b2528"}, + {file = "yarl-1.9.2-cp310-cp310-win32.whl", hash = "sha256:8ea48e0a2f931064469bdabca50c2f578b565fc446f302a79ba6cc0ee7f384d3"}, + {file = "yarl-1.9.2-cp310-cp310-win_amd64.whl", hash = "sha256:50f33040f3836e912ed16d212f6cc1efb3231a8a60526a407aeb66c1c1956dde"}, + {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:646d663eb2232d7909e6601f1a9107e66f9791f290a1b3dc7057818fe44fc2b6"}, + {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aff634b15beff8902d1f918012fc2a42e0dbae6f469fce134c8a0dc51ca423bb"}, + {file = "yarl-1.9.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:a83503934c6273806aed765035716216cc9ab4e0364f7f066227e1aaea90b8d0"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b25322201585c69abc7b0e89e72790469f7dad90d26754717f3310bfe30331c2"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:22a94666751778629f1ec4280b08eb11815783c63f52092a5953faf73be24191"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8ec53a0ea2a80c5cd1ab397925f94bff59222aa3cf9c6da938ce05c9ec20428d"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:159d81f22d7a43e6eabc36d7194cb53f2f15f498dbbfa8edc8a3239350f59fe7"}, + {file = "yarl-1.9.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:832b7e711027c114d79dffb92576acd1bd2decc467dec60e1cac96912602d0e6"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:95d2ecefbcf4e744ea952d073c6922e72ee650ffc79028eb1e320e732898d7e8"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:d4e2c6d555e77b37288eaf45b8f60f0737c9efa3452c6c44626a5455aeb250b9"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:783185c75c12a017cc345015ea359cc801c3b29a2966c2655cd12b233bf5a2be"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:b8cc1863402472f16c600e3e93d542b7e7542a540f95c30afd472e8e549fc3f7"}, + {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:822b30a0f22e588b32d3120f6d41e4ed021806418b4c9f0bc3048b8c8cb3f92a"}, + {file = "yarl-1.9.2-cp311-cp311-win32.whl", hash = "sha256:a60347f234c2212a9f0361955007fcf4033a75bf600a33c88a0a8e91af77c0e8"}, + {file = "yarl-1.9.2-cp311-cp311-win_amd64.whl", hash = "sha256:be6b3fdec5c62f2a67cb3f8c6dbf56bbf3f61c0f046f84645cd1ca73532ea051"}, + {file = "yarl-1.9.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:38a3928ae37558bc1b559f67410df446d1fbfa87318b124bf5032c31e3447b74"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ac9bb4c5ce3975aeac288cfcb5061ce60e0d14d92209e780c93954076c7c4367"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3da8a678ca8b96c8606bbb8bfacd99a12ad5dd288bc6f7979baddd62f71c63ef"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:13414591ff516e04fcdee8dc051c13fd3db13b673c7a4cb1350e6b2ad9639ad3"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf74d08542c3a9ea97bb8f343d4fcbd4d8f91bba5ec9d5d7f792dbe727f88938"}, + {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6e7221580dc1db478464cfeef9b03b95c5852cc22894e418562997df0d074ccc"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:494053246b119b041960ddcd20fd76224149cfea8ed8777b687358727911dd33"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:52a25809fcbecfc63ac9ba0c0fb586f90837f5425edfd1ec9f3372b119585e45"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:e65610c5792870d45d7b68c677681376fcf9cc1c289f23e8e8b39c1485384185"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1b1bba902cba32cdec51fca038fd53f8beee88b77efc373968d1ed021024cc04"}, + {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:662e6016409828ee910f5d9602a2729a8a57d74b163c89a837de3fea050c7582"}, + {file = "yarl-1.9.2-cp37-cp37m-win32.whl", hash = "sha256:f364d3480bffd3aa566e886587eaca7c8c04d74f6e8933f3f2c996b7f09bee1b"}, + {file = "yarl-1.9.2-cp37-cp37m-win_amd64.whl", hash = "sha256:6a5883464143ab3ae9ba68daae8e7c5c95b969462bbe42e2464d60e7e2698368"}, + {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5610f80cf43b6202e2c33ba3ec2ee0a2884f8f423c8f4f62906731d876ef4fac"}, + {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b9a4e67ad7b646cd6f0938c7ebfd60e481b7410f574c560e455e938d2da8e0f4"}, + {file = "yarl-1.9.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:83fcc480d7549ccebe9415d96d9263e2d4226798c37ebd18c930fce43dfb9574"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5fcd436ea16fee7d4207c045b1e340020e58a2597301cfbcfdbe5abd2356c2fb"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:84e0b1599334b1e1478db01b756e55937d4614f8654311eb26012091be109d59"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:3458a24e4ea3fd8930e934c129b676c27452e4ebda80fbe47b56d8c6c7a63a9e"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:838162460b3a08987546e881a2bfa573960bb559dfa739e7800ceeec92e64417"}, + {file = "yarl-1.9.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f4e2d08f07a3d7d3e12549052eb5ad3eab1c349c53ac51c209a0e5991bbada78"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:de119f56f3c5f0e2fb4dee508531a32b069a5f2c6e827b272d1e0ff5ac040333"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:149ddea5abf329752ea5051b61bd6c1d979e13fbf122d3a1f9f0c8be6cb6f63c"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:674ca19cbee4a82c9f54e0d1eee28116e63bc6fd1e96c43031d11cbab8b2afd5"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:9b3152f2f5677b997ae6c804b73da05a39daa6a9e85a512e0e6823d81cdad7cc"}, + {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:5415d5a4b080dc9612b1b63cba008db84e908b95848369aa1da3686ae27b6d2b"}, + {file = "yarl-1.9.2-cp38-cp38-win32.whl", hash = "sha256:f7a3d8146575e08c29ed1cd287068e6d02f1c7bdff8970db96683b9591b86ee7"}, + {file = "yarl-1.9.2-cp38-cp38-win_amd64.whl", hash = "sha256:63c48f6cef34e6319a74c727376e95626f84ea091f92c0250a98e53e62c77c72"}, + {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:75df5ef94c3fdc393c6b19d80e6ef1ecc9ae2f4263c09cacb178d871c02a5ba9"}, + {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c027a6e96ef77d401d8d5a5c8d6bc478e8042f1e448272e8d9752cb0aff8b5c8"}, + {file = "yarl-1.9.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f3b078dbe227f79be488ffcfc7a9edb3409d018e0952cf13f15fd6512847f3f7"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:59723a029760079b7d991a401386390c4be5bfec1e7dd83e25a6a0881859e716"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b03917871bf859a81ccb180c9a2e6c1e04d2f6a51d953e6a5cdd70c93d4e5a2a"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c1012fa63eb6c032f3ce5d2171c267992ae0c00b9e164efe4d73db818465fac3"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a74dcbfe780e62f4b5a062714576f16c2f3493a0394e555ab141bf0d746bb955"}, + {file = "yarl-1.9.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8c56986609b057b4839968ba901944af91b8e92f1725d1a2d77cbac6972b9ed1"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:2c315df3293cd521033533d242d15eab26583360b58f7ee5d9565f15fee1bef4"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:b7232f8dfbd225d57340e441d8caf8652a6acd06b389ea2d3222b8bc89cbfca6"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:53338749febd28935d55b41bf0bcc79d634881195a39f6b2f767870b72514caf"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:066c163aec9d3d073dc9ffe5dd3ad05069bcb03fcaab8d221290ba99f9f69ee3"}, + {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:8288d7cd28f8119b07dd49b7230d6b4562f9b61ee9a4ab02221060d21136be80"}, + {file = "yarl-1.9.2-cp39-cp39-win32.whl", hash = "sha256:b124e2a6d223b65ba8768d5706d103280914d61f5cae3afbc50fc3dfcc016623"}, + {file = "yarl-1.9.2-cp39-cp39-win_amd64.whl", hash = "sha256:61016e7d582bc46a5378ffdd02cd0314fb8ba52f40f9cf4d9a5e7dbef88dee18"}, + {file = "yarl-1.9.2.tar.gz", hash = "sha256:04ab9d4b9f587c06d801c2abfe9317b77cdf996c65a90d5e84ecc45010823571"}, +] + +[package.dependencies] +idna = ">=2.0" +multidict = ">=4.0" + +[metadata] +lock-version = "2.0" +python-versions = "3.9.15" +content-hash = "9f4fbf1058f6ec31e437c39104086dac3ead6792fb7b266eca4f5594ea5cd24c" diff --git a/libs/libapi/poetry.toml b/libs/libapi/poetry.toml new file mode 100644 index 00000000..5fcef8cd --- /dev/null +++ b/libs/libapi/poetry.toml @@ -0,0 +1,3 @@ +[virtualenvs] +in-project = true +prefer-active-python = true diff --git a/libs/libapi/pyproject.toml b/libs/libapi/pyproject.toml new file mode 100644 index 00000000..adc11924 --- /dev/null +++ b/libs/libapi/pyproject.toml @@ -0,0 +1,54 @@ +[tool.poetry] +authors = ["The HuggingFace Authors."] +description = "Library for the API services" +name = "libapi" +version = "0.1.0" +license = "Apache-2.0" + +[tool.poetry.dependencies] +cryptography = "^41.0.1" +environs = "^9.5.0" +libcommon = {path = "../../libs/libcommon", develop = true} +orjson = "^3.8.6" +pyjwt = { extras = ["crypto"], version = "^2.6.0" } +python = "3.9.15" +requests = "^2.28.2" +starlette = "^0.28.0" +starlette-prometheus = "^0.9.0" + +[tool.poetry.group.dev.dependencies] +bandit = "^1.7.4" +black = "^22.12.0" +flake8 = "^3.9.2" +isort = "^5.12.0" +mypy = "^1.0.0" +pip-audit = "^2.5.4" +pytest = "^7.2.1" +pytest-cov = "^2.12.1" +pytest-httpserver = "^1.0.6" +types-requests = "^2.28.11" + +[build-system] +build-backend = "poetry.core.masonry.api" +requires = ["poetry-core>=1.0.0"] + +[tool.pytest.ini_options] +filterwarnings = ["ignore::DeprecationWarning"] +[tool.coverage.run] +source = ["libapi"] + +[tool.isort] +profile = "black" + +[tool.black] +line-length = 119 +preview = true + +[tool.mypy] +strict = true + +[[tool.mypy.overrides]] +module = [ + "prometheus_client.*", +] +ignore_missing_imports = true diff --git a/libs/libapi/src/libapi/__init__.py b/libs/libapi/src/libapi/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/libs/libapi/src/libapi/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/services/api/src/api/authentication.py b/libs/libapi/src/libapi/authentication.py similarity index 98% rename from services/api/src/api/authentication.py rename to libs/libapi/src/libapi/authentication.py index 1308657d..b423a07a 100644 --- a/services/api/src/api/authentication.py +++ b/libs/libapi/src/libapi/authentication.py @@ -13,2 +13 @@ from starlette.requests import Request -from api.jwt_token import is_jwt_valid -from api.utils import ( +from libapi.exceptions import ( @@ -18,0 +18 @@ from api.utils import ( +from libapi.jwt_token import is_jwt_valid diff --git a/libs/libapi/src/libapi/config.py b/libs/libapi/src/libapi/config.py new file mode 100644 index 00000000..a533253a --- /dev/null +++ b/libs/libapi/src/libapi/config.py @@ -0,0 +1,67 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from dataclasses import dataclass +from typing import Optional + +from environs import Env + +API_UVICORN_HOSTNAME = "localhost" +API_UVICORN_NUM_WORKERS = 2 +API_UVICORN_PORT = 8000 + + +@dataclass(frozen=True) +class UvicornConfig: + hostname: str = API_UVICORN_HOSTNAME + num_workers: int = API_UVICORN_NUM_WORKERS + port: int = API_UVICORN_PORT + + @classmethod + def from_env(cls) -> "UvicornConfig": + env = Env(expand_vars=True) + with env.prefixed("API_UVICORN_"): + return cls( + hostname=env.str(name="HOSTNAME", default=API_UVICORN_HOSTNAME), + num_workers=env.int(name="NUM_WORKERS", default=API_UVICORN_NUM_WORKERS), + port=env.int(name="PORT", default=API_UVICORN_PORT), + ) + + +API_EXTERNAL_AUTH_URL = None +API_HF_AUTH_PATH = "/api/datasets/%s/auth-check" +API_HF_JWT_PUBLIC_KEY_URL = None +API_HF_JWT_ALGORITHM = "EdDSA" +API_HF_TIMEOUT_SECONDS = 0.2 +API_HF_WEBHOOK_SECRET = None +API_MAX_AGE_LONG = 120 # 2 minutes +API_MAX_AGE_SHORT = 10 # 10 seconds + + +@dataclass(frozen=True) +class ApiConfig: + external_auth_url: Optional[str] = API_EXTERNAL_AUTH_URL # not documented + hf_auth_path: str = API_HF_AUTH_PATH + hf_jwt_public_key_url: Optional[str] = API_HF_JWT_PUBLIC_KEY_URL + hf_jwt_algorithm: Optional[str] = API_HF_JWT_ALGORITHM + hf_timeout_seconds: Optional[float] = API_HF_TIMEOUT_SECONDS + hf_webhook_secret: Optional[str] = API_HF_WEBHOOK_SECRET + max_age_long: int = API_MAX_AGE_LONG + max_age_short: int = API_MAX_AGE_SHORT + + @classmethod + def from_env(cls, hf_endpoint: str) -> "ApiConfig": + env = Env(expand_vars=True) + with env.prefixed("API_"): + hf_auth_path = env.str(name="HF_AUTH_PATH", default=API_HF_AUTH_PATH) + external_auth_url = None if hf_auth_path is None else f"{hf_endpoint}{hf_auth_path}" + return cls( + external_auth_url=external_auth_url, + hf_auth_path=hf_auth_path, + hf_jwt_public_key_url=env.str(name="HF_JWT_PUBLIC_KEY_URL", default=API_HF_JWT_PUBLIC_KEY_URL), + hf_jwt_algorithm=env.str(name="HF_JWT_ALGORITHM", default=API_HF_JWT_ALGORITHM), + hf_timeout_seconds=env.float(name="HF_TIMEOUT_SECONDS", default=API_HF_TIMEOUT_SECONDS), + hf_webhook_secret=env.str(name="HF_WEBHOOK_SECRET", default=API_HF_WEBHOOK_SECRET), + max_age_long=env.int(name="MAX_AGE_LONG", default=API_MAX_AGE_LONG), + max_age_short=env.int(name="MAX_AGE_SHORT", default=API_MAX_AGE_SHORT), + ) diff --git a/libs/libapi/src/libapi/exceptions.py b/libs/libapi/src/libapi/exceptions.py new file mode 100644 index 00000000..b867f9de --- /dev/null +++ b/libs/libapi/src/libapi/exceptions.py @@ -0,0 +1,106 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging +from http import HTTPStatus +from typing import Literal, Optional + +from libcommon.exceptions import CustomError + +ApiErrorCode = Literal[ + "AuthCheckHubRequestError", + "ExternalAuthenticatedError", + "ExternalUnauthenticatedError", + "InvalidParameter", + "JWKError", + "MissingProcessingStepsError", + "MissingRequiredParameter", + "ResponseNotFound", + "ResponseNotReady", + "UnexpectedApiError", +] + + +class ApiError(CustomError): + """Base class for exceptions raised by an API service.""" + + def __init__( + self, + message: str, + status_code: HTTPStatus, + code: ApiErrorCode, + cause: Optional[BaseException] = None, + disclose_cause: bool = False, + ): + super().__init__( + message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause + ) + + +class AuthCheckHubRequestError(ApiError): + """Raised when the external authentication check failed or timed out.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__( + message, HTTPStatus.INTERNAL_SERVER_ERROR, "AuthCheckHubRequestError", cause=cause, disclose_cause=False + ) + + +class ExternalAuthenticatedError(ApiError): + """Raised when the external authentication check failed while the user was authenticated. + + Even if the external authentication server returns 403 in that case, we return 404 because + we don't know if the dataset exist or not. It's also coherent with how the Hugging Face Hub works.""" + + def __init__(self, message: str): + super().__init__(message, HTTPStatus.NOT_FOUND, "ExternalAuthenticatedError") + + +class ExternalUnauthenticatedError(ApiError): + """Raised when the external authentication check failed while the user was unauthenticated.""" + + def __init__(self, message: str): + super().__init__(message, HTTPStatus.UNAUTHORIZED, "ExternalUnauthenticatedError") + + +class InvalidParameterError(ApiError): + """Raised when a parameter has an invalid value.""" + + def __init__(self, message: str): + super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "InvalidParameter") + + +class JWKError(ApiError): + """Raised when the JWT key (JWK) could not be fetched or parsed.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "JWKError", cause=cause, disclose_cause=False) + + +class MissingRequiredParameterError(ApiError): + """Raised when a required parameter is missing.""" + + def __init__(self, message: str): + super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "MissingRequiredParameter") + + +class ResponseNotFoundError(ApiError): + """Raised when the response has not been found.""" + + def __init__(self, message: str): + super().__init__(message, HTTPStatus.NOT_FOUND, "ResponseNotFound") + + +class ResponseNotReadyError(ApiError): + """Raised when the response has not been processed yet.""" + + def __init__(self, message: str): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseNotReady") + + +class UnexpectedApiError(ApiError): + """Raised when the server raised an unexpected error.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + logging.error(message, exc_info=cause) + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "UnexpectedApiError", cause) diff --git a/services/api/src/api/jwt_token.py b/libs/libapi/src/libapi/jwt_token.py similarity index 99% rename from services/api/src/api/jwt_token.py rename to libs/libapi/src/libapi/jwt_token.py index 26d7ae1f..1fcf3707 100644 --- a/services/api/src/api/jwt_token.py +++ b/libs/libapi/src/libapi/jwt_token.py @@ -31 +31 @@ from jwt.algorithms import ( -from api.utils import JWKError +from libapi.exceptions import JWKError diff --git a/libs/libapi/src/libapi/py.typed b/libs/libapi/src/libapi/py.typed new file mode 100644 index 00000000..e69de29b diff --git a/libs/libapi/src/libapi/routes/__init__.py b/libs/libapi/src/libapi/routes/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/libs/libapi/src/libapi/routes/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/services/api/src/api/routes/healthcheck.py b/libs/libapi/src/libapi/routes/healthcheck.py similarity index 100% rename from services/api/src/api/routes/healthcheck.py rename to libs/libapi/src/libapi/routes/healthcheck.py diff --git a/services/api/src/api/routes/metrics.py b/libs/libapi/src/libapi/routes/metrics.py similarity index 94% rename from services/api/src/api/routes/metrics.py rename to libs/libapi/src/libapi/routes/metrics.py index ef637fe4..92ffc9bc 100644 --- a/services/api/src/api/routes/metrics.py +++ b/libs/libapi/src/libapi/routes/metrics.py @@ -11 +11 @@ from starlette.responses import Response -from api.utils import Endpoint +from libapi.utils import Endpoint diff --git a/services/api/src/api/utils.py b/libs/libapi/src/libapi/utils.py similarity index 54% rename from services/api/src/api/utils.py rename to libs/libapi/src/libapi/utils.py index f6defeee..2f2480de 100644 --- a/services/api/src/api/utils.py +++ b/libs/libapi/src/libapi/utils.py @@ -4 +3,0 @@ -import logging @@ -6 +5 @@ from http import HTTPStatus -from typing import Any, Callable, Coroutine, List, Literal, Optional +from typing import Any, Callable, Coroutine, List, Optional @@ -16,95 +15 @@ from starlette.responses import JSONResponse, Response -ApiErrorCode = Literal[ - "AuthCheckHubRequestError", - "ExternalAuthenticatedError", - "ExternalUnauthenticatedError", - "InvalidParameter", - "JWKError", - "MissingRequiredParameter", - "MissingProcessingStepsError", - "ResponseNotFound", - "ResponseNotReady", - "UnexpectedError", # also in libcommon.exceptions -] - - -class ApiCustomError(CustomError): - """Base class for exceptions in this module.""" - - def __init__( - self, - message: str, - status_code: HTTPStatus, - code: ApiErrorCode, - cause: Optional[BaseException] = None, - disclose_cause: bool = False, - ): - super().__init__(message, status_code, str(code), cause, disclose_cause) - - -class MissingRequiredParameterError(ApiCustomError): - """Raised when a required parameter is missing.""" - - def __init__(self, message: str): - super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "MissingRequiredParameter") - - -class InvalidParameterError(ApiCustomError): - """Raised when a parameter has an invalid value.""" - - def __init__(self, message: str): - super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "InvalidParameter") - - -class ResponseNotReadyError(ApiCustomError): - """Raised when the response has not been processed yet.""" - - def __init__(self, message: str): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseNotReady") - - -class ResponseNotFoundError(ApiCustomError): - """Raised when the response has not been found.""" - - def __init__(self, message: str): - super().__init__(message, HTTPStatus.NOT_FOUND, "ResponseNotFound") - - -class UnexpectedError(ApiCustomError): - """Raised when the server raised an unexpected error.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "UnexpectedError", cause) - logging.error(message, exc_info=cause) - - -class ExternalUnauthenticatedError(ApiCustomError): - """Raised when the external authentication check failed while the user was unauthenticated.""" - - def __init__(self, message: str): - super().__init__(message, HTTPStatus.UNAUTHORIZED, "ExternalUnauthenticatedError") - - -class ExternalAuthenticatedError(ApiCustomError): - """Raised when the external authentication check failed while the user was authenticated. - - Even if the external authentication server returns 403 in that case, we return 404 because - we don't know if the dataset exist or not. It's also coherent with how the Hugging Face Hub works.""" - - def __init__(self, message: str): - super().__init__(message, HTTPStatus.NOT_FOUND, "ExternalAuthenticatedError") - - -class JWKError(ApiCustomError): - """Raised when the JWT key (JWK) could not be fetched or parsed.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "JWKError", cause=cause, disclose_cause=False) - - -class AuthCheckHubRequestError(ApiCustomError): - """Raised when the external authentication check failed or timed out.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__( - message, HTTPStatus.INTERNAL_SERVER_ERROR, "AuthCheckHubRequestError", cause=cause, disclose_cause=False - ) +from libapi.exceptions import ResponseNotFoundError, ResponseNotReadyError @@ -154 +59 @@ def get_json_error_response( -def get_json_api_error_response(error: ApiCustomError, max_age: int = 0, revision: Optional[str] = None) -> Response: +def get_json_api_error_response(error: CustomError, max_age: int = 0, revision: Optional[str] = None) -> Response: diff --git a/libs/libapi/tests/__init__.py b/libs/libapi/tests/__init__.py new file mode 100644 index 00000000..fa0c50f2 --- /dev/null +++ b/libs/libapi/tests/__init__.py @@ -0,0 +1,2 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. diff --git a/libs/libapi/tests/conftest.py b/libs/libapi/tests/conftest.py new file mode 100644 index 00000000..dcd155a6 --- /dev/null +++ b/libs/libapi/tests/conftest.py @@ -0,0 +1,36 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from pytest import fixture + +from libapi.config import ApiConfig + + +@fixture(scope="session") +def hostname() -> str: + return "localhost" + + +@fixture(scope="session") +def port() -> str: + return "8888" + + +@fixture(scope="session") +def httpserver_listen_address(hostname: str, port: int) -> tuple[str, int]: + return (hostname, port) + + +@fixture(scope="session") +def hf_endpoint(hostname: str, port: int) -> str: + return f"http://{hostname}:{port}" + + +@fixture(scope="session") +def api_config(hf_endpoint: str) -> ApiConfig: + return ApiConfig.from_env(hf_endpoint=hf_endpoint) + + +@fixture(scope="session") +def hf_auth_path(api_config: ApiConfig) -> str: + return api_config.hf_auth_path diff --git a/services/api/tests/test_authentication.py b/libs/libapi/tests/test_authentication.py similarity index 78% rename from services/api/tests/test_authentication.py rename to libs/libapi/tests/test_authentication.py index ab32c1dd..e0451dd6 100644 --- a/services/api/tests/test_authentication.py +++ b/libs/libapi/tests/test_authentication.py @@ -3,0 +4 @@ +import datetime @@ -17,2 +18,2 @@ from werkzeug.wrappers import Response as WerkzeugResponse -from api.authentication import auth_check -from api.utils import ( +from libapi.authentication import auth_check +from libapi.exceptions import ( @@ -24,8 +25,31 @@ from api.utils import ( -from .test_jwt_token import ( - algorithm_rs256, - dataset_ok, - payload_ok, - private_key, - public_key, -) -from .utils import auth_callback +private_key = """-----BEGIN RSA PRIVATE KEY----- +MIIBOQIBAAJAZTmplhS/Jd73ycVut7TglMObheQqXM7RZYlwazLU4wpfIVIwOh9I +sCZGSgLyFq42KWIikKLEs/yqx3pRGfq+rwIDAQABAkAMyF9WCICq86Eu5bO5lynV +H26AVfPTjHp87AI6R00C7p9n8hO/DhHaHpc3InOSsXsw9d2hmz37jwwBFiwMHMMh +AiEAtbttHlIO+yO29oXw4P6+yO11lMy1UpT1sPVTnR9TXbUCIQCOl7Zuyy2ZY9ZW +pDhW91x/14uXjnLXPypgY9bcfggJUwIhAJQG1LzrzjQWRUPMmgZKuhBkC3BmxhM8 +LlwzmCXVjEw5AiA7JnAFEb9+q82T71d3q/DxD0bWvb6hz5ASoBfXK2jGBQIgbaQp +h4Tk6UJuj1xgKNs75Pk3pG2tj8AQiuBk3l62vRU= +-----END RSA PRIVATE KEY-----""" +public_key = """-----BEGIN PUBLIC KEY----- +MFswDQYJKoZIhvcNAQEBBQADSgAwRwJAZTmplhS/Jd73ycVut7TglMObheQqXM7R +ZYlwazLU4wpfIVIwOh9IsCZGSgLyFq42KWIikKLEs/yqx3pRGfq+rwIDAQAB +-----END PUBLIC KEY-----""" + +dataset_ok = "dataset" +exp_ok = datetime.datetime.now().timestamp() + 1000 +read_ok = True +sub_ok = f"datasets/{dataset_ok}" +payload_ok = {"sub": sub_ok, "read": read_ok, "exp": exp_ok} +algorithm_rs256 = "RS256" + + +def auth_callback(request: WerkzeugRequest) -> WerkzeugResponse: + # return 401 if a cookie has been provided, 404 if a token has been provided, + # and 200 if none has been provided + # + # caveat: the returned status codes don't simulate the reality + # they're just used to check every case + return WerkzeugResponse( + status=401 if request.headers.get("cookie") else 404 if request.headers.get("authorization") else 200 + ) diff --git a/services/api/tests/test_jwt_token.py b/libs/libapi/tests/test_jwt_token.py similarity index 98% rename from services/api/tests/test_jwt_token.py rename to libs/libapi/tests/test_jwt_token.py index 4a9f6835..0060cfb0 100644 --- a/services/api/tests/test_jwt_token.py +++ b/libs/libapi/tests/test_jwt_token.py @@ -11 +11 @@ import pytest -from api.jwt_token import is_jwt_valid, parse_jwt_public_key +from libapi.jwt_token import is_jwt_valid, parse_jwt_public_key diff --git a/services/api/Dockerfile b/services/api/Dockerfile index 9a42a9d5..437a2367 100644 --- a/services/api/Dockerfile +++ b/services/api/Dockerfile @@ -28,0 +29 @@ COPY libs/libcommon ./libs/libcommon +COPY libs/libapi ./libs/libapi diff --git a/services/api/README.md b/services/api/README.md index 3a147851..45239b60 100644 --- a/services/api/README.md +++ b/services/api/README.md @@ -3 +3 @@ -> API to get the first rows of 🤗 datasets +> API on 🤗 datasets @@ -11,21 +11 @@ The worker can be configured using environment variables. They are grouped by sc -Set environment variables to configure the application (`API_` prefix): - -- `API_HF_AUTH_PATH`: the path of the external authentication service, on the hub (see `HF_ENDPOINT`). The string must contain `%s` which will be replaced with the dataset name. The external authentication service must return 200, 401, 403 or 404. Defaults to "/api/datasets/%s/auth-check". -- `API_HF_JWT_PUBLIC_KEY_URL`: the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format. It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored. Defaults to empty. -- `API_HF_JWT_ALGORITHM`: the algorithm used to encode the JWT. Defaults to `"EdDSA"`. -- `API_HF_TIMEOUT_SECONDS`: the timeout in seconds for the requests to the Hugging Face Hub. Defaults to `0.2` (200 ms). -- `API_HF_WEBHOOK_SECRET`: a shared secret sent by the Hub in the "X-Webhook-Secret" header of POST requests sent to /webhook, to authenticate the originator and bypass some validation of the content (avoiding roundtrip to the Hub). If not set, all the validations are done. Defaults to empty. -- `API_MAX_AGE_LONG`: number of seconds to set in the `max-age` header on data endpoints. Defaults to `120` (2 minutes). -- `API_MAX_AGE_SHORT`: number of seconds to set in the `max-age` header on technical endpoints. Defaults to `10` (10 seconds). - -### Uvicorn - -The following environment variables are used to configure the Uvicorn server (`API_UVICORN_` prefix): - -- `API_UVICORN_HOSTNAME`: the hostname. Defaults to `"localhost"`. -- `API_UVICORN_NUM_WORKERS`: the number of uvicorn workers. Defaults to `2`. -- `API_UVICORN_PORT`: the port. Defaults to `8000`. - -### Prometheus - -- `PROMETHEUS_MULTIPROC_DIR`: the directory where the uvicorn workers share their prometheus metrics. See https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn. Defaults to empty, in which case every worker manages its own metrics, and the /metrics endpoint returns the metrics of a random worker. +See [../../libs/libapi/README.md](../../libs/libapi/README.md) for more information about the API configuration. diff --git a/services/api/dev.Dockerfile b/services/api/dev.Dockerfile index 7af4bd54..09196638 100644 --- a/services/api/dev.Dockerfile +++ b/services/api/dev.Dockerfile @@ -27,0 +28,2 @@ COPY libs/libcommon/pyproject.toml ./libs/libcommon/pyproject.toml +COPY libs/libapi/poetry.lock ./libs/libapi/poetry.lock +COPY libs/libapi/pyproject.toml ./libs/libapi/pyproject.toml @@ -34,0 +37,3 @@ RUN mkdir ./libs/libcommon/src && mkdir ./libs/libcommon/src/libcommon && touch +# Initialize an empty libapi +# Mapping a volume to ./libs/libapi/src is required when running this image. +RUN mkdir ./libs/libapi/src && mkdir ./libs/libapi/src/libapi && touch ./libs/libapi/src/libapi/__init__.py diff --git a/services/api/poetry.lock b/services/api/poetry.lock index d688e743..c60d328a 100644 --- a/services/api/poetry.lock +++ b/services/api/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. +# This file is automatically @generated by Poetry and should not be changed by hand. @@ -1133,0 +1134,24 @@ test = ["pytest (>=7.2)", "pytest-cov (>=4.0)"] +[[package]] +name = "libapi" +version = "0.1.0" +description = "Library for the API services" +category = "main" +optional = false +python-versions = "3.9.15" +files = [] +develop = true + +[package.dependencies] +cryptography = "^41.0.1" +environs = "^9.5.0" +libcommon = {path = "../../libs/libcommon", develop = true} +orjson = "^3.8.6" +pyjwt = {version = "^2.6.0", extras = ["crypto"]} +requests = "^2.28.2" +starlette = "^0.28.0" +starlette-prometheus = "^0.9.0" + +[package.source] +type = "directory" +url = "../../libs/libapi" + @@ -1263,60 +1286,0 @@ testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"] -[[package]] -name = "markupsafe" -version = "2.1.3" -description = "Safely add untrusted strings to HTML/XML markup." -category = "dev" -optional = false -python-versions = ">=3.7" -files = [ - {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cd0f502fe016460680cd20aaa5a76d241d6f35a1c3350c474bac1273803893fa"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e09031c87a1e51556fdcb46e5bd4f59dfb743061cf93c4d6831bf894f125eb57"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:68e78619a61ecf91e76aa3e6e8e33fc4894a2bebe93410754bd28fce0a8a4f9f"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65c1a9bcdadc6c28eecee2c119465aebff8f7a584dd719facdd9e825ec61ab52"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:525808b8019e36eb524b8c68acdd63a37e75714eac50e988180b169d64480a00"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:962f82a3086483f5e5f64dbad880d31038b698494799b097bc59c2edf392fce6"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:aa7bd130efab1c280bed0f45501b7c8795f9fdbeb02e965371bbef3523627779"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c9c804664ebe8f83a211cace637506669e7890fec1b4195b505c214e50dd4eb7"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-win32.whl", hash = "sha256:10bbfe99883db80bdbaff2dcf681dfc6533a614f700da1287707e8a5d78a8431"}, - {file = "MarkupSafe-2.1.3-cp310-cp310-win_amd64.whl", hash = "sha256:1577735524cdad32f9f694208aa75e422adba74f1baee7551620e43a3141f559"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:ad9e82fb8f09ade1c3e1b996a6337afac2b8b9e365f926f5a61aacc71adc5b3c"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3c0fae6c3be832a0a0473ac912810b2877c8cb9d76ca48de1ed31e1c68386575"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b076b6226fb84157e3f7c971a47ff3a679d837cf338547532ab866c57930dbee"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bfce63a9e7834b12b87c64d6b155fdd9b3b96191b6bd334bf37db7ff1fe457f2"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:338ae27d6b8745585f87218a3f23f1512dbf52c26c28e322dbe54bcede54ccb9"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:e4dd52d80b8c83fdce44e12478ad2e85c64ea965e75d66dbeafb0a3e77308fcc"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:df0be2b576a7abbf737b1575f048c23fb1d769f267ec4358296f31c2479db8f9"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:5bbe06f8eeafd38e5d0a4894ffec89378b6c6a625ff57e3028921f8ff59318ac"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-win32.whl", hash = "sha256:dd15ff04ffd7e05ffcb7fe79f1b98041b8ea30ae9234aed2a9168b5797c3effb"}, - {file = "MarkupSafe-2.1.3-cp311-cp311-win_amd64.whl", hash = "sha256:134da1eca9ec0ae528110ccc9e48041e0828d79f24121a1a146161103c76e686"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8e254ae696c88d98da6555f5ace2279cf7cd5b3f52be2b5cf97feafe883b58d2"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cb0932dc158471523c9637e807d9bfb93e06a95cbf010f1a38b98623b929ef2b"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9402b03f1a1b4dc4c19845e5c749e3ab82d5078d16a2a4c2cd2df62d57bb0707"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca379055a47383d02a5400cb0d110cef0a776fc644cda797db0c5696cfd7e18e"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7ff0f54cb4ff66dd38bebd335a38e2c22c41a8ee45aa608efc890ac3e3931bc"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c011a4149cfbcf9f03994ec2edffcb8b1dc2d2aede7ca243746df97a5d41ce48"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:56d9f2ecac662ca1611d183feb03a3fa4406469dafe241673d521dd5ae92a155"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-win32.whl", hash = "sha256:8758846a7e80910096950b67071243da3e5a20ed2546e6392603c096778d48e0"}, - {file = "MarkupSafe-2.1.3-cp37-cp37m-win_amd64.whl", hash = "sha256:787003c0ddb00500e49a10f2844fac87aa6ce977b90b0feaaf9de23c22508b24"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:2ef12179d3a291be237280175b542c07a36e7f60718296278d8593d21ca937d4"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:2c1b19b3aaacc6e57b7e25710ff571c24d6c3613a45e905b1fde04d691b98ee0"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8afafd99945ead6e075b973fefa56379c5b5c53fd8937dad92c662da5d8fd5ee"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8c41976a29d078bb235fea9b2ecd3da465df42a562910f9022f1a03107bd02be"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d080e0a5eb2529460b30190fcfcc4199bd7f827663f858a226a81bc27beaa97e"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:69c0f17e9f5a7afdf2cc9fb2d1ce6aabdb3bafb7f38017c0b77862bcec2bbad8"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:504b320cd4b7eff6f968eddf81127112db685e81f7e36e75f9f84f0df46041c3"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:42de32b22b6b804f42c5d98be4f7e5e977ecdd9ee9b660fda1a3edf03b11792d"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-win32.whl", hash = "sha256:ceb01949af7121f9fc39f7d27f91be8546f3fb112c608bc4029aef0bab86a2a5"}, - {file = "MarkupSafe-2.1.3-cp38-cp38-win_amd64.whl", hash = "sha256:1b40069d487e7edb2676d3fbdb2b0829ffa2cd63a2ec26c4938b2d34391b4ecc"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:8023faf4e01efadfa183e863fefde0046de576c6f14659e8782065bcece22198"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6b2b56950d93e41f33b4223ead100ea0fe11f8e6ee5f641eb753ce4b77a7042b"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9dcdfd0eaf283af041973bff14a2e143b8bd64e069f4c383416ecd79a81aab58"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:05fb21170423db021895e1ea1e1f3ab3adb85d1c2333cbc2310f2a26bc77272e"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:282c2cb35b5b673bbcadb33a585408104df04f14b2d9b01d4c345a3b92861c2c"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ab4a0df41e7c16a1392727727e7998a467472d0ad65f3ad5e6e765015df08636"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:7ef3cb2ebbf91e330e3bb937efada0edd9003683db6b57bb108c4001f37a02ea"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:0a4e4a1aff6c7ac4cd55792abf96c915634c2b97e3cc1c7129578aa68ebd754e"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-win32.whl", hash = "sha256:fec21693218efe39aa7f8599346e90c705afa52c5b31ae019b2e57e8f6542bb2"}, - {file = "MarkupSafe-2.1.3-cp39-cp39-win_amd64.whl", hash = "sha256:3fd4abcb888d15a94f32b75d8fd18ee162ca0c064f35b11134be77050296d6ba"}, - {file = "MarkupSafe-2.1.3.tar.gz", hash = "sha256:af598ed32d6ae86f1b747b82783958b1a4ab8f617b06fe68795c7f026abbdcad"}, -] - @@ -2528,15 +2491,0 @@ testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtuale -[[package]] -name = "pytest-httpserver" -version = "1.0.8" -description = "pytest-httpserver is a httpserver for pytest" -category = "dev" -optional = false -python-versions = ">=3.8,<4.0" -files = [ - {file = "pytest_httpserver-1.0.8-py3-none-any.whl", hash = "sha256:24cd3d9f6a0b927c7bfc400d0b3fda7442721b8267ce29942bf307b190f0bb09"}, - {file = "pytest_httpserver-1.0.8.tar.gz", hash = "sha256:e052f69bc8a9073db02484681e8e47004dd1fb3763b0ae833bd899e5895c559a"}, -] - -[package.dependencies] -Werkzeug = ">=2.0.0" - @@ -2851,0 +2801 @@ files = [ + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -2911 +2861 @@ name = "starlette" -version = "0.27.0" +version = "0.28.0" @@ -2917,2 +2867,2 @@ files = [ - {file = "starlette-0.27.0-py3-none-any.whl", hash = "sha256:918416370e846586541235ccd38a474c08b80443ed31c578a418e2209b3eef91"}, - {file = "starlette-0.27.0.tar.gz", hash = "sha256:6a6b0d042acb8d469a01eba54e9cda6cbd24ac602c4cd016723117d6a7e73b75"}, + {file = "starlette-0.28.0-py3-none-any.whl", hash = "sha256:e58b9fc402c579950260fbb6d57173395c4e62804c40d3ede7e9ef1074f0c579"}, + {file = "starlette-0.28.0.tar.gz", hash = "sha256:7bf3da5e997e796cc202cef2bd3f96a7d9b1e1943203c2fe2b42e020bc658482"}, @@ -3195,18 +3144,0 @@ files = [ -[[package]] -name = "werkzeug" -version = "2.3.4" -description = "The comprehensive WSGI web application library." -category = "dev" -optional = false -python-versions = ">=3.8" -files = [ - {file = "Werkzeug-2.3.4-py3-none-any.whl", hash = "sha256:48e5e61472fee0ddee27ebad085614ebedb7af41e88f687aaf881afb723a162f"}, - {file = "Werkzeug-2.3.4.tar.gz", hash = "sha256:1d5a58e0377d1fe39d061a5de4469e414e78ccb1e1e59c0f5ad6fa1c36c52b76"}, -] - -[package.dependencies] -MarkupSafe = ">=2.1.1" - -[package.extras] -watchdog = ["watchdog (>=2.3)"] - @@ -3412 +3344 @@ python-versions = "3.9.15" -content-hash = "781ea6fdf22b868dd68ea1f37735c621a9064e2ee616b3e2538fa8924697b1b2" +content-hash = "c46efd608bb9fde12d1d6aec4385e7b26da6d08011d84e287d5a1663f3c93217" diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml index 7fc589ed..46450593 100644 --- a/services/api/pyproject.toml +++ b/services/api/pyproject.toml @@ -9 +8,0 @@ license = "Apache-2.0" -cryptography = "^41.0.1" @@ -12,2 +11 @@ jsonschema = "^4.17.0" -libcommon = {path = "../../libs/libcommon", develop = true} -pyjwt = { extras = ["crypto"], version = "^2.6.0" } +libapi = {path = "../../libs/libapi", develop = true} @@ -16 +13,0 @@ python = "3.9.15" -requests = "^2.28.2" @@ -18,2 +14,0 @@ soundfile = ">=0.12.1" -starlette = "^0.27.0" -starlette-prometheus = "^0.9.0" @@ -34 +28,0 @@ pytest-cov = "^2.12.1" -pytest-httpserver = "^1.0.6" @@ -36 +29,0 @@ types-psutil = "^5.9.5" -types-requests = "^2.28.11" @@ -37,0 +31 @@ types-jsonschema = "^4.17.0.4" +types-requests = "^2.28.11" diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index a124b5bb..8f6992b2 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -4,0 +5,4 @@ import uvicorn +from libapi.config import UvicornConfig +from libapi.jwt_token import fetch_jwt_public_key +from libapi.routes.healthcheck import healthcheck_endpoint +from libapi.routes.metrics import create_metrics_endpoint @@ -16,2 +20 @@ from starlette_prometheus import PrometheusMiddleware -from api.config import AppConfig, EndpointConfig, UvicornConfig -from api.jwt_token import fetch_jwt_public_key +from api.config import AppConfig, EndpointConfig @@ -19,2 +21,0 @@ from api.routes.endpoint import EndpointsDefinition, create_endpoint -from api.routes.healthcheck import healthcheck_endpoint -from api.routes.metrics import create_metrics_endpoint diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index e8518a51..28a62fc7 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -5 +5 @@ from dataclasses import dataclass, field -from typing import List, Mapping, Optional +from typing import List, Mapping @@ -7 +7 @@ from typing import List, Mapping, Optional -from environs import Env +from libapi.config import ApiConfig @@ -19,60 +18,0 @@ from libcommon.processing_graph import InputType -API_UVICORN_HOSTNAME = "localhost" -API_UVICORN_NUM_WORKERS = 2 -API_UVICORN_PORT = 8000 - - -@dataclass(frozen=True) -class UvicornConfig: - hostname: str = API_UVICORN_HOSTNAME - num_workers: int = API_UVICORN_NUM_WORKERS - port: int = API_UVICORN_PORT - - @classmethod - def from_env(cls) -> "UvicornConfig": - env = Env(expand_vars=True) - with env.prefixed("API_UVICORN_"): - return cls( - hostname=env.str(name="HOSTNAME", default=API_UVICORN_HOSTNAME), - num_workers=env.int(name="NUM_WORKERS", default=API_UVICORN_NUM_WORKERS), - port=env.int(name="PORT", default=API_UVICORN_PORT), - ) - - -API_EXTERNAL_AUTH_URL = None -API_HF_AUTH_PATH = "/api/datasets/%s/auth-check" -API_HF_JWT_PUBLIC_KEY_URL = None -API_HF_JWT_ALGORITHM = "EdDSA" -API_HF_TIMEOUT_SECONDS = 0.2 -API_HF_WEBHOOK_SECRET = None -API_MAX_AGE_LONG = 120 # 2 minutes -API_MAX_AGE_SHORT = 10 # 10 seconds - - -@dataclass(frozen=True) -class ApiConfig: - external_auth_url: Optional[str] = API_EXTERNAL_AUTH_URL # not documented - hf_auth_path: str = API_HF_AUTH_PATH - hf_jwt_public_key_url: Optional[str] = API_HF_JWT_PUBLIC_KEY_URL - hf_jwt_algorithm: Optional[str] = API_HF_JWT_ALGORITHM - hf_timeout_seconds: Optional[float] = API_HF_TIMEOUT_SECONDS - hf_webhook_secret: Optional[str] = API_HF_WEBHOOK_SECRET - max_age_long: int = API_MAX_AGE_LONG - max_age_short: int = API_MAX_AGE_SHORT - - @classmethod - def from_env(cls, common_config: CommonConfig) -> "ApiConfig": - env = Env(expand_vars=True) - with env.prefixed("API_"): - hf_auth_path = env.str(name="HF_AUTH_PATH", default=API_HF_AUTH_PATH) - external_auth_url = None if hf_auth_path is None else f"{common_config.hf_endpoint}{hf_auth_path}" - return cls( - external_auth_url=external_auth_url, - hf_auth_path=hf_auth_path, - hf_jwt_public_key_url=env.str(name="HF_JWT_PUBLIC_KEY_URL", default=API_HF_JWT_PUBLIC_KEY_URL), - hf_jwt_algorithm=env.str(name="HF_JWT_ALGORITHM", default=API_HF_JWT_ALGORITHM), - hf_timeout_seconds=env.float(name="HF_TIMEOUT_SECONDS", default=API_HF_TIMEOUT_SECONDS), - hf_webhook_secret=env.str(name="HF_WEBHOOK_SECRET", default=API_HF_WEBHOOK_SECRET), - max_age_long=env.int(name="MAX_AGE_LONG", default=API_MAX_AGE_LONG), - max_age_short=env.int(name="MAX_AGE_SHORT", default=API_MAX_AGE_SHORT), - ) - @@ -101 +41 @@ class AppConfig: - api=ApiConfig.from_env(common_config=common_config), + api=ApiConfig.from_env(hf_endpoint=common_config.hf_endpoint), diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py index 71f6f00e..e01d22b4 100644 --- a/services/api/src/api/routes/endpoint.py +++ b/services/api/src/api/routes/endpoint.py @@ -8,0 +9,14 @@ from typing import List, Mapping, Optional, Tuple, TypedDict +from libapi.authentication import auth_check +from libapi.exceptions import ( + ApiError, + MissingRequiredParameterError, + UnexpectedApiError, +) +from libapi.utils import ( + Endpoint, + are_valid_parameters, + get_json_api_error_response, + get_json_error_response, + get_json_ok_response, + try_backfill_dataset, +) @@ -19 +32,0 @@ from starlette.responses import Response -from api.authentication import auth_check @@ -21,11 +33,0 @@ from api.config import EndpointConfig -from api.utils import ( - ApiCustomError, - Endpoint, - MissingRequiredParameterError, - UnexpectedError, - are_valid_parameters, - get_json_api_error_response, - get_json_error_response, - get_json_ok_response, - try_backfill_dataset, -) @@ -319 +321 @@ def create_endpoint( - error = e if isinstance(e, ApiCustomError) else UnexpectedError("Unexpected error.", e) + error = e if isinstance(e, ApiError) else UnexpectedApiError("Unexpected error.", e) diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 18430cb7..2529260f 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -13,0 +14,14 @@ from fsspec.implementations.http import HTTPFileSystem +from libapi.authentication import auth_check +from libapi.exceptions import ( + ApiError, + InvalidParameterError, + MissingRequiredParameterError, + UnexpectedApiError, +) +from libapi.utils import ( + Endpoint, + are_valid_parameters, + get_json_api_error_response, + get_json_ok_response, + try_backfill_dataset, +) @@ -26,13 +39,0 @@ from starlette.responses import Response -from api.authentication import auth_check -from api.utils import ( - ApiCustomError, - Endpoint, - InvalidParameterError, - MissingRequiredParameterError, - UnexpectedError, - are_valid_parameters, - get_json_api_error_response, - get_json_ok_response, - try_backfill_dataset, -) - @@ -381 +382 @@ def create_rows_endpoint( - error = e if isinstance(e, ApiCustomError) else UnexpectedError("Unexpected error.", e) + error = e if isinstance(e, ApiError) else UnexpectedApiError("Unexpected error.", e) diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py index 029eaa97..10517797 100644 --- a/services/api/src/api/routes/valid.py +++ b/services/api/src/api/routes/valid.py @@ -7,0 +8,2 @@ from typing import List, Set, TypedDict +from libapi.exceptions import UnexpectedApiError +from libapi.utils import Endpoint, get_json_api_error_response, get_json_ok_response @@ -14,7 +15,0 @@ from starlette.responses import Response -from api.utils import ( - Endpoint, - UnexpectedError, - get_json_api_error_response, - get_json_ok_response, -) - @@ -79 +74,3 @@ def create_valid_endpoint( - return get_json_api_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age_short) + return get_json_api_error_response( + UnexpectedApiError("Unexpected error.", e), max_age=max_age_short + ) diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py index 96f2c810..5c43bef8 100644 --- a/services/api/src/api/routes/webhook.py +++ b/services/api/src/api/routes/webhook.py @@ -7,0 +8 @@ from jsonschema import ValidationError, validate +from libapi.utils import Endpoint, get_response @@ -16,2 +16,0 @@ from starlette.responses import Response -from api.utils import Endpoint, get_response - diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py index 603e16bf..cbc59c89 100644 --- a/services/api/tests/conftest.py +++ b/services/api/tests/conftest.py @@ -6,0 +7 @@ from typing import Iterator +from libapi.config import UvicornConfig @@ -14 +15 @@ from pytest import MonkeyPatch, fixture -from api.config import AppConfig, EndpointConfig, UvicornConfig +from api.config import AppConfig, EndpointConfig diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index e691f92e..37f8f820 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -6,0 +7 @@ from unittest.mock import patch +from libapi.exceptions import ResponseNotReadyError @@ -15 +15,0 @@ from api.routes.endpoint import EndpointsDefinition, get_cache_entry_from_steps -from api.utils import ResponseNotReadyError diff --git a/services/api/tests/routes/test_rows.py b/services/api/tests/routes/test_rows.py index 7613d52a..e8c944db 100644 --- a/services/api/tests/routes/test_rows.py +++ b/services/api/tests/routes/test_rows.py @@ -0,0 +1,3 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py index 4827c75d..dd0f3a96 100644 --- a/services/api/tests/routes/test_valid.py +++ b/services/api/tests/routes/test_valid.py @@ -0,0 +1,3 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + diff --git a/services/api/tests/utils.py b/services/api/tests/utils.py deleted file mode 100644 index b7cc894b..00000000 --- a/services/api/tests/utils.py +++ /dev/null @@ -1,16 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -from werkzeug.wrappers.request import Request -from werkzeug.wrappers.response import Response - - -def auth_callback(request: Request) -> Response: - # return 401 if a cookie has been provided, 404 if a token has been provided, - # and 200 if none has been provided - # - # caveat: the returned status codes don't simulate the reality - # they're just used to check every case - return Response( - status=401 if request.headers.get("cookie") else 404 if request.headers.get("authorization") else 200 - ) diff --git a/tools/Python.mk b/tools/Python.mk index 8632f448..91e40366 100644 --- a/tools/Python.mk +++ b/tools/Python.mk @@ -31 +31 @@ pip-audit: - bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' |sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d')" + bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' |sed '/^torch @/d' | sed '/^libapi @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d')"
e1368b7dd349730d1e94a9bf94064919de2ee6d1
Quentin Lhoest
2023-07-04T12:30:09
optional num_bytes_original_files (#1478)
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py index 5bdfcd35..ad008a4b 100644 --- a/services/worker/src/worker/dtos.py +++ b/services/worker/src/worker/dtos.py @@ -143 +143,3 @@ class ConfigSize(TypedDict): - num_bytes_original_files: int + num_bytes_original_files: Optional[ + int + ] # optional because partial parquet conversion can't provide the size of the original data @@ -197 +199 @@ class DatasetSize(TypedDict): - num_bytes_original_files: int + num_bytes_original_files: Optional[int] diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py index 9f47241b..785ea8f9 100644 --- a/services/worker/src/worker/job_runners/config/size.py +++ b/services/worker/src/worker/job_runners/config/size.py @@ -68 +68 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - "num_bytes_original_files": config_info["download_size"], + "num_bytes_original_files": config_info["download_size"] if not content["partial"] else None, diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py index 71217c1c..62c62704 100644 --- a/services/worker/src/worker/job_runners/dataset/size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -6 +6 @@ from http import HTTPStatus -from typing import Tuple +from typing import Optional, Tuple @@ -93,0 +94,7 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: + num_bytes_original_files: Optional[int] = 0 + for config_size in config_sizes: + if num_bytes_original_files is not None and isinstance(config_size["num_bytes_original_files"], int): + num_bytes_original_files += config_size["num_bytes_original_files"] + else: + num_bytes_original_files = None + break @@ -96 +103 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - "num_bytes_original_files": sum(config_size["num_bytes_original_files"] for config_size in config_sizes), + "num_bytes_original_files": num_bytes_original_files,
00c0a16807736717bff563b7eb9db2e82e84374a
Quentin Lhoest
2023-07-04T09:38:22
unblock oscar (#1477)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index d787905d..953aeb10 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -100 +100 @@ parquetAndInfo: - blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member" + blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
efb74b8330ebbe5a6832c566cd4f05873166e9ee
Quentin Lhoest
2023-07-04T09:38:14
More workers (#1476)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 1566f1fa..d787905d 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -124 +124 @@ cacheMaintenance: - action: "backfill" + action: "skip" @@ -129 +129 @@ cacheMaintenance: - error_codes_to_retry: "DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError" + error_codes_to_retry: "" @@ -272 +272 @@ workers: - replicas: 20 + replicas: 80 @@ -287 +287 @@ workers: - replicas: 4 + replicas: 12
10ecac78d4c640228d53ccddb83314c239252980
Quentin Lhoest
2023-07-03T18:31:46
fix stream_convert_to_parquet for GeneratorBasedBuilder (#1473)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 880b991a..cae1b12c 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -808,0 +809,3 @@ def stream_convert_to_parquet(builder: DatasetBuilder, max_dataset_size: int) -> + prepare_split_kwargs: dict[str, Any] = ( + {"check_duplicate_keys": True} if isinstance(builder, datasets.builder.GeneratorBasedBuilder) else {} + ) @@ -813 +816,3 @@ def stream_convert_to_parquet(builder: DatasetBuilder, max_dataset_size: int) -> - builder._prepare_split(split_generator=splits_generators[split], file_format="parquet") + builder._prepare_split( + split_generator=splits_generators[split], file_format="parquet", **prepare_split_kwargs + ) diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 6c4f72f2..a78ba301 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -840 +840 @@ def test_get_delete_operations( -def test_stream_convert_to_parquet( +def test_stream_convert_to_parquet_arrowbasedbuilder( @@ -866,0 +867,38 @@ def test_stream_convert_to_parquet( [email protected]( + "max_dataset_size,expected_num_shards", + [ + (1, 1), + (150, 19), + (300, 38), + (9999999, 1000), + ], +) +def test_stream_convert_to_parquet_generatorbasedbuilder( + max_dataset_size: int, expected_num_shards: int, tmp_path: Path +) -> None: + num_rows = 1000 + + def long_generator() -> Iterator[Dict[str, int]]: + for i in range(num_rows): + yield {"foo": i} + + cache_dir = str(tmp_path / "test_limit_parquet_writes_cache_dir") + builder = ParametrizedGeneratorBasedBuilder(generator=long_generator, cache_dir=cache_dir) + with patch("worker.job_runners.config.parquet_and_info.get_writer_batch_size", lambda ds_config_info: 1): + with patch.object(datasets.config, "MAX_SHARD_SIZE", 1): + parquet_operations, partial = stream_convert_to_parquet(builder, max_dataset_size=max_dataset_size) + num_shards = len(parquet_operations) + assert num_shards == expected_num_shards + assert partial == (expected_num_shards < num_rows) + assert all(isinstance(op.path_or_fileobj, str) for op in parquet_operations) + parquet_files = list_generated_parquet_files(builder, partial=partial) + assert len(parquet_files) == expected_num_shards + assert all(os.path.isfile(parquet_file.local_file) for parquet_file in parquet_files) + one_sample_max_size = 100 + expected_max_dataset_size = max_dataset_size + one_sample_max_size + assert ( + sum(pq.ParquetFile(parquet_file.local_file).read().nbytes for parquet_file in parquet_files) + < expected_max_dataset_size + ) + +
9189407c5e423a1cdbeeae270fa5637439a3563a
Quentin Lhoest
2023-07-03T17:20:29
Add partial to subsequent parquet-and-info jobs (#1471)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py index c1f0b674..c4958fd5 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py @@ -20 +20 @@ class MigrationAddPartialToCacheResponse(Migration): - " config-parquet-and-info" + " config-parquet-and-info and subsequent steps" @@ -25 +25,12 @@ class MigrationAddPartialToCacheResponse(Migration): - "kind": "config-parquet-and-info", + "kind": { + "$in": [ + "config-parquet-and-info", + "config-parquet", + "dataset-parquet", + "config-parquet-metadata", + "config-info", + "dataset-info", + "config-size", + "dataset-size", + ] + }, @@ -36 +47,16 @@ class MigrationAddPartialToCacheResponse(Migration): - {"kind": "config-parquet-and-info", "http_status": 200}, {"$unset": {"content.partial": ""}} + { + "kind": { + "$in": [ + "config-parquet-and-info", + "config-parquet", + "dataset-parquet", + "config-parquet-metadata", + "config-info", + "dataset-info", + "config-size", + "dataset-size", + ] + }, + "http_status": 200, + }, + {"$unset": {"content.partial": ""}}, diff --git a/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py index 85e1a036..6f4eab65 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py @@ -2,0 +3 @@ +from typing import Any, Dict, List @@ -13 +14 @@ from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_con -def assert_partial(dataset: str) -> None: +def assert_partial(dataset: str, kind: str) -> None: @@ -15 +16 @@ def assert_partial(dataset: str) -> None: - entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset}) + entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset, "kind": kind}) @@ -20 +21 @@ def assert_partial(dataset: str) -> None: -def assert_unchanged(dataset: str) -> None: +def assert_unchanged(dataset: str, kind: str) -> None: @@ -22 +23 @@ def assert_unchanged(dataset: str) -> None: - entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset}) + entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset, "kind": kind}) @@ -27,5 +28,8 @@ def assert_unchanged(dataset: str) -> None: -def test_cache_add_partial(mongo_host: str) -> None: - with MongoResource(database="test_cache_add_partial", host=mongo_host, mongoengine_alias="cache"): - db = get_db(CACHE_MONGOENGINE_ALIAS) - db[CACHE_COLLECTION_RESPONSES].insert_many( - [ +cache: List[Dict[str, Any]] = [ + { + "config": "lhoestq--demo1", + "dataset": "lhoestq/demo1", + "kind": "config-parquet-and-info", + "split": None, + "content": { + "parquet_files": [ @@ -33 +36,0 @@ def test_cache_add_partial(mongo_host: str) -> None: - "config": "lhoestq--demo1", @@ -35,90 +38,5 @@ def test_cache_add_partial(mongo_host: str) -> None: - "kind": "config-parquet-and-info", - "split": None, - "content": { - "parquet_files": [ - { - "dataset": "lhoestq/demo1", - "config": "lhoestq--demo1", - "split": "test", - "url": "https://huggingface.co/.../csv-test.parquet", - "filename": "csv-test.parquet", - "size": 4415, - }, - { - "dataset": "lhoestq/demo1", - "config": "lhoestq--demo1", - "split": "train", - "url": "https://huggingface.co/.../csv-train.parquet", - "filename": "csv-train.parquet", - "size": 5038, - }, - ], - "dataset_info": { - "description": "", - "citation": "", - "homepage": "", - "license": "", - "features": {}, - "builder_name": "csv", - "config_name": "lhoestq--demo1", - "version": {}, - "splits": {}, - "download_checksums": {}, - "download_size": 2340, - "dataset_size": 2464, - "size_in_bytes": 4804, - }, - }, - "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11", - "details": None, - "error_code": None, - "http_status": 200, - "job_runner_version": 3, - "progress": 1.0, - }, - { - "config": "lhoestq--demo2", - "dataset": "lhoestq/demo2", - "kind": "config-parquet-and-info", - "split": None, - "content": { - "parquet_files": [ - { - "dataset": "lhoestq/demo2", - "config": "lhoestq--demo2", - "split": "test", - "url": "https://huggingface.co/.../csv-test.parquet", - "filename": "csv-test.parquet", - "size": 4415, - }, - { - "dataset": "lhoestq/demo2", - "config": "lhoestq--demo2", - "split": "train", - "url": "https://huggingface.co/.../csv-train.parquet", - "filename": "csv-train.parquet", - "size": 5038, - }, - ], - "dataset_info": { - "description": "", - "citation": "", - "homepage": "", - "license": "", - "features": {}, - "builder_name": "csv", - "config_name": "lhoestq--demo2", - "version": {}, - "splits": {}, - "download_checksums": {}, - "download_size": 2340, - "dataset_size": 2464, - "size_in_bytes": 4804, - }, - }, - "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11", - "details": None, - "error_code": None, - "http_status": 200, - "job_runner_version": 3, - "progress": 1.0, + "config": "lhoestq--demo1", + "split": "test", + "url": "https://huggingface.co/.../csv-test.parquet", + "filename": "csv-test.parquet", + "size": 4415, @@ -127,46 +45,6 @@ def test_cache_add_partial(mongo_host: str) -> None: - "config": "lhoestq--error", - "dataset": "lhoestq/error", - "kind": "config-parquet-and-info", - "split": None, - "content": {"error": "Streaming is not supported for lhoestq/error"}, - "dataset_git_revision": "ec3c8d414af3dfe600399f5e6ef2c682938676f3", - "details": { - "error": "Streaming is not supported for lhoestq/error", - "cause_exception": "TypeError", - "cause_message": "Streaming is not supported for lhoestq/error", - "cause_traceback": [ - "Traceback (most recent call last):\n", - ( - ' File "/src/services/worker/src/worker/job_manager.py", line 163, in process\n ' - " job_result = self.job_runner.compute()\n" - ), - ( - ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' - " 932, in compute\n compute_config_parquet_and_info_response(\n" - ), - ( - ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' - " 825, in compute_config_parquet_and_info_response\n raise_if_not_supported(\n" - ), - ( - ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' - " 367, in raise_if_not_supported\n raise_if_too_big_from_external_data_files(\n" - ), - ( - ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' - " 447, in raise_if_too_big_from_external_data_files\n " - " builder._split_generators(mock_dl_manager)\n" - ), - ( - " File" - ' "/tmp/modules-cache/datasets_modules/.../error.py",' - ' line 190, in _split_generators\n raise TypeError("Streaming is not supported for' - ' lhoestq/error")\n' - ), - "TypeError: Streaming is not supported for lhoestq/error\n", - ], - }, - "error_code": "UnexpectedError", - "http_status": 500, - "job_runner_version": 3, - "progress": None, + "dataset": "lhoestq/demo1", + "config": "lhoestq--demo1", + "split": "train", + "url": "https://huggingface.co/.../csv-train.parquet", + "filename": "csv-train.parquet", + "size": 5038, @@ -174,2 +52,99 @@ def test_cache_add_partial(mongo_host: str) -> None: - ] - ) + ], + "dataset_info": { + "description": "", + "citation": "", + "homepage": "", + "license": "", + "features": {}, + "builder_name": "csv", + "config_name": "lhoestq--demo1", + "version": {}, + "splits": {}, + "download_checksums": {}, + "download_size": 2340, + "dataset_size": 2464, + "size_in_bytes": 4804, + }, + }, + "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11", + "details": None, + "error_code": None, + "http_status": 200, + "job_runner_version": 3, + "progress": 1.0, + }, + { + "config": "lhoestq--error", + "dataset": "lhoestq/error", + "kind": "config-parquet-and-info", + "split": None, + "content": {"error": "Streaming is not supported for lhoestq/error"}, + "dataset_git_revision": "ec3c8d414af3dfe600399f5e6ef2c682938676f3", + "details": { + "error": "Streaming is not supported for lhoestq/error", + "cause_exception": "TypeError", + "cause_message": "Streaming is not supported for lhoestq/error", + "cause_traceback": [ + "Traceback (most recent call last):\n", + ( + ' File "/src/services/worker/src/worker/job_manager.py", line 163, in process\n ' + " job_result = self.job_runner.compute()\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 932, in compute\n compute_config_parquet_and_info_response(\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 825, in compute_config_parquet_and_info_response\n raise_if_not_supported(\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 367, in raise_if_not_supported\n raise_if_too_big_from_external_data_files(\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 447, in raise_if_too_big_from_external_data_files\n " + " builder._split_generators(mock_dl_manager)\n" + ), + ( + " File" + ' "/tmp/modules-cache/datasets_modules/.../error.py",' + ' line 190, in _split_generators\n raise TypeError("Streaming is not supported for' + ' lhoestq/error")\n' + ), + "TypeError: Streaming is not supported for lhoestq/error\n", + ], + }, + "error_code": "UnexpectedError", + "http_status": 500, + "job_runner_version": 3, + "progress": None, + }, +] +cache2: List[Dict[str, Any]] = [ + { + "config": "lhoestq--demo2", + "dataset": "lhoestq/demo2", + "kind": kind, + "split": None, + "content": {}, + "http_status": 200, + } + for kind in [ + "config-parquet-and-info", + "config-parquet", + "dataset-parquet", + "config-parquet-metadata", + "config-info", + "dataset-info", + "config-size", + "dataset-size", + ] +] + + +def test_cache_add_partial(mongo_host: str) -> None: + with MongoResource(database="test_cache_add_partial", host=mongo_host, mongoengine_alias="cache"): + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many(cache + cache2) @@ -183,3 +158,13 @@ def test_cache_add_partial(mongo_host: str) -> None: - assert_partial("lhoestq/demo1") - assert_partial("lhoestq/demo2") - assert_unchanged("lhoestq/error") + assert_partial("lhoestq/demo1", kind="config-parquet-and-info") + assert_unchanged("lhoestq/error", kind="config-parquet-and-info") + for kind in [ + "config-parquet-and-info", + "config-parquet", + "dataset-parquet", + "config-parquet-metadata", + "config-info", + "dataset-info", + "config-size", + "dataset-size", + ]: + assert_partial("lhoestq/demo2", kind=kind) @@ -188,3 +173,13 @@ def test_cache_add_partial(mongo_host: str) -> None: - assert_unchanged("lhoestq/demo1") - assert_unchanged("lhoestq/demo2") - assert_unchanged("lhoestq/error") + assert_unchanged("lhoestq/demo1", kind="config-parquet-and-info") + assert_unchanged("lhoestq/error", kind="config-parquet-and-info") + for kind in [ + "config-parquet-and-info", + "config-parquet", + "dataset-parquet", + "config-parquet-metadata", + "config-info", + "dataset-info", + "config-size", + "dataset-size", + ]: + assert_unchanged("lhoestq/demo2", kind=kind) diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py index ca029d38..5bdfcd35 100644 --- a/services/worker/src/worker/dtos.py +++ b/services/worker/src/worker/dtos.py @@ -112,0 +113 @@ class ConfigInfoResponse(TypedDict): + partial: bool @@ -130,0 +132 @@ class ConfigParquetMetadataResponse(TypedDict): + partial: bool @@ -134,0 +137 @@ class ConfigParquetResponse(TypedDict): + partial: bool @@ -160,0 +164 @@ class ConfigSizeResponse(TypedDict): + partial: bool @@ -175,0 +180 @@ class DatasetInfoResponse(TypedDict): + partial: bool @@ -186,0 +192 @@ class DatasetParquetResponse(TypedDict): + partial: bool @@ -206,0 +213 @@ class DatasetSizeResponse(TypedDict): + partial: bool diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py index 33d68219..ca254aff 100644 --- a/services/worker/src/worker/job_runners/config/info.py +++ b/services/worker/src/worker/job_runners/config/info.py @@ -33,0 +34 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons + partial = content["partial"] @@ -45 +46 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons - return ConfigInfoResponse(dataset_info=config_info) + return ConfigInfoResponse(dataset_info=config_info, partial=partial) diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py index a6fc6833..fa614c97 100644 --- a/services/worker/src/worker/job_runners/config/parquet.py +++ b/services/worker/src/worker/job_runners/config/parquet.py @@ -44,0 +45 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse + partial = content["partial"] @@ -47 +48 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - return ConfigParquetResponse(parquet_files=parquet_files) + return ConfigParquetResponse(parquet_files=parquet_files, partial=partial) diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py index 55de2fe3..0129a70e 100644 --- a/services/worker/src/worker/job_runners/config/parquet_metadata.py +++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py @@ -3,0 +4 @@ +import functools @@ -5 +5,0 @@ import logging -from functools import partial @@ -71,0 +72 @@ def compute_parquet_metadata_response( + partial = config_parquet_best_response.response["content"]["partial"] @@ -80 +81,5 @@ def compute_parquet_metadata_response( - partial(get_parquet_file, fs=fs, hf_token=hf_token), source_urls, desc=desc, unit="pq", disable=True + functools.partial(get_parquet_file, fs=fs, hf_token=hf_token), + source_urls, + desc=desc, + unit="pq", + disable=True, @@ -108 +113 @@ def compute_parquet_metadata_response( - return ConfigParquetMetadataResponse(parquet_files_metadata=parquet_files_metadata) + return ConfigParquetMetadataResponse(parquet_files_metadata=parquet_files_metadata, partial=partial) diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py index 2610f994..9f47241b 100644 --- a/services/worker/src/worker/job_runners/config/size.py +++ b/services/worker/src/worker/job_runners/config/size.py @@ -76,0 +77 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons + partial = content["partial"] @@ -85 +86,2 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons - } + }, + "partial": partial, diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py index 1e8802c9..852af3e9 100644 --- a/services/worker/src/worker/job_runners/dataset/info.py +++ b/services/worker/src/worker/job_runners/dataset/info.py @@ -48,0 +49 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl + partial = False @@ -76,0 +78 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl + partial = partial or config_response["content"]["partial"] @@ -83 +85 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - return DatasetInfoResponse(dataset_info=config_infos, pending=pending, failed=failed), progress + return DatasetInfoResponse(dataset_info=config_infos, pending=pending, failed=failed, partial=partial), progress diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index 9425c249..4b04f122 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -52,0 +53 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa + partial = False @@ -84 +85,3 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa - config_parquet_content = ConfigParquetResponse(parquet_files=response["content"]["parquet_files"]) + config_parquet_content = ConfigParquetResponse( + parquet_files=response["content"]["parquet_files"], partial=response["content"]["partial"] + ) @@ -85,0 +89 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa + partial = partial or config_parquet_content["partial"] @@ -92,5 +96 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa - DatasetParquetResponse( - parquet_files=parquet_files, - pending=pending, - failed=failed, - ), + DatasetParquetResponse(parquet_files=parquet_files, pending=pending, failed=failed, partial=partial), diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py index 7019388d..71217c1c 100644 --- a/services/worker/src/worker/job_runners/dataset/size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -55,0 +56 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: + partial = False @@ -87 +88,3 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - config_size_content = ConfigSizeResponse(size=response["content"]["size"]) + config_size_content = ConfigSizeResponse( + size=response["content"]["size"], partial=response["content"]["partial"] + ) @@ -89,0 +93 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: + partial = partial or config_size_content["partial"] @@ -111,0 +116 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: + "partial": partial, diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py index e0aac1b4..49be6df2 100644 --- a/services/worker/tests/job_runners/config/test_info.py +++ b/services/worker/tests/job_runners/config/test_info.py @@ -181,4 +181 @@ def get_job_runner( - { - "parquet_files": PARQUET_FILES, - "dataset_info": CONFIG_INFO_1, - }, + {"parquet_files": PARQUET_FILES, "dataset_info": CONFIG_INFO_1, "partial": False}, @@ -186 +183 @@ def get_job_runner( - {"dataset_info": CONFIG_INFO_1}, + {"dataset_info": CONFIG_INFO_1, "partial": False}, diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index c2e9f007..1189453f 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -96 +96,2 @@ def get_job_runner( - ] + ], + partial=False, @@ -179 +180,2 @@ def get_job_runner( - ] + ], + partial=False, diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py index 96f1cfbc..571a8f8e 100644 --- a/services/worker/tests/job_runners/config/test_parquet_metadata.py +++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py @@ -105,0 +106 @@ def get_job_runner( + partial=False, @@ -130 +131,2 @@ def get_job_runner( - ] + ], + partial=False, diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py index b55b679f..df89b2e2 100644 --- a/services/worker/tests/job_runners/config/test_size.py +++ b/services/worker/tests/job_runners/config/test_size.py @@ -122,0 +123 @@ def get_job_runner( + "partial": False, @@ -156 +157,2 @@ def get_job_runner( - } + }, + "partial": False, diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py index 26aeb369..0d6b8158 100644 --- a/services/worker/tests/job_runners/dataset/test_info.py +++ b/services/worker/tests/job_runners/dataset/test_info.py @@ -49 +49 @@ UPSTREAM_RESPONSE_CONFIG_INFO_1: UpstreamResponse = UpstreamResponse( - content={"dataset_info": CONFIG_INFO_1}, + content={"dataset_info": CONFIG_INFO_1, "partial": False}, @@ -57 +57 @@ UPSTREAM_RESPONSE_CONFIG_INFO_2: UpstreamResponse = UpstreamResponse( - content={"dataset_info": CONFIG_INFO_2}, + content={"dataset_info": CONFIG_INFO_2, "partial": False}, @@ -64,0 +65 @@ EXPECTED_OK = ( + "partial": False, @@ -82,0 +84 @@ EXPECTED_PARTIAL_PENDING = ( + "partial": False, @@ -100,0 +103 @@ EXPECTED_PARTIAL_FAILED = ( + "partial": False, diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index 8f63b188..28c04d3f 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -100 +100,2 @@ def get_job_runner( - ] + ], + partial=False, @@ -118 +119,2 @@ def get_job_runner( - ] + ], + partial=False, @@ -133,0 +136 @@ def get_job_runner( + partial=False, diff --git a/services/worker/tests/job_runners/dataset/test_size.py b/services/worker/tests/job_runners/dataset/test_size.py index 49490e44..76326d1d 100644 --- a/services/worker/tests/job_runners/dataset/test_size.py +++ b/services/worker/tests/job_runners/dataset/test_size.py @@ -120 +120,2 @@ def get_job_runner( - } + }, + "partial": False, @@ -159 +160,2 @@ def get_job_runner( - } + }, + "partial": False, @@ -233,0 +236 @@ def get_job_runner( + "partial": False,
cb4142773a2b46122049c34435dfb063ec656869
Polina Kazakova
2023-07-03T17:11:05
Rename `/dataset-info` endpoint to `/info` (#1468)
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml index b4dbd129..bd3bb312 100644 --- a/docs/source/_toctree.yml +++ b/docs/source/_toctree.yml @@ -14,0 +15,2 @@ + - local: info + title: Get dataset information diff --git a/docs/source/info.mdx b/docs/source/info.mdx new file mode 100644 index 00000000..dd9bc8c6 --- /dev/null +++ b/docs/source/info.mdx @@ -0,0 +1,139 @@ +# Get dataset information + +Datasets Server provides an `/info` endpoint for exploring the general information about dataset, including such fields as description, citation, homepage, license and features. + +The `/info` endpoint accepts two query parameters: + +- `dataset`: the dataset name +- `config`: the configuration name + +<inferencesnippet> +<python> +```python +import requests +headers = {"Authorization": f"Bearer {API_TOKEN}"} +API_URL = "https://datasets-server.huggingface.co/info?dataset=duorc&config=SelfRC" +def query(): + response = requests.get(API_URL, headers=headers) + return response.json() +data = query() +``` +</python> +<js> +```js +import fetch from "node-fetch"; +async function query(data) { + const response = await fetch( + "https://datasets-server.huggingface.co/info?dataset=duorc&config=SelfRC", + { + headers: { Authorization: `Bearer ${API_TOKEN}` }, + method: "GET" + } + ); + const result = await response.json(); + return result; +} +query().then((response) => { + console.log(JSON.stringify(response)); +}); +``` +</js> +<curl> +```curl +curl https://datasets-server.huggingface.co/info?dataset=duorc&config=SelfRC \ + -X GET \ + -H "Authorization: Bearer ${API_TOKEN}" +``` +</curl> +</inferencesnippet> + +The endpoint response is a JSON with the `dataset_info` key. Its structure and content correspond to [DatasetInfo](https://huggingface.co/docs/datasets/package_reference/main_classes#datasets.DatasetInfo) object of the `datasets` library. + +```json +{ + "dataset_info": { + "description": "DuoRC contains 186,089 unique question-answer pairs created from a collection of 7680 pairs of movie plots where each pair in the collection reflects two versions of the same movie.\n", + "citation": "@inproceedings{DuoRC,\nauthor = { Amrita Saha and Rahul Aralikatte and Mitesh M. Khapra and Karthik Sankaranarayanan},title = {{DuoRC: Towards Complex Language Understanding with Paraphrased Reading Comprehension}},\nbooktitle = {Meeting of the Association for Computational Linguistics (ACL)},\nyear = {2018}\n}\n", + "homepage": "https://duorc.github.io/", + "license": "https://raw.githubusercontent.com/duorc/duorc/master/LICENSE", + "features": { + "plot_id": { + "dtype": "string", + "_type": "Value" + }, + "plot": { + "dtype": "string", + "_type": "Value" + }, + "title": { + "dtype": "string", + "_type": "Value" + }, + "question_id": { + "dtype": "string", + "_type": "Value" + }, + "question": { + "dtype": "string", + "_type": "Value" + }, + "answers": { + "feature": { + "dtype": "string", + "_type": "Value" + }, + "_type": "Sequence" + }, + "no_answer": { + "dtype": "bool", + "_type": "Value" + } + }, + "builder_name": "duorc", + "config_name": "SelfRC", + "version": { + "version_str": "1.0.0", + "major": 1, + "minor": 0, + "patch": 0 + }, + "splits": { + "train": { + "name": "train", + "num_bytes": 239852729, + "num_examples": 60721, + "dataset_name": "duorc" + }, + "validation": { + "name": "validation", + "num_bytes": 51662519, + "num_examples": 12961, + "dataset_name": "duorc" + }, + "test": { + "name": "test", + "num_bytes": 49142710, + "num_examples": 12559, + "dataset_name": "duorc" + } + }, + "download_checksums": { + "https://raw.githubusercontent.com/duorc/duorc/master/dataset/SelfRC_train.json": { + "num_bytes": 24388192, + "checksum": null + }, + "https://raw.githubusercontent.com/duorc/duorc/master/dataset/SelfRC_dev.json": { + "num_bytes": 5051240, + "checksum": null + }, + "https://raw.githubusercontent.com/duorc/duorc/master/dataset/SelfRC_test.json": { + "num_bytes": 5023228, + "checksum": null + } + }, + "download_size": 34462660, + "dataset_size": 340657958, + "size_in_bytes": 375120618 + } +} +``` \ No newline at end of file diff --git a/e2e/tests/test_11_api.py b/e2e/tests/test_11_api.py index 7732ea1f..48e60d68 100644 --- a/e2e/tests/test_11_api.py +++ b/e2e/tests/test_11_api.py @@ -58,2 +58,2 @@ def test_auth_e2e( - ("/dataset-info", "dataset"), - ("/dataset-info", "config"), + ("/info", "dataset"), + ("/info", "config"), diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py index 5bb0f533..e8518a51 100644 --- a/services/api/src/api/config.py +++ b/services/api/src/api/config.py @@ -138 +138 @@ class EndpointConfig: - "/dataset-info": {"dataset": ["dataset-info"], "config": ["config-info"]}, + "/info": {"dataset": ["dataset-info"], "config": ["config-info"]}, diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index a3a98220..e691f92e 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -65 +65 @@ def test_endpoints_definition() -> None: - dataset_info = definition["/dataset-info"] + dataset_info = definition["/info"] @@ -90,0 +91,4 @@ def test_endpoints_definition() -> None: + # assert old endpoints don't exist + with raises(KeyError): + _ = definition["/dataset-info"] +
793ce6a17d88315e6cc09075278606926257d526
Sylvain Lesage
2023-07-03T15:55:04
feat: 🎸 backfill cache entries older than 90 days (#1469)
diff --git a/chart/templates/_envCache.tpl b/chart/templates/_envCache.tpl index 4701fbfd..f4974fd6 100644 --- a/chart/templates/_envCache.tpl +++ b/chart/templates/_envCache.tpl @@ -4,0 +5,2 @@ +- name: CACHE_MAX_DAYS + value: {{ .Values.cache.maxDays | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index 98cc505b..212ccc51 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -114,0 +115,2 @@ cache: + # Maximum number of days to keep the cached API responses + maxDays: 90 diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py index de3bedf1..a64bbad6 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py +++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py @@ -15,0 +16 @@ def backfill_cache( + cache_max_days: int, @@ -50,0 +52 @@ def backfill_cache( + cache_max_days=cache_max_days, diff --git a/jobs/cache_maintenance/src/cache_maintenance/main.py b/jobs/cache_maintenance/src/cache_maintenance/main.py index e9acb982..a6926853 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/main.py +++ b/jobs/cache_maintenance/src/cache_maintenance/main.py @@ -64,0 +65 @@ def run_job() -> None: + cache_max_days=job_config.cache.max_days, diff --git a/libs/libcommon/README.md b/libs/libcommon/README.md index 833ec48b..1b16130a 100644 --- a/libs/libcommon/README.md +++ b/libs/libcommon/README.md @@ -28,0 +29 @@ Set environment variables to configure the storage of precomputed API responses +- `CACHE_MAX_DAYS`: maximum number of days to keep the cache entries. Defaults to `90`. diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 99c6bebc..3a713a1f 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -178,0 +179 @@ class LogConfig: +CACHE_MAX_DAYS = 90 # 3 months @@ -184,0 +186 @@ class CacheConfig: + max_days: int = CACHE_MAX_DAYS @@ -192,0 +195 @@ class CacheConfig: + max_days=env.int(name="MAX_DAYS", default=CACHE_MAX_DAYS), diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py index 0797eef8..4a0b4603 100644 --- a/libs/libcommon/src/libcommon/operations.py +++ b/libs/libcommon/src/libcommon/operations.py @@ -15,0 +16 @@ def backfill_dataset( + cache_max_days: int, @@ -24,0 +26 @@ def backfill_dataset( + cache_max_days (int): the number of days to keep the cache @@ -31 +33 @@ def backfill_dataset( - revision=revision, priority=priority, error_codes_to_retry=[] + revision=revision, priority=priority, error_codes_to_retry=[], cache_max_days=cache_max_days diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py index 1bbab28c..1a5669b0 100644 --- a/libs/libcommon/src/libcommon/orchestrator.py +++ b/libs/libcommon/src/libcommon/orchestrator.py @@ -33,0 +34 @@ class CacheStatus: + cache_is_old: Dict[str, ArtifactState] = field(default_factory=dict) @@ -42,0 +44 @@ class CacheStatus: + "cache_is_old": sorted(self.cache_is_old.keys()), @@ -309,0 +312 @@ class DatasetBackfillPlan(Plan): + cache_max_days: maximum number of days to keep the cache @@ -317,0 +321 @@ class DatasetBackfillPlan(Plan): + cache_max_days: int @@ -478,0 +483,5 @@ class DatasetBackfillPlan(Plan): + # is an old entry? + if artifact_state.cache_state.is_old(days=self.cache_max_days): + cache_status.cache_is_old[artifact_state.id] = artifact_state + continue + @@ -537,0 +547 @@ class DatasetBackfillPlan(Plan): + + list(self.cache_status.cache_is_old.values()) @@ -573 +583,3 @@ class DatasetOrchestrator: - def set_revision(self, revision: str, priority: Priority, error_codes_to_retry: List[str]) -> None: + def set_revision( + self, revision: str, priority: Priority, error_codes_to_retry: List[str], cache_max_days: int + ) -> None: @@ -583,0 +596 @@ class DatasetOrchestrator: + cache_max_days (int): The maximum number of days for which the cache is considered valid. @@ -614,0 +628 @@ class DatasetOrchestrator: + cache_max_days=cache_max_days, @@ -723 +737,3 @@ class DatasetOrchestrator: - def backfill(self, revision: str, priority: Priority, error_codes_to_retry: Optional[List[str]] = None) -> int: + def backfill( + self, revision: str, priority: Priority, cache_max_days: int, error_codes_to_retry: Optional[List[str]] = None + ) -> int: @@ -729,0 +746 @@ class DatasetOrchestrator: + cache_max_days (int): The maximum number of days to keep the cache. @@ -752,0 +770 @@ class DatasetOrchestrator: + cache_max_days=cache_max_days, diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py index b047fbe5..7a7f250e 100644 --- a/libs/libcommon/src/libcommon/state.py +++ b/libs/libcommon/src/libcommon/state.py @@ -12,0 +13 @@ from libcommon.simple_cache import CacheEntryMetadata, fetch_names +from libcommon.utils import get_datetime @@ -91,0 +93,7 @@ class CacheState: + def is_old(self, days: int) -> bool: + if self.cache_entry_metadata is None or days <= 0: + return False + return self.cache_entry_metadata["updated_at"] < get_datetime(days).replace(tzinfo=None) + # ^ we remove the timezone to avoid comparing timezone-aware and timezone-naive datetimes + # could be done better, but we don't need more precision + diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py index b921ea78..ac79f831 100644 --- a/libs/libcommon/src/libcommon/utils.py +++ b/libs/libcommon/src/libcommon/utils.py @@ -7 +7 @@ import mimetypes -from datetime import datetime, timezone +from datetime import datetime, timedelta, timezone @@ -91,2 +91,5 @@ def orjson_dumps(content: Any) -> bytes: -def get_datetime() -> datetime: - return datetime.now(timezone.utc) +def get_datetime(days: Optional[int] = None) -> datetime: + date = datetime.now(timezone.utc) + if days is not None: + date = date - timedelta(days=days) + return date diff --git a/libs/libcommon/tests/test_backfill.py b/libs/libcommon/tests/test_backfill.py index 404f35a0..d69c231a 100644 --- a/libs/libcommon/tests/test_backfill.py +++ b/libs/libcommon/tests/test_backfill.py @@ -12 +12 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.utils import Priority, Status +from libcommon.utils import Priority, Status, get_datetime @@ -86,0 +87 @@ def test_initial_state( + "cache_is_old": [], @@ -121,0 +123 @@ def test_da_is_computed( + "cache_is_old": [], @@ -155,0 +158 @@ def test_ca_1_is_computed( + "cache_is_old": [], @@ -199,0 +203 @@ def test_plan_one_job_creation_and_termination( + "cache_is_old": [], @@ -218,0 +223 @@ def test_plan_one_job_creation_and_termination( + "cache_is_old": [], @@ -237,0 +243 @@ def test_plan_one_job_creation_and_termination( + "cache_is_old": [], @@ -282,0 +289 @@ def test_plan_all_job_creation_and_termination(processing_graph: ProcessingGraph + "cache_is_old": [], @@ -299,0 +307 @@ def test_plan_all_job_creation_and_termination(processing_graph: ProcessingGraph + "cache_is_old": [], @@ -343,0 +352 @@ def test_plan_compute_all(processing_graph: ProcessingGraph, up_to_date: List[st + "cache_is_old": [], @@ -383,0 +393 @@ def test_plan_retry_error_and_outdated_by_parent( + "cache_is_old": [], @@ -394,0 +405,29 @@ def test_plan_retry_error_and_outdated_by_parent( [email protected]( + "days_ago,is_old", + [(10, False), (30, True)], +) +def test_plan_old(days_ago: int, is_old: bool) -> None: + compute_all(processing_graph=PROCESSING_GRAPH_ONE_STEP) + + CACHE_MAX_DAYS = 20 + put_cache(step=STEP_DA, dataset=DATASET_NAME, revision=REVISION_NAME, updated_at=get_datetime(days_ago)) + + dataset_backfill_plan = get_dataset_backfill_plan( + processing_graph=PROCESSING_GRAPH_ONE_STEP, cache_max_days=CACHE_MAX_DAYS + ) + assert_dataset_backfill_plan( + dataset_backfill_plan=dataset_backfill_plan, + cache_status={ + "cache_has_different_git_revision": [], + "cache_is_old": [ARTIFACT_DA] if is_old else [], + "cache_is_outdated_by_parent": [], + "cache_is_empty": [], + "cache_is_error_to_retry": [], + "cache_is_job_runner_obsolete": [], + "up_to_date": [] if is_old else [ARTIFACT_DA], + }, + queue_status={"in_process": []}, + tasks=["CreateJobs,1"] if is_old else [], + ) + + @@ -428,0 +468 @@ def test_plan_outdated_by_parent( + "cache_is_old": [], @@ -472,0 +513 @@ def test_plan_job_runner_version_and_outdated_by_parent( + "cache_is_old": [], @@ -516,0 +558 @@ def test_plan_git_revision_and_outdated_by_parent( + "cache_is_old": [], @@ -562,0 +605 @@ def test_plan_fan_in_updated( + "cache_is_old": [], @@ -652,0 +696 @@ def test_plan_incoherent_state( + "cache_is_old": [], @@ -669,0 +714 @@ def test_plan_incoherent_state( + "cache_is_old": [], @@ -795,0 +841 @@ def test_delete_jobs( + "cache_is_old": [], @@ -827,0 +874 @@ def test_multiple_revisions() -> None: + "cache_is_old": [], @@ -848,0 +896 @@ def test_multiple_revisions() -> None: + "cache_is_old": [], @@ -866,0 +915 @@ def test_multiple_revisions() -> None: + "cache_is_old": [], @@ -884,0 +934 @@ def test_multiple_revisions() -> None: + "cache_is_old": [], diff --git a/libs/libcommon/tests/test_backfill_on_real_graph.py b/libs/libcommon/tests/test_backfill_on_real_graph.py index dfc46497..5cb1cf36 100644 --- a/libs/libcommon/tests/test_backfill_on_real_graph.py +++ b/libs/libcommon/tests/test_backfill_on_real_graph.py @@ -49,0 +50 @@ def test_plan_job_creation_and_termination() -> None: + "cache_is_old": [], @@ -81,0 +83 @@ def test_plan_job_creation_and_termination() -> None: + "cache_is_old": [], @@ -137,0 +140 @@ def test_plan_job_creation_and_termination() -> None: + "cache_is_old": [], diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py index 3131bc1a..32a687c4 100644 --- a/libs/libcommon/tests/test_orchestrator.py +++ b/libs/libcommon/tests/test_orchestrator.py @@ -40,0 +41,2 @@ from .utils import ( +CACHE_MAX_DAYS = 90 + @@ -205 +207,3 @@ def test_set_revision( - dataset_orchestrator.set_revision(revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[]) + dataset_orchestrator.set_revision( + revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[], cache_max_days=CACHE_MAX_DAYS + ) @@ -239 +243,3 @@ def test_set_revision_handle_existing_jobs( - dataset_orchestrator.set_revision(revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[]) + dataset_orchestrator.set_revision( + revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[], cache_max_days=CACHE_MAX_DAYS + ) diff --git a/libs/libcommon/tests/utils.py b/libs/libcommon/tests/utils.py index f6bd9197..e4addd22 100644 --- a/libs/libcommon/tests/utils.py +++ b/libs/libcommon/tests/utils.py @@ -3,0 +4 @@ +from datetime import datetime @@ -28,0 +30 @@ SPLIT_NAMES_CONTENT = { +CACHE_MAX_DAYS = 90 @@ -211,0 +214 @@ def get_dataset_backfill_plan( + cache_max_days: Optional[int] = None, @@ -217,0 +221 @@ def get_dataset_backfill_plan( + cache_max_days=CACHE_MAX_DAYS if cache_max_days is None else cache_max_days, @@ -265,0 +270 @@ def put_cache( + updated_at: Optional[datetime] = None, @@ -298,0 +304 @@ def put_cache( + updated_at=updated_at, diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py index 120725b5..33855962 100644 --- a/services/admin/src/admin/app.py +++ b/services/admin/src/admin/app.py @@ -87,0 +88 @@ def create_app() -> Starlette: + cache_max_days=app_config.cache.max_days, @@ -99,0 +101 @@ def create_app() -> Starlette: + cache_max_days=app_config.cache.max_days, diff --git a/services/admin/src/admin/routes/dataset_backfill.py b/services/admin/src/admin/routes/dataset_backfill.py index 3604391f..7df678fa 100644 --- a/services/admin/src/admin/routes/dataset_backfill.py +++ b/services/admin/src/admin/routes/dataset_backfill.py @@ -28,0 +29 @@ def create_dataset_backfill_endpoint( + cache_max_days: int, @@ -53 +54,3 @@ def create_dataset_backfill_endpoint( - dataset_orchestrator.backfill(revision=dataset_git_revision, priority=Priority.LOW) + dataset_orchestrator.backfill( + revision=dataset_git_revision, priority=Priority.LOW, cache_max_days=cache_max_days + ) diff --git a/services/admin/src/admin/routes/dataset_backfill_plan.py b/services/admin/src/admin/routes/dataset_backfill_plan.py index ea9d071b..908a8b48 100644 --- a/services/admin/src/admin/routes/dataset_backfill_plan.py +++ b/services/admin/src/admin/routes/dataset_backfill_plan.py @@ -28,0 +29 @@ def create_dataset_backfill_plan_endpoint( + cache_max_days: int, @@ -55,0 +57 @@ def create_dataset_backfill_plan_endpoint( + cache_max_days=cache_max_days, diff --git a/services/admin/src/admin/routes/dataset_state.py b/services/admin/src/admin/routes/dataset_state.py deleted file mode 100644 index 7daae82e..00000000 --- a/services/admin/src/admin/routes/dataset_state.py +++ /dev/null @@ -1,63 +0,0 @@ -# SPDX-License-Identifier: Apache-2.0 -# Copyright 2022 The HuggingFace Authors. - -import logging -from typing import Optional - -from libcommon.dataset import get_dataset_git_revision -from libcommon.orchestrator import DatasetBackfillPlan -from libcommon.processing_graph import ProcessingGraph -from starlette.requests import Request -from starlette.responses import Response - -from admin.authentication import auth_check -from admin.utils import ( - AdminCustomError, - Endpoint, - MissingRequiredParameterError, - UnexpectedError, - are_valid_parameters, - get_json_admin_error_response, - get_json_ok_response, -) - - -def create_dataset_backfill_plan_endpoint( - processing_graph: ProcessingGraph, - max_age: int, - hf_endpoint: str, - external_auth_url: Optional[str] = None, - organization: Optional[str] = None, - hf_token: Optional[str] = None, - hf_timeout_seconds: Optional[float] = None, -) -> Endpoint: - async def dataset_backfill_plan_endpoint(request: Request) -> Response: - try: - dataset = request.query_params.get("dataset") - if not are_valid_parameters([dataset]) or not dataset: - raise MissingRequiredParameterError("Parameter 'dataset' is required") - logging.info(f"/dataset-backfill-plan, dataset={dataset}") - - # if auth_check fails, it will raise an exception that will be caught below - auth_check( - external_auth_url=external_auth_url, - request=request, - organization=organization, - hf_timeout_seconds=hf_timeout_seconds, - ) - - dataset_git_revision = get_dataset_git_revision( - dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds - ) - dataset_backfill_plan = DatasetBackfillPlan( - dataset=dataset, - processing_graph=processing_graph, - revision=dataset_git_revision, - ) - return get_json_ok_response(dataset_backfill_plan.as_response(), max_age=max_age) - except AdminCustomError as e: - return get_json_admin_error_response(e, max_age=max_age) - except Exception as e: - return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age) - - return dataset_backfill_plan_endpoint diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index 13479c33..a124b5bb 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -82,0 +83 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + cache_max_days=app_config.cache.max_days, @@ -102 +103,3 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi - processing_graph=processing_graph, hf_webhook_secret=app_config.api.hf_webhook_secret + processing_graph=processing_graph, + hf_webhook_secret=app_config.api.hf_webhook_secret, + cache_max_days=app_config.cache.max_days, @@ -121,0 +125 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + cache_max_days=app_config.cache.max_days, diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py index 4ddb8bc5..71f6f00e 100644 --- a/services/api/src/api/routes/endpoint.py +++ b/services/api/src/api/routes/endpoint.py @@ -63,0 +64 @@ def get_cache_entry_from_steps( + cache_max_days: int, @@ -88,0 +90 @@ def get_cache_entry_from_steps( + cache_max_days=cache_max_days, @@ -223,0 +226 @@ def create_endpoint( + cache_max_days: int, @@ -296,0 +300 @@ def create_endpoint( + cache_max_days=cache_max_days, diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index d0689d8e..18430cb7 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -256,0 +257 @@ def create_rows_endpoint( + cache_max_days: int, @@ -331,0 +333 @@ def create_rows_endpoint( + cache_max_days=cache_max_days, diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py index ecb375a9..96f2c810 100644 --- a/services/api/src/api/routes/webhook.py +++ b/services/api/src/api/routes/webhook.py @@ -65,0 +66 @@ def process_payload( + cache_max_days: int, @@ -84 +85,5 @@ def process_payload( - dataset=dataset, revision=revision, processing_graph=processing_graph, priority=Priority.NORMAL + dataset=dataset, + revision=revision, + processing_graph=processing_graph, + priority=Priority.NORMAL, + cache_max_days=cache_max_days, @@ -90 +95,5 @@ def process_payload( - dataset=moved_to, revision=revision, processing_graph=processing_graph, priority=Priority.NORMAL + dataset=moved_to, + revision=revision, + processing_graph=processing_graph, + priority=Priority.NORMAL, + cache_max_days=cache_max_days, @@ -95 +104,3 @@ def process_payload( -def create_webhook_endpoint(processing_graph: ProcessingGraph, hf_webhook_secret: Optional[str] = None) -> Endpoint: +def create_webhook_endpoint( + processing_graph: ProcessingGraph, cache_max_days: int, hf_webhook_secret: Optional[str] = None +) -> Endpoint: @@ -130 +141,6 @@ def create_webhook_endpoint(processing_graph: ProcessingGraph, hf_webhook_secret - process_payload(processing_graph=processing_graph, payload=payload, trust_sender=trust_sender) + process_payload( + processing_graph=processing_graph, + payload=payload, + trust_sender=trust_sender, + cache_max_days=cache_max_days, + ) diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py index 079529ee..f6defeee 100644 --- a/services/api/src/api/utils.py +++ b/services/api/src/api/utils.py @@ -175,0 +176 @@ def try_backfill_dataset( + cache_max_days: int, @@ -195 +196,3 @@ def try_backfill_dataset( - dataset_orchestrator.set_revision(revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[]) + dataset_orchestrator.set_revision( + revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[], cache_max_days=cache_max_days + ) diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index 69e0dfd2..a3a98220 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -16,0 +17,2 @@ from api.utils import ResponseNotReadyError +CACHE_MAX_DAYS = 90 + @@ -123,6 +125,7 @@ def test_get_cache_entry_from_steps() -> None: - [step_without_error, step_with_error], - dataset, - config, - None, - processing_graph, - app_config.common.hf_endpoint, + processing_steps=[step_without_error, step_with_error], + dataset=dataset, + config=config, + split=None, + processing_graph=processing_graph, + hf_endpoint=app_config.common.hf_endpoint, + cache_max_days=CACHE_MAX_DAYS, @@ -135,6 +138,7 @@ def test_get_cache_entry_from_steps() -> None: - [step_with_error, step_without_error], - dataset, - config, - None, - processing_graph, - app_config.common.hf_endpoint, + processing_steps=[step_with_error, step_without_error], + dataset=dataset, + config=config, + split=None, + processing_graph=processing_graph, + hf_endpoint=app_config.common.hf_endpoint, + cache_max_days=CACHE_MAX_DAYS, @@ -147 +151,7 @@ def test_get_cache_entry_from_steps() -> None: - [step_with_error, step_with_error], dataset, config, None, processing_graph, app_config.common.hf_endpoint + processing_steps=[step_with_error, step_with_error], + dataset=dataset, + config=config, + split=None, + processing_graph=processing_graph, + hf_endpoint=app_config.common.hf_endpoint, + cache_max_days=CACHE_MAX_DAYS, @@ -160 +170,7 @@ def test_get_cache_entry_from_steps() -> None: - [non_existent_step], dataset, None, None, processing_graph, app_config.common.hf_endpoint + processing_steps=[non_existent_step], + dataset=dataset, + config=None, + split=None, + processing_graph=processing_graph, + hf_endpoint=app_config.common.hf_endpoint, + cache_max_days=CACHE_MAX_DAYS, diff --git a/tools/docker-compose-base.yml b/tools/docker-compose-base.yml index 0fdef1d0..19ab22aa 100644 --- a/tools/docker-compose-base.yml +++ b/tools/docker-compose-base.yml @@ -12,0 +13 @@ services: + CACHE_MAX_DAYS: ${CACHE_MAX_DAYS-90} diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index b9ca8e9e..59ed778e 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -12,0 +13 @@ services: + CACHE_MAX_DAYS: ${CACHE_MAX_DAYS-90}
288a5de35f9f6c7cd05cbbb25bea5f38bd459992
Quentin Lhoest
2023-07-03T15:40:31
Stream convert to parquet (#1448)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index b0dc9cf6..1566f1fa 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -124 +124 @@ cacheMaintenance: - action: "skip" + action: "backfill" @@ -129 +129 @@ cacheMaintenance: - error_codes_to_retry: "" + error_codes_to_retry: "DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError" diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 6d18f4da..eaf51193 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -49,0 +50,3 @@ from mongodb_migration.migrations._20230622131500_lock_add_owner import ( +from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_config_parquet_and_info import ( + MigrationAddPartialToCacheResponse, +) @@ -237,0 +241,3 @@ class MigrationsCollector: + MigrationAddPartialToCacheResponse( + version="20230703110100", description="add 'partial' field to config-parquet-and-info" + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py new file mode 100644 index 00000000..c1f0b674 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py @@ -0,0 +1,42 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS +from libcommon.simple_cache import CachedResponseDocument +from mongoengine.connection import get_db + +from mongodb_migration.check import check_documents +from mongodb_migration.migration import Migration + + +# connection already occurred in the main.py (caveat: we use globals) +class MigrationAddPartialToCacheResponse(Migration): + def up(self) -> None: + # See https://docs.mongoengine.org/guide/migration.html#example-1-addition-of-a-field + logging.info( + "If missing, add the partial field with the default value (false) to the cached results of" + " config-parquet-and-info" + ) + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many( + { + "kind": "config-parquet-and-info", + "http_status": 200, + "content.partial": {"$exists": False}, + }, + {"$set": {"content.partial": False}}, + ) + + def down(self) -> None: + logging.info("Remove the partial field from all the cached results") + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].update_many( + {"kind": "config-parquet-and-info", "http_status": 200}, {"$unset": {"content.partial": ""}} + ) + + def validate(self) -> None: + logging.info("Ensure that a random selection of cached results have the 'partial' field") + + check_documents(DocCls=CachedResponseDocument, sample_size=10) diff --git a/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py new file mode 100644 index 00000000..85e1a036 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py @@ -0,0 +1,192 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_config_parquet_and_info import ( + MigrationAddPartialToCacheResponse, +) + + +def assert_partial(dataset: str) -> None: + db = get_db(CACHE_MONGOENGINE_ALIAS) + entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset}) + assert entry is not None + assert entry["content"]["partial"] is False + + +def assert_unchanged(dataset: str) -> None: + db = get_db(CACHE_MONGOENGINE_ALIAS) + entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset}) + assert entry is not None + assert "partial" not in entry["content"] + + +def test_cache_add_partial(mongo_host: str) -> None: + with MongoResource(database="test_cache_add_partial", host=mongo_host, mongoengine_alias="cache"): + db = get_db(CACHE_MONGOENGINE_ALIAS) + db[CACHE_COLLECTION_RESPONSES].insert_many( + [ + { + "config": "lhoestq--demo1", + "dataset": "lhoestq/demo1", + "kind": "config-parquet-and-info", + "split": None, + "content": { + "parquet_files": [ + { + "dataset": "lhoestq/demo1", + "config": "lhoestq--demo1", + "split": "test", + "url": "https://huggingface.co/.../csv-test.parquet", + "filename": "csv-test.parquet", + "size": 4415, + }, + { + "dataset": "lhoestq/demo1", + "config": "lhoestq--demo1", + "split": "train", + "url": "https://huggingface.co/.../csv-train.parquet", + "filename": "csv-train.parquet", + "size": 5038, + }, + ], + "dataset_info": { + "description": "", + "citation": "", + "homepage": "", + "license": "", + "features": {}, + "builder_name": "csv", + "config_name": "lhoestq--demo1", + "version": {}, + "splits": {}, + "download_checksums": {}, + "download_size": 2340, + "dataset_size": 2464, + "size_in_bytes": 4804, + }, + }, + "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11", + "details": None, + "error_code": None, + "http_status": 200, + "job_runner_version": 3, + "progress": 1.0, + }, + { + "config": "lhoestq--demo2", + "dataset": "lhoestq/demo2", + "kind": "config-parquet-and-info", + "split": None, + "content": { + "parquet_files": [ + { + "dataset": "lhoestq/demo2", + "config": "lhoestq--demo2", + "split": "test", + "url": "https://huggingface.co/.../csv-test.parquet", + "filename": "csv-test.parquet", + "size": 4415, + }, + { + "dataset": "lhoestq/demo2", + "config": "lhoestq--demo2", + "split": "train", + "url": "https://huggingface.co/.../csv-train.parquet", + "filename": "csv-train.parquet", + "size": 5038, + }, + ], + "dataset_info": { + "description": "", + "citation": "", + "homepage": "", + "license": "", + "features": {}, + "builder_name": "csv", + "config_name": "lhoestq--demo2", + "version": {}, + "splits": {}, + "download_checksums": {}, + "download_size": 2340, + "dataset_size": 2464, + "size_in_bytes": 4804, + }, + }, + "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11", + "details": None, + "error_code": None, + "http_status": 200, + "job_runner_version": 3, + "progress": 1.0, + }, + { + "config": "lhoestq--error", + "dataset": "lhoestq/error", + "kind": "config-parquet-and-info", + "split": None, + "content": {"error": "Streaming is not supported for lhoestq/error"}, + "dataset_git_revision": "ec3c8d414af3dfe600399f5e6ef2c682938676f3", + "details": { + "error": "Streaming is not supported for lhoestq/error", + "cause_exception": "TypeError", + "cause_message": "Streaming is not supported for lhoestq/error", + "cause_traceback": [ + "Traceback (most recent call last):\n", + ( + ' File "/src/services/worker/src/worker/job_manager.py", line 163, in process\n ' + " job_result = self.job_runner.compute()\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 932, in compute\n compute_config_parquet_and_info_response(\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 825, in compute_config_parquet_and_info_response\n raise_if_not_supported(\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 367, in raise_if_not_supported\n raise_if_too_big_from_external_data_files(\n" + ), + ( + ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line' + " 447, in raise_if_too_big_from_external_data_files\n " + " builder._split_generators(mock_dl_manager)\n" + ), + ( + " File" + ' "/tmp/modules-cache/datasets_modules/.../error.py",' + ' line 190, in _split_generators\n raise TypeError("Streaming is not supported for' + ' lhoestq/error")\n' + ), + "TypeError: Streaming is not supported for lhoestq/error\n", + ], + }, + "error_code": "UnexpectedError", + "http_status": 500, + "job_runner_version": 3, + "progress": None, + }, + ] + ) + + migration = MigrationAddPartialToCacheResponse( + version="20230703110100", + description="add partial field to config-parquet-and-info", + ) + migration.up() + + assert_partial("lhoestq/demo1") + assert_partial("lhoestq/demo2") + assert_unchanged("lhoestq/error") + + migration.down() + assert_unchanged("lhoestq/demo1") + assert_unchanged("lhoestq/demo2") + assert_unchanged("lhoestq/error") + + db[CACHE_COLLECTION_RESPONSES].drop() diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py index 46e66228..623f85a4 100644 --- a/libs/libcommon/src/libcommon/exceptions.py +++ b/libs/libcommon/src/libcommon/exceptions.py @@ -86,4 +85,0 @@ CacheableErrorCode = Literal[ - "DatasetTooBigFromDatasetsError", - "DatasetTooBigFromHubError", - "DatasetWithTooBigExternalFilesError", - "DatasetWithTooManyExternalFilesError", @@ -225,21 +220,0 @@ class DatasetRevisionNotFoundError(CacheableError): -class DatasetTooBigFromDatasetsError(CacheableError): - """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromDatasetsError", cause, False) - - -class DatasetTooBigFromHubError(CacheableError): - """Raised when the dataset size (sum of files on the Hub) is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromHubError", cause, False) - - -class DatasetWithTooBigExternalFilesError(CacheableError): - """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooBigExternalFilesError", cause, True) - - @@ -253,7 +227,0 @@ class DatasetWithTooManyConfigsError(CacheableError): -class DatasetWithTooManyExternalFilesError(CacheableError): - """Raised when the number of external data files of a dataset is too big.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyExternalFilesError", cause, True) - - diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py index 17a6a41d..ca029d38 100644 --- a/services/worker/src/worker/dtos.py +++ b/services/worker/src/worker/dtos.py @@ -117,0 +118 @@ class ConfigParquetAndInfoResponse(TypedDict): + partial: bool diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 14da1e9e..880b991a 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -4 +4 @@ -import glob +import functools @@ -5,0 +6 @@ import logging +import os @@ -7 +8 @@ import re -from functools import partial +from contextlib import ExitStack @@ -10 +11,14 @@ from pathlib import Path -from typing import Any, List, Optional, Set, Tuple +from types import TracebackType +from typing import ( + Any, + Callable, + Generator, + List, + Optional, + Set, + Tuple, + Type, + TypeVar, + Union, +) +from unittest.mock import patch @@ -20,0 +35 @@ from datasets import DownloadConfig, Features, load_dataset_builder +from datasets.arrow_writer import ParquetWriter @@ -56,4 +70,0 @@ from libcommon.exceptions import ( - DatasetTooBigFromDatasetsError, - DatasetTooBigFromHubError, - DatasetWithTooBigExternalFilesError, - DatasetWithTooManyExternalFilesError, @@ -91,0 +103,8 @@ MAX_OPERATIONS_PER_COMMIT = 500 +T = TypeVar("T") + + +def repo_file_rfilename_sort_key(repo_file: RepoFile) -> str: + if not isinstance(repo_file.rfilename, str): # check type for mypy + raise ValueError(f"Expected a string for repo_file.rfilename, but got a '{type(repo_file.rfilename)}'.") + return repo_file.rfilename + @@ -94 +113,3 @@ class ParquetFile: - def __init__(self, local_file: str, local_dir: str, config: str): + def __init__( + self, local_file: str, local_dir: str, config: str, split: str, shard_idx: int, partial: bool = False + ): @@ -96,0 +118,5 @@ class ParquetFile: + if shard_idx >= MAX_FILES_PER_DIRECTORY: + raise DatasetWithTooManyParquetFilesError( + "The dataset has too many parquet files and can't be uploaded in the parquet directory " + f"because it exceeds the maximum number of files per directory ({MAX_FILES_PER_DIRECTORY})." + ) @@ -99,0 +126,3 @@ class ParquetFile: + self.split = split + self.shard_idx = shard_idx + self.partial = partial @@ -103 +132,5 @@ class ParquetFile: - return f'{self.config}/{self.local_file.removeprefix(f"{self.local_dir}/")}' + if self.partial: + # Using 4 digits is ok since MAX_FILES_PER_DIRECTORY == 10_000 + return f"{self.config}/partial/{self.split}/{self.shard_idx:04d}.parquet" + else: + return f'{self.config}/{self.local_file.removeprefix(f"{self.local_dir}/")}' @@ -106 +139,2 @@ class ParquetFile: -p = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>\w+(\.\w+)*?)(-[0-9]{5}-of-[0-9]{5})?.parquet") +filename_pattern = re.compile("^[0-9]{4}\\.parquet$") +legacy_filename_pattern = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>\w+(\.\w+)*?)(-[0-9]{5}-of-[0-9]{5})?.parquet") @@ -109,0 +144,12 @@ def parse_repo_filename(filename: str) -> Tuple[str, str]: + if not filename_pattern.match(os.path.basename(filename)): + return parse_legacy_repo_filename(filename) + parts = filename.split("/") + if len(parts) == 4 and parts[1] == "partial": + parts.pop(1) + if len(parts) != 3: + raise ValueError(f"Invalid filename: {filename}") + config, split, _ = parts + return config, split + + +def parse_legacy_repo_filename(filename: str) -> Tuple[str, str]: @@ -114 +160 @@ def parse_repo_filename(filename: str) -> Tuple[str, str]: - m = p.match(fname) + m = legacy_filename_pattern.match(fname) @@ -186 +232 @@ def is_parquet_builder_with_hub_files(builder: DatasetBuilder, hf_endpoint: str) -def raise_if_too_big_from_hub( +def _is_too_big_from_hub( @@ -189 +235 @@ def raise_if_too_big_from_hub( -) -> None: +) -> bool: @@ -199,6 +244,0 @@ def raise_if_too_big_from_hub( - Returns: - `None` - Raises the following errors: - - [`libcommon.exceptions.DatasetTooBigFromHubError`] - If the dataset is too big to be converted to parquet, as measured by the sum of the repository - files sizes given by the Hub. @@ -207,5 +247 @@ def raise_if_too_big_from_hub( - if dataset_size > max_dataset_size: - raise DatasetTooBigFromHubError( - f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes. " - f"Current size of files on the hub is {dataset_size} bytes." - ) + return bool(dataset_size > max_dataset_size) @@ -214 +250 @@ def raise_if_too_big_from_hub( -def raise_if_too_big_from_datasets( +def _is_too_big_from_datasets( @@ -217 +253 @@ def raise_if_too_big_from_datasets( -) -> None: +) -> bool: @@ -227,8 +262,0 @@ def raise_if_too_big_from_datasets( - Returns: - `None` - Raises the following errors: - - [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) - If the datasets.config.HF_ENDPOINT is not set to the expected value - - [`libcommon.exceptions.DatasetTooBigFromDatasetsError`] - If the dataset is too big to be converted to parquet, as measured by the sum of the configs - sizes given by the datasets library. @@ -237,6 +265 @@ def raise_if_too_big_from_datasets( - if dataset_size > max_dataset_size: - raise DatasetTooBigFromDatasetsError( - f"The dataset is too big to be converted to Parquet. The size of the dataset ({dataset_size} B, as given" - f" per the datasets library) exceeds the maximum supported size ({max_dataset_size} B). Please report the" - " issue." - ) + return bool(dataset_size > max_dataset_size) @@ -285 +308 @@ def raise_if_requires_manual_download( -def raise_if_not_supported( +def is_dataset_too_big( @@ -292 +315 @@ def raise_if_not_supported( -) -> None: +) -> bool: @@ -294,4 +317,4 @@ def raise_if_not_supported( - Raise an error if the dataset is not supported: - - if the dataset is in the list of blocked datasets - - if the dataset cannot be accessed (does not exist, private) - - if the dataset is too big, and not in the list of supported datasets + Check: + - the size of the dataset repository + - the size in dataset info + - the size and number of external files @@ -320,14 +342,0 @@ def raise_if_not_supported( - - [`libcommon.exceptions.DatasetManualDownloadError`]: - If the dataset requires manual download. - - [`libcommon.exceptions.DatasetRevisionNotFoundError`] - If the revision does not exist or cannot be accessed using the token. - - [`libcommon.exceptions.DatasetTooBigFromDatasetsError`] - If the dataset is too big to be converted to parquet, as measured by the sum of the configs - sizes given by the datasets library. - - [`libcommon.exceptions.DatasetTooBigFromHubError`] - If the dataset is too big to be converted to parquet, as measured by the sum of the repository - files sizes given by the Hub. - - [`libcommon.exceptions.DatasetWithTooManyExternalFilesError`] - If the dataset has too many external files to be converted to parquet - - [`libcommon.exceptions.DatasetWithTooBigExternalFilesError`] - If the dataset is too big external files be converted to parquet @@ -352,15 +361,12 @@ def raise_if_not_supported( - raise_if_requires_manual_download( - builder=builder, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - ) - raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size) - raise_if_too_big_from_external_data_files( - builder=builder, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - hf_token=hf_token, - ) - raise_if_too_big_from_datasets( - builder.info, - max_dataset_size=max_dataset_size, + return ( + _is_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size) + or _is_too_big_from_datasets( + builder.info, + max_dataset_size=max_dataset_size, + ) + or _is_too_big_from_external_data_files( + builder=builder, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + hf_token=hf_token, + ) @@ -426 +432 @@ class _MockStreamingDownloadManager(StreamingDownloadManager): # type: ignore -def raise_if_too_big_from_external_data_files( +def _is_too_big_from_external_data_files( @@ -428 +434 @@ def raise_if_too_big_from_external_data_files( -) -> None: +) -> bool: @@ -432 +438 @@ def raise_if_too_big_from_external_data_files( - return + return False @@ -488,4 +494 @@ def raise_if_too_big_from_external_data_files( - raise DatasetWithTooManyExternalFilesError( - f"The conversion to parquet is limited to datasets with less than {max_external_data_files} files. " - f"However it uses {len(ext_data_files)} data files." - ) + return True @@ -496 +499 @@ def raise_if_too_big_from_external_data_files( - get_size = partial(_request_size, hf_token=hf_token) + get_size = functools.partial(_request_size, hf_token=hf_token) @@ -500,6 +503,2 @@ def raise_if_too_big_from_external_data_files( - if total_size > max_dataset_size: - raise DatasetWithTooBigExternalFilesError( - f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes." - f" However {i + 1} data files of {len(ext_data_files)} are already bigger than" - f" {total_size} bytes." - ) + return total_size > max_dataset_size + return False @@ -542,0 +542 @@ def raise_if_too_big_from_external_data_files( + return False @@ -648 +648 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None: - partial(retry_get_parquet_file_and_size, fs=fs, hf_token=hf_token), + functools.partial(retry_get_parquet_file_and_size, fs=fs, hf_token=hf_token), @@ -668,0 +669,160 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None: +class limit_parquet_writes: + """ + Context manager that limits the number of bytes a `DatasetBuilder` can write to parquet. + + It works by monitoring the calls to `pq.ParquetWriter.write_table` and stopping + the `GeneratorBasedBuilder._generate_examples` and `ArrowBasedBuilder._generate_tables` + generators once we reach the maximum number of bytes. + + Since the generator is stopped after we reach the maximum number of bytes, the actual + number of bytes generated might be slightly higher than the requested limit. + + Example of usage: + + ```python + builder = load_dataset_builder("squad") + max_dataset_size = 10_000_000 + with limit_parquet_writes(builder, max_dataset_size=max_dataset_size) as limiter: + builder.download_and_prepare(file_format="parquet") + assert builder.info.dataset_size == limiter.total_bytes < max_dataset_size + epsilon + ``` + + The limiter is usually used with a `StreamingDownloadManager` to not have to download + the full dataset: + + ```python + builder = load_dataset_builder("squad") + max_dataset_size = 10_000_000 + dl_manager = StreamingDownloadManager(...) + for split_generator in builder._split_generators(dl_manager): + with limit_parquet_writes(builder, max_dataset_size=max_dataset_size): + builder._prepare_split(split_generator=split_generator, file_format="parquet") + ``` + """ + + def __init__( + self, + builder: Union[datasets.builder.GeneratorBasedBuilder, datasets.builder.ArrowBasedBuilder], + max_dataset_size: int, + ) -> None: + self.total_bytes = 0 + self.builder = builder + self.max_dataset_size = max_dataset_size + self.exit_stack = ExitStack() + + def __enter__(self) -> "limit_parquet_writes": + limiter = self + + class _TrackedParquetWriter(pq.ParquetWriter): # type: ignore + """Count on-the-fly how many bytes are written""" + + def track_write_table(self, pa_table: pa.Table) -> None: + limiter.total_bytes += pa_table.nbytes + + def write_table(self, pa_table: pa.Table, row_group_size: Optional[int] = None) -> None: + self.track_write_table(pa_table) + super().write_table(pa_table, row_group_size=row_group_size) + + def limited_generator( + generator: Callable[..., Generator[T, None, None]] + ) -> Callable[..., Generator[T, None, None]]: + """Stop the underlying generator once we reach the maximum dataset size""" + + @functools.wraps(generator) + def wrapped(*args: Any, **kwargs: Any) -> Generator[T, None, None]: + for item in generator(*args, **kwargs): + if limiter.total_bytes < limiter.max_dataset_size: + yield item + else: + break + + return wrapped + + self.exit_stack.enter_context(patch.object(ParquetWriter, "_WRITER_CLASS", _TrackedParquetWriter)) + if isinstance(self.builder, datasets.builder.GeneratorBasedBuilder): + self.exit_stack.enter_context( + patch.object(self.builder, "_generate_examples", limited_generator(self.builder._generate_examples)) + ) + else: + self.exit_stack.enter_context( + patch.object(self.builder, "_generate_tables", limited_generator(self.builder._generate_tables)) + ) + return self + + def __exit__( + self, + exc_type: Optional[Type[BaseException]], + exc_value: Optional[BaseException], + traceback: Optional[TracebackType], + ) -> None: + return self.exit_stack.close() + + +def list_generated_parquet_files(builder: DatasetBuilder, partial: bool = False) -> List[ParquetFile]: + """List the parquet files generated by `builder.download_and_prepare` in the `builder.cache_dir`.""" + if not builder.info.splits: + raise EmptyDatasetError("No split found after generating parquet files") + split_dict = builder.info.splits + local_parquet_files: List[ParquetFile] = [] + for split, split_info in split_dict.items(): + # We know the `datasets` library uses a template for the shards names: + # - {builder.name}-{split}.parquet if there is only one shard + # - {builder.name}-{split}-{shard_idx:05d}-of-{num_shards:05d}.parquet otherwise + num_shards = len(split_info.shard_lengths) if isinstance(split_info.shard_lengths, list) else 1 + fname_prefix = f"{builder.name}-{split}" + local_parquet_files.extend( + [ + ParquetFile( + local_file=os.path.join( + builder.cache_dir, + fname_prefix + + (f"-{shard_idx:05d}-of-{num_shards:05d}.parquet" if num_shards > 1 else ".parquet"), + ), + local_dir=builder.cache_dir, + config=builder.config.name, + split=split, + shard_idx=shard_idx, + partial=partial, + ) + for shard_idx in range(num_shards) + ] + ) + return local_parquet_files + + +def stream_convert_to_parquet(builder: DatasetBuilder, max_dataset_size: int) -> Tuple[List[CommitOperationAdd], bool]: + """Stream and prepare the dataset as parquet files and fills the builder info.""" + writer_batch_size = get_writer_batch_size(builder.info) + if writer_batch_size is not None and ( + builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size + ): + builder._writer_batch_size = writer_batch_size + dl_manager = StreamingDownloadManager( + base_path=builder.base_path, + download_config=DownloadConfig(use_auth_token=builder.use_auth_token, storage_options=builder.storage_options), + dataset_name=builder.name, + data_dir=builder.config.data_dir, + ) + os.makedirs(builder.cache_dir, exist_ok=True) + split_dict = SplitDict(dataset_name=builder.name) + splits_generators = {sg.name: sg for sg in builder._split_generators(dl_manager)} + partial = False + for split in splits_generators: + split_dict.add(splits_generators[split].split_info) + with limit_parquet_writes(builder, max_dataset_size=max_dataset_size) as limiter: + builder._prepare_split(split_generator=splits_generators[split], file_format="parquet") + partial = partial or limiter.total_bytes >= max_dataset_size + builder.info.splits = split_dict + builder.info.dataset_size = sum(split.num_bytes for split in builder.info.splits.values()) + builder.info.download_size = None + builder.info.size_in_bytes = None + + # send the files to the target revision + local_parquet_files = list_generated_parquet_files(builder, partial=partial) + parquet_operations: List[CommitOperationAdd] = [ + CommitOperationAdd(path_in_repo=parquet_file.path_in_repo, path_or_fileobj=parquet_file.local_file) + for parquet_file in local_parquet_files + ] + return parquet_operations, partial + + @@ -670 +830 @@ def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]: - """Download and prepare the dataset as parquet files and fills the builder info""" + """Download and prepare the dataset as parquet files and fills the builder info.""" @@ -680,4 +840 @@ def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]: - local_parquet_files = [ - ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=builder.config.name) - for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet") - ] + local_parquet_files = list_generated_parquet_files(builder) @@ -1016,0 +1174 @@ def compute_config_parquet_and_info_response( + partial = False @@ -1020,0 +1179,5 @@ def compute_config_parquet_and_info_response( + raise_if_requires_manual_download( + builder=builder, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + ) @@ -1024,10 +1187,11 @@ def compute_config_parquet_and_info_response( - if dataset not in supported_datasets: - raise_if_not_supported( - dataset_info=dataset_info, - builder=builder, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - ) - parquet_operations = convert_to_parquet(builder) + if is_dataset_too_big( + dataset_info=dataset_info, + builder=builder, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + ): + parquet_operations, partial = stream_convert_to_parquet(builder, max_dataset_size=max_dataset_size) + else: + parquet_operations = convert_to_parquet(builder) @@ -1071,0 +1236 @@ def compute_config_parquet_and_info_response( + repo_files.sort(key=repo_file_rfilename_sort_key) @@ -1088,0 +1254 @@ def compute_config_parquet_and_info_response( + partial=partial, diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 11e9d9a0..1ceb456a 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -374,0 +375,18 @@ def create_dataset_info_response_for_csv(dataset: str, config: str) -> Any: +def create_dataset_info_response_for_partially_generated_big_csv(config: str) -> Any: + # Dataset is partially converted to parquet: the first 10KB instead of the full 5MB + # Missing fields: + # - download_size: not applicable, because the dataset is generated using partially downloaded files + return { + "description": "", + "citation": "", + "homepage": "", + "license": "", + "features": BIG_cols, + "builder_name": "csv", + "config_name": config, + "version": {"version_str": "0.0.0", "major": 0, "minor": 0, "patch": 0}, + "splits": {"train": {"name": "train", "num_bytes": 12380, "num_examples": 10, "dataset_name": "csv"}}, + "dataset_size": 12380, + } + + @@ -419 +437,3 @@ def create_parquet_and_info_response( - dataset: str, data_type: Literal["csv", "audio", "big_parquet", "big_parquet_no_info"] + dataset: str, + data_type: Literal["csv", "big-csv", "audio", "big_parquet", "big_parquet_no_info"], + partial: bool = False, @@ -423,2 +443,13 @@ def create_parquet_and_info_response( - filename = "csv-train.parquet" if data_type == "csv" else "parquet-train.parquet" - size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE if data_type == "audio" else BIG_PARQUET_FILE + if partial: + filename = "0000.parquet" + else: + filename = "csv-train.parquet" if "csv" in data_type else "parquet-train.parquet" + size = ( + CSV_PARQUET_SIZE + if data_type == "csv" + else PARTIAL_CSV_PARQUET_SIZE + if data_type == "big-csv" + else AUDIO_PARQUET_SIZE + if data_type == "audio" + else BIG_PARQUET_FILE + ) @@ -427,0 +459,2 @@ def create_parquet_and_info_response( + else create_dataset_info_response_for_partially_generated_big_csv(config) + if data_type == "big-csv" @@ -441 +474,3 @@ def create_parquet_and_info_response( - repo_id=f"datasets/{dataset}", revision="refs%2Fconvert%2Fparquet", filename=f"{config}/{filename}" + repo_id=f"datasets/{dataset}", + revision="refs%2Fconvert%2Fparquet", + filename=f"{config}/partial/{split}/{filename}" if partial else f"{config}/{filename}", @@ -447,0 +483 @@ def create_parquet_and_info_response( + "partial": partial, @@ -451,0 +488 @@ CSV_PARQUET_SIZE = 1_866 +PARTIAL_CSV_PARQUET_SIZE = 8_188 @@ -748 +785,3 @@ def hub_reponses_big_csv(hub_public_big_csv: str) -> HubDatasetTest: - "parquet_and_info_response": None, + "parquet_and_info_response": create_parquet_and_info_response( + dataset=hub_public_big_csv, data_type="big-csv", partial=True + ), diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index c7db50e4..c2e9f007 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -84,0 +85 @@ def get_job_runner( + partial=False, @@ -148,0 +150 @@ def get_job_runner( + partial=False, diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index ec3bc9c2..6c4f72f2 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -4,0 +5 @@ import io +import os @@ -11 +12,2 @@ from pathlib import Path -from typing import Any, Callable, Iterator, List, Optional, Set, TypedDict +from typing import Any, Callable, Dict, Iterator, List, Optional, Set, TypedDict +from unittest.mock import patch @@ -13,0 +16 @@ import datasets.builder +import datasets.config @@ -15,0 +19 @@ import pandas as pd +import pyarrow.parquet as pq @@ -18,0 +23,3 @@ from datasets import Audio, Features, Image, Value, load_dataset_builder +from datasets.packaged_modules.generator.generator import ( + Generator as ParametrizedGeneratorBasedBuilder, +) @@ -25,4 +31,0 @@ from libcommon.exceptions import ( - DatasetTooBigFromDatasetsError, - DatasetTooBigFromHubError, - DatasetWithTooBigExternalFilesError, - DatasetWithTooManyExternalFilesError, @@ -40,0 +44,3 @@ from worker.job_runners.config.parquet_and_info import ( + _is_too_big_from_datasets, + _is_too_big_from_external_data_files, + _is_too_big_from_hub, @@ -43,0 +50,2 @@ from worker.job_runners.config.parquet_and_info import ( + limit_parquet_writes, + list_generated_parquet_files, @@ -47,3 +55 @@ from worker.job_runners.config.parquet_and_info import ( - raise_if_too_big_from_datasets, - raise_if_too_big_from_external_data_files, - raise_if_too_big_from_hub, + stream_convert_to_parquet, @@ -112 +118 @@ def assert_content_is_equal(content: Any, expected: Any) -> None: - assert set(content) == {"parquet_files", "dataset_info"}, content + assert set(content) == {"parquet_files", "dataset_info", "partial"}, content @@ -120,0 +127 @@ def assert_content_is_equal(content: Any, expected: Any) -> None: + assert content["partial"] == expected["partial"], content @@ -241 +248 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_ - "name,raises", + "name,expected", @@ -244 +251 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_ -def test_raise_if_too_big_from_hub( +def test__is_too_big_from_hub( @@ -248 +255 @@ def test_raise_if_too_big_from_hub( - raises: bool, + expected: bool, @@ -259,9 +266,4 @@ def test_raise_if_too_big_from_hub( - if raises: - with pytest.raises(DatasetTooBigFromHubError): - raise_if_too_big_from_hub( - dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size - ) - else: - raise_if_too_big_from_hub( - dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size - ) + assert ( + _is_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size) + == expected + ) @@ -271 +273 @@ def test_raise_if_too_big_from_hub( - "name,raises", + "name,expected", @@ -274 +276 @@ def test_raise_if_too_big_from_hub( -def test_raise_if_too_big_from_datasets( +def test__is_too_big_from_datasets( @@ -278 +280 @@ def test_raise_if_too_big_from_datasets( - raises: bool, + expected: bool, @@ -283,8 +285,2 @@ def test_raise_if_too_big_from_datasets( - if raises: - with pytest.raises(DatasetTooBigFromDatasetsError): - raise_if_too_big_from_datasets( - info=builder.info, - max_dataset_size=app_config.parquet_and_info.max_dataset_size, - ) - else: - raise_if_too_big_from_datasets( + assert ( + _is_too_big_from_datasets( @@ -293,0 +290,2 @@ def test_raise_if_too_big_from_datasets( + == expected + ) @@ -297 +295 @@ def test_raise_if_too_big_from_datasets( - "max_dataset_size,max_external_data_files,raises", + "max_dataset_size,max_external_data_files,expected", @@ -303 +301 @@ def test_raise_if_too_big_from_datasets( -def test_raise_if_too_big_external_files( +def test__is_too_big_external_files( @@ -305 +303 @@ def test_raise_if_too_big_external_files( - raises: bool, + expected: bool, @@ -312,10 +310,2 @@ def test_raise_if_too_big_external_files( - if raises: - with pytest.raises(DatasetWithTooBigExternalFilesError): - raise_if_too_big_from_external_data_files( - builder=external_files_dataset_builder, - hf_token=app_config.common.hf_token, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - ) - else: - raise_if_too_big_from_external_data_files( + assert ( + _is_too_big_from_external_data_files( @@ -326,0 +317,2 @@ def test_raise_if_too_big_external_files( + == expected + ) @@ -330 +322 @@ def test_raise_if_too_big_external_files( - "max_dataset_size,max_external_data_files,raises", + "max_dataset_size,max_external_data_files,expected", @@ -338 +330 @@ def test_raise_if_too_many_external_files( - raises: bool, + expected: bool, @@ -345,10 +337,2 @@ def test_raise_if_too_many_external_files( - if raises: - with pytest.raises(DatasetWithTooManyExternalFilesError): - raise_if_too_big_from_external_data_files( - builder=external_files_dataset_builder, - hf_token=app_config.common.hf_token, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - ) - else: - raise_if_too_big_from_external_data_files( + assert ( + _is_too_big_from_external_data_files( @@ -359,0 +344,2 @@ def test_raise_if_too_many_external_files( + == expected + ) @@ -387 +373 @@ def test_supported_if_big_parquet( -def test_not_supported_if_big_non_parquet( +def test_partially_converted_if_big_non_parquet( @@ -403,3 +389,14 @@ def test_not_supported_if_big_non_parquet( - with pytest.raises(CustomError) as e: - job_runner.compute() - assert e.typename == "DatasetTooBigFromHubError" + from datasets.packaged_modules.csv.csv import CsvConfig + + # Set a small chunk size to yield more than one Arrow Table in _generate_tables + # to be able to stop the generation mid-way. + with patch.object(CsvConfig, "pd_read_csv_kwargs", {"chunksize": 10}): + response = job_runner.compute() + assert response + content = response.content + assert content + assert len(content["parquet_files"]) == 1 + assert_content_is_equal(content, hub_reponses_big_csv["parquet_and_info_response"]) + # dataset is partially generated + assert content["parquet_files"][0]["size"] < app_config.parquet_and_info.max_dataset_size + assert content["parquet_files"][0]["url"].endswith("/partial/train/0000.parquet") @@ -597,0 +595,3 @@ def test_previous_step_error( + ("config/partial/split/0000.parquet", "split", "config", False), + ("config/partial/split.with.dots/0000.parquet", "split.with.dots", "config", False), + ("config/partial/toomanyzeros/00000.parquet", "toomanyzeros", "config", True), @@ -828,0 +829,57 @@ def test_get_delete_operations( + + [email protected]( + "max_dataset_size,expected_num_shards", + [ + (1, 1), + (150, 2), + (300, 4), + (9999999, 10), + ], +) +def test_stream_convert_to_parquet( + csv_path: str, max_dataset_size: int, expected_num_shards: int, tmp_path: Path +) -> None: + num_data_files = 10 + builder = load_dataset_builder( + "csv", + data_files={"train": [csv_path] * num_data_files}, + cache_dir=str(tmp_path / f"test_stream_convert_to_parquet-{max_dataset_size=}"), + ) + with patch("worker.job_runners.config.parquet_and_info.get_writer_batch_size", lambda ds_config_info: 1): + with patch.object(datasets.config, "MAX_SHARD_SIZE", 1): + parquet_operations, partial = stream_convert_to_parquet(builder, max_dataset_size=max_dataset_size) + num_shards = len(parquet_operations) + assert num_shards == expected_num_shards + assert partial == (expected_num_shards < num_data_files) + assert all(isinstance(op.path_or_fileobj, str) for op in parquet_operations) + parquet_files = list_generated_parquet_files(builder, partial=partial) + assert len(parquet_files) == expected_num_shards + assert all(os.path.isfile(parquet_file.local_file) for parquet_file in parquet_files) + one_sample_max_size = 100 + expected_max_dataset_size = max_dataset_size + one_sample_max_size + assert ( + sum(pq.ParquetFile(parquet_file.local_file).read().nbytes for parquet_file in parquet_files) + < expected_max_dataset_size + ) + + +def test_limit_parquet_writes(tmp_path: Path) -> None: + num_examples = 0 + + def long_generator() -> Iterator[Dict[str, int]]: + nonlocal num_examples + for i in range(10_000_000): + yield {"foo": i} + num_examples += 1 + + one_sample_size = 8 + max_dataset_size = 50_000 + expected_max_dataset_size = max_dataset_size + datasets.config.DEFAULT_MAX_BATCH_SIZE * one_sample_size + expected_max_num_examples = 1 + max_dataset_size // one_sample_size + datasets.config.DEFAULT_MAX_BATCH_SIZE + cache_dir = str(tmp_path / "test_limit_parquet_writes_cache_dir") + builder = ParametrizedGeneratorBasedBuilder(generator=long_generator, cache_dir=cache_dir) + with limit_parquet_writes(builder, max_dataset_size=max_dataset_size) as limiter: + builder.download_and_prepare(file_format="parquet") + assert builder.info.dataset_size == limiter.total_bytes <= expected_max_dataset_size + assert builder.info.splits["train"].num_examples == num_examples < expected_max_num_examples
2996e47b6f4abca258b1a59d942dd1019aac2c60
Sylvain Lesage
2023-07-03T15:26:54
feat: 🎸 use Normal priority only for API (#1470)
diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py index 37ee3ee2..0797eef8 100644 --- a/libs/libcommon/src/libcommon/operations.py +++ b/libs/libcommon/src/libcommon/operations.py @@ -16 +16 @@ def backfill_dataset( - priority: Priority = Priority.NORMAL, + priority: Priority = Priority.LOW, @@ -25 +25 @@ def backfill_dataset( - priority (Priority, optional): The priority of the job. Defaults to Priority.NORMAL. + priority (Priority, optional): The priority of the job. Defaults to Priority.LOW. diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 40898dcf..57031369 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -129 +129 @@ class JobDocument(Document): - priority (`Priority`, optional): The priority of the job. Defaults to Priority.NORMAL. + priority (`Priority`, optional): The priority of the job. Defaults to Priority.LOW. @@ -169 +169 @@ class JobDocument(Document): - priority = EnumField(Priority, default=Priority.NORMAL) + priority = EnumField(Priority, default=Priority.LOW) @@ -350 +350 @@ class Queue: - priority: Priority = Priority.NORMAL, + priority: Priority = Priority.LOW, @@ -363 +363 @@ class Queue: - priority (`Priority`, optional): The priority of the job. Defaults to Priority.NORMAL. + priority (`Priority`, optional): The priority of the job. Defaults to Priority.LOW. diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index fc81b103..792905ee 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -130,2 +130,2 @@ def test_priority_logic_creation_order() -> None: - check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL) - check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL) + check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.LOW) + check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.LOW) @@ -143,2 +143,2 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None: - check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL) - check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.NORMAL) + check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.LOW) + check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.LOW) @@ -146 +146 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None: - check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL) + check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.LOW) @@ -161,6 +161,2 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None: - check_job( - queue=queue, expected_dataset="org1/dataset1", expected_split="split1", expected_priority=Priority.NORMAL - ) - check_job( - queue=queue, expected_dataset="org2/dataset2", expected_split="split1", expected_priority=Priority.NORMAL - ) + check_job(queue=queue, expected_dataset="org1/dataset1", expected_split="split1", expected_priority=Priority.LOW) + check_job(queue=queue, expected_dataset="org2/dataset2", expected_split="split1", expected_priority=Priority.LOW) @@ -168,6 +164,2 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None: - check_job( - queue=queue, expected_dataset="no_org_dataset3", expected_split="split1", expected_priority=Priority.NORMAL - ) - check_job( - queue=queue, expected_dataset="org1/dataset2", expected_split="split1", expected_priority=Priority.NORMAL - ) + check_job(queue=queue, expected_dataset="no_org_dataset3", expected_split="split1", expected_priority=Priority.LOW) + check_job(queue=queue, expected_dataset="org1/dataset2", expected_split="split1", expected_priority=Priority.LOW) @@ -188 +179,0 @@ def test_priority_logic_priority_order() -> None: - priority=Priority.LOW, @@ -190 +181,8 @@ def test_priority_logic_priority_order() -> None: - queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1") + queue.add_job( + job_type=test_type, + dataset="dataset2", + revision=test_revision, + config="config", + split="split1", + priority=Priority.NORMAL, + ) diff --git a/services/admin/src/admin/routes/dataset_backfill.py b/services/admin/src/admin/routes/dataset_backfill.py index 4d388344..3604391f 100644 --- a/services/admin/src/admin/routes/dataset_backfill.py +++ b/services/admin/src/admin/routes/dataset_backfill.py @@ -53 +53 @@ def create_dataset_backfill_endpoint( - dataset_orchestrator.backfill(revision=dataset_git_revision, priority=Priority.NORMAL) + dataset_orchestrator.backfill(revision=dataset_git_revision, priority=Priority.LOW) diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py index 439f4dde..3f37f937 100644 --- a/services/admin/src/admin/routes/force_refresh.py +++ b/services/admin/src/admin/routes/force_refresh.py @@ -10,0 +11 @@ from libcommon.queue import Queue +from libcommon.utils import Priority @@ -62 +63,8 @@ def create_force_refresh_endpoint( - Queue().add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split) + Queue().add_job( + job_type=job_type, + dataset=dataset, + revision=revision, + config=config, + split=split, + priority=Priority.LOW, + )
dc604553180cd0e30ce2cce7a30e9d1b6f67fe65
Andrea Francis Soria Jimenez
2023-07-03T12:56:22
Adding retry to create duckdb index commit (#1466)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 8e82aa94..14da1e9e 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -80 +80,7 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas -from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url, retry +from worker.utils import ( + HF_HUB_HTTP_ERROR_RETRY_SLEEPS, + LOCK_GIT_BRANCH_RETRY_SLEEPS, + create_branch, + hf_hub_url, + retry, +) @@ -753 +758,0 @@ def create_commits( - sleeps = [1, 1, 1, 10, 10, 10] @@ -757 +762 @@ def create_commits( - retry_create_commit = retry(on=[HfHubHTTPError], sleeps=sleeps)(hf_api.create_commit) + retry_create_commit = retry(on=[HfHubHTTPError], sleeps=HF_HUB_HTTP_ERROR_RETRY_SLEEPS)(hf_api.create_commit) @@ -772 +777 @@ def create_commits( - f" {len(sleeps)} attempts)." + f" {len(HF_HUB_HTTP_ERROR_RETRY_SLEEPS)} attempts)." diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py index 1ed21f5d..97fbb673 100644 --- a/services/worker/src/worker/job_runners/split/duckdb_index.py +++ b/services/worker/src/worker/job_runners/split/duckdb_index.py @@ -15 +15 @@ from huggingface_hub.hf_api import HfApi -from huggingface_hub.utils._errors import RepositoryNotFoundError +from huggingface_hub.utils._errors import HfHubHTTPError, RepositoryNotFoundError @@ -19,0 +20 @@ from libcommon.exceptions import ( + CreateCommitError, @@ -38 +39,7 @@ from worker.job_runners.split.split_job_runner import SplitJobRunnerWithCache -from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url +from worker.utils import ( + HF_HUB_HTTP_ERROR_RETRY_SLEEPS, + LOCK_GIT_BRANCH_RETRY_SLEEPS, + create_branch, + hf_hub_url, + retry, +) @@ -187,7 +194,2 @@ def compute_index_rows( - committer_hf_api.create_commit( - repo_id=dataset, - repo_type=DATASET_TYPE, - revision=target_revision, - operations=delete_operations + add_operations, - commit_message=commit_message, - parent_commit=target_dataset_info.sha, + retry_create_commit = retry(on=[HfHubHTTPError], sleeps=HF_HUB_HTTP_ERROR_RETRY_SLEEPS)( + committer_hf_api.create_commit @@ -194,0 +197,20 @@ def compute_index_rows( + try: + retry_create_commit( + repo_id=dataset, + repo_type=DATASET_TYPE, + revision=target_revision, + operations=delete_operations + add_operations, + commit_message=commit_message, + parent_commit=target_dataset_info.sha, + ) + except RuntimeError as e: + if e.__cause__ and isinstance(e.__cause__, HfHubHTTPError): + raise CreateCommitError( + message=( + f"Commit {commit_message} could not be created on the Hub (after" + f" {len(HF_HUB_HTTP_ERROR_RETRY_SLEEPS)} attempts)." + ), + cause=e.__cause__, + ) from e.__cause__ + raise e + diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index c3b985d8..56308fc3 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -338,0 +339 @@ DATASET_TYPE = "dataset" +HF_HUB_HTTP_ERROR_RETRY_SLEEPS = [1, 1, 1, 10, 10, 10]
812df7632bd074bf1fd46ec2e25e35ee5e2492ca
geethika-123
2023-07-01T15:49:32
Rename classes to indicate inheritance (#1428)
diff --git a/jobs/cache_maintenance/src/cache_maintenance/metrics.py b/jobs/cache_maintenance/src/cache_maintenance/metrics.py index 2eeb4f0b..2aae8f11 100644 --- a/jobs/cache_maintenance/src/cache_maintenance/metrics.py +++ b/jobs/cache_maintenance/src/cache_maintenance/metrics.py @@ -6 +6 @@ import logging -from libcommon.metrics import CacheTotalMetric, JobTotalMetric +from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument @@ -17 +17 @@ def collect_metrics(processing_graph: ProcessingGraph) -> None: - JobTotalMetric.objects(queue=processing_step.job_type, status=status).upsert_one(total=total) + JobTotalMetricDocument.objects(queue=processing_step.job_type, status=status).upsert_one(total=total) @@ -21 +21 @@ def collect_metrics(processing_graph: ProcessingGraph) -> None: - CacheTotalMetric.objects( + CacheTotalMetricDocument.objects( diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py index 7f8cf4bd..aed47435 100644 --- a/jobs/cache_maintenance/tests/test_collect_metrics.py +++ b/jobs/cache_maintenance/tests/test_collect_metrics.py @@ -6 +6 @@ from http import HTTPStatus -from libcommon.metrics import CacheTotalMetric, JobTotalMetric +from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument @@ -41 +41 @@ def test_collect_metrics() -> None: - cache_metrics = CacheTotalMetric.objects() + cache_metrics = CacheTotalMetricDocument.objects() @@ -45 +45 @@ def test_collect_metrics() -> None: - job_metrics = JobTotalMetric.objects() + job_metrics = JobTotalMetricDocument.objects() diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py index 46cc56f5..a87c27df 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py @@ -7 +7 @@ import logging -from libcommon.simple_cache import CachedResponse +from libcommon.simple_cache import CachedResponseDocument @@ -95 +95 @@ class MigrationMoveToGenericCachedResponse(Migration): - check_documents(DocCls=CachedResponse, sample_size=10) + check_documents(DocCls=CachedResponseDocument, sample_size=10) @@ -106 +106 @@ class MigrationMoveToGenericCachedResponse(Migration): - cached_responses_count = CachedResponse.objects.count() + cached_responses_count = CachedResponseDocument.objects.count() diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py index 28e4ddf5..b7cc73e2 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py @@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS -from libcommon.queue import Job +from libcommon.queue import JobDocument @@ -30 +30 @@ class MigrationAddPriorityToJob(Migration): - check_documents(DocCls=Job, sample_size=10) + check_documents(DocCls=JobDocument, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py index a30d5236..d63eaea5 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py @@ -7 +7 @@ from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_AL -from libcommon.simple_cache import CachedResponse +from libcommon.simple_cache import CachedResponseDocument @@ -30 +30 @@ class MigrationAddProgressToCacheResponse(Migration): - check_documents(DocCls=CachedResponse, sample_size=10) + check_documents(DocCls=CachedResponseDocument, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py index e052a46c..6ea4e74e 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py @@ -7 +7 @@ from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_AL -from libcommon.simple_cache import CachedResponse +from libcommon.simple_cache import CachedResponseDocument @@ -46 +46 @@ class MigrationAddJobRunnerVersionToCacheResponse(Migration): - check_documents(DocCls=CachedResponse, sample_size=10) + check_documents(DocCls=CachedResponseDocument, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py index 976b3e7d..28a9b9b1 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py @@ -7 +7 @@ from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_AL -from libcommon.simple_cache import CachedResponse +from libcommon.simple_cache import CachedResponseDocument @@ -27 +27 @@ class MigrationRemoveWorkerVersionFromCachedResponse(Migration): - check_documents(DocCls=CachedResponse, sample_size=10) + check_documents(DocCls=CachedResponseDocument, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py index 6628d45b..c08d0590 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py @@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS -from libcommon.queue import Job +from libcommon.queue import JobDocument @@ -27 +27 @@ class MigrationRemoveForceFromJob(Migration): - check_documents(DocCls=Job, sample_size=10) + check_documents(DocCls=JobDocument, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py index 4f6dea35..fb2cec30 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py @@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS -from libcommon.queue import Job +from libcommon.queue import JobDocument @@ -31 +31 @@ class MigrationDeleteSkippedJobs(Migration): - if not isinstance(doc, Job): + if not isinstance(doc, JobDocument): @@ -36 +36 @@ class MigrationDeleteSkippedJobs(Migration): - check_documents(DocCls=Job, sample_size=10, custom_validation=custom_validation) + check_documents(DocCls=JobDocument, sample_size=10, custom_validation=custom_validation) diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py index c1cee0a3..89fdd00e 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py @@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS -from libcommon.queue import Job +from libcommon.queue import JobDocument @@ -32 +32 @@ class MigrationQueueAddRevisionToJob(Migration): - check_documents(DocCls=Job, sample_size=10) + check_documents(DocCls=JobDocument, sample_size=10) diff --git a/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py b/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py index ce0ca4b5..b16aff74 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py +++ b/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py @@ -7,2 +7,2 @@ from typing import Optional -from libcommon.queue import Job -from libcommon.simple_cache import CachedResponse +from libcommon.queue import JobDocument +from libcommon.simple_cache import CachedResponseDocument @@ -47 +47 @@ class CacheRenamingMigration(CacheMigration): - check_documents(DocCls=CachedResponse, sample_size=10) + check_documents(DocCls=CachedResponseDocument, sample_size=10) @@ -114 +114 @@ class QueueRenamingMigration(QueueMigration): - check_documents(DocCls=Job, sample_size=10) + check_documents(DocCls=JobDocument, sample_size=10) diff --git a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py index b64bf12e..7fced24f 100644 --- a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py +++ b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py @@ -5 +5 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS -from libcommon.queue import Job +from libcommon.queue import JobDocument @@ -21 +21 @@ def test_queue_delete_indexes_with_force(mongo_host: str) -> None: - Job( + JobDocument( diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index 09cb6301..aa8b7027 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -13 +13 @@ from libcommon.constants import ( -from libcommon.queue import Job +from libcommon.queue import JobDocument @@ -121 +121 @@ def test_queue_delete_ttl_index(mongo_host: str) -> None: - Job( + JobDocument( diff --git a/libs/libcommon/src/libcommon/metrics.py b/libs/libcommon/src/libcommon/metrics.py index 8c4343ce..b0833d61 100644 --- a/libs/libcommon/src/libcommon/metrics.py +++ b/libs/libcommon/src/libcommon/metrics.py @@ -39 +39 @@ class QuerySetManager(Generic[U]): -class JobTotalMetric(Document): +class JobTotalMetricDocument(Document): @@ -60 +60 @@ class JobTotalMetric(Document): - objects = QuerySetManager["JobTotalMetric"]() + objects = QuerySetManager["JobTotalMetricDocument"]() @@ -63 +63 @@ class JobTotalMetric(Document): -class CacheTotalMetric(Document): +class CacheTotalMetricDocument(Document): @@ -86 +86 @@ class CacheTotalMetric(Document): - objects = QuerySetManager["CacheTotalMetric"]() + objects = QuerySetManager["CacheTotalMetricDocument"]() @@ -91,2 +91,2 @@ def _clean_metrics_database() -> None: - CacheTotalMetric.drop_collection() # type: ignore - JobTotalMetric.drop_collection() # type: ignore + CacheTotalMetricDocument.drop_collection() # type: ignore + JobTotalMetricDocument.drop_collection() # type: ignore diff --git a/libs/libcommon/src/libcommon/prometheus.py b/libs/libcommon/src/libcommon/prometheus.py index 44a1c88a..27fa882a 100644 --- a/libs/libcommon/src/libcommon/prometheus.py +++ b/libs/libcommon/src/libcommon/prometheus.py @@ -19 +19 @@ from psutil import disk_usage -from libcommon.metrics import CacheTotalMetric, JobTotalMetric +from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument @@ -67 +67 @@ def update_queue_jobs_total() -> None: - for job_metric in JobTotalMetric.objects(): + for job_metric in JobTotalMetricDocument.objects(): @@ -72 +72 @@ def update_responses_in_cache_total() -> None: - for cache_metric in CacheTotalMetric.objects(): + for cache_metric in CacheTotalMetricDocument.objects(): diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 6aa66820..40898dcf 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -117 +117 @@ class NoWaitingJobError(Exception): -class Job(Document): +class JobDocument(Document): @@ -193 +193 @@ class Job(Document): - objects = QuerySetManager["Job"]() + objects = QuerySetManager["JobDocument"]() @@ -211 +211 @@ class Job(Document): - def get(cls, job_id: str) -> "Job": + def get(cls, job_id: str) -> "JobDocument": @@ -351 +351 @@ class Queue: - ) -> Job: + ) -> JobDocument: @@ -367 +367 @@ class Queue: - return Job( + return JobDocument( @@ -393 +393 @@ class Queue: - Job( + JobDocument( @@ -412 +412 @@ class Queue: - job_ids = Job.objects.insert(jobs, load_bulk=False) + job_ids = JobDocument.objects.insert(jobs, load_bulk=False) @@ -429 +429 @@ class Queue: - existing = Job.objects(pk__in=job_ids) + existing = JobDocument.objects(pk__in=job_ids) @@ -440 +440 @@ class Queue: - ) -> Job: + ) -> JobDocument: @@ -467 +467 @@ class Queue: - started_jobs = Job.objects(status=Status.STARTED, **filters) + started_jobs = JobDocument.objects(status=Status.STARTED, **filters) @@ -473 +473 @@ class Queue: - Job.objects( + JobDocument.objects( @@ -506 +506 @@ class Queue: - Job.objects( + JobDocument.objects( @@ -524 +524 @@ class Queue: - ) -> Job: + ) -> JobDocument: @@ -549 +549 @@ class Queue: - def _start_newest_job_and_cancel_others(self, job: Job) -> Job: + def _start_newest_job_and_cancel_others(self, job: JobDocument) -> JobDocument: @@ -570 +570 @@ class Queue: - waiting_jobs = Job.objects( + waiting_jobs = JobDocument.objects( @@ -644 +644 @@ class Queue: - def get_job_with_id(self, job_id: str) -> Job: + def get_job_with_id(self, job_id: str) -> JobDocument: @@ -655 +655 @@ class Queue: - return Job.objects(pk=job_id).get() + return JobDocument.objects(pk=job_id).get() @@ -671 +671 @@ class Queue: - def _get_started_job(self, job_id: str) -> Job: + def _get_started_job(self, job_id: str) -> JobDocument: @@ -681 +681 @@ class Queue: - job = Job.objects(pk=job_id).get() + job = JobDocument.objects(pk=job_id).get() @@ -751 +751 @@ class Queue: - Job.objects( + JobDocument.objects( @@ -764 +764 @@ class Queue: - for job in Job.objects(type=job_type, status=Status.STARTED.value): + for job in JobDocument.objects(type=job_type, status=Status.STARTED.value): @@ -807 +807 @@ class Queue: - for job in Job.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters) + for job in JobDocument.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters) @@ -815 +815 @@ class Queue: - return Job.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters).count() > 0 + return JobDocument.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters).count() > 0 @@ -827 +827 @@ class Queue: - return Job.objects(type=job_type, status=status.value).count() + return JobDocument.objects(type=job_type, status=status.value).count() @@ -856 +856 @@ class Queue: - return [d.to_dict() for d in Job.objects(type=job_type, status=status.value)] + return [d.to_dict() for d in JobDocument.objects(type=job_type, status=status.value)] @@ -875 +875 @@ class Queue: - for d in Job.objects( + for d in JobDocument.objects( @@ -899 +899 @@ class Queue: - started_jobs = Job.objects(status=Status.STARTED) + started_jobs = JobDocument.objects(status=Status.STARTED) @@ -923 +923 @@ def _clean_queue_database() -> None: - Job.drop_collection() # type: ignore + JobDocument.drop_collection() # type: ignore diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index 096b8b2d..2d2d897a 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -70 +70 @@ class SplitFullName(NamedTuple): -class CachedResponse(Document): +class CachedResponseDocument(Document): @@ -116 +116 @@ class CachedResponse(Document): - objects = QuerySetManager["CachedResponse"]() + objects = QuerySetManager["CachedResponseDocument"]() @@ -123,2 +123,2 @@ class CachedResponse(Document): -CachedResponse.config.required = False # type: ignore -CachedResponse.split.required = False # type: ignore +CachedResponseDocument.config.required = False # type: ignore +CachedResponseDocument.split.required = False # type: ignore @@ -146 +146 @@ def upsert_response( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split).upsert_one( + CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split).upsert_one( @@ -188 +188 @@ def delete_response( - return CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split).delete() + return CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split).delete() @@ -192 +192 @@ def delete_dataset_responses(dataset: str) -> Optional[int]: - return CachedResponse.objects(dataset=dataset).delete() + return CachedResponseDocument.objects(dataset=dataset).delete() @@ -209 +209 @@ def get_response_without_content( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split) @@ -240 +240 @@ def get_response_metadata( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split) @@ -300 +300 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split) @@ -322 +322 @@ def get_response_with_details( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split) @@ -444 +444 @@ def get_valid_datasets(kind: str) -> Set[str]: - return set(CachedResponse.objects(kind=kind, http_status=HTTPStatus.OK).distinct("dataset")) + return set(CachedResponseDocument.objects(kind=kind, http_status=HTTPStatus.OK).distinct("dataset")) @@ -448 +448 @@ def is_valid_for_kinds(dataset: str, kinds: List[str]) -> bool: - return CachedResponse.objects(dataset=dataset, kind__in=kinds, http_status=HTTPStatus.OK).count() > 0 + return CachedResponseDocument.objects(dataset=dataset, kind__in=kinds, http_status=HTTPStatus.OK).count() > 0 @@ -466 +466 @@ def get_responses_count_by_kind_status_and_error_code() -> List[CountEntry]: - entries = CachedResponse.objects().only("kind", "http_status", "error_code") + entries = CachedResponseDocument.objects().only("kind", "http_status", "error_code") @@ -535 +535 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo - queryset = CachedResponse.objects(kind=kind) + queryset = CachedResponseDocument.objects(kind=kind) @@ -538 +538 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo - queryset = CachedResponse.objects(kind=kind, id__gt=ObjectId(cursor)) + queryset = CachedResponseDocument.objects(kind=kind, id__gt=ObjectId(cursor)) @@ -566 +566 @@ def get_outdated_split_full_names_for_step(kind: str, current_version: int) -> L - responses = CachedResponse.objects(kind=kind, job_runner_version__lt=current_version).only( + responses = CachedResponseDocument.objects(kind=kind, job_runner_version__lt=current_version).only( @@ -575 +575 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L - responses = CachedResponse.objects(kind=kind, dataset=dataset).exclude("content") + responses = CachedResponseDocument.objects(kind=kind, dataset=dataset).exclude("content") @@ -628 +628 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int) - queryset = CachedResponse.objects(kind=kind) + queryset = CachedResponseDocument.objects(kind=kind) @@ -631 +631 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int) - queryset = CachedResponse.objects(kind=kind, id__gt=ObjectId(cursor)) + queryset = CachedResponseDocument.objects(kind=kind, id__gt=ObjectId(cursor)) @@ -706 +706 @@ def get_cache_entries_df(dataset: str, cache_kinds: Optional[List[str]] = None) - for response in CachedResponse.objects(dataset=dataset, **filters).only( + for response in CachedResponseDocument.objects(dataset=dataset, **filters).only( @@ -723 +723 @@ def has_some_cache(dataset: str) -> bool: - return CachedResponse.objects(dataset=dataset).count() > 0 + return CachedResponseDocument.objects(dataset=dataset).count() > 0 @@ -760 +760 @@ def _clean_cache_database() -> None: - CachedResponse.drop_collection() # type: ignore + CachedResponseDocument.drop_collection() # type: ignore diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py index 82848306..3131bc1a 100644 --- a/libs/libcommon/tests/test_orchestrator.py +++ b/libs/libcommon/tests/test_orchestrator.py @@ -11 +11 @@ from libcommon.processing_graph import Artifact, ProcessingGraph -from libcommon.queue import Job, Queue +from libcommon.queue import JobDocument, Queue @@ -13 +13 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import CachedResponse, upsert_response_params +from libcommon.simple_cache import CachedResponseDocument, upsert_response_params @@ -169 +169 @@ def test_finish_job( - assert Job.objects(dataset=DATASET_NAME).count() == 1 + len(artifacts_to_create) + assert JobDocument.objects(dataset=DATASET_NAME).count() == 1 + len(artifacts_to_create) @@ -171 +171 @@ def test_finish_job( - done_job = Job.objects(dataset=DATASET_NAME, status=Status.SUCCESS) + done_job = JobDocument.objects(dataset=DATASET_NAME, status=Status.SUCCESS) @@ -174 +174 @@ def test_finish_job( - waiting_jobs = Job.objects(dataset=DATASET_NAME, status=Status.WAITING) + waiting_jobs = JobDocument.objects(dataset=DATASET_NAME, status=Status.WAITING) @@ -178,2 +178,2 @@ def test_finish_job( - assert CachedResponse.objects(dataset=DATASET_NAME).count() == 1 - cached_response = CachedResponse.objects(dataset=DATASET_NAME).first() + assert CachedResponseDocument.objects(dataset=DATASET_NAME).count() == 1 + cached_response = CachedResponseDocument.objects(dataset=DATASET_NAME).first() diff --git a/libs/libcommon/tests/test_prometheus.py b/libs/libcommon/tests/test_prometheus.py index 50e9d0ff..59c6a524 100644 --- a/libs/libcommon/tests/test_prometheus.py +++ b/libs/libcommon/tests/test_prometheus.py @@ -10 +10 @@ import pytest -from libcommon.metrics import CacheTotalMetric, JobTotalMetric +from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument @@ -151 +151 @@ def test_cache_metrics(metrics_mongo_resource: MetricsMongoResource) -> None: - collection = CacheTotalMetric._get_collection() + collection = CacheTotalMetricDocument._get_collection() @@ -184 +184 @@ def test_queue_metrics(metrics_mongo_resource: MetricsMongoResource) -> None: - collection = JobTotalMetric._get_collection() + collection = JobTotalMetricDocument._get_collection() diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 34626226..fc81b103 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -18 +18 @@ from libcommon.constants import QUEUE_TTL_SECONDS -from libcommon.queue import EmptyQueueError, Job, Lock, Queue, lock +from libcommon.queue import EmptyQueueError, JobDocument, Lock, Queue, lock @@ -320 +320 @@ def test_has_ttl_index_on_finished_at_field() -> None: - for name, value in Job._get_collection().index_information().items() + for name, value in JobDocument._get_collection().index_information().items() @@ -326 +326 @@ def test_has_ttl_index_on_finished_at_field() -> None: - assert Job._get_collection().index_information()[ttl_index_name]["expireAfterSeconds"] == QUEUE_TTL_SECONDS + assert JobDocument._get_collection().index_information()[ttl_index_name]["expireAfterSeconds"] == QUEUE_TTL_SECONDS diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index 75f9845d..0a71e793 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -15 +15 @@ from libcommon.simple_cache import ( - CachedResponse, + CachedResponseDocument, @@ -56 +56 @@ def test_insert_null_values() -> None: - CachedResponse.objects(kind=kind, dataset=dataset_a, config=config, split=split).upsert_one( + CachedResponseDocument.objects(kind=kind, dataset=dataset_a, config=config, split=split).upsert_one( @@ -60,2 +60,2 @@ def test_insert_null_values() -> None: - assert CachedResponse.objects.count() == 1 - cached_response = CachedResponse.objects.get() + assert CachedResponseDocument.objects.count() == 1 + cached_response = CachedResponseDocument.objects.get() @@ -67 +67 @@ def test_insert_null_values() -> None: - CachedResponse( + CachedResponseDocument( @@ -70,2 +70,2 @@ def test_insert_null_values() -> None: - assert CachedResponse.objects.count() == 2 - cached_response = CachedResponse.objects(dataset=dataset_b).get() + assert CachedResponseDocument.objects.count() == 2 + cached_response = CachedResponseDocument.objects(dataset=dataset_b).get() @@ -76 +76 @@ def test_insert_null_values() -> None: - coll = CachedResponse._get_collection() + coll = CachedResponseDocument._get_collection() @@ -87,2 +87,2 @@ def test_insert_null_values() -> None: - assert CachedResponse.objects.count() == 3 - cached_response = CachedResponse.objects(dataset=dataset_c).get() + assert CachedResponseDocument.objects.count() == 3 + cached_response = CachedResponseDocument.objects(dataset=dataset_c).get() diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py index d83cbf26..3e30ca26 100644 --- a/services/worker/tests/test_executor.py +++ b/services/worker/tests/test_executor.py @@ -15 +15 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.queue import Job, JobDoesNotExistError, Queue +from libcommon.queue import JobDocument, JobDoesNotExistError, Queue @@ -17 +17 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import CachedResponse +from libcommon.simple_cache import CachedResponseDocument @@ -86 +86 @@ def start_worker_loop_with_long_job() -> None: - current_job = Job.objects(pk=current_job_info["job_id"]).get() + current_job = JobDocument.objects(pk=current_job_info["job_id"]).get() @@ -107 +107 @@ def set_worker_state(worker_state_file_path: str) -> Iterator[WorkerState]: -def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]: +def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[JobDocument]: @@ -112 +112 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I - Job.get(job_id=job_info["job_id"]).delete() + JobDocument.get(job_id=job_info["job_id"]).delete() @@ -116 +116 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I - job = Job( + job = JobDocument( @@ -136 +136,3 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I -def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: QueueMongoResource) -> Iterator[Job]: +def set_long_running_job_in_queue( + app_config: AppConfig, queue_mongo_resource: QueueMongoResource +) -> Iterator[JobDocument]: @@ -141 +143 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q - Job.get(job_id=job_info["job_id"]).delete() + JobDocument.get(job_id=job_info["job_id"]).delete() @@ -146 +148 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q - job = Job( + job = JobDocument( @@ -167 +169 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q -def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]: +def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[JobDocument]: @@ -172 +174 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato - Job.get(job_id=job_info["job_id"]).delete() + JobDocument.get(job_id=job_info["job_id"]).delete() @@ -176 +178 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato - job = Job( + job = JobDocument( @@ -234 +236 @@ def test_executor_heartbeat( - set_just_started_job_in_queue: Job, + set_just_started_job_in_queue: JobDocument, @@ -248,3 +250,3 @@ def test_executor_kill_zombies( - set_just_started_job_in_queue: Job, - set_long_running_job_in_queue: Job, - set_zombie_job_in_queue: Job, + set_just_started_job_in_queue: JobDocument, + set_long_running_job_in_queue: JobDocument, + set_zombie_job_in_queue: JobDocument, @@ -259,3 +261,3 @@ def test_executor_kill_zombies( - assert Job.objects(pk=zombie.pk).get().status in [Status.ERROR, Status.CANCELLED, Status.SUCCESS] - assert Job.objects(pk=normal_job.pk).get().status == Status.STARTED - response = CachedResponse.objects()[0] + assert JobDocument.objects(pk=zombie.pk).get().status in [Status.ERROR, Status.CANCELLED, Status.SUCCESS] + assert JobDocument.objects(pk=normal_job.pk).get().status == Status.STARTED + response = CachedResponseDocument.objects()[0] @@ -273 +275 @@ def test_executor_kill_zombies( - CachedResponse.objects().delete() + CachedResponseDocument.objects().delete() @@ -279,2 +281,2 @@ def test_executor_start( - set_just_started_job_in_queue: Job, - set_zombie_job_in_queue: Job, + set_just_started_job_in_queue: JobDocument, + set_zombie_job_in_queue: JobDocument, @@ -299 +301 @@ def test_executor_start( - assert Job.objects(pk=set_just_started_job_in_queue.pk).get().last_heartbeat is not None + assert JobDocument.objects(pk=set_just_started_job_in_queue.pk).get().last_heartbeat is not None @@ -301 +303,5 @@ def test_executor_start( - assert Job.objects(pk=set_zombie_job_in_queue.pk).get().status in [Status.ERROR, Status.CANCELLED, Status.SUCCESS] + assert JobDocument.objects(pk=set_zombie_job_in_queue.pk).get().status in [ + Status.ERROR, + Status.CANCELLED, + Status.SUCCESS, + ] @@ -328,2 +334,2 @@ def test_executor_stops_on_long_job( - set_long_running_job_in_queue: Job, - set_just_started_job_in_queue: Job, + set_long_running_job_in_queue: JobDocument, + set_just_started_job_in_queue: JobDocument, @@ -350 +356 @@ def test_executor_stops_on_long_job( - responses = CachedResponse.objects() + responses = CachedResponseDocument.objects() @@ -368 +374 @@ def test_executor_stops_on_long_job( - CachedResponse.objects().delete() + CachedResponseDocument.objects().delete() diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py index 0e912bdb..a6e54257 100644 --- a/services/worker/tests/test_job_manager.py +++ b/services/worker/tests/test_job_manager.py @@ -8 +8 @@ from libcommon.processing_graph import ProcessingGraph, ProcessingStep -from libcommon.queue import Job, Queue +from libcommon.queue import JobDocument, Queue @@ -10 +10 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.simple_cache import CachedResponse, get_response, upsert_response +from libcommon.simple_cache import CachedResponseDocument, get_response, upsert_response @@ -127 +127 @@ def test_backfill(priority: Priority, app_config: AppConfig) -> None: - assert Job.objects().count() == 0 + assert JobDocument.objects().count() == 0 @@ -197 +197 @@ def test_job_runner_set_crashed( - assert Job.objects().count() == 0 + assert JobDocument.objects().count() == 0 @@ -219 +219 @@ def test_job_runner_set_crashed( - response = CachedResponse.objects()[0] + response = CachedResponseDocument.objects()[0]
e26ef0fd08298579346325193ba2d2405b93f43c
Sylvain Lesage
2023-06-30T20:02:12
Ensure parquet shards are sorted (#1465)
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json index 9c3ed661..832fd7a7 100644 --- a/chart/static-files/openapi.json +++ b/chart/static-files/openapi.json @@ -2786 +2786 @@ - "description": "The dataset is converted to the parquet format. The endpoint gives the list of the parquet files.", + "description": "The dataset is converted to the parquet format. The endpoint gives the list of the parquet files. The same split can have multiple parquet files, called shards. They are sorted by their shard index.", diff --git a/docs/source/parquet.mdx b/docs/source/parquet.mdx index a789265e..927a5a1e 100644 --- a/docs/source/parquet.mdx +++ b/docs/source/parquet.mdx @@ -116 +116 @@ The endpoint also gives the filename and size of each file: -Big datasets are partitioned into Parquet files (shards) of about 500MB each. The filename contains the name of the dataset, the split, the shard index, and the total number of shards (`dataset-name-train-0000-of-0004.parquet`). For example, the `train` split of the [`amazon_polarity`](https://datasets-server.huggingface.co/parquet?dataset=amazon_polarity) dataset is partitioned into 4 shards: +Big datasets are partitioned into Parquet files (shards) of about 500MB each. The filename contains the name of the dataset, the split, the shard index, and the total number of shards (`dataset-name-train-0000-of-0004.parquet`). For a given split, the elements in the list are sorted by their shard index, in ascending order. For example, the `train` split of the [`amazon_polarity`](https://datasets-server.huggingface.co/parquet?dataset=amazon_polarity) dataset is partitioned into 4 shards: diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 7f57d086..5bb181a5 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -20,4 +19,0 @@ DEFAULT_JOB_RUNNER_VERSION = 1 -PROCESSING_STEP_DATASET_CONFIG_NAMES_VERSION = 1 -PROCESSING_STEP_CONFIG_PARQUET_VERSION = 4 -PROCESSING_STEP_CONFIG_PARQUET_METADATA_VERSION = 1 -PROCESSING_STEP_CONFIG_SIZE_VERSION = 2 @@ -25,0 +22,7 @@ PROCESSING_STEP_CONFIG_OPT_IN_OUT_URLS_COUNT_VERSION = 3 +PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 3 +PROCESSING_STEP_CONFIG_PARQUET_METADATA_VERSION = 1 +PROCESSING_STEP_CONFIG_PARQUET_VERSION = 5 +PROCESSING_STEP_CONFIG_SIZE_VERSION = 2 +PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_INFO_VERSION = 3 +PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_STREAMING_VERSION = 3 +PROCESSING_STEP_DATASET_CONFIG_NAMES_VERSION = 1 @@ -30,0 +34,2 @@ PROCESSING_STEP_DATASET_SIZE_VERSION = 2 +PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 3 +PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION = 1 @@ -32 +36,0 @@ PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION = 2 -PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 3 @@ -34,3 +38 @@ PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION = 3 -PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_INFO_VERSION = 3 -PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_STREAMING_VERSION = 3 -PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 3 +PROCESSING_STEP_SPLIT_IMAGE_URL_COLUMNS_VERSION = 1 @@ -39,2 +40,0 @@ PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION = 4 -PROCESSING_STEP_SPLIT_IMAGE_URL_COLUMNS_VERSION = 1 -PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION = 1 diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py index 572df22c..a6fc6833 100644 --- a/services/worker/src/worker/job_runners/config/parquet.py +++ b/services/worker/src/worker/job_runners/config/parquet.py @@ -38,2 +37,0 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - if "parquet_files" not in content: - raise PreviousStepFormatError("Previous step did not return the expected content: 'parquet_files'.") @@ -41 +39,8 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse - parquet_files = [parquet_file for parquet_file in content["parquet_files"] if parquet_file.get("config") == config] + try: + parquet_files = [ + parquet_file for parquet_file in content["parquet_files"] if parquet_file.get("config") == config + ] + # sort by filename, which ensures the shards are in order: 00000, 00001, 00002, ... + parquet_files.sort(key=lambda x: x["filename"]) # type: ignore + except KeyError as e: + raise PreviousStepFormatError("Previous step did not return the expected content: 'parquet_files'.", e) from e diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index 000388ff..9425c249 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -26 +26 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float]: +def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, float]: @@ -111 +111 @@ class DatasetParquetJobRunner(DatasetJobRunner): - response_content, progress = compute_sizes_response(dataset=self.dataset) + response_content, progress = compute_parquet_response(dataset=self.dataset) diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index e314000a..c7db50e4 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -116,0 +117,64 @@ def get_job_runner( + ( + "shards_order", + "config_1", + HTTPStatus.OK, + ConfigParquetAndInfoResponse( + parquet_files=[ + SplitHubFile( + dataset="ok", + config="config_1", + split="train", + url="url2", + filename="parquet-train-00001-of-05534.parquet", + size=0, + ), + SplitHubFile( + dataset="ok", + config="config_1", + split="train", + url="url1", + filename="parquet-train-00000-of-05534.parquet", + size=0, + ), + SplitHubFile( + dataset="ok", + config="config_1", + split="test", + url="url2", + filename="parquet-test-00000-of-00001.parquet", + size=0, + ), + ], + dataset_info={"description": "value", "dataset_size": 10}, + ), + None, + ConfigParquetResponse( + parquet_files=[ + SplitHubFile( + dataset="ok", + config="config_1", + split="test", + url="url2", + filename="parquet-test-00000-of-00001.parquet", + size=0, + ), + SplitHubFile( + dataset="ok", + config="config_1", + split="train", + url="url1", + filename="parquet-train-00000-of-05534.parquet", + size=0, + ), + SplitHubFile( + dataset="ok", + config="config_1", + split="train", + url="url2", + filename="parquet-train-00001-of-05534.parquet", + size=0, + ), + ] + ), + False, + ),
f9d5c9ff873ab93369ec605644b5e72f595ac6e2
Andrea Francis Soria Jimenez
2023-06-29T19:06:02
Disable backfill (#1461)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 62a4355f..b0dc9cf6 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -124 +124 @@ cacheMaintenance: - action: "backfill" + action: "skip" @@ -129 +129 @@ cacheMaintenance: - error_codes_to_retry: "AskAccessHubRequestError,CachedResponseNotFound,DatasetInfoHubRequestError,DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError,ExternalFilesSizeRequestConnectionError,ExternalFilesSizeRequestTimeoutError,FileSystemError,GatedExtraFieldsError,JobManagerExceededMaximumDurationError,JobRunnerCrashedError,JobRunnerExceededMaximumDurationError,PreviousStepFormatError,PreviousStepStatusError" + error_codes_to_retry: ""
e6aae54dc8591379eeca3f36f9068f7d24abe729
Andrea Francis Soria Jimenez
2023-06-29T15:54:38
Adding fix (#1459)
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py index 01874c4c..1ed21f5d 100644 --- a/services/worker/src/worker/job_runners/split/duckdb_index.py +++ b/services/worker/src/worker/job_runners/split/duckdb_index.py @@ -45 +45 @@ CREATE_SEQUENCE_COMMAND = "CREATE OR REPLACE SEQUENCE serial START 1;" -CREATE_INDEX_COMMAND = "PRAGMA create_fts_index('data', '__hf_index_id', '*', overwrite=1);" +CREATE_INDEX_COMMAND = "PRAGMA create_fts_index('data', '__hf_index_id', {columns}, overwrite=1);" @@ -111 +111 @@ def compute_index_rows( - column_names = ",".join(list(features.keys())) + column_names = ",".join('"' + column + '"' for column in list(features.keys())) @@ -152,2 +152,3 @@ def compute_index_rows( - logging.debug(CREATE_INDEX_COMMAND) - con.sql(CREATE_INDEX_COMMAND) + create_index_sql = CREATE_INDEX_COMMAND.format(columns=column_names) + logging.debug(create_index_sql) + con.sql(create_index_sql) diff --git a/services/worker/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py index 6e987e20..7e7a28ff 100644 --- a/services/worker/tests/fixtures/datasets.py +++ b/services/worker/tests/fixtures/datasets.py @@ -158 +158,8 @@ def datasets() -> Mapping[str, Dataset]: - ] + ], + "column with spaces": [ + "a", + "b", + "c", + "d", + "e", + ],
7a83fc631a34627264f28a4af2b5e28549550e56
Sylvain Lesage
2023-06-29T15:31:04
feat: 🎸 backfill the datasets (#1460)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index b0dc9cf6..62a4355f 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -124 +124 @@ cacheMaintenance: - action: "skip" + action: "backfill" @@ -129 +129 @@ cacheMaintenance: - error_codes_to_retry: "" + error_codes_to_retry: "AskAccessHubRequestError,CachedResponseNotFound,DatasetInfoHubRequestError,DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError,ExternalFilesSizeRequestConnectionError,ExternalFilesSizeRequestTimeoutError,FileSystemError,GatedExtraFieldsError,JobManagerExceededMaximumDurationError,JobRunnerCrashedError,JobRunnerExceededMaximumDurationError,PreviousStepFormatError,PreviousStepStatusError"
e8a5be6f6b43860ae9613d18a64c0062216266c0
Sylvain Lesage
2023-06-29T14:13:14
Replace valid with preview and viewer (#1452)
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json index 59a058d9..9c3ed661 100644 --- a/chart/static-files/openapi.json +++ b/chart/static-files/openapi.json @@ -898,4 +897,0 @@ - "preview": { - "type": "array", - "items": { "type": "string" } - }, @@ -906 +902 @@ - "valid": { + "preview": { @@ -908,2 +904 @@ - "items": { "type": "string" }, - "deprecated": true + "items": { "type": "string" } @@ -915 +910 @@ - "required": ["valid"], + "required": ["preview", "viewer"], @@ -917 +912,4 @@ - "valid": { + "viewer": { + "type": "boolean" + }, + "preview": { @@ -3190,7 +3187,0 @@ - ], - "valid": [ - "0n1xus/codexglue", - "0n1xus/pytorrent-standalone", - "0x7194633/rupile", - "51la5/keyword-extraction", - "AHussain0418/day2_data" @@ -3303 +3294,9 @@ - "valid": true + "viewer": true, + "preview": true + } + }, + "preview": { + "summary": "dataset with only preview", + "value": { + "viewer": false, + "preview": true @@ -3309 +3308,2 @@ - "valid": false + "viewer": false, + "preview": false diff --git a/docs/source/valid.mdx b/docs/source/valid.mdx index 3acd4e3d..4bbc0a67 100644 --- a/docs/source/valid.mdx +++ b/docs/source/valid.mdx @@ -5,2 +5,2 @@ Before you download a dataset from the Hub, it is helpful to know which datasets -* `/valid` returns a list of all the datasets that work without any errors. -* `/is-valid` checks if a specific dataset works without any errors. +- `/valid` returns a list of all the datasets that work without any errors. +- `/is-valid` checks if a specific dataset works without any errors. @@ -11,5 +11,6 @@ The API endpoints will return an error for datasets that cannot be loaded with t - Currently, only {" "} - <a href="https://huggingface.co/docs/datasets/stream">streamable</a> datasets - are supported so Datasets Server can extract the 100 first rows without downloading the - whole dataset. This is especially useful for previewing large datasets where downloading - the whole dataset may take hours! + The largest datasets are partially supported by Datasets Server. If they are{" "} + <a href="https://huggingface.co/docs/datasets/stream">streamable</a>, Datasets Server can + extract the first 100 rows without downloading the whole dataset. This is + especially useful for previewing large datasets where downloading the whole + dataset may take hours! See the <code>preview</code> field in the response of{" "} + <code>/valid</code> to view these partially supported datasets. @@ -61 +62,4 @@ curl https://datasets-server.huggingface.co/valid \ -The endpoint response is a JSON containing a list valid datasets nested under the `valid` key: +The endpoint response is a JSON containing lists of datasets nested under the following keys: + +- `viewer`: the dataset is fully-supported and the Dataset Viewer is working on the Hub dataset page. It also means the dataset has been auto-converted to Parquet. +- `preview`: the dataset is partially supported and the Dataset Viewer on the Hub dataset page shows a preview of the first 100 rows obtained by streaming. @@ -65,6 +69,14 @@ The endpoint response is a JSON containing a list valid datasets nested under th - "valid": [ - "0n1xus/codexglue", - "0n1xus/pytorrent-standalone", - "0x7194633/rupile", - "51la5/keyword-extraction", - ... + "viewer": [ + "0-hero/OIG-small-chip2", + "000alen/semantic","04-07-22/wep-probes", + "0721boy/nva-pic", + "0Tick/Danbooru-Random-Posts-Scrape", + "0Tick/E621-Random-PostsTag-Scrape", + "0n1xus/codexglue" + "..." + ], + "preview": [ + "0x7194633/GCRL-flibusta", + "0xJustin/Dungeons_and_Diffusion_uncropped", + "0xaryan/music-classifier","13GP/training", + "..." @@ -122 +134,13 @@ The response looks like this if a dataset is valid: -{"valid": true} +{ + "viewer": true, + "preview": true +} +``` + +If only the first rows of a dataset are available, then the response looks like: + +```json +{ + "viewer": false, + "preview": true +} @@ -125 +149 @@ The response looks like this if a dataset is valid: -If a dataset is not valid, then the response looks like: +Finally, if the dataset is not valid at all, then the response is: @@ -128 +152,4 @@ If a dataset is not valid, then the response looks like: -{"valid": false} +{ + "viewer": false, + "preview": false +} @@ -131,0 +159 @@ Some cases where a dataset is not valid are: + @@ -138 +166,2 @@ Some cases where a dataset is not valid are: - Remember if a dataset is <a href="./quick_start#gated-datasets">gated</a>, you'll need to provide your user token to submit a successful query! + Remember if a dataset is <a href="./quick_start#gated-datasets">gated</a>, + you'll need to provide your user token to submit a successful query! diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index b36d4535..99c6bebc 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -322,0 +323 @@ class ProcessingGraphConfig: + # special case: triggered by all the steps that have "enables_preview" or "enables_viewer" @@ -324 +325 @@ class ProcessingGraphConfig: - "dataset-split-names", + "config-size", diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 66a609b6..7f57d086 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -27 +27 @@ PROCESSING_STEP_DATASET_INFO_VERSION = 2 -PROCESSING_STEP_DATASET_IS_VALID_VERSION = 3 +PROCESSING_STEP_DATASET_IS_VALID_VERSION = 4 diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index 13582b42..096b8b2d 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -447,11 +447,2 @@ def get_valid_datasets(kind: str) -> Set[str]: -def get_validity_by_kind(dataset: str, kinds: Optional[List[str]] = None) -> Mapping[str, bool]: - # TODO: rework with aggregate - entries = ( - CachedResponse.objects(dataset=dataset) - if kinds is None - else CachedResponse.objects(dataset=dataset, kind__in=kinds) - ).only("kind", "http_status") - return { - str(kind): entries(kind=kind, http_status=HTTPStatus.OK).first() is not None - for kind in sorted(entries.distinct("kind")) - } +def is_valid_for_kinds(dataset: str, kinds: List[str]) -> bool: + return CachedResponse.objects(dataset=dataset, kind__in=kinds, http_status=HTTPStatus.OK).count() > 0 diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py index ea672c6f..0a15fda6 100644 --- a/libs/libcommon/tests/test_processing_graph.py +++ b/libs/libcommon/tests/test_processing_graph.py @@ -115 +115 @@ def graph() -> ProcessingGraph: - ["dataset-is-valid"], + [], @@ -182 +182 @@ def graph() -> ProcessingGraph: - ["dataset-size"], + ["dataset-is-valid", "dataset-size"], @@ -196 +196 @@ def graph() -> ProcessingGraph: - "dataset-split-names", + "config-size", @@ -203 +202,0 @@ def graph() -> ProcessingGraph: - "dataset-split-names", @@ -205,0 +205 @@ def graph() -> ProcessingGraph: + "config-size", diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index e3e9d851..75f9845d 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -34 +34 @@ from libcommon.simple_cache import ( - get_validity_by_kind, + is_valid_for_kinds, @@ -270,2 +270,2 @@ def test_get_valid_dataset_names_only_invalid_responses() -> None: -def test_get_validity_by_kind_empty() -> None: - assert not get_validity_by_kind(dataset="dataset") +def test_is_valid_for_kinds_empty() -> None: + assert not is_valid_for_kinds(dataset="dataset", kinds=[]) @@ -274 +274 @@ def test_get_validity_by_kind_empty() -> None: -def test_get_validity_by_kind_two_valid_datasets() -> None: +def test_is_valid_for_kinds_two_valid_datasets() -> None: @@ -281,5 +281,4 @@ def test_get_validity_by_kind_two_valid_datasets() -> None: - assert get_validity_by_kind(dataset=dataset_a) == {kind: True} - assert get_validity_by_kind(dataset=dataset_b) == {kind: True} - assert get_validity_by_kind(dataset=dataset_b, kinds=[kind]) == {kind: True} - assert not get_validity_by_kind(dataset=dataset_b, kinds=[other_kind]) - assert get_validity_by_kind(dataset=dataset_b, kinds=[kind, other_kind]) == {kind: True} + assert is_valid_for_kinds(dataset=dataset_a, kinds=[kind]) + assert is_valid_for_kinds(dataset=dataset_b, kinds=[kind]) + assert not is_valid_for_kinds(dataset=dataset_b, kinds=[other_kind]) + assert is_valid_for_kinds(dataset=dataset_b, kinds=[kind, other_kind]) @@ -288 +287 @@ def test_get_validity_by_kind_two_valid_datasets() -> None: -def test_get_validity_by_kind_two_valid_kinds() -> None: +def test_is_valid_for_kinds_two_valid_kinds() -> None: @@ -294 +293 @@ def test_get_validity_by_kind_two_valid_kinds() -> None: - assert get_validity_by_kind(dataset=dataset) == {kind_a: True, kind_b: True} + assert is_valid_for_kinds(dataset=dataset, kinds=[kind_a, kind_b]) @@ -297 +296 @@ def test_get_validity_by_kind_two_valid_kinds() -> None: -def test_get_validity_by_kind_at_least_one_valid_response() -> None: +def test_is_valid_for_kinds_at_least_one_valid_response() -> None: @@ -306 +305 @@ def test_get_validity_by_kind_at_least_one_valid_response() -> None: - assert get_validity_by_kind(dataset=dataset) == {kind: True} + assert is_valid_for_kinds(dataset=dataset, kinds=[kind]) @@ -309 +308 @@ def test_get_validity_by_kind_at_least_one_valid_response() -> None: -def test_get_validity_by_kind_only_invalid_responses() -> None: +def test_is_valid_for_kinds_only_invalid_responses() -> None: @@ -320 +319 @@ def test_get_validity_by_kind_only_invalid_responses() -> None: - assert get_validity_by_kind(dataset=dataset) == {kind: False} + assert not is_valid_for_kinds(dataset=dataset, kinds=[kind]) diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py index 606353e4..029eaa97 100644 --- a/services/api/src/api/routes/valid.py +++ b/services/api/src/api/routes/valid.py @@ -23,2 +22,0 @@ class ValidContent(TypedDict): - valid: List[str] - preview: List[str] @@ -25,0 +24 @@ class ValidContent(TypedDict): + preview: List[str] @@ -40 +38,0 @@ class ValidDatasets: - _valid_set = set.union(_viewer_set, _preview_set) @@ -42,2 +39,0 @@ class ValidDatasets: - valid=sorted(_valid_set), - preview=sorted(_preview_set), @@ -44,0 +41 @@ class ValidDatasets: + preview=sorted(_preview_set), diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py index 38fb758f..4827c75d 100644 --- a/services/api/tests/routes/test_valid.py +++ b/services/api/tests/routes/test_valid.py @@ -35 +35 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) -> - assert valid_datasets.content == {"valid": [], "preview": [], "viewer": []} + assert valid_datasets.content == {"preview": [], "viewer": []} @@ -39 +39 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) -> - "processing_graph_specification,expected_preview,expected_viewer,expected_valid", + "processing_graph_specification,expected_preview,expected_viewer", @@ -41,6 +41,6 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) -> - ({step_1: {}, step_2: {}}, [], [], []), - ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset"], [], ["dataset"]), - ({step_1: {}, step_2: {"enables_preview": True}}, [], [], []), - ({step_1: {"enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]), - ({step_1: {}, step_2: {"enables_viewer": True}}, [], [], []), - ({step_1: {"enables_preview": True, "enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]), + ({step_1: {}, step_2: {}}, [], []), + ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset"], []), + ({step_1: {}, step_2: {"enables_preview": True}}, [], []), + ({step_1: {"enables_viewer": True}, step_2: {}}, [], ["dataset"]), + ({step_1: {}, step_2: {"enables_viewer": True}}, [], []), + ({step_1: {"enables_preview": True, "enables_viewer": True}, step_2: {}}, [], ["dataset"]), @@ -53 +52,0 @@ def test_one_dataset( - expected_valid: List[str], @@ -61 +59,0 @@ def test_one_dataset( - "valid": expected_valid, @@ -68 +66 @@ def test_one_dataset( - "processing_graph_specification,expected_preview,expected_viewer,expected_valid", + "processing_graph_specification,expected_preview,expected_viewer", @@ -70,3 +68,3 @@ def test_one_dataset( - ({step_1: {}, step_2: {}}, [], [], []), - ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset1"], [], ["dataset1"]), - ({step_1: {}, step_2: {"enables_preview": True}}, ["dataset2"], [], ["dataset2"]), + ({step_1: {}, step_2: {}}, [], []), + ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset1"], []), + ({step_1: {}, step_2: {"enables_preview": True}}, ["dataset2"], []), @@ -77 +74,0 @@ def test_one_dataset( - ["dataset1", "dataset2"], @@ -83 +79,0 @@ def test_one_dataset( - ["dataset1", "dataset2"], @@ -91 +86,0 @@ def test_two_datasets( - expected_valid: List[str], @@ -108 +102,0 @@ def test_two_datasets( - "valid": expected_valid, @@ -115 +109 @@ def test_two_datasets( - "processing_graph_specification,expected_preview,expected_viewer,expected_valid", + "processing_graph_specification,expected_preview,expected_viewer", @@ -125 +118,0 @@ def test_two_datasets( - [], @@ -135 +127,0 @@ def test_two_datasets( - ["dataset"], @@ -149 +140,0 @@ def test_two_datasets( - ["dataset"], @@ -159 +149,0 @@ def test_two_datasets( - ["dataset"], @@ -173 +162,0 @@ def test_two_datasets( - ["dataset"], @@ -187 +175,0 @@ def test_two_datasets( - ["dataset"], @@ -195 +182,0 @@ def test_three_steps( - expected_valid: List[str], @@ -224 +210,0 @@ def test_three_steps( - "valid": expected_valid, @@ -241,4 +226,0 @@ def test_errors() -> None: - "valid": [ - dataset_a, - dataset_b, - ], diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py index a883d39b..3e270714 100644 --- a/services/api/tests/test_app.py +++ b/services/api/tests/test_app.py @@ -51 +50,0 @@ def test_get_valid_datasets(client: TestClient) -> None: - assert "valid" in response.json() diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py index 5eb630d1..17a6a41d 100644 --- a/services/worker/src/worker/dtos.py +++ b/services/worker/src/worker/dtos.py @@ -178 +178,2 @@ class DatasetIsValidResponse(TypedDict): - valid: bool + preview: bool + viewer: bool diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 87c48d90..493d7ba3 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -184,0 +185 @@ class JobRunnerFactory(BaseJobRunnerFactory): + processing_graph=self.processing_graph, diff --git a/services/worker/src/worker/job_runners/dataset/is_valid.py b/services/worker/src/worker/job_runners/dataset/is_valid.py index cb45bf8f..caf65b0a 100644 --- a/services/worker/src/worker/job_runners/dataset/is_valid.py +++ b/services/worker/src/worker/job_runners/dataset/is_valid.py @@ -5 +4,0 @@ import logging -from typing import Tuple @@ -8 +7,3 @@ from libcommon.constants import PROCESSING_STEP_DATASET_IS_VALID_VERSION -from libcommon.simple_cache import get_validity_by_kind +from libcommon.processing_graph import ProcessingGraph, ProcessingStep +from libcommon.simple_cache import is_valid_for_kinds +from libcommon.utils import JobInfo @@ -10 +11,2 @@ from libcommon.simple_cache import get_validity_by_kind -from worker.dtos import DatasetIsValidResponse, JobResult +from worker.config import AppConfig +from worker.dtos import CompleteJobResult, DatasetIsValidResponse, JobResult @@ -13,2 +14,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -SPLIT_KINDS = ["config-split-names-from-streaming", "config-split-names-from-info"] -FIRST_ROWS_KINDS = ["split-first-rows-from-streaming", "split-first-rows-from-parquet"] @@ -16,2 +16 @@ FIRST_ROWS_KINDS = ["split-first-rows-from-streaming", "split-first-rows-from-pa - -def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, float]: +def compute_is_valid_response(dataset: str, processing_graph: ProcessingGraph) -> DatasetIsValidResponse: @@ -21,3 +20,4 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo - A dataset is valid if: - - /splits is valid for at least one of the configs - - /first-rows is valid for at least one of the splits + + A dataset is valid if at least one response of any of the artifacts for any of the + steps (for viewer and preview) is valid. + The deprecated `valid` field is an "or" of the `preview` and `viewer` fields. @@ -28,0 +29,3 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo + processing_graph (`ProcessingGraph`): + The processing graph. In particular, it must provide the list of + processing steps that enable the viewer and the preview. @@ -30 +33 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo - `DatasetIsValidResponse`: An object with the is_valid_response. + `DatasetIsValidResponse`: The response (viewer, preview). @@ -34,3 +37,5 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo - validity_by_kind = get_validity_by_kind(dataset=dataset, kinds=SPLIT_KINDS + FIRST_ROWS_KINDS) - is_valid = any(validity_by_kind[kind] for kind in SPLIT_KINDS if kind in validity_by_kind) and any( - validity_by_kind[kind] for kind in FIRST_ROWS_KINDS if kind in validity_by_kind + viewer = is_valid_for_kinds( + dataset=dataset, kinds=[step.cache_kind for step in processing_graph.get_processing_steps_enables_viewer()] + ) + preview = is_valid_for_kinds( + dataset=dataset, kinds=[step.cache_kind for step in processing_graph.get_processing_steps_enables_preview()] @@ -39 +44 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo - return (DatasetIsValidResponse({"valid": is_valid}), 1.0) + return DatasetIsValidResponse({"viewer": viewer, "preview": preview}) @@ -50,0 +56,14 @@ class DatasetIsValidJobRunner(DatasetJobRunner): + def __init__( + self, + job_info: JobInfo, + app_config: AppConfig, + processing_step: ProcessingStep, + processing_graph: ProcessingGraph, + ) -> None: + super().__init__( + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + ) + self.processing_graph = processing_graph + @@ -54,2 +73,2 @@ class DatasetIsValidJobRunner(DatasetJobRunner): - response_content, progress = compute_is_valid_response(dataset=self.dataset) - return JobResult(response_content, progress=progress) + response_content = compute_is_valid_response(dataset=self.dataset, processing_graph=self.processing_graph) + return CompleteJobResult(response_content) diff --git a/services/worker/tests/job_runners/dataset/test_is_valid.py b/services/worker/tests/job_runners/dataset/test_is_valid.py index b03d2647..a6c64001 100644 --- a/services/worker/tests/job_runners/dataset/test_is_valid.py +++ b/services/worker/tests/job_runners/dataset/test_is_valid.py @@ -26,0 +27 @@ GetJobRunner = Callable[[str, AppConfig], DatasetIsValidJobRunner] +DATASET = "dataset" @@ -28,2 +29,2 @@ GetJobRunner = Callable[[str, AppConfig], DatasetIsValidJobRunner] -UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING: UpstreamResponse = UpstreamResponse( - kind="config-split-names-from-streaming", dataset="dataset_ok", config=None, http_status=HTTPStatus.OK, content={} +UPSTREAM_RESPONSE_CONFIG_SIZE: UpstreamResponse = UpstreamResponse( + kind="config-size", dataset=DATASET, config="config", http_status=HTTPStatus.OK, content={} @@ -31,2 +32,7 @@ UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING: UpstreamResponse = UpstreamRespons -UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO: UpstreamResponse = UpstreamResponse( - kind="config-split-names-from-info", dataset="dataset_ok", config=None, http_status=HTTPStatus.OK, content={} +UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET: UpstreamResponse = UpstreamResponse( + kind="split-first-rows-from-parquet", + dataset=DATASET, + config="config", + split="split", + http_status=HTTPStatus.OK, + content={}, @@ -36 +42 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING: UpstreamResponse = UpstreamRe - dataset="dataset_ok", + dataset=DATASET, @@ -37,0 +44 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING: UpstreamResponse = UpstreamRe + split="split", @@ -41,2 +48,2 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING: UpstreamResponse = UpstreamRe -UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET: UpstreamResponse = UpstreamResponse( - kind="split-first-rows-from-parquet", dataset="dataset_ok", config="config", http_status=HTTPStatus.OK, content={} +UPSTREAM_RESPONSE_CONFIG_SIZE_ERROR: UpstreamResponse = UpstreamResponse( + kind="config-size", dataset=DATASET, config="config", http_status=HTTPStatus.INTERNAL_SERVER_ERROR, content={} @@ -44,4 +51,5 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET: UpstreamResponse = UpstreamResp -UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO_ERROR: UpstreamResponse = UpstreamResponse( - kind="config-split-names-from-info", - dataset="dataset_ok", - config=None, +UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR: UpstreamResponse = UpstreamResponse( + kind="split-first-rows-from-parquet", + dataset=DATASET, + config="config", + split="split", @@ -53 +61 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR: UpstreamResponse = Upst - dataset="dataset_ok", + dataset=DATASET, @@ -54,0 +63 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR: UpstreamResponse = Upst + split="split", @@ -58,2 +67,2 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR: UpstreamResponse = Upst -EXPECTED_OK = ( - {"valid": True}, +EXPECTED_ERROR = ( + {"viewer": False, "preview": False}, @@ -62,2 +71,10 @@ EXPECTED_OK = ( -EXPECTED_ERROR = ( - {"valid": False}, +EXPECTED_VIEWER_OK = ( + {"viewer": True, "preview": False}, + 1.0, +) +EXPECTED_PREVIEW_OK = ( + {"viewer": False, "preview": True}, + 1.0, +) +EXPECTED_BOTH_OK = ( + {"viewer": True, "preview": True}, @@ -78,8 +95 @@ def get_job_runner( - processing_graph = ProcessingGraph( - { - processing_step_name: { - "input_type": "dataset", - "job_runner_version": DatasetIsValidJobRunner.get_job_runner_version(), - } - } - ) + processing_graph = ProcessingGraph(app_config.processing_graph.specification) @@ -99,0 +110 @@ def get_job_runner( + processing_graph=processing_graph, @@ -106 +117 @@ def get_job_runner( - "dataset,upstream_responses,expected_error_code,expected,should_raise", + "upstream_responses,expected", @@ -109 +119,0 @@ def get_job_runner( - "dataset_ok", @@ -111 +121,2 @@ def get_job_runner( - UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING, + UPSTREAM_RESPONSE_CONFIG_SIZE, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET, @@ -114,3 +125 @@ def get_job_runner( - None, - EXPECTED_OK, - False, + EXPECTED_BOTH_OK, @@ -119 +127,0 @@ def get_job_runner( - "dataset_ok", @@ -121,2 +129,2 @@ def get_job_runner( - UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING, - UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO, + UPSTREAM_RESPONSE_CONFIG_SIZE_ERROR, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET, @@ -123,0 +132,5 @@ def get_job_runner( + ], + EXPECTED_PREVIEW_OK, + ), + ( + [ @@ -124,0 +138 @@ def get_job_runner( + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING, @@ -126,3 +140 @@ def get_job_runner( - None, - EXPECTED_OK, - False, + EXPECTED_PREVIEW_OK, @@ -130,3 +141,0 @@ def get_job_runner( - ("dataset_ok", [], None, EXPECTED_ERROR, False), - ("dataset_ok", [UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO], None, EXPECTED_ERROR, False), - ("dataset_ok", [UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING], None, EXPECTED_ERROR, False), @@ -134,5 +143,6 @@ def get_job_runner( - "dataset_ok", - [UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO_ERROR, UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING], - None, - EXPECTED_ERROR, - False, + [ + UPSTREAM_RESPONSE_CONFIG_SIZE, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING, + ], + EXPECTED_BOTH_OK, @@ -141,5 +151,6 @@ def get_job_runner( - "dataset_ok", - [UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO, UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR], - None, - EXPECTED_ERROR, - False, + [ + UPSTREAM_RESPONSE_CONFIG_SIZE, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR, + ], + EXPECTED_VIEWER_OK, @@ -148 +158,0 @@ def get_job_runner( - "dataset_ok", @@ -150,3 +160,8 @@ def get_job_runner( - UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING, - UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO_ERROR, - UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET, + UPSTREAM_RESPONSE_CONFIG_SIZE, + ], + EXPECTED_VIEWER_OK, + ), + ( + [ + UPSTREAM_RESPONSE_CONFIG_SIZE_ERROR, + UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR, @@ -155,3 +170,5 @@ def get_job_runner( - None, - EXPECTED_OK, - False, + EXPECTED_ERROR, + ), + ( + [], + EXPECTED_ERROR, @@ -164 +180,0 @@ def test_compute( - dataset: str, @@ -166 +181,0 @@ def test_compute( - expected_error_code: str, @@ -168 +182,0 @@ def test_compute( - should_raise: bool, @@ -169,0 +184 @@ def test_compute( + dataset = DATASET @@ -173,8 +188,3 @@ def test_compute( - if should_raise: - with pytest.raises(Exception) as e: - job_runner.compute() - assert e.typename == expected_error_code - else: - compute_result = job_runner.compute() - assert compute_result.content == expected[0] - assert compute_result.progress == expected[1] + compute_result = job_runner.compute() + assert compute_result.content == expected[0] + assert compute_result.progress == expected[1] @@ -187 +197 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No - assert compute_result.content == {"valid": False} + assert compute_result.content == {"viewer": False, "preview": False} diff --git a/services/worker/tests/job_runners/utils.py b/services/worker/tests/job_runners/utils.py index 30e58fdf..04d80f76 100644 --- a/services/worker/tests/job_runners/utils.py +++ b/services/worker/tests/job_runners/utils.py @@ -8 +8 @@ from typing import Any, Mapping, Optional, TypedDict -class UpstreamResponse(TypedDict): +class _UpstreamResponse(TypedDict): @@ -11 +10,0 @@ class UpstreamResponse(TypedDict): - config: Optional[str] @@ -13,0 +13,5 @@ class UpstreamResponse(TypedDict): + + +class UpstreamResponse(_UpstreamResponse, total=False): + config: Optional[str] + split: Optional[str]
ee9072dedf7130cc64ffbbe608d9a055003d37e4
Sylvain Lesage
2023-06-29T07:27:39
feat: 🎸 reduce the number of workers back to 20 (#1457)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index a0f96bd5..b0dc9cf6 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -272 +272 @@ workers: - replicas: 100 + replicas: 20
0bb49fbb7fe6bec3237f34a760d1a4f02d85e01b
Andrea Francis Soria Jimenez
2023-06-28T20:42:34
Disable backfill (#1456)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 774978a2..a0f96bd5 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -124 +124 @@ cacheMaintenance: - action: "backfill" + action: "skip" @@ -129 +129 @@ cacheMaintenance: - error_codes_to_retry: "CreateCommitError,LockedDatasetTimeoutError" + error_codes_to_retry: ""
f1c6f82d05510f06c666b182bddcb39c42fcdfec
Polina Kazakova
2023-06-28T18:33:53
Update quality target in Makefile for /chart (#1455)
diff --git a/chart/Makefile b/chart/Makefile index 57716f53..779c0593 100644 --- a/chart/Makefile +++ b/chart/Makefile @@ -8 +8 @@ quality: - helm lint --values env/dev.yaml + helm lint --values env/staging.yaml
7ab76da7a74c7f2ddc2656c3c3953ddb1b591d1d
Andrea Francis Soria Jimenez
2023-06-28T15:53:06
Enable backfill (#1454)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 22a72830..774978a2 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -124,2 +124,2 @@ cacheMaintenance: - action: "skip" - # ^ allowed values are {skip,backfill,upgrade} + action: "backfill" + # ^ allowed values are {backfill, collect-metrics,skip} @@ -129 +129 @@ cacheMaintenance: - error_codes_to_retry: "" + error_codes_to_retry: "CreateCommitError,LockedDatasetTimeoutError"
5d4d7bd731590ce56ac5448bce35cd5d8c581f7d
Andrea Francis Soria Jimenez
2023-06-28T15:31:04
Increase resources (#1453)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 96acf9f7..22a72830 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -272 +272 @@ workers: - replicas: 20 + replicas: 100 @@ -283 +283 @@ workers: - workerJobTypesBlocked: "dataset-config-names,config-split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan" + workerJobTypesBlocked: "dataset-config-names,config-split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan,split-duckdb-index"
2f6dbe2423a22c7749f34d73f3e54e42dd93f228
Andrea Francis Soria Jimenez
2023-06-28T14:50:36
Change commiter key (#1451)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 1a35e169..732dcf41 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.14.0 +version: 1.14.1 diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index 68b16da1..18e1497c 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -98 +98 @@ - key: HF_TOKEN + key: PARQUET_CONVERTER_HF_TOKEN
d9c9d42d9a9a3666272c45279a171041a43deab4
Andrea Francis Soria Jimenez
2023-06-28T14:18:17
Adding debug logs for split-duckdb-index (#1449)
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py index daaa0d25..01874c4c 100644 --- a/services/worker/src/worker/job_runners/split/duckdb_index.py +++ b/services/worker/src/worker/job_runners/split/duckdb_index.py @@ -163,0 +164 @@ def compute_index_rows( + logging.debug(f"try to create branch for {dataset=} with {target_revision=} on {hf_endpoint=}") @@ -170,0 +172 @@ def compute_index_rows( + logging.debug(f"get dataset info for {dataset=} with {target_revision=}") @@ -175,0 +178 @@ def compute_index_rows( + logging.debug(f"delete operations for {dataset=} {delete_operations=}") @@ -180,0 +184 @@ def compute_index_rows( + logging.debug(f"add operations for {dataset=} {add_operations=}") @@ -189,0 +194 @@ def compute_index_rows( + logging.debug(f"create commit {commit_message} for {dataset=} {add_operations=}") @@ -192,0 +198 @@ def compute_index_rows( + logging.debug(f"dataset info for {dataset=} {target_dataset_info=}")
f17eb0f6371bd6df1dd5f79eb27340e049c46d54
Andrea Francis Soria Jimenez
2023-06-27T20:01:09
Adding other processing steps (#1441)
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 7d454058..b36d4535 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -362,0 +363,2 @@ class ProcessingGraphConfig: + "config-split-names-from-streaming", + "config-parquet-and-info", diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py index c75d2104..ea672c6f 100644 --- a/libs/libcommon/tests/test_processing_graph.py +++ b/libs/libcommon/tests/test_processing_graph.py @@ -85,0 +86 @@ def graph() -> ProcessingGraph: + "split-duckdb-index", @@ -106,0 +108 @@ def graph() -> ProcessingGraph: + "split-duckdb-index", @@ -298 +300 @@ def graph() -> ProcessingGraph: - ["config-split-names-from-info"], + ["config-split-names-from-info", "config-split-names-from-streaming", "config-parquet-and-info"], @@ -300,0 +303 @@ def graph() -> ProcessingGraph: + "config-split-names-from-streaming",
228334075f2368885e7a1ed1ae77573165fa12d1
Andrea Francis Soria Jimenez
2023-06-27T19:12:46
Try to fix Duckdb extensions (#1440)
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py index 822ccb50..daaa0d25 100644 --- a/services/worker/src/worker/job_runners/split/duckdb_index.py +++ b/services/worker/src/worker/job_runners/split/duckdb_index.py @@ -129,0 +130,4 @@ def compute_index_rows( + # index all columns + db_path = duckdb_index_file_directory.resolve() / DUCKDB_DEFAULT_INDEX_FILENAME + con = duckdb.connect(str(db_path.resolve())) + @@ -132 +136 @@ def compute_index_rows( - duckdb.execute(SET_EXTENSIONS_DIRECTORY_COMMAND.format(directory=extensions_directory)) + con.execute(SET_EXTENSIONS_DIRECTORY_COMMAND.format(directory=extensions_directory)) @@ -134,4 +138,4 @@ def compute_index_rows( - duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="httpfs")) - duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="httpfs")) - duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="fts")) - duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="fts")) + con.execute(INSTALL_EXTENSION_COMMAND.format(extension="httpfs")) + con.execute(LOAD_EXTENSION_COMMAND.format(extension="httpfs")) + con.execute(INSTALL_EXTENSION_COMMAND.format(extension="fts")) + con.execute(LOAD_EXTENSION_COMMAND.format(extension="fts")) @@ -139,4 +142,0 @@ def compute_index_rows( - # index all columns - db_path = duckdb_index_file_directory.resolve() / DUCKDB_DEFAULT_INDEX_FILENAME - - con = duckdb.connect(str(db_path.resolve()))
b2a91994683ef8107197d2f0e12cbcd17df42011
Andrea Francis Soria Jimenez
2023-06-27T16:45:36
Set Duckdb extensions install directory (#1439)
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index 77a03377..68b16da1 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -110,0 +111,2 @@ +- name: DUCKDB_INDEX_EXTENSIONS_DIRECTORY + value: "/tmp/duckdb-extensions" diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 231ec1f5..7d454058 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -113,0 +114 @@ DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s" +DUCKDB_INDEX_EXTENSIONS_DIRECTORY: Optional[str] = None @@ -123,0 +125 @@ class DuckDbIndexConfig: + extensions_directory: Optional[str] = DUCKDB_INDEX_EXTENSIONS_DIRECTORY @@ -137,0 +140 @@ class DuckDbIndexConfig: + extensions_directory=env.str(name="EXTENSIONS_DIRECTORY", default=DUCKDB_INDEX_EXTENSIONS_DIRECTORY), diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py index b715de60..822ccb50 100644 --- a/services/worker/src/worker/job_runners/split/duckdb_index.py +++ b/services/worker/src/worker/job_runners/split/duckdb_index.py @@ -48,0 +49 @@ LOAD_EXTENSION_COMMAND = "LOAD '{extension}';" +SET_EXTENSIONS_DIRECTORY_COMMAND = "SET extension_directory='{directory}';" @@ -62,0 +64 @@ def compute_index_rows( + extensions_directory: Optional[str], @@ -128,0 +131,3 @@ def compute_index_rows( + if extensions_directory is not None: + duckdb.execute(SET_EXTENSIONS_DIRECTORY_COMMAND.format(directory=extensions_directory)) + @@ -259,0 +265 @@ class SplitDuckDbIndexJobRunner(SplitJobRunnerWithCache): + extensions_directory=self.duckdb_index_config.extensions_directory,
be96d0a6693b8405e95e850d5181246c476839db
Andrea Francis Soria Jimenez
2023-06-27T15:31:40
Increase chart version (#1438)
diff --git a/chart/Chart.yaml b/chart/Chart.yaml index 13b7046f..1a35e169 100644 --- a/chart/Chart.yaml +++ b/chart/Chart.yaml @@ -21 +21 @@ type: application -version: 1.13.0 +version: 1.14.0
b988e218f04a9340d88d1afc6dbd0176b943b92a
Andrea Francis Soria Jimenez
2023-06-27T13:41:16
feat: Index the (text) datasets contents to enable full-text search - DuckDB (#1296)
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json index 11b6b1fe..59a058d9 100644 --- a/chart/static-files/openapi.json +++ b/chart/static-files/openapi.json @@ -928 +928 @@ - "items": { "$ref": "#/components/schemas/ParquetFileItem" } + "items": { "$ref": "#/components/schemas/SplitHubFile" } @@ -932 +932 @@ - "ParquetFileItem": { + "SplitHubFile": { diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index 6f2d1411..77a03377 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -89,0 +90,21 @@ +# specific to 'split-duckdb-index' job runner +- name: DUCKDB_INDEX_COMMIT_MESSAGE + value: {{ .Values.duckDBIndex.commitMessage | quote }} +- name: DUCKDB_INDEX_COMMITTER_HF_TOKEN + {{- if .Values.secrets.appParquetConverterHfToken.fromSecret }} + valueFrom: + secretKeyRef: + name: {{ .Values.secrets.appParquetConverterHfToken.secretName | quote }} + key: HF_TOKEN + optional: false + {{- else }} + value: {{ .Values.secrets.appParquetConverterHfToken.value }} + {{- end }} +- name: DUCKDB_INDEX_TARGET_REVISION + value: {{ .Values.duckDBIndex.targetRevision | quote }} +- name: DUCKDB_INDEX_URL_TEMPLATE + value: {{ .Values.duckDBIndex.urlTemplate | quote }} +- name: DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES + value: {{ .Values.duckDBIndex.maxParquetSizeBytes | quote }} +- name: DUCKDB_INDEX_STORAGE_DIRECTORY + value: {{ .Values.duckDBIndex.storageDirectory | quote }} diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl index 12aeec4c..ac9370e6 100644 --- a/chart/templates/_helpers.tpl +++ b/chart/templates/_helpers.tpl @@ -171,0 +172,9 @@ The parquet-metadata/ subpath in the NFS +{{/* +The duckdb-index/ subpath in the NFS +- in a subdirectory named as the chart (datasets-server/), and below it, +- in a subdirectory named as the Release, so that Releases will not share the same dir +*/}} +{{- define "duckDBIndex.subpath" -}} +{{- printf "%s/%s/%s/" .Chart.Name .Release.Name "duckdb-index" }} +{{- end }} + diff --git a/chart/templates/_initContainerDuckDBIndex.tpl b/chart/templates/_initContainerDuckDBIndex.tpl new file mode 100644 index 00000000..ed7cb43b --- /dev/null +++ b/chart/templates/_initContainerDuckDBIndex.tpl @@ -0,0 +1,21 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +{{- define "initContainerDuckDBIndex" -}} +- name: prepare-duckdb-index + image: ubuntu:focal + imagePullPolicy: {{ .Values.images.pullPolicy }} + command: ["/bin/sh", "-c"] + args: + - chown {{ .Values.uid }}:{{ .Values.gid }} /mounted-path; + volumeMounts: + - mountPath: /mounted-path + mountPropagation: None + name: data + subPath: "{{ include "duckDBIndex.subpath" . }}" + readOnly: false + securityContext: + runAsNonRoot: false + runAsUser: 0 + runAsGroup: 0 +{{- end -}} diff --git a/chart/templates/_volumeMountDuckDBIndex.tpl b/chart/templates/_volumeMountDuckDBIndex.tpl new file mode 100644 index 00000000..01c37b89 --- /dev/null +++ b/chart/templates/_volumeMountDuckDBIndex.tpl @@ -0,0 +1,10 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +{{- define "volumeMountDuckDBIndexRW" -}} +- mountPath: {{ .Values.duckDBIndex.storageDirectory | quote }} + mountPropagation: None + name: data + subPath: "{{ include "duckDBIndex.subpath" . }}" + readOnly: false +{{- end -}} diff --git a/chart/templates/worker/_container.tpl b/chart/templates/worker/_container.tpl index 9f83bad8..f9b86817 100644 --- a/chart/templates/worker/_container.tpl +++ b/chart/templates/worker/_container.tpl @@ -26,0 +27 @@ + {{ include "volumeMountDuckDBIndexRW" . | nindent 2 }} diff --git a/chart/templates/worker/_deployment.yaml b/chart/templates/worker/_deployment.yaml index e06d319c..03a70646 100644 --- a/chart/templates/worker/_deployment.yaml +++ b/chart/templates/worker/_deployment.yaml @@ -28,0 +29 @@ spec: + {{ include "initContainerDuckDBIndex" . | nindent 8 }} diff --git a/chart/values.yaml b/chart/values.yaml index 4a57fc7a..98cc505b 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -216,0 +217,11 @@ parquetMetadata: +duckDBIndex: + # Directory on the shared storage (used temporarily to prepare the duckdb indexes before sending to the Hub) + storageDirectory: "/duckdb-index" + # the git commit message when the duckdb index file is uploaded to the Hub. Defaults to `Update duckdb index files`. + commitMessage: "Update duckdb index files" + # the git revision of the dataset where to store the duckdb index file. Defaults to `refs/convert/parquet`. + targetRevision: "refs/convert/parquet" + # the URL template to build the duckdb index file URL. Defaults to `/datasets/%s/resolve/%s/%s`. + urlTemplate: "/datasets/%s/resolve/%s/%s" + # the maximum size of the split parquets. + maxParquetSizeBytes: "100_000_000" diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index e4d63e70..231ec1f5 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -26,0 +27 @@ from libcommon.constants import ( + PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION, @@ -106,0 +108,33 @@ class ParquetMetadataConfig: +DUCKDB_INDEX_STORAGE_DIRECTORY = None +DUCKDB_INDEX_COMMIT_MESSAGE = "Update duckdb index file" +DUCKDB_INDEX_COMMITTER_HF_TOKEN = None +DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES = 100_000_000 +DUCKDB_INDEX_TARGET_REVISION = "refs/convert/parquet" +DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s" + + +@dataclass(frozen=True) +class DuckDbIndexConfig: + storage_directory: Optional[str] = DUCKDB_INDEX_STORAGE_DIRECTORY + commit_message: str = DUCKDB_INDEX_COMMIT_MESSAGE + committer_hf_token: Optional[str] = DUCKDB_INDEX_COMMITTER_HF_TOKEN + target_revision: str = DUCKDB_INDEX_TARGET_REVISION + url_template: str = DUCKDB_INDEX_URL_TEMPLATE + max_parquet_size_bytes: int = DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES + + @classmethod + def from_env(cls) -> "DuckDbIndexConfig": + env = Env(expand_vars=True) + with env.prefixed("DUCKDB_INDEX_"): + return cls( + storage_directory=env.str(name="STORAGE_DIRECTORY", default=DUCKDB_INDEX_STORAGE_DIRECTORY), + commit_message=env.str(name="COMMIT_MESSAGE", default=DUCKDB_INDEX_COMMIT_MESSAGE), + committer_hf_token=env.str(name="COMMITTER_HF_TOKEN", default=DUCKDB_INDEX_COMMITTER_HF_TOKEN), + target_revision=env.str(name="TARGET_REVISION", default=DUCKDB_INDEX_TARGET_REVISION), + url_template=env.str(name="URL_TEMPLATE", default=DUCKDB_INDEX_URL_TEMPLATE), + max_parquet_size_bytes=env.int( + name="MAX_PARQUET_SIZE_BYTES", default=DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES + ), + ) + + @@ -321,0 +356,7 @@ class ProcessingGraphConfig: + "split-duckdb-index": { + "input_type": "split", + "triggered_by": [ + "config-split-names-from-info", + ], + "job_runner_version": PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION, + }, diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index cd41126f..66a609b6 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -8,0 +9 @@ PARQUET_METADATA_CACHE_APPNAME = "datasets_server_parquet_metadata" +DUCKDB_INDEX_CACHE_APPNAME = "datasets_server_duckdb_index" @@ -38,0 +40 @@ PROCESSING_STEP_SPLIT_IMAGE_URL_COLUMNS_VERSION = 1 +PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION = 1 diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py index f9de8f3f..46e66228 100644 --- a/libs/libcommon/src/libcommon/exceptions.py +++ b/libs/libcommon/src/libcommon/exceptions.py @@ -75,0 +76 @@ CacheableErrorCode = Literal[ + "CacheDirectoryNotInitializedError", @@ -91,0 +93 @@ CacheableErrorCode = Literal[ + "DuckDBIndexFileNotFoundError", @@ -104,0 +107 @@ CacheableErrorCode = Literal[ + "NoIndexableColumnsError", @@ -114,0 +118 @@ CacheableErrorCode = Literal[ + "SplitWithTooBigParquetError", @@ -138,0 +143,7 @@ class CacheableError(CustomError): +class CacheDirectoryNotInitializedError(CacheableError): + """Raised when the cache directory has not been initialized before job compute.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "CacheDirectoryNotInitializedError", cause, True) + + @@ -234,0 +246,7 @@ class DatasetWithTooBigExternalFilesError(CacheableError): +class DatasetWithTooManyConfigsError(CacheableError): + """Raised when the number of configs of a dataset exceeded the limit.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyConfigsError", cause, True) + + @@ -249,2 +267,2 @@ class DatasetWithTooManyParquetFilesError(CacheableError): -class LockedDatasetTimeoutError(CacheableError): - """Raised when a dataset is locked by another job.""" +class DuckDBIndexFileNotFoundError(CacheableError): + """Raised when no duckdb index file was found for split.""" @@ -253 +271 @@ class LockedDatasetTimeoutError(CacheableError): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "LockedDatasetTimeoutError", cause, True) + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "DuckDBIndexFileNotFoundError", cause, False) @@ -357,0 +376,7 @@ class JobManagerExceededMaximumDurationError(CacheableError): +class LockedDatasetTimeoutError(CacheableError): + """Raised when a dataset is locked by another job.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "LockedDatasetTimeoutError", cause, True) + + @@ -371,0 +397,7 @@ class NormalRowsError(CacheableError): +class NoIndexableColumnsError(CacheableError): + """Raised when split does not have string columns to index.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "NoIndexableColumnsError", cause, True) + + @@ -452,0 +485,7 @@ class SplitNotFoundError(CacheableError): +class SplitWithTooBigParquetError(CacheableError): + """Raised when the split parquet size (sum of parquet sizes given) is too big.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "SplitWithTooBigParquetError", cause, False) + + @@ -499,7 +537,0 @@ class UnsupportedExternalFilesError(CacheableError): - - -class DatasetWithTooManyConfigsError(CacheableError): - """Raised when the number of configs of a dataset exceeded the limit.""" - - def __init__(self, message: str, cause: Optional[BaseException] = None): - super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyConfigsError", cause, True) diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py index 0f8ca107..351cb65e 100644 --- a/libs/libcommon/src/libcommon/parquet_utils.py +++ b/libs/libcommon/src/libcommon/parquet_utils.py @@ -22,0 +23 @@ from libcommon.simple_cache import get_previous_step_or_raise +from libcommon.utils import SplitHubFile @@ -39,9 +39,0 @@ class FileSystemError(Exception): -class ParquetFileItem(TypedDict): - dataset: str - config: str - split: str - url: str - filename: str - size: int - - @@ -137 +129 @@ class ParquetIndexWithoutMetadata: - parquet_file_items: List[ParquetFileItem], + parquet_file_items: List[SplitHubFile], diff --git a/libs/libcommon/src/libcommon/storage.py b/libs/libcommon/src/libcommon/storage.py index bbef1442..63d9c108 100644 --- a/libs/libcommon/src/libcommon/storage.py +++ b/libs/libcommon/src/libcommon/storage.py @@ -14,0 +15 @@ from libcommon.constants import ( + DUCKDB_INDEX_CACHE_APPNAME, @@ -83,0 +85,14 @@ def init_parquet_metadata_dir(directory: Optional[StrPath] = None) -> StrPath: +def init_duckdb_index_cache_dir(directory: Optional[StrPath] = None) -> StrPath: + """Initialize the duckdb index directory. + + If directory is None, it will be set to the default duckdb index location on the machine. + + Args: + directory (Optional[Union[str, PathLike[str]]], optional): The directory to initialize. Defaults to None. + + Returns: + Union[str, PathLike[str]]: The directory. + """ + return init_dir(directory, appname=DUCKDB_INDEX_CACHE_APPNAME) + + diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py index 301a7609..b921ea78 100644 --- a/libs/libcommon/src/libcommon/utils.py +++ b/libs/libcommon/src/libcommon/utils.py @@ -67,0 +68,9 @@ class JobResult(TypedDict): +class SplitHubFile(TypedDict): + dataset: str + config: str + split: str + url: str + filename: str + size: int + + diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py index 49bd37fb..c75d2104 100644 --- a/libs/libcommon/tests/test_processing_graph.py +++ b/libs/libcommon/tests/test_processing_graph.py @@ -95,0 +96 @@ def graph() -> ProcessingGraph: + "split-duckdb-index", @@ -102 +103,5 @@ def graph() -> ProcessingGraph: - ["split-first-rows-from-streaming", "dataset-split-names", "config-opt-in-out-urls-count"], + [ + "split-first-rows-from-streaming", + "dataset-split-names", + "config-opt-in-out-urls-count", + ], @@ -289,0 +295,11 @@ def graph() -> ProcessingGraph: + ( + "split-duckdb-index", + [], + ["config-split-names-from-info"], + [ + "config-split-names-from-info", + "config-parquet-and-info", + "config-info", + "dataset-config-names", + ], + ), diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 0ece7030..f2363ed9 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -996,0 +997,62 @@ idna = ["idna (>=2.1)"] +[[package]] +name = "duckdb" +version = "0.8.1" +description = "DuckDB embedded database" +category = "main" +optional = false +python-versions = "*" +files = [ + {file = "duckdb-0.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:14781d21580ee72aba1f5dcae7734674c9b6c078dd60470a08b2b420d15b996d"}, + {file = "duckdb-0.8.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:f13bf7ab0e56ddd2014ef762ae4ee5ea4df5a69545ce1191b8d7df8118ba3167"}, + {file = "duckdb-0.8.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e4032042d8363e55365bbca3faafc6dc336ed2aad088f10ae1a534ebc5bcc181"}, + {file = "duckdb-0.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:31a71bd8f0b0ca77c27fa89b99349ef22599ffefe1e7684ae2e1aa2904a08684"}, + {file = "duckdb-0.8.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:24568d6e48f3dbbf4a933109e323507a46b9399ed24c5d4388c4987ddc694fd0"}, + {file = "duckdb-0.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:297226c0dadaa07f7c5ae7cbdb9adba9567db7b16693dbd1b406b739ce0d7924"}, + {file = "duckdb-0.8.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5792cf777ece2c0591194006b4d3e531f720186102492872cb32ddb9363919cf"}, + {file = "duckdb-0.8.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:12803f9f41582b68921d6b21f95ba7a51e1d8f36832b7d8006186f58c3d1b344"}, + {file = "duckdb-0.8.1-cp310-cp310-win32.whl", hash = "sha256:d0953d5a2355ddc49095e7aef1392b7f59c5be5cec8cdc98b9d9dc1f01e7ce2b"}, + {file = "duckdb-0.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:6e6583c98a7d6637e83bcadfbd86e1f183917ea539f23b6b41178f32f813a5eb"}, + {file = "duckdb-0.8.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:fad7ed0d4415f633d955ac24717fa13a500012b600751d4edb050b75fb940c25"}, + {file = "duckdb-0.8.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:81ae602f34d38d9c48dd60f94b89f28df3ef346830978441b83c5b4eae131d08"}, + {file = "duckdb-0.8.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:7d75cfe563aaa058d3b4ccaaa371c6271e00e3070df5de72361fd161b2fe6780"}, + {file = "duckdb-0.8.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8dbb55e7a3336f2462e5e916fc128c47fe1c03b6208d6bd413ac11ed95132aa0"}, + {file = "duckdb-0.8.1-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a6df53efd63b6fdf04657385a791a4e3c4fb94bfd5db181c4843e2c46b04fef5"}, + {file = "duckdb-0.8.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b188b80b70d1159b17c9baaf541c1799c1ce8b2af4add179a9eed8e2616be96"}, + {file = "duckdb-0.8.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:5ad481ee353f31250b45d64b4a104e53b21415577943aa8f84d0af266dc9af85"}, + {file = "duckdb-0.8.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d1d1b1729993611b1892509d21c21628917625cdbe824a61ce891baadf684b32"}, + {file = "duckdb-0.8.1-cp311-cp311-win32.whl", hash = "sha256:2d8f9cc301e8455a4f89aa1088b8a2d628f0c1f158d4cf9bc78971ed88d82eea"}, + {file = "duckdb-0.8.1-cp311-cp311-win_amd64.whl", hash = "sha256:07457a43605223f62d93d2a5a66b3f97731f79bbbe81fdd5b79954306122f612"}, + {file = "duckdb-0.8.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:d2c8062c3e978dbcd80d712ca3e307de8a06bd4f343aa457d7dd7294692a3842"}, + {file = "duckdb-0.8.1-cp36-cp36m-win32.whl", hash = "sha256:fad486c65ae944eae2de0d590a0a4fb91a9893df98411d66cab03359f9cba39b"}, + {file = "duckdb-0.8.1-cp36-cp36m-win_amd64.whl", hash = "sha256:86fa4506622c52d2df93089c8e7075f1c4d0ba56f4bf27faebde8725355edf32"}, + {file = "duckdb-0.8.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:60e07a62782f88420046e30cc0e3de842d0901c4fd5b8e4d28b73826ec0c3f5e"}, + {file = "duckdb-0.8.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f18563675977f8cbf03748efee0165b4c8ef64e0cbe48366f78e2914d82138bb"}, + {file = "duckdb-0.8.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16e179443832bea8439ae4dff93cf1e42c545144ead7a4ef5f473e373eea925a"}, + {file = "duckdb-0.8.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a413d5267cb41a1afe69d30dd6d4842c588256a6fed7554c7e07dad251ede095"}, + {file = "duckdb-0.8.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:3784680df59eadd683b0a4c2375d451a64470ca54bd171c01e36951962b1d332"}, + {file = "duckdb-0.8.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:67a1725c2b01f9b53571ecf3f92959b652f60156c1c48fb35798302e39b3c1a2"}, + {file = "duckdb-0.8.1-cp37-cp37m-win32.whl", hash = "sha256:197d37e2588c5ad063e79819054eedb7550d43bf1a557d03ba8f8f67f71acc42"}, + {file = "duckdb-0.8.1-cp37-cp37m-win_amd64.whl", hash = "sha256:3843feb79edf100800f5037c32d5d5a5474fb94b32ace66c707b96605e7c16b2"}, + {file = "duckdb-0.8.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:624c889b0f2d656794757b3cc4fc58030d5e285f5ad2ef9fba1ea34a01dab7fb"}, + {file = "duckdb-0.8.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:fcbe3742d77eb5add2d617d487266d825e663270ef90253366137a47eaab9448"}, + {file = "duckdb-0.8.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47516c9299d09e9dbba097b9fb339b389313c4941da5c54109df01df0f05e78c"}, + {file = "duckdb-0.8.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cf1ba718b7522d34399446ebd5d4b9fcac0b56b6ac07bfebf618fd190ec37c1d"}, + {file = "duckdb-0.8.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e36e35d38a9ae798fe8cf6a839e81494d5b634af89f4ec9483f4d0a313fc6bdb"}, + {file = "duckdb-0.8.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:23493313f88ce6e708a512daacad13e83e6d1ea0be204b175df1348f7fc78671"}, + {file = "duckdb-0.8.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:1fb9bf0b6f63616c8a4b9a6a32789045e98c108df100e6bac783dc1e36073737"}, + {file = "duckdb-0.8.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:12fc13ecd5eddd28b203b9e3999040d3a7374a8f4b833b04bd26b8c5685c2635"}, + {file = "duckdb-0.8.1-cp38-cp38-win32.whl", hash = "sha256:a12bf4b18306c9cb2c9ba50520317e6cf2de861f121d6f0678505fa83468c627"}, + {file = "duckdb-0.8.1-cp38-cp38-win_amd64.whl", hash = "sha256:e4e809358b9559c00caac4233e0e2014f3f55cd753a31c4bcbbd1b55ad0d35e4"}, + {file = "duckdb-0.8.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:7acedfc00d97fbdb8c3d120418c41ef3cb86ef59367f3a9a30dff24470d38680"}, + {file = "duckdb-0.8.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:99bfe264059cdc1e318769103f656f98e819cd4e231cd76c1d1a0327f3e5cef8"}, + {file = "duckdb-0.8.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:538b225f361066231bc6cd66c04a5561de3eea56115a5dd773e99e5d47eb1b89"}, + {file = "duckdb-0.8.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ae0be3f71a18cd8492d05d0fc1bc67d01d5a9457b04822d025b0fc8ee6efe32e"}, + {file = "duckdb-0.8.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cd82ba63b58672e46c8ec60bc9946aa4dd7b77f21c1ba09633d8847ad9eb0d7b"}, + {file = "duckdb-0.8.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:780a34559aaec8354e83aa4b7b31b3555f1b2cf75728bf5ce11b89a950f5cdd9"}, + {file = "duckdb-0.8.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:01f0d4e9f7103523672bda8d3f77f440b3e0155dd3b2f24997bc0c77f8deb460"}, + {file = "duckdb-0.8.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:31f692decb98c2d57891da27180201d9e93bb470a3051fcf413e8da65bca37a5"}, + {file = "duckdb-0.8.1-cp39-cp39-win32.whl", hash = "sha256:e7fe93449cd309bbc67d1bf6f6392a6118e94a9a4479ab8a80518742e855370a"}, + {file = "duckdb-0.8.1-cp39-cp39-win_amd64.whl", hash = "sha256:81d670bc6807672f038332d9bf587037aabdd741b0810de191984325ed307abd"}, + {file = "duckdb-0.8.1.tar.gz", hash = "sha256:a54d37f4abc2afc4f92314aaa56ecf215a411f40af4bffe1e86bd25e62aceee9"}, +] + @@ -5594 +5656 @@ python-versions = "3.9.15" -content-hash = "109af95b92d54671ee4da9ddd322e3b03b5f9882ec38c9ace064afcbe37b5a2b" +content-hash = "3aa60ce2866418d5594a71e79a63dbd8e2bd3991c079c53bc055a7c584b3f69e" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index f5feb938..a2e6034d 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -14,0 +15 @@ conllu = "^4.5.2" +duckdb = "^0.8.1" diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index bc3fd397..0a57557e 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -11,0 +12 @@ from libcommon.config import ( + DuckDbIndexConfig, @@ -251,0 +253 @@ class AppConfig: + duckdb_index: DuckDbIndexConfig = field(default_factory=DuckDbIndexConfig) @@ -269,0 +272 @@ class AppConfig: + duckdb_index=DuckDbIndexConfig.from_env(), diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py index ef4a4cc0..5eb630d1 100644 --- a/services/worker/src/worker/dtos.py +++ b/services/worker/src/worker/dtos.py @@ -6,0 +7,2 @@ from typing import Any, Dict, List, Mapping, Optional, TypedDict +from libcommon.utils import SplitHubFile + @@ -113,6 +114,0 @@ class ConfigInfoResponse(TypedDict): -class ParquetFileItem(SplitItem): - url: str - filename: str - size: int - - @@ -120 +116 @@ class ConfigParquetAndInfoResponse(TypedDict): - parquet_files: List[ParquetFileItem] + parquet_files: List[SplitHubFile] @@ -137 +133 @@ class ConfigParquetResponse(TypedDict): - parquet_files: List[ParquetFileItem] + parquet_files: List[SplitHubFile] @@ -186 +182 @@ class DatasetParquetResponse(TypedDict): - parquet_files: List[ParquetFileItem] + parquet_files: List[SplitHubFile] diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 2352944c..87c48d90 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -36,0 +37 @@ from worker.job_runners.dataset.split_names import DatasetSplitNamesJobRunner +from worker.job_runners.split.duckdb_index import SplitDuckDbIndexJobRunner @@ -75,0 +77 @@ class JobRunnerFactory(BaseJobRunnerFactory): + duckdb_index_cache_directory: StrPath @@ -217,0 +220,8 @@ class JobRunnerFactory(BaseJobRunnerFactory): + if job_type == SplitDuckDbIndexJobRunner.get_job_type(): + return SplitDuckDbIndexJobRunner( + job_info=job_info, + app_config=self.app_config, + processing_step=processing_step, + duckdb_index_cache_directory=self.duckdb_index_cache_directory, + ) + @@ -236,0 +247 @@ class JobRunnerFactory(BaseJobRunnerFactory): + SplitDuckDbIndexJobRunner.get_job_type(), diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 8df7ea7a..8e82aa94 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -11 +11 @@ from typing import Any, List, Optional, Set, Tuple -from urllib.parse import quote, unquote +from urllib.parse import unquote @@ -74 +74 @@ from libcommon.simple_cache import get_previous_step_or_raise -from libcommon.utils import JobInfo +from libcommon.utils import JobInfo, SplitHubFile @@ -78 +78 @@ from worker.config import AppConfig, ParquetAndInfoConfig -from worker.dtos import CompleteJobResult, ConfigParquetAndInfoResponse, ParquetFileItem +from worker.dtos import CompleteJobResult, ConfigParquetAndInfoResponse @@ -80 +80 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas -from worker.utils import retry +from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url, retry @@ -100,6 +99,0 @@ class ParquetFile: -# TODO: use huggingface_hub's hf_hub_url after -# https://github.com/huggingface/huggingface_hub/issues/1082 -def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url_template: str) -> str: - return (hf_endpoint + url_template) % (repo_id, quote(revision, safe=""), filename) - - @@ -128 +122 @@ def create_parquet_file_item( -) -> ParquetFileItem: +) -> SplitHubFile: @@ -1037 +1030,0 @@ def compute_config_parquet_and_info_response( - sleeps = [1, 1, 1, 1, 1, 10, 10, 10, 10, 100] * 3 @@ -1039 +1032,3 @@ def compute_config_parquet_and_info_response( - with lock.git_branch(dataset=dataset, branch=target_revision, owner=job_id, sleeps=sleeps): + with lock.git_branch( + dataset=dataset, branch=target_revision, owner=job_id, sleeps=LOCK_GIT_BRANCH_RETRY_SLEEPS + ): @@ -1042,2 +1037,5 @@ def compute_config_parquet_and_info_response( - refs = retry(on=[requests.exceptions.ConnectionError], sleeps=[1, 1, 1, 10, 10])(hf_api.list_repo_refs)( - repo_id=dataset, repo_type=DATASET_TYPE + create_branch( + dataset=dataset, + target_revision=target_revision, + hf_api=hf_api, + committer_hf_api=committer_hf_api, @@ -1045,9 +1043 @@ def compute_config_parquet_and_info_response( - if all(ref.ref != target_revision for ref in refs.converts): - initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id - committer_hf_api.create_branch( - repo_id=dataset, - branch=target_revision, - repo_type=DATASET_TYPE, - revision=initial_commit, - exist_ok=True, - ) + diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py index 29a34f9a..55de2fe3 100644 --- a/services/worker/src/worker/job_runners/config/parquet_metadata.py +++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py @@ -8 +7,0 @@ from typing import List, Optional -from datasets.utils.file_utils import get_authentication_headers_for_url @@ -19 +18 @@ from libcommon.storage import StrPath -from libcommon.utils import JobInfo +from libcommon.utils import JobInfo, SplitHubFile @@ -28 +26,0 @@ from worker.dtos import ( - ParquetFileItem, @@ -32,5 +30 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner - - -def get_parquet_file(url: str, fs: HTTPFileSystem, hf_token: Optional[str]) -> ParquetFile: - headers = get_authentication_headers_for_url(url, use_auth_token=hf_token) - return ParquetFile(fs.open(url, headers=headers)) +from worker.utils import get_parquet_file @@ -73 +67 @@ def compute_parquet_metadata_response( - parquet_file_items: List[ParquetFileItem] = [ + parquet_file_items: List[SplitHubFile] = [ diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index 9ba4eddc..000388ff 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -14,0 +15 @@ from libcommon.simple_cache import ( +from libcommon.utils import SplitHubFile @@ -20 +20,0 @@ from worker.dtos import ( - ParquetFileItem, @@ -49 +49 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - parquet_files: list[ParquetFileItem] = [] + parquet_files: list[SplitHubFile] = [] diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py new file mode 100644 index 00000000..b715de60 --- /dev/null +++ b/services/worker/src/worker/job_runners/split/duckdb_index.py @@ -0,0 +1,265 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +import logging +from pathlib import Path +from typing import List, Optional, Set + +import duckdb +from huggingface_hub._commit_api import ( + CommitOperation, + CommitOperationAdd, + CommitOperationDelete, +) +from huggingface_hub.hf_api import HfApi +from huggingface_hub.utils._errors import RepositoryNotFoundError +from libcommon.config import DuckDbIndexConfig +from libcommon.constants import PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION +from libcommon.exceptions import ( + CacheDirectoryNotInitializedError, + DatasetNotFoundError, + DuckDBIndexFileNotFoundError, + LockedDatasetTimeoutError, + NoIndexableColumnsError, + ParquetResponseEmptyError, + PreviousStepFormatError, + SplitNotFoundError, + SplitWithTooBigParquetError, +) +from libcommon.processing_graph import ProcessingStep +from libcommon.queue import lock +from libcommon.simple_cache import get_previous_step_or_raise +from libcommon.storage import StrPath +from libcommon.utils import JobInfo, SplitHubFile + +from worker.config import AppConfig +from worker.dtos import CompleteJobResult +from worker.job_runners.split.split_job_runner import SplitJobRunnerWithCache +from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url + +DATASET_TYPE = "dataset" +STRING_FEATURE_DTYPE = "string" +VALUE_FEATURE_TYPE = "Value" +DUCKDB_DEFAULT_INDEX_FILENAME = "index.duckdb" +CREATE_SEQUENCE_COMMAND = "CREATE OR REPLACE SEQUENCE serial START 1;" +CREATE_INDEX_COMMAND = "PRAGMA create_fts_index('data', '__hf_index_id', '*', overwrite=1);" +CREATE_TABLE_COMMAND = "CREATE OR REPLACE TABLE data AS SELECT nextval('serial') AS __hf_index_id, {columns} FROM" +INSTALL_EXTENSION_COMMAND = "INSTALL '{extension}';" +LOAD_EXTENSION_COMMAND = "LOAD '{extension}';" + + +def compute_index_rows( + job_id: str, + dataset: str, + config: str, + split: str, + duckdb_index_file_directory: Path, + target_revision: str, + hf_endpoint: str, + commit_message: str, + url_template: str, + hf_token: Optional[str], + max_parquet_size_bytes: int, + committer_hf_token: Optional[str], +) -> SplitHubFile: + logging.info(f"get split-duckdb-index for dataset={dataset} config={config} split={split}") + + # validate split + split_names_best_response = get_previous_step_or_raise( + kinds=["config-split-names-from-streaming", "config-split-names-from-info"], dataset=dataset, config=config + ) + try: + splits_content = split_names_best_response.response["content"]["splits"] + except Exception as e: + raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e + + if split not in [split_item["split"] for split_item in splits_content]: + raise SplitNotFoundError(f"The split '{split}' does not exist for the config '{config}' of the dataset.") + + # get parquet urls and dataset_info + config_parquet_and_info_step = "config-parquet-and-info" + parquet_and_info_best_response = get_previous_step_or_raise( + kinds=[config_parquet_and_info_step], + dataset=dataset, + config=config, + ) + content_parquet_and_info = parquet_and_info_best_response.response["content"] + try: + split_parquet_files = [ + parquet_file + for parquet_file in content_parquet_and_info["parquet_files"] + if parquet_file["config"] == config and parquet_file["split"] == split + ] + + split_parquets_size = sum(parquet_file["size"] for parquet_file in split_parquet_files) + + if split_parquets_size > max_parquet_size_bytes: + raise SplitWithTooBigParquetError( + f"The indexing is limited to split parquets under {max_parquet_size_bytes} bytes. " + f"Current size of sum of split parquets is {split_parquets_size} bytes." + ) + + parquet_urls = [parquet_file["url"] for parquet_file in split_parquet_files] + + if not parquet_urls: + raise ParquetResponseEmptyError("No parquet files found.") + + # get the features + features = content_parquet_and_info["dataset_info"]["features"] + column_names = ",".join(list(features.keys())) + + # look for string columns + string_columns = [ + column + for column, feature in features.items() + if "dtype" in feature + and "_type" in feature + and feature["dtype"] == STRING_FEATURE_DTYPE + and feature["_type"] == VALUE_FEATURE_TYPE + ] + if not string_columns: + raise NoIndexableColumnsError("No string columns available to index.") + + except KeyError as e: + raise PreviousStepFormatError( + f"Previous step '{config_parquet_and_info_step}' did not return the expected content.", e + ) from e + + # configure duckdb extensions + duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="httpfs")) + duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="httpfs")) + duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="fts")) + duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="fts")) + + # index all columns + db_path = duckdb_index_file_directory.resolve() / DUCKDB_DEFAULT_INDEX_FILENAME + + con = duckdb.connect(str(db_path.resolve())) + logging.debug(CREATE_SEQUENCE_COMMAND) + con.sql(CREATE_SEQUENCE_COMMAND) + + create_command_sql = f"{CREATE_TABLE_COMMAND.format(columns=column_names)} read_parquet({parquet_urls});" + logging.debug(create_command_sql) + con.sql(create_command_sql) + + # TODO: by default, 'porter' stemmer is being used, use a specific one by dataset language in the future + # see https://duckdb.org/docs/extensions/full_text_search.html for more details about 'stemmer' parameter + logging.debug(CREATE_INDEX_COMMAND) + con.sql(CREATE_INDEX_COMMAND) + con.close() + + hf_api = HfApi(endpoint=hf_endpoint, token=hf_token) + committer_hf_api = HfApi(endpoint=hf_endpoint, token=committer_hf_token) + index_file_location = f"{config}/{split}/{DUCKDB_DEFAULT_INDEX_FILENAME}" + + try: + with lock.git_branch( + dataset=dataset, branch=target_revision, owner=job_id, sleeps=LOCK_GIT_BRANCH_RETRY_SLEEPS + ): + create_branch( + dataset=dataset, + target_revision=target_revision, + hf_api=hf_api, + committer_hf_api=committer_hf_api, + ) + + target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False) + all_repo_files: Set[str] = {f.rfilename for f in target_dataset_info.siblings} + delete_operations: List[CommitOperation] = [] + if index_file_location in all_repo_files: + delete_operations.append(CommitOperationDelete(path_in_repo=index_file_location)) + + # send the files to the target revision + add_operations: List[CommitOperation] = [ + CommitOperationAdd(path_in_repo=index_file_location, path_or_fileobj=db_path.resolve()) + ] + + committer_hf_api.create_commit( + repo_id=dataset, + repo_type=DATASET_TYPE, + revision=target_revision, + operations=delete_operations + add_operations, + commit_message=commit_message, + parent_commit=target_dataset_info.sha, + ) + + # call the API again to get the index file + target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True) + except TimeoutError as err: + raise LockedDatasetTimeoutError("the dataset is currently locked, please try again later.") from err + except RepositoryNotFoundError as err: + raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err + + repo_files = [ + repo_file for repo_file in target_dataset_info.siblings if repo_file.rfilename == index_file_location + ] + + if not repo_files or len(repo_files) != 1: + logging.warning(f"Found {len(repo_files)} index files, should be only 1") + raise DuckDBIndexFileNotFoundError("No index file was found") + + repo_file = repo_files[0] + if repo_file.size is None: + raise ValueError(f"Cannot get size of {repo_file.rfilename}") + + return SplitHubFile( + dataset=dataset, + config=config, + split=split, + url=hf_hub_url( + repo_id=dataset, + filename=repo_file.rfilename, + hf_endpoint=hf_endpoint, + revision=target_revision, + url_template=url_template, + ), + filename=Path(repo_file.rfilename).name, + size=repo_file.size, + ) + + +class SplitDuckDbIndexJobRunner(SplitJobRunnerWithCache): + duckdb_index_config: DuckDbIndexConfig + + def __init__( + self, + job_info: JobInfo, + app_config: AppConfig, + processing_step: ProcessingStep, + duckdb_index_cache_directory: StrPath, + ) -> None: + super().__init__( + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + cache_directory=Path(duckdb_index_cache_directory), + ) + self.duckdb_index_config = app_config.duckdb_index + + @staticmethod + def get_job_type() -> str: + return "split-duckdb-index" + + @staticmethod + def get_job_runner_version() -> int: + return PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION + + def compute(self) -> CompleteJobResult: + if self.cache_subdirectory is None: + raise CacheDirectoryNotInitializedError("Cache directory has not been initialized.") + return CompleteJobResult( + compute_index_rows( + job_id=self.job_info["job_id"], + dataset=self.dataset, + config=self.config, + split=self.split, + duckdb_index_file_directory=self.cache_subdirectory, + hf_token=self.app_config.common.hf_token, + url_template=self.duckdb_index_config.url_template, + commit_message=self.duckdb_index_config.commit_message, + committer_hf_token=self.duckdb_index_config.committer_hf_token, + hf_endpoint=self.app_config.common.hf_endpoint, + target_revision=self.duckdb_index_config.target_revision, + max_parquet_size_bytes=self.duckdb_index_config.max_parquet_size_bytes, + ) + ) diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py index da297ccc..4d207280 100644 --- a/services/worker/src/worker/main.py +++ b/services/worker/src/worker/main.py @@ -9 +9,5 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.storage import init_assets_dir, init_parquet_metadata_dir +from libcommon.storage import ( + init_assets_dir, + init_duckdb_index_cache_dir, + init_parquet_metadata_dir, +) @@ -29,0 +34 @@ if __name__ == "__main__": + duckdb_index_cache_directory = init_duckdb_index_cache_dir(directory=app_config.duckdb_index.storage_directory) @@ -56,0 +62 @@ if __name__ == "__main__": + duckdb_index_cache_directory=duckdb_index_cache_directory, diff --git a/services/worker/src/worker/start_worker_loop.py b/services/worker/src/worker/start_worker_loop.py index 5aa4246d..d5e69a82 100644 --- a/services/worker/src/worker/start_worker_loop.py +++ b/services/worker/src/worker/start_worker_loop.py @@ -9 +9,5 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource -from libcommon.storage import init_assets_dir, init_parquet_metadata_dir +from libcommon.storage import ( + init_assets_dir, + init_duckdb_index_cache_dir, + init_parquet_metadata_dir, +) @@ -28,0 +33 @@ if __name__ == "__main__": + duckdb_index_cache_directory = init_duckdb_index_cache_dir(directory=app_config.duckdb_index.storage_directory) @@ -55,0 +61 @@ if __name__ == "__main__": + duckdb_index_cache_directory=duckdb_index_cache_directory, diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index fab6b031..c3b985d8 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -20,0 +21 @@ from typing import ( +from urllib.parse import quote @@ -22,0 +24 @@ import PIL +import requests @@ -31 +33,9 @@ from datasets import ( -from libcommon.exceptions import NormalRowsError, StreamingRowsError +from datasets.utils.file_utils import get_authentication_headers_for_url +from fsspec.implementations.http import HTTPFileSystem +from huggingface_hub.hf_api import HfApi +from huggingface_hub.utils._errors import RepositoryNotFoundError +from libcommon.exceptions import ( + DatasetNotFoundError, + NormalRowsError, + StreamingRowsError, +) @@ -32,0 +43 @@ from libcommon.utils import orjson_dumps +from pyarrow.parquet import ParquetFile @@ -312,0 +324,31 @@ def get_rows_or_raise( + + +# TODO: use huggingface_hub's hf_hub_url after +# https://github.com/huggingface/huggingface_hub/issues/1082 +def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url_template: str) -> str: + return (hf_endpoint + url_template) % (repo_id, quote(revision, safe=""), filename) + + +def get_parquet_file(url: str, fs: HTTPFileSystem, hf_token: Optional[str]) -> ParquetFile: + headers = get_authentication_headers_for_url(url, use_auth_token=hf_token) + return ParquetFile(fs.open(url, headers=headers)) + + +DATASET_TYPE = "dataset" + +LIST_REPO_REFS_RETRY_SLEEPS = [1, 1, 1, 10, 10] +LOCK_GIT_BRANCH_RETRY_SLEEPS = [1, 1, 1, 1, 1, 10, 10, 10, 10, 100] * 3 + + +def create_branch(dataset: str, target_revision: str, hf_api: HfApi, committer_hf_api: HfApi) -> None: + try: + refs = retry(on=[requests.exceptions.ConnectionError], sleeps=LIST_REPO_REFS_RETRY_SLEEPS)( + hf_api.list_repo_refs + )(repo_id=dataset, repo_type=DATASET_TYPE) + if all(ref.ref != target_revision for ref in refs.converts): + initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id + committer_hf_api.create_branch( + repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit, exist_ok=True + ) + except RepositoryNotFoundError as err: + raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 851a99c6..fab26487 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -11 +11,6 @@ from libcommon.simple_cache import _clean_cache_database -from libcommon.storage import StrPath, init_assets_dir, init_parquet_metadata_dir +from libcommon.storage import ( + StrPath, + init_assets_dir, + init_duckdb_index_cache_dir, + init_parquet_metadata_dir, +) @@ -67,0 +73 @@ def set_env_vars( + mp.setenv("DUCKDB_INDEX_COMMITTER_HF_TOKEN", CI_PARQUET_CONVERTER_APP_TOKEN) @@ -121,0 +128,5 @@ def parquet_metadata_directory(app_config: AppConfig) -> StrPath: +@fixture +def duckdb_index_cache_directory(app_config: AppConfig) -> StrPath: + return init_duckdb_index_cache_dir(app_config.duckdb_index.storage_directory) + + diff --git a/services/worker/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py index 18edaad1..6e987e20 100644 --- a/services/worker/tests/fixtures/datasets.py +++ b/services/worker/tests/fixtures/datasets.py @@ -145,0 +146,17 @@ def datasets() -> Mapping[str, Dataset]: + "duckdb_index": Dataset.from_pandas( + pd.DataFrame( + { + "text": [ + ( + "Grand Moff Tarkin and Lord Vader are interrupted in their discussion by the buzz of the" + " comlink" + ), + "There goes another one.", + "Vader turns round and round in circles as his ship spins into space.", + "We count thirty Rebel ships, Lord Vader.", + "The wingman spots the pirateship coming at him and warns the Dark Lord", + ] + }, + dtype=pd.StringDtype(storage="python"), + ) + ), diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index f34ca587..11e9d9a0 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -283,0 +284,7 @@ def hub_public_spawning_opt_in_out(datasets: Mapping[str, Dataset]) -> Iterator[ [email protected](scope="session") +def hub_public_duckdb_index(datasets: Mapping[str, Dataset]) -> Iterator[str]: + repo_id = create_hub_dataset_repo(prefix="duckdb_index", dataset=datasets["duckdb_index"]) + yield repo_id + delete_hub_dataset_repo(repo_id=repo_id) + + @@ -766,0 +774,13 @@ def hub_reponses_spawning_opt_in_out(hub_public_spawning_opt_in_out: str) -> Hub + + [email protected] +def hub_responses_duckdb_index(hub_public_duckdb_index: str) -> HubDatasetTest: + return { + "name": hub_public_duckdb_index, + "config_names_response": create_config_names_response(hub_public_duckdb_index), + "splits_response": create_splits_response(hub_public_duckdb_index), + "first_rows_response": create_first_rows_response(hub_public_duckdb_index, TEXT_cols, TEXT_rows), + "parquet_and_info_response": create_parquet_and_info_response( + dataset=hub_public_duckdb_index, data_type="csv" + ), + } diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index 031fee96..e314000a 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -12 +12 @@ from libcommon.simple_cache import CachedArtifactError, upsert_response -from libcommon.utils import Priority +from libcommon.utils import Priority, SplitHubFile @@ -15,5 +15 @@ from worker.config import AppConfig -from worker.dtos import ( - ConfigParquetAndInfoResponse, - ConfigParquetResponse, - ParquetFileItem, -) +from worker.dtos import ConfigParquetAndInfoResponse, ConfigParquetResponse @@ -81 +77 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( @@ -84 +80 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( @@ -93 +89 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( @@ -96 +92 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py index 613d616a..96f1cfbc 100644 --- a/services/worker/tests/job_runners/config/test_parquet_metadata.py +++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py @@ -22 +22 @@ from libcommon.storage import StrPath -from libcommon.utils import Priority +from libcommon.utils import Priority, SplitHubFile @@ -28 +27,0 @@ from worker.dtos import ( - ParquetFileItem, @@ -100 +99 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( @@ -103 +102 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index 61cfbfbc..8f63b188 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -12 +12 @@ from libcommon.simple_cache import CachedArtifactError, upsert_response -from libcommon.utils import Priority +from libcommon.utils import Priority, SplitHubFile @@ -15 +15 @@ from worker.config import AppConfig -from worker.dtos import ConfigParquetResponse, DatasetParquetResponse, ParquetFileItem +from worker.dtos import ConfigParquetResponse, DatasetParquetResponse @@ -92 +92 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( @@ -110 +110 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( @@ -125 +125 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( @@ -128 +128 @@ def get_job_runner( - ParquetFileItem( + SplitHubFile( diff --git a/services/worker/tests/job_runners/split/test_duckdb_index.py b/services/worker/tests/job_runners/split/test_duckdb_index.py new file mode 100644 index 00000000..a8fe40cf --- /dev/null +++ b/services/worker/tests/job_runners/split/test_duckdb_index.py @@ -0,0 +1,242 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +import os +from dataclasses import replace +from http import HTTPStatus +from typing import Callable, Optional + +import duckdb +import pytest +import requests +from libcommon.processing_graph import ProcessingGraph +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.simple_cache import upsert_response +from libcommon.storage import StrPath +from libcommon.utils import Priority + +from worker.config import AppConfig +from worker.job_runners.config.parquet_and_info import ConfigParquetAndInfoJobRunner +from worker.job_runners.split.duckdb_index import SplitDuckDbIndexJobRunner +from worker.resources import LibrariesResource + +from ...fixtures.hub import HubDatasetTest + +GetJobRunner = Callable[[str, str, str, AppConfig], SplitDuckDbIndexJobRunner] + +GetParquetJobRunner = Callable[[str, str, AppConfig], ConfigParquetAndInfoJobRunner] + + [email protected] +def get_job_runner( + duckdb_index_cache_directory: StrPath, + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + split: str, + app_config: AppConfig, + ) -> SplitDuckDbIndexJobRunner: + processing_step_name = SplitDuckDbIndexJobRunner.get_job_type() + processing_graph = ProcessingGraph( + { + "dataset-step": {"input_type": "dataset"}, + "config-parquet": { + "input_type": "config", + "triggered_by": "dataset-step", + "provides_config_parquet": True, + }, + "config-split-names-from-streaming": { + "input_type": "config", + "triggered_by": "dataset-step", + }, + processing_step_name: { + "input_type": "dataset", + "job_runner_version": SplitDuckDbIndexJobRunner.get_job_runner_version(), + "triggered_by": ["config-parquet", "config-split-names-from-streaming"], + }, + } + ) + return SplitDuckDbIndexJobRunner( + job_info={ + "type": SplitDuckDbIndexJobRunner.get_job_type(), + "params": { + "dataset": dataset, + "revision": "revision", + "config": config, + "split": split, + }, + "job_id": "job_id", + "priority": Priority.NORMAL, + }, + app_config=app_config, + processing_step=processing_graph.get_processing_step(processing_step_name), + duckdb_index_cache_directory=duckdb_index_cache_directory, + ) + + return _get_job_runner + + [email protected] +def get_parquet_job_runner( + libraries_resource: LibrariesResource, + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetParquetJobRunner: + def _get_job_runner( + dataset: str, + config: str, + app_config: AppConfig, + ) -> ConfigParquetAndInfoJobRunner: + processing_step_name = ConfigParquetAndInfoJobRunner.get_job_type() + processing_graph = ProcessingGraph( + { + "dataset-level": {"input_type": "dataset"}, + processing_step_name: { + "input_type": "config", + "job_runner_version": ConfigParquetAndInfoJobRunner.get_job_runner_version(), + "triggered_by": "dataset-level", + }, + } + ) + return ConfigParquetAndInfoJobRunner( + job_info={ + "type": ConfigParquetAndInfoJobRunner.get_job_type(), + "params": { + "dataset": dataset, + "revision": "revision", + "config": config, + "split": None, + }, + "job_id": "job_id", + "priority": Priority.NORMAL, + }, + app_config=app_config, + processing_step=processing_graph.get_processing_step(processing_step_name), + hf_datasets_cache=libraries_resource.hf_datasets_cache, + ) + + return _get_job_runner + + [email protected]( + "hub_dataset_name,max_parquet_size_bytes,expected_error_code", + [ + ("duckdb_index", None, None), + ("duckdb_index", 1_000, "SplitWithTooBigParquetError"), # parquet size is 2812 + ("public", None, "NoIndexableColumnsError"), # dataset does not have string columns to index + ], +) +def test_compute( + get_parquet_job_runner: GetParquetJobRunner, + get_job_runner: GetJobRunner, + app_config: AppConfig, + hub_responses_public: HubDatasetTest, + hub_responses_duckdb_index: HubDatasetTest, + hub_dataset_name: str, + max_parquet_size_bytes: Optional[int], + expected_error_code: str, +) -> None: + hub_datasets = {"public": hub_responses_public, "duckdb_index": hub_responses_duckdb_index} + dataset = hub_datasets[hub_dataset_name]["name"] + config_names = hub_datasets[hub_dataset_name]["config_names_response"] + config = hub_datasets[hub_dataset_name]["config_names_response"]["config_names"][0]["config"] + splits_response = hub_datasets[hub_dataset_name]["splits_response"] + split = "train" + + upsert_response( + "dataset-config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=config_names, + ) + + upsert_response( + "config-split-names-from-streaming", + dataset=dataset, + config=config, + http_status=HTTPStatus.OK, + content=splits_response, + ) + + app_config = ( + app_config + if max_parquet_size_bytes is None + else replace( + app_config, duckdb_index=replace(app_config.duckdb_index, max_parquet_size_bytes=max_parquet_size_bytes) + ) + ) + + parquet_job_runner = get_parquet_job_runner(dataset, config, app_config) + parquet_response = parquet_job_runner.compute() + config_parquet = parquet_response.content + + # simulate more than one parquet file to index + extra_parquet_file = config_parquet["parquet_files"][0] + config_parquet["parquet_files"].append(extra_parquet_file) + + upsert_response( + "config-parquet-and-info", + dataset=dataset, + config=config, + http_status=HTTPStatus.OK, + content=config_parquet, + ) + + assert parquet_response + job_runner = get_job_runner(dataset, config, split, app_config) + job_runner.pre_compute() + + if expected_error_code: + with pytest.raises(Exception) as e: + job_runner.compute() + assert e.typename == expected_error_code + else: + job_runner.pre_compute() + response = job_runner.compute() + assert response + content = response.content + url = content["url"] + file_name = content["filename"] + assert url is not None + assert file_name is not None + job_runner.post_compute() + + # download locally duckdb index file + duckdb_file = requests.get(url) + with open(file_name, "wb") as f: + f.write(duckdb_file.content) + + duckdb.execute("INSTALL 'fts';") + duckdb.execute("LOAD 'fts';") + con = duckdb.connect(file_name) + + # validate number of inserted records + record_count = con.sql("SELECT COUNT(*) FROM data;").fetchall() + assert record_count is not None + assert isinstance(record_count, list) + assert record_count[0] == (10,) # dataset has 5 rows but since parquet file was duplicate it is 10 + + # perform a search to validate fts feature + query = "Lord Vader" + result = con.execute( + "SELECT text FROM data WHERE fts_main_data.match_bm25(__hf_index_id, ?) IS NOT NULL;", + [query], + ) + rows = result.df() + assert rows is not None + assert (rows["text"].eq("Vader turns round and round in circles as his ship spins into space.")).any() + assert (rows["text"].eq("The wingman spots the pirateship coming at him and warns the Dark Lord")).any() + assert (rows["text"].eq("We count thirty Rebel ships, Lord Vader.")).any() + assert ( + rows["text"].eq( + "Grand Moff Tarkin and Lord Vader are interrupted in their discussion by the buzz of the comlink" + ) + ).any() + assert not (rows["text"].eq("There goes another one.")).any() + + con.close() + os.remove(file_name) + job_runner.post_compute() diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py index 87f44cd5..d83cbf26 100644 --- a/services/worker/tests/test_executor.py +++ b/services/worker/tests/test_executor.py @@ -201,0 +202 @@ def job_runner_factory( + duckdb_index_cache_directory: StrPath, @@ -209,0 +211 @@ def job_runner_factory( + duckdb_index_cache_directory=duckdb_index_cache_directory, diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py index 3ed3b0e7..e10bc8c0 100644 --- a/services/worker/tests/test_job_runner_factory.py +++ b/services/worker/tests/test_job_runner_factory.py @@ -41,0 +42 @@ def test_create_job_runner( + duckdb_index_cache_directory: StrPath, @@ -50,0 +52 @@ def test_create_job_runner( + duckdb_index_cache_directory=duckdb_index_cache_directory, diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml index d6b2bd92..37b1c87d 100644 --- a/tools/docker-compose-datasets-server.yml +++ b/tools/docker-compose-datasets-server.yml @@ -93,0 +94 @@ services: + - duckdb-index:${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index}:rw @@ -114,0 +116,6 @@ services: + DUCKDB_INDEX_STORAGE_DIRECTORY: ${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index} + DUCKDB_INDEX_COMMIT_MESSAGE: ${DUCKDB_INDEX_COMMIT_MESSAGE-Update duckdb index file} + DUCKDB_INDEX_COMMITTER_HF_TOKEN: ${DUCKDB_INDEX_COMMITTER_HF_TOKEN-} + DUCKDB_INDEX_TARGET_REVISION: ${DUCKDB_INDEX_TARGET_REVISION-refs/convert/parquet} + DUCKDB_INDEX_URL_TEMPLATE: ${DUCKDB_INDEX_URL_TEMPLATE-/datasets/%s/resolve/%s/%s} + DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES: ${DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES-100_000_000} @@ -147,0 +155 @@ volumes: + duckdb-index: diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index c50d781f..233e90f2 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -97,0 +98 @@ services: + - duckdb-index:${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index}:rw @@ -118,0 +120,6 @@ services: + DUCKDB_INDEX_STORAGE_DIRECTORY: ${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index} + DUCKDB_INDEX_COMMIT_MESSAGE: ${DUCKDB_INDEX_COMMIT_MESSAGE-Update duckdb index files} + DUCKDB_INDEX_COMMITTER_HF_TOKEN: ${DUCKDB_INDEX_COMMITTER_HF_TOKEN-} + DUCKDB_INDEX_TARGET_REVISION: ${DUCKDB_INDEX_TARGET_REVISION-refs/convert/parquet} + DUCKDB_INDEX_URL_TEMPLATE: ${DUCKDB_INDEX_URL_TEMPLATE-/datasets/%s/resolve/%s/%s} + DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES: ${DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES-100_000_000} @@ -149,0 +157 @@ volumes: + duckdb-index:
04f53d629c65a6d280b4ea79daf87c5822fe5785
Sylvain Lesage
2023-06-27T13:04:00
refactor: 💡 remove dead code (#1435)
diff --git a/services/worker/src/worker/job_manager.py b/services/worker/src/worker/job_manager.py index d4807508..5a5a4f63 100644 --- a/services/worker/src/worker/job_manager.py +++ b/services/worker/src/worker/job_manager.py @@ -30,3 +29,0 @@ from worker.job_runner import JobRunner -# List of error codes that should trigger a retry. -ERROR_CODES_TO_RETRY: list[str] = ["ClientConnectionError"] -
1928052b47b1619a516f96ab5c1df781df4af36a
Quentin Lhoest
2023-06-27T12:12:19
unqote path and revision (#1434)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index c5c879be..8df7ea7a 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -11 +11 @@ from typing import Any, List, Optional, Set, Tuple -from urllib.parse import quote +from urllib.parse import quote, unquote @@ -586,0 +587,2 @@ def copy_parquet_files(builder: DatasetBuilder) -> List[CommitOperationCopy]: + src_revision = unquote(src_revision) + src_path_in_repo = unquote(src_path_in_repo)
64e84e792c755525d3a6c48425567617931e001c
Quentin Lhoest
2023-06-26T17:23:12
Fix auth in rows again (#1430)
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py index f95a63f8..0f8ca107 100644 --- a/libs/libcommon/src/libcommon/parquet_utils.py +++ b/libs/libcommon/src/libcommon/parquet_utils.py @@ -265,0 +266 @@ class ParquetIndexWithMetadata: + **self.httpfs.kwargs, diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 177865f5..d0689d8e 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -274 +274 @@ def create_rows_endpoint( - httpfs=HTTPFileSystem(storage_options={"headers": {"authorization": f"Bearer {hf_token}"}}), + httpfs=HTTPFileSystem(headers={"authorization": f"Bearer {hf_token}"}), diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py index 56a3815d..613d616a 100644 --- a/services/worker/tests/job_runners/config/test_parquet_metadata.py +++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py @@ -7 +7 @@ from pathlib import Path -from typing import Any, Callable +from typing import Any, Callable, Mapping, Optional @@ -13 +13,3 @@ import pytest -from fsspec.implementations.http import HTTPFileSystem +from datasets import Dataset, Features +from fsspec.implementations.http import HTTPFile, HTTPFileSystem +from huggingface_hub import hf_hub_url @@ -14,0 +17 @@ from libcommon.exceptions import PreviousStepFormatError +from libcommon.parquet_utils import ParquetIndexWithMetadata @@ -29,0 +33,3 @@ from worker.job_runners.config.parquet_metadata import ConfigParquetMetadataJobR +from ...constants import CI_USER_TOKEN +from ...fixtures.hub import hf_api + @@ -188,0 +195,71 @@ def test_compute( + + +class AuthenticatedHTTPFile(HTTPFile): # type: ignore + last_url: Optional[str] = None + + def __init__( # type: ignore + self, + fs, + url, + session=None, + block_size=None, + mode="rb", + cache_type="bytes", + cache_options=None, + size=None, + loop=None, + asynchronous=False, + **kwargs, + ) -> None: + super().__init__( + fs, + url, + session=session, + block_size=block_size, + mode=mode, + cache_type=cache_type, + cache_options=cache_options, + size=size, + loop=loop, + asynchronous=asynchronous, + **kwargs, + ) + assert self.kwargs == {"headers": {"authorization": f"Bearer {CI_USER_TOKEN}"}} + AuthenticatedHTTPFile.last_url = url + + +def test_ParquetIndexWithMetadata_query( + datasets: Mapping[str, Dataset], hub_public_big: str, tmp_path_factory: pytest.TempPathFactory +) -> None: + ds = datasets["big"] + httpfs = HTTPFileSystem(headers={"authorization": f"Bearer {CI_USER_TOKEN}"}) + filename = next( + iter( + repo_file + for repo_file in hf_api.list_repo_files(repo_id=hub_public_big, repo_type="dataset") + if repo_file.endswith(".parquet") + ) + ) + url = hf_hub_url(repo_id=hub_public_big, filename=filename, repo_type="dataset") + metadata_path = str(tmp_path_factory.mktemp("test_ParquetIndexWithMetadata_query") / "metadata.parquet") + with httpfs.open(url) as f: + num_bytes = f.size + pf = pq.ParquetFile(url, filesystem=httpfs) + num_rows = pf.metadata.num_rows + features = Features.from_arrow_schema(pf.schema_arrow) + pf.metadata.write_metadata_file(metadata_path) + index = ParquetIndexWithMetadata( + features=features, + supported_columns=list(features), + unsupported_columns=[], + parquet_files_urls=[url], + metadata_paths=[metadata_path], + num_rows=[num_rows], + num_bytes=[num_bytes], + httpfs=httpfs, + hf_token=CI_USER_TOKEN, + ) + with patch("libcommon.parquet_utils.HTTPFile", AuthenticatedHTTPFile): + out = index.query(offset=0, length=2).to_pydict() + assert out == ds[:2] + assert AuthenticatedHTTPFile.last_url == url
a249894dd692ea6da078912aed9c462d3a99b51c
Sylvain Lesage
2023-06-26T16:17:02
feat: 🎸 don't insert a new lock when releasing (#1432)
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index e3b7df0e..6aa66820 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -296 +295,0 @@ class lock(contextlib.AbstractContextManager["lock"]): - upsert=True,
a0479f8a8ac7868c94344370a7965575872c05e3
Sylvain Lesage
2023-06-26T15:56:46
fix: 🐛 remove the "required" constraint on created_at in Lock (#1431)
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index e7c92ebd..e3b7df0e 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -239 +239 @@ class Lock(Document): - created_at = DateTimeField(required=True) + created_at = DateTimeField()
97cc53ff804ca606e05744eb699ae240b5842eff
Sylvain Lesage
2023-06-26T15:37:26
Ensure only one job is started for the same unicity_id (#1420)
diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py index eb58797e..7f8cf4bd 100644 --- a/jobs/cache_maintenance/tests/test_collect_metrics.py +++ b/jobs/cache_maintenance/tests/test_collect_metrics.py @@ -27 +27 @@ def test_collect_metrics() -> None: - queue.upsert_job( + queue.add_job( diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 4b6dba13..6d18f4da 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -46,0 +47,3 @@ from mongodb_migration.migrations._20230516101600_queue_delete_index_without_rev +from mongodb_migration.migrations._20230622131500_lock_add_owner import ( + MigrationAddOwnerToQueueLock, +) @@ -231,0 +235,3 @@ class MigrationsCollector: + MigrationAddOwnerToQueueLock( + version="20230622131800", description="add 'owner' field copying the job_id value" + ), diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230622131500_lock_add_owner.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230622131500_lock_add_owner.py new file mode 100644 index 00000000..e539be50 --- /dev/null +++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230622131500_lock_add_owner.py @@ -0,0 +1,32 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +import logging + +from libcommon.constants import QUEUE_COLLECTION_LOCKS, QUEUE_MONGOENGINE_ALIAS +from libcommon.queue import Lock +from mongoengine.connection import get_db + +from mongodb_migration.check import check_documents +from mongodb_migration.migration import Migration + + +# connection already occurred in the main.py (caveat: we use globals) +class MigrationAddOwnerToQueueLock(Migration): + def up(self) -> None: + # See https://docs.mongoengine.org/guide/migration.html#example-1-addition-of-a-field + logging.info("If missing, add the owner field with the same value as the field job_id to the locks") + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_LOCKS].update_many( + {"owner": {"$exists": False}}, [{"$set": {"owner": "$job_id"}}] # type: ignore + ) + + def down(self) -> None: + logging.info("Remove the owner field from all the locks") + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_LOCKS].update_many({}, {"$unset": {"owner": ""}}) + + def validate(self) -> None: + logging.info("Ensure that a random selection of locks have the 'owner' field") + + check_documents(DocCls=Lock, sample_size=10) diff --git a/jobs/mongodb_migration/tests/migrations/test_20230622131500_lock_add_owner.py b/jobs/mongodb_migration/tests/migrations/test_20230622131500_lock_add_owner.py new file mode 100644 index 00000000..f97639a9 --- /dev/null +++ b/jobs/mongodb_migration/tests/migrations/test_20230622131500_lock_add_owner.py @@ -0,0 +1,64 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from typing import Optional + +from libcommon.constants import QUEUE_COLLECTION_LOCKS, QUEUE_MONGOENGINE_ALIAS +from libcommon.resources import MongoResource +from mongoengine.connection import get_db + +from mongodb_migration.migrations._20230622131500_lock_add_owner import ( + MigrationAddOwnerToQueueLock, +) + + +def assert_owner(key: str, owner: Optional[str]) -> None: + db = get_db(QUEUE_MONGOENGINE_ALIAS) + entry = db[QUEUE_COLLECTION_LOCKS].find_one({"key": key}) + assert entry is not None + if owner is None: + assert "owner" not in entry or entry["owner"] is None + else: + assert entry["owner"] == owner + + +def test_lock_add_owner(mongo_host: str) -> None: + with MongoResource(database="test_lock_add_owner", host=mongo_host, mongoengine_alias="queue"): + db = get_db(QUEUE_MONGOENGINE_ALIAS) + db[QUEUE_COLLECTION_LOCKS].insert_many( + [ + { + "key": "key1", + "job_id": "job_id1", + "created_at": "2022-01-01T00:00:00.000000Z", + }, + { + "key": "key2", + "job_id": None, + "created_at": "2022-01-01T00:00:00.000000Z", + }, + { + "key": "key3", + "job_id": "job_id3", + "owner": "owner3", + "created_at": "2022-01-01T00:00:00.000000Z", + }, + ] + ) + + migration = MigrationAddOwnerToQueueLock( + version="20230622131500", + description="add owner field to locks", + ) + migration.up() + + assert_owner("key1", "job_id1") + assert_owner("key2", None) + assert_owner("key3", "owner3") + + migration.down() + assert_owner("key1", None) + assert_owner("key2", None) + assert_owner("key3", None) + + db[QUEUE_COLLECTION_LOCKS].drop() diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 55aea4de..e7c92ebd 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -98,0 +99,12 @@ class JobDoesNotExistError(DoesNotExist): +class AlreadyStartedJobError(Exception): + pass + + +class LockTimeoutError(Exception): + pass + + +class NoWaitingJobError(Exception): + pass + + @@ -222 +234 @@ class Lock(Document): - meta = {"collection": QUEUE_COLLECTION_LOCKS, "db_alias": QUEUE_MONGOENGINE_ALIAS, "indexes": [("key", "job_id")]} + meta = {"collection": QUEUE_COLLECTION_LOCKS, "db_alias": QUEUE_MONGOENGINE_ALIAS, "indexes": [("key", "owner")]} @@ -224 +236,2 @@ class Lock(Document): - job_id = StringField() + owner = StringField() + job_id = StringField() # deprecated @@ -234,2 +247,3 @@ class lock(contextlib.AbstractContextManager["lock"]): - Provides a simple way of inter-worker communication using a MongoDB lock. - A lock is used to indicate another worker of your application that a resource + Provides a simple way of inter-applications communication using a MongoDB lock. + + An example usage is to another worker of your application that a resource @@ -242 +256 @@ class lock(contextlib.AbstractContextManager["lock"]): - with lock(key=key, job_id=job.pk): + with lock(key=key, owner=job.pk): @@ -251 +265 @@ class lock(contextlib.AbstractContextManager["lock"]): - lock(key=key, job_id=job.pk).acquire() + lock(key=key, owner=job.pk).acquire() @@ -259 +273 @@ class lock(contextlib.AbstractContextManager["lock"]): - def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> None: + def __init__(self, key: str, owner: str, sleeps: Sequence[float] = _default_sleeps) -> None: @@ -261 +275 @@ class lock(contextlib.AbstractContextManager["lock"]): - self.job_id = job_id + self.owner = owner @@ -267 +281 @@ class lock(contextlib.AbstractContextManager["lock"]): - Lock.objects(key=self.key, job_id__in=[None, self.job_id]).update( + Lock.objects(key=self.key, owner__in=[None, self.owner]).update( @@ -271 +285 @@ class lock(contextlib.AbstractContextManager["lock"]): - job_id=self.job_id, + owner=self.owner, @@ -276 +290 @@ class lock(contextlib.AbstractContextManager["lock"]): - logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'") + logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for owner='{self.owner}'") @@ -281 +295 @@ class lock(contextlib.AbstractContextManager["lock"]): - Lock.objects(key=self.key, job_id=self.job_id).update( + Lock.objects(key=self.key, owner=self.owner).update( @@ -285 +299 @@ class lock(contextlib.AbstractContextManager["lock"]): - job_id=None, + owner=None, @@ -300 +314 @@ class lock(contextlib.AbstractContextManager["lock"]): - def git_branch(cls, dataset: str, branch: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> "lock": + def git_branch(cls, dataset: str, branch: str, owner: str, sleeps: Sequence[float] = _default_sleeps) -> "lock": @@ -307 +321 @@ class lock(contextlib.AbstractContextManager["lock"]): - job_id (`str`): the current job id that holds the lock + owner (`str`): the current job id that holds the lock @@ -311 +325 @@ class lock(contextlib.AbstractContextManager["lock"]): - return cls(key=key, job_id=job_id, sleeps=sleeps) + return cls(key=key, owner=owner, sleeps=sleeps) @@ -330 +344 @@ class Queue: - def _add_job( + def add_job( @@ -341 +355,2 @@ class Queue: - This method should not be called directly. Use `upsert_job` instead. + Note that the same "unicity_id" can have multiple jobs in the waiting state, with the same or different + revisions and or priorities. @@ -366,38 +380,0 @@ class Queue: - def upsert_job( - self, - job_type: str, - dataset: str, - revision: str, - config: Optional[str] = None, - split: Optional[str] = None, - priority: Priority = Priority.NORMAL, - ) -> Job: - """Add, or update, a job to the queue in the waiting state. - - If jobs already exist with the same parameters in the waiting state, they are cancelled and replaced by a new - one. - Note that the new job inherits the highest priority of the previous waiting jobs. - - Args: - job_type (`str`): The type of the job - dataset (`str`): The dataset on which to apply the job. - revision (`str`): The git revision of the dataset. - config (`str`, optional): The config on which to apply the job. - split (`str`, optional): The config on which to apply the job. - priority (`Priority`, optional): The priority of the job. Defaults to Priority.NORMAL. - - Returns: the job - """ - canceled_jobs = self.cancel_jobs( - job_type=job_type, - dataset=dataset, - config=config, - split=split, - statuses_to_cancel=[Status.WAITING], - ) - if any(job["priority"] == Priority.NORMAL for job in canceled_jobs): - priority = Priority.NORMAL - return self._add_job( - job_type=job_type, dataset=dataset, revision=revision, config=config, split=split, priority=priority - ) - @@ -441,39 +417,0 @@ class Queue: - def cancel_jobs( - self, - job_type: str, - dataset: str, - config: Optional[str] = None, - split: Optional[str] = None, - statuses_to_cancel: Optional[List[Status]] = None, - ) -> List[JobDict]: - """Cancel jobs from the queue. - - Note that the jobs for all the revisions are canceled. - - Returns the list of canceled jobs (as JobDict, before they are canceled, to be able to know their previous - status) - - Args: - job_type (`str`): The type of the job - dataset (`str`): The dataset on which to apply the job. - config (`str`, optional): The config on which to apply the job. - split (`str`, optional): The config on which to apply the job. - statuses_to_cancel (`list[Status]`, optional): The list of statuses to cancel. Defaults to - [Status.WAITING, Status.STARTED]. - - Returns: - `list[JobDict]`: The list of canceled jobs - """ - if statuses_to_cancel is None: - statuses_to_cancel = [Status.WAITING, Status.STARTED] - existing = Job.objects( - type=job_type, - dataset=dataset, - config=config, - split=split, - status__in=statuses_to_cancel, - ) - job_dicts = [job.to_dict() for job in existing] - existing.update(finished_at=get_datetime(), status=Status.CANCELLED) - return job_dicts - @@ -540 +478 @@ class Queue: - .only("type", "dataset", "revision", "config", "split", "priority") + .only("type", "dataset", "revision", "config", "split", "priority", "unicity_id") @@ -577 +515 @@ class Queue: - .only("type", "dataset", "revision", "config", "split", "priority") + .only("type", "dataset", "revision", "config", "split", "priority", "unicity_id") @@ -612 +550,15 @@ class Queue: - def _start_job(self, job: Job) -> Job: + def _start_newest_job_and_cancel_others(self, job: Job) -> Job: + """Start a job (the newest one for unicity_id) and cancel the other ones. + + A lock is used to ensure that the job is not started by another worker. + + Args: + job: the job to start + + Returns: + the started job + + Raises: + AlreadyStartedJobError: if a started job already exist for the same unicity_id. + LockTimeoutError: if the lock could not be acquired after 20 retries. + """ @@ -614,2 +566,40 @@ class Queue: - job.update(started_at=get_datetime(), status=Status.STARTED) - return job + RETRIES = 20 + try: + # retry for 2 seconds + with lock(key=job.unicity_id, owner=str(job.pk), sleeps=[0.1] * RETRIES): + # get all the pending jobs for the same unicity_id + waiting_jobs = Job.objects( + unicity_id=job.unicity_id, status__in=[Status.WAITING, Status.STARTED] + ).order_by("-created_at") + datetime = get_datetime() + # raise if any job has already been started for unicity_id + num_started_jobs = waiting_jobs(status=Status.STARTED).count() + if num_started_jobs > 0: + if num_started_jobs > 1: + logging.critical(f"job {job.unicity_id} has been started {num_started_jobs} times. Max is 1.") + raise AlreadyStartedJobError(f"job {job.unicity_id} has been started by another worker") + # get the most recent one + first_job = waiting_jobs.first() + if not first_job: + raise NoWaitingJobError(f"no waiting job could be found for {job.unicity_id}") + # start it + first_job.update( + started_at=datetime, + status=Status.STARTED, + write_concern={"w": "majority", "fsync": True}, + read_concern={"level": "majority"}, + ) + # and cancel the other ones, if any + for job in waiting_jobs.skip(1): + # TODO: try to do a bulk update (not sure it's possible) + job.update( + finished_at=datetime, + status=Status.CANCELLED, + write_concern={"w": "majority", "fsync": True}, + read_concern={"level": "majority"}, + ) + return first_job.reload() + except TimeoutError as err: + raise LockTimeoutError( + f"could not acquire the lock for job {job.unicity_id} after {RETRIES} retries." + ) from err @@ -622 +612,2 @@ class Queue: - The job is moved from the waiting state to the started state. + The job is moved from the waiting state to the started state. A lock is used to ensure that only one worker + can start a job at a time. @@ -630,0 +622,2 @@ class Queue: + AlreadyStartedJobError: if a started job already exist for the same unicity_id + LockTimeoutError: if the lock cannot be acquired @@ -633,0 +627 @@ class Queue: + @@ -640 +633,0 @@ class Queue: - self._start_job(next_waiting_job) @@ -649 +642,2 @@ class Queue: - return next_waiting_job.info() + started_job = self._start_newest_job_and_cancel_others(job=next_waiting_job) + return started_job.info() @@ -773 +767 @@ class Queue: - self.upsert_job( + self.add_job( diff --git a/libs/libcommon/tests/test_backfill.py b/libs/libcommon/tests/test_backfill.py index 7e445658..404f35a0 100644 --- a/libs/libcommon/tests/test_backfill.py +++ b/libs/libcommon/tests/test_backfill.py @@ -770 +770 @@ def test_delete_jobs( - job = queue._add_job(job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority) + job = queue.add_job(job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority) @@ -775 +775,3 @@ def test_delete_jobs( - queue._start_job(job) + job.status = Status.STARTED + job.started_at = datetime.now() + job.save() diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py index 82e476d5..82848306 100644 --- a/libs/libcommon/tests/test_orchestrator.py +++ b/libs/libcommon/tests/test_orchestrator.py @@ -145 +145 @@ def test_finish_job( - Queue()._add_job( + Queue().add_job( diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 82b930fb..34626226 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -24 +24,3 @@ def get_old_datetime() -> datetime: - return get_datetime() - timedelta(days=1) + # Beware: the TTL index is set to 10 minutes. So it will delete the finished jobs after 10 minutes. + # We have to use a datetime that is not older than 10 minutes. + return get_datetime() - timedelta(seconds=(QUEUE_TTL_SECONDS / 2)) @@ -32 +34 @@ def queue_mongo_resource_autouse(queue_mongo_resource: QueueMongoResource) -> Qu -def test__add_job() -> None: +def test_add_job() -> None: @@ -39 +41 @@ def test__add_job() -> None: - queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + job1 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) @@ -41 +43 @@ def test__add_job() -> None: - queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + job2 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) @@ -43 +45 @@ def test__add_job() -> None: - # get and start the first job + # get and start a job the second one should have been picked @@ -44,0 +47 @@ def test__add_job() -> None: + assert job2.reload().status == Status.STARTED @@ -49,0 +53,2 @@ def test__add_job() -> None: + # and the first job should have been cancelled + assert job1.reload().status == Status.CANCELLED @@ -53 +58,2 @@ def test__add_job() -> None: - queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + job3 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + assert job3.status == Status.WAITING @@ -61,4 +67 @@ def test__add_job() -> None: - # process the second job - job_info = queue.start_job() - queue.finish_job(job_id=job_info["job_id"], is_success=True) - # and the third one + # process the third job @@ -78,78 +80,0 @@ def test__add_job() -> None: -def test_upsert_job() -> None: - test_type = "test_type" - test_dataset = "test_dataset" - test_revision_1 = "test_revision_1" - test_revision_2 = "test_revision_2" - # get the queue - queue = Queue() - # upsert a job - queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_1) - # a second call creates a second waiting job, and the first one is cancelled - queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_1) - # a third call, with a different revision, creates a third waiting job, and the second one is cancelled - # because the unicity_id is the same - queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_2) - assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2) - # get and start the last job - job_info = queue.start_job() - assert job_info["type"] == test_type - assert job_info["params"]["dataset"] == test_dataset - assert job_info["params"]["revision"] == test_revision_2 - assert job_info["params"]["config"] is None - assert job_info["params"]["split"] is None - assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2) - # adding the job while the first one has not finished yet adds a new waiting job - queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_2) - with pytest.raises(EmptyQueueError): - # but: it's not possible to start two jobs with the same arguments - queue.start_job() - # finish the first job - queue.finish_job(job_id=job_info["job_id"], is_success=True) - # the queue is not empty - assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2) - # process the second job - job_info = queue.start_job() - queue.finish_job(job_id=job_info["job_id"], is_success=True) - # the queue is empty - assert not queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2) - with pytest.raises(EmptyQueueError): - # an error is raised if we try to start a job - queue.start_job() - - [email protected]( - "statuses_to_cancel, expected_remaining_number", - [ - (None, 0), - ([Status.WAITING], 1), - ([Status.WAITING, Status.STARTED], 0), - ([Status.STARTED], 1), - ([Status.SUCCESS], 2), - ], -) -def test_cancel_jobs(statuses_to_cancel: Optional[List[Status]], expected_remaining_number: int) -> None: - test_type = "test_type" - test_dataset = "test_dataset" - test_revision_1 = "test_revision_1" - test_revision_2 = "test_revision_2" - queue = Queue() - queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision_1) - queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision_2) - queue.start_job() - - canceled_job_dicts = queue.cancel_jobs( - job_type=test_type, dataset=test_dataset, statuses_to_cancel=statuses_to_cancel - ) - assert len(canceled_job_dicts) == 2 - expected_remaining_number - - if expected_remaining_number == 0: - with pytest.raises(EmptyQueueError): - queue.start_job() - assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_1) == ( - statuses_to_cancel is not None and Status.STARTED not in statuses_to_cancel - ) - assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2) == ( - statuses_to_cancel is not None and Status.WAITING not in statuses_to_cancel - ) - - @@ -173 +98 @@ def test_cancel_jobs_by_job_id( - job = queue._add_job(job_type=test_type, dataset=job_id, revision="test_revision") + job = queue.add_job(job_type=test_type, dataset=job_id, revision="test_revision") @@ -199 +124,13 @@ def check_job(queue: Queue, expected_dataset: str, expected_split: str, expected -def test_priority_logic() -> None: +def test_priority_logic_creation_order() -> None: + test_type = "test_type" + test_revision = "test_revision" + queue = Queue() + queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1") + queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2") + check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL) + check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL) + with pytest.raises(EmptyQueueError): + queue.start_job() + + +def test_priority_logic_started_jobs_per_dataset_order() -> None: @@ -203,3 +140,20 @@ def test_priority_logic() -> None: - queue.upsert_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1") - queue.upsert_job( - job_type=test_type, dataset="dataset1/dataset", revision=test_revision, config="config", split="split1" + queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1") + queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2") + queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1") + check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL) + check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.NORMAL) + # ^ before, even if the creation date is after, because the dataset is different and has no started job + check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL) + with pytest.raises(EmptyQueueError): + queue.start_job() + + +def test_priority_logic_started_jobs_per_namespace_order() -> None: + test_type = "test_type" + test_revision = "test_revision" + queue = Queue() + queue.add_job(job_type=test_type, dataset="org1/dataset1", revision=test_revision, config="config", split="split1") + queue.add_job(job_type=test_type, dataset="org1/dataset2", revision=test_revision, config="config", split="split1") + queue.add_job(job_type=test_type, dataset="org2/dataset2", revision=test_revision, config="config", split="split1") + queue.add_job( + job_type=test_type, dataset="no_org_dataset3", revision=test_revision, config="config", split="split1" @@ -207,8 +161,2 @@ def test_priority_logic() -> None: - queue.upsert_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2") - queue.upsert_job( - job_type=test_type, - dataset="dataset2", - revision=test_revision, - config="config", - split="split1", - priority=Priority.LOW, + check_job( + queue=queue, expected_dataset="org1/dataset1", expected_split="split1", expected_priority=Priority.NORMAL @@ -216,7 +164,2 @@ def test_priority_logic() -> None: - queue.upsert_job( - job_type=test_type, - dataset="dataset2/dataset", - revision=test_revision, - config="config", - split="split1", - priority=Priority.LOW, + check_job( + queue=queue, expected_dataset="org2/dataset2", expected_split="split1", expected_priority=Priority.NORMAL @@ -224,9 +167,6 @@ def test_priority_logic() -> None: - queue.upsert_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split2") - queue.upsert_job(job_type=test_type, dataset="dataset3", revision=test_revision, config="config", split="split1") - queue.upsert_job( - job_type=test_type, - dataset="dataset3", - revision=test_revision, - config="config", - split="split1", - priority=Priority.LOW, + # ^ before, even if the creation date is after, because the namespace is different and has no started job + check_job( + queue=queue, expected_dataset="no_org_dataset3", expected_split="split1", expected_priority=Priority.NORMAL + ) + check_job( + queue=queue, expected_dataset="org1/dataset2", expected_split="split1", expected_priority=Priority.NORMAL @@ -234,2 +174,9 @@ def test_priority_logic() -> None: - queue.upsert_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1") - queue.upsert_job( + with pytest.raises(EmptyQueueError): + queue.start_job() + + +def test_priority_logic_priority_order() -> None: + test_type = "test_type" + test_revision = "test_revision" + queue = Queue() + queue.add_job( @@ -237 +184 @@ def test_priority_logic() -> None: - dataset="dataset2", + dataset="dataset1", @@ -243,16 +190,4 @@ def test_priority_logic() -> None: - check_job( - queue=queue, expected_dataset="dataset1/dataset", expected_split="split1", expected_priority=Priority.NORMAL - ) - check_job(queue=queue, expected_dataset="dataset2", expected_split="split2", expected_priority=Priority.NORMAL) - check_job(queue=queue, expected_dataset="dataset3", expected_split="split1", expected_priority=Priority.NORMAL) - # ^ before the other "dataset3" jobs because its priority is higher (it inherited Priority.NORMAL in upsert_job) - check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL) - # ^ same namespace as dataset1/dataset, goes after namespaces without any started job - check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL) - # ^ comes after the other "dataset1" jobs because the last upsert_job call moved its creation date - check_job( - queue=queue, expected_dataset="dataset2/dataset", expected_split="split1", expected_priority=Priority.LOW - ) - # ^ comes after the other "dataset2" jobs because its priority is lower - check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.LOW) - # ^ the rest of the rules apply for Priority.LOW jobs + queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1") + check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.NORMAL) + # ^ before, even if the creation date is after, because the priority is higher + check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.LOW) @@ -284 +219 @@ def test_job_types_only( - queue.upsert_job(job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None) + queue.add_job(job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None) @@ -309 +244 @@ def test_count_by_status() -> None: - queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision) + queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision) @@ -314 +249 @@ def test_count_by_status() -> None: - queue.upsert_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision) + queue.add_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision) @@ -333 +268 @@ def test_get_dataset_pending_jobs_for_type() -> None: - queue.upsert_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) + queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) @@ -339 +274 @@ def test_get_dataset_pending_jobs_for_type() -> None: - queue.upsert_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) + queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) @@ -344 +279 @@ def test_get_dataset_pending_jobs_for_type() -> None: - queue.upsert_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) + queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None) @@ -356 +291 @@ def test_queue_heartbeat() -> None: - job = queue.upsert_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1") + job = queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1") @@ -370 +305 @@ def test_queue_get_zombies() -> None: - zombie = queue.upsert_job( + zombie = queue.add_job( @@ -374 +309 @@ def test_queue_get_zombies() -> None: - queue.upsert_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split2") + queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split2") @@ -411 +346 @@ def locked_increment(tmp_file: Path) -> None: - with lock(key="test_lock", job_id=str(os.getpid()), sleeps=sleeps): + with lock(key="test_lock", owner=str(os.getpid()), sleeps=sleeps): @@ -434 +369 @@ def git_branch_locked_increment(tmp_file: Path) -> None: - with lock.git_branch(dataset=dataset, branch=branch, job_id=str(os.getpid()), sleeps=sleeps): + with lock.git_branch(dataset=dataset, branch=branch, owner=str(os.getpid()), sleeps=sleeps): @@ -452 +387 @@ def test_lock_git_branch(tmp_path_factory: pytest.TempPathFactory, queue_mongo_r - assert Lock.objects().get().job_id is None + assert Lock.objects().get().owner is None diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py index 47e48854..439f4dde 100644 --- a/services/admin/src/admin/routes/force_refresh.py +++ b/services/admin/src/admin/routes/force_refresh.py @@ -62 +62 @@ def create_force_refresh_endpoint( - Queue().upsert_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split) + Queue().add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split) diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index ab566739..69e0dfd2 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -154 +154 @@ def test_get_cache_entry_from_steps() -> None: - queue.upsert_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config) + queue.add_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config) diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 704e6a5a..c5c879be 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -1037 +1037 @@ def compute_config_parquet_and_info_response( - with lock.git_branch(dataset=dataset, branch=target_revision, job_id=job_id, sleeps=sleeps): + with lock.git_branch(dataset=dataset, branch=target_revision, owner=job_id, sleeps=sleeps): diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py index a1dc8d9f..c0303b21 100644 --- a/services/worker/src/worker/loop.py +++ b/services/worker/src/worker/loop.py @@ -14 +14,7 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.queue import EmptyQueueError, Queue +from libcommon.queue import ( + AlreadyStartedJobError, + EmptyQueueError, + LockTimeoutError, + NoWaitingJobError, + Queue, +) @@ -129 +135 @@ class Loop: - except EmptyQueueError: + except (EmptyQueueError, AlreadyStartedJobError, LockTimeoutError, NoWaitingJobError) as e: @@ -131 +137 @@ class Loop: - logging.debug("no job in the queue") + logging.debug(e) diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py index e4eb41dd..0e912bdb 100644 --- a/services/worker/tests/test_job_manager.py +++ b/services/worker/tests/test_job_manager.py @@ -128 +128 @@ def test_backfill(priority: Priority, app_config: AppConfig) -> None: - queue.upsert_job( + queue.add_job( @@ -198 +198 @@ def test_job_runner_set_crashed( - queue.upsert_job( + queue.add_job( diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py index 35fc1c6a..2a766b92 100644 --- a/services/worker/tests/test_loop.py +++ b/services/worker/tests/test_loop.py @@ -72 +72 @@ def test_process_next_job( - loop.queue.upsert_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split) + loop.queue.add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split)
b7a9721cf1120b11604bff978d8f28554c467e80
Albert Villanova del Moral
2023-06-26T13:01:25
Remove too restrictive __all__ definitions (#1423)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py index 9be31dc5..7c7b13e0 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py +++ b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py @@ -7 +7 @@ from typing import Generic, Type, TypeVar -from mongoengine import Document, DoesNotExist +from mongoengine import Document @@ -58,4 +57,0 @@ def _clean_maintenance_database() -> None: - - -# explicit re-export -__all__ = ["DoesNotExist"] diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 550b1a52..55aea4de 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -18,2 +18,2 @@ import pytz -from mongoengine import Document, DoesNotExist -from mongoengine.errors import NotUniqueError +from mongoengine import Document +from mongoengine.errors import DoesNotExist, NotUniqueError @@ -94,0 +95,4 @@ class EmptyQueueError(Exception): +class JobDoesNotExistError(DoesNotExist): + pass + + @@ -193,0 +198,7 @@ class Job(Document): + @classmethod + def get(cls, job_id: str) -> "Job": + try: + return cls.objects(pk=job_id).get() + except DoesNotExist as e: + raise JobDoesNotExistError(f"Job does not exist: {job_id=}") from e + @@ -921,4 +931,0 @@ def _clean_queue_database() -> None: - - -# explicit re-export -__all__ = ["DoesNotExist"] diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index f537eb28..13582b42 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -25 +25,2 @@ from bson.errors import InvalidId -from mongoengine import Document, DoesNotExist +from mongoengine import Document +from mongoengine.errors import DoesNotExist @@ -125,0 +127,4 @@ CachedResponse.split.required = False # type: ignore +class CacheEntryDoesNotExistError(DoesNotExist): + pass + + @@ -198 +203 @@ class CacheEntryWithoutContent(TypedDict): -# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them +# Note: we let the exceptions throw: it's the responsibility of the caller to manage them @@ -202,5 +207,8 @@ def get_response_without_content( - response = ( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) - .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress") - .get() - ) + try: + response = ( + CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress") + .get() + ) + except DoesNotExist as e: + raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e @@ -226 +234 @@ class CacheEntryMetadata(CacheEntryWithoutContent): -# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them +# Note: we let the exceptions throw: it's the responsibility of the caller to manage them @@ -230,5 +238,8 @@ def get_response_metadata( - response = ( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) - .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "updated_at") - .get() - ) + try: + response = ( + CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "updated_at") + .get() + ) + except DoesNotExist as e: + raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e @@ -285 +296 @@ class CachedArtifactError(Exception): -# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them +# Note: we let the exceptions throw: it's the responsibility of the caller to manage them @@ -287,5 +298,8 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O - response = ( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) - .only("content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress") - .get() - ) + try: + response = ( + CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + .only("content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress") + .get() + ) + except DoesNotExist as e: + raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e @@ -302 +316 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O -# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them +# Note: we let the exceptions throw: it's the responsibility of the caller to manage them @@ -306,4 +320,13 @@ def get_response_with_details( - response = ( - CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) - .only( - "content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "details" + try: + response = ( + CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split) + .only( + "content", + "http_status", + "error_code", + "job_runner_version", + "dataset_git_revision", + "progress", + "details", + ) + .get() @@ -311,2 +334,2 @@ def get_response_with_details( - .get() - ) + except DoesNotExist as e: + raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e @@ -332 +355 @@ def get_response_or_missing_error( - except DoesNotExist: + except CacheEntryDoesNotExistError: @@ -747,4 +769,0 @@ def _clean_cache_database() -> None: - - -# explicit re-export -__all__ = ["DoesNotExist"] diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py index d78a0fa1..e3e9d851 100644 --- a/libs/libcommon/tests/test_simple_cache.py +++ b/libs/libcommon/tests/test_simple_cache.py @@ -15,0 +16 @@ from libcommon.simple_cache import ( + CacheEntryDoesNotExistError, @@ -18 +18,0 @@ from libcommon.simple_cache import ( - DoesNotExist, @@ -141 +141 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None: - with pytest.raises(DoesNotExist): + with pytest.raises(CacheEntryDoesNotExistError): @@ -181 +181 @@ def test_delete_response() -> None: - with pytest.raises(DoesNotExist): + with pytest.raises(CacheEntryDoesNotExistError): @@ -200 +200 @@ def test_delete_dataset_responses() -> None: - with pytest.raises(DoesNotExist): + with pytest.raises(CacheEntryDoesNotExistError): @@ -202 +202 @@ def test_delete_dataset_responses() -> None: - with pytest.raises(DoesNotExist): + with pytest.raises(CacheEntryDoesNotExistError): diff --git a/services/worker/src/worker/job_manager.py b/services/worker/src/worker/job_manager.py index 17ee0782..d4807508 100644 --- a/services/worker/src/worker/job_manager.py +++ b/services/worker/src/worker/job_manager.py @@ -22 +22 @@ from libcommon.simple_cache import ( - DoesNotExist, + CacheEntryDoesNotExistError, @@ -146 +146 @@ class JobManager: - except DoesNotExist: + except CacheEntryDoesNotExistError: diff --git a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py index 524ff628..7de09d90 100644 --- a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py @@ -11 +11 @@ from libcommon.simple_cache import ( - DoesNotExist, + CacheEntryDoesNotExistError, @@ -48 +48 @@ def compute_opt_in_out_urls_scan_response(dataset: str, config: str) -> Tuple[Op - except DoesNotExist: + except CacheEntryDoesNotExistError: diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py index d5251d9b..1e8802c9 100644 --- a/services/worker/src/worker/job_runners/dataset/info.py +++ b/services/worker/src/worker/job_runners/dataset/info.py @@ -11 +11 @@ from libcommon.simple_cache import ( - DoesNotExist, + CacheEntryDoesNotExistError, @@ -54 +54 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl - except DoesNotExist: + except CacheEntryDoesNotExistError: diff --git a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py index 06533df7..4d1a1e46 100644 --- a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py @@ -11 +11 @@ from libcommon.simple_cache import ( - DoesNotExist, + CacheEntryDoesNotExistError, @@ -42 +42 @@ def compute_opt_in_out_urls_count_response(dataset: str) -> Tuple[OptInOutUrlsCo - except DoesNotExist: + except CacheEntryDoesNotExistError: diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index 2ebb9bac..9ba4eddc 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -11 +11 @@ from libcommon.simple_cache import ( - DoesNotExist, + CacheEntryDoesNotExistError, @@ -58 +58 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float] - except DoesNotExist: + except CacheEntryDoesNotExistError: diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py index b48c0f6e..7019388d 100644 --- a/services/worker/src/worker/job_runners/dataset/size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -11 +11 @@ from libcommon.simple_cache import ( - DoesNotExist, + CacheEntryDoesNotExistError, @@ -61 +61 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]: - except DoesNotExist: + except CacheEntryDoesNotExistError: diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py index a34fa1f3..87f44cd5 100644 --- a/services/worker/tests/test_executor.py +++ b/services/worker/tests/test_executor.py @@ -15 +15 @@ from libcommon.processing_graph import ProcessingGraph -from libcommon.queue import DoesNotExist, Job, Queue +from libcommon.queue import Job, JobDoesNotExistError, Queue @@ -112,2 +112,2 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I - Job.objects(pk=job_info["job_id"]).get().delete() - except DoesNotExist: + Job.get(job_id=job_info["job_id"]).delete() + except JobDoesNotExistError: @@ -141,2 +141,2 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q - Job.objects(pk=job_info["job_id"]).get().delete() - except DoesNotExist: + Job.get(job_id=job_info["job_id"]).delete() + except JobDoesNotExistError: @@ -172,2 +172,2 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato - Job.objects(pk=job_info["job_id"]).get().delete() - except DoesNotExist: + Job.get(job_id=job_info["job_id"]).delete() + except JobDoesNotExistError:
1a549c7cf593ebcf18644a1668741b9425e61c76
Andrea Francis Soria Jimenez
2023-06-26T11:04:12
Move dtos to its own file (#1426)
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py new file mode 100644 index 00000000..ef4a4cc0 --- /dev/null +++ b/services/worker/src/worker/dtos.py @@ -0,0 +1,208 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +from dataclasses import dataclass, field +from typing import Any, Dict, List, Mapping, Optional, TypedDict + + +class JobRunnerInfo(TypedDict): + job_type: str + job_runner_version: int + + +@dataclass +class JobResult: + content: Mapping[str, Any] + progress: float + + def __post_init__(self) -> None: + if self.progress < 0.0 or self.progress > 1.0: + raise ValueError(f"Progress should be between 0 and 1, but got {self.progress}") + + +@dataclass +class CompleteJobResult(JobResult): + content: Mapping[str, Any] + progress: float = field(init=False, default=1.0) + + +class DatasetItem(TypedDict): + dataset: str + + +class ConfigItem(DatasetItem): + config: Optional[str] + + +class SplitItem(ConfigItem): + split: Optional[str] + + +class SplitsList(TypedDict): + splits: List[SplitItem] + + +class FailedConfigItem(ConfigItem): + error: Mapping[str, Any] + + +class DatasetSplitNamesResponse(TypedDict): + splits: List[SplitItem] + pending: List[ConfigItem] + failed: List[FailedConfigItem] + + +class PreviousJob(SplitItem): + kind: str + + +class FeatureItem(TypedDict): + feature_idx: int + name: str + type: Mapping[str, Any] + + +class RowItem(TypedDict): + row_idx: int + row: Mapping[str, Any] + truncated_cells: List[str] + + +class SplitFirstRowsResponse(SplitItem): + features: List[FeatureItem] + rows: List[RowItem] + + +class OptUrl(TypedDict): + url: str + row_idx: int + column_name: str + + +class OptInOutUrlsCountResponse(TypedDict): + urls_columns: List[str] + num_opt_in_urls: int + num_opt_out_urls: int + num_urls: int + num_scanned_rows: int + has_urls_columns: bool + full_scan: Optional[bool] + + +class OptInOutUrlsScanResponse(OptInOutUrlsCountResponse): + opt_in_urls: List[OptUrl] + opt_out_urls: List[OptUrl] + + +class ImageUrlColumnsResponse(TypedDict): + columns: List[str] + + +Row = Mapping[str, Any] + + +class RowsContent(TypedDict): + rows: List[Row] + all_fetched: bool + + +class ConfigInfoResponse(TypedDict): + dataset_info: Dict[str, Any] + + +class ParquetFileItem(SplitItem): + url: str + filename: str + size: int + + +class ConfigParquetAndInfoResponse(TypedDict): + parquet_files: List[ParquetFileItem] + dataset_info: Dict[str, Any] + + +class ParquetFileMetadataItem(SplitItem): + url: str + filename: str + size: int + num_rows: int + parquet_metadata_subpath: str + + +class ConfigParquetMetadataResponse(TypedDict): + parquet_files_metadata: List[ParquetFileMetadataItem] + + +class ConfigParquetResponse(TypedDict): + parquet_files: List[ParquetFileItem] + + +class ConfigSize(TypedDict): + dataset: str + config: str + num_bytes_original_files: int + num_bytes_parquet_files: int + num_bytes_memory: int + num_rows: int + num_columns: int + + +class SplitSize(SplitItem): + num_bytes_parquet_files: int + num_bytes_memory: int + num_rows: int + num_columns: int + + +class ConfigSizeContent(TypedDict): + config: ConfigSize + splits: list[SplitSize] + + +class ConfigSizeResponse(TypedDict): + size: ConfigSizeContent + + +class ConfigNameItem(TypedDict): + dataset: str + config: str + + +class DatasetConfigNamesResponse(TypedDict): + config_names: List[ConfigNameItem] + + +class DatasetInfoResponse(TypedDict): + dataset_info: Dict[str, Any] + pending: List[PreviousJob] + failed: List[PreviousJob] + + +class DatasetIsValidResponse(TypedDict): + valid: bool + + +class DatasetParquetResponse(TypedDict): + parquet_files: List[ParquetFileItem] + pending: list[PreviousJob] + failed: list[PreviousJob] + + +class DatasetSize(TypedDict): + dataset: str + num_bytes_original_files: int + num_bytes_parquet_files: int + num_bytes_memory: int + num_rows: int + + +class DatasetSizeContent(TypedDict): + dataset: DatasetSize + configs: list[ConfigSize] + splits: list[SplitSize] + + +class DatasetSizeResponse(TypedDict): + size: DatasetSizeContent + pending: list[PreviousJob] + failed: list[PreviousJob] diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py index d4c9d839..8a99f94a 100644 --- a/services/worker/src/worker/job_runner.py +++ b/services/worker/src/worker/job_runner.py @@ -11 +11 @@ from worker.config import AppConfig -from worker.utils import JobResult, JobRunnerInfo +from worker.dtos import JobResult, JobRunnerInfo diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py index ae0e2760..33d68219 100644 --- a/services/worker/src/worker/job_runners/config/info.py +++ b/services/worker/src/worker/job_runners/config/info.py @@ -2 +1,0 @@ import logging -from typing import Any, Dict, TypedDict @@ -7,0 +7 @@ from libcommon.simple_cache import get_previous_step_or_raise +from worker.dtos import CompleteJobResult, ConfigInfoResponse @@ -9,5 +8,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import CompleteJobResult - - -class ConfigInfoResponse(TypedDict): - dataset_info: Dict[str, Any] diff --git a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py index 9c9306c4..524ff628 100644 --- a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py @@ -15,0 +16 @@ from libcommon.simple_cache import ( +from worker.dtos import JobResult, OptInOutUrlsCountResponse @@ -17 +17,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import JobResult, OptInOutUrlsCountResponse diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py index eae1d780..572df22c 100644 --- a/services/worker/src/worker/job_runners/config/parquet.py +++ b/services/worker/src/worker/job_runners/config/parquet.py @@ -5 +4,0 @@ import logging -from typing import List, TypedDict @@ -10,0 +10 @@ from libcommon.simple_cache import get_previous_step_or_raise +from worker.dtos import CompleteJobResult, ConfigParquetResponse @@ -12,6 +11,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.job_runners.config.parquet_and_info import ParquetFileItem -from worker.utils import CompleteJobResult - - -class ConfigParquetResponse(TypedDict): - parquet_files: List[ParquetFileItem] diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 313166d5..704e6a5a 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -10 +10 @@ from pathlib import Path -from typing import Any, Dict, List, Optional, Set, Tuple, TypedDict +from typing import Any, List, Optional, Set, Tuple @@ -77,0 +78 @@ from worker.config import AppConfig, ParquetAndInfoConfig +from worker.dtos import CompleteJobResult, ConfigParquetAndInfoResponse, ParquetFileItem @@ -79,16 +80 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas -from worker.utils import CompleteJobResult, retry - - -class ParquetFileItem(TypedDict): - dataset: str - config: str - split: str - url: str - filename: str - size: int - - -class ConfigParquetAndInfoResponse(TypedDict): - parquet_files: List[ParquetFileItem] - dataset_info: Dict[str, Any] - +from worker.utils import retry diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py index e1cb6ef5..29a34f9a 100644 --- a/services/worker/src/worker/job_runners/config/parquet_metadata.py +++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py @@ -6 +6 @@ from functools import partial -from typing import List, Optional, TypedDict +from typing import List, Optional @@ -24,0 +25,6 @@ from worker.config import AppConfig +from worker.dtos import ( + CompleteJobResult, + ConfigParquetMetadataResponse, + ParquetFileItem, + ParquetFileMetadataItem, +) @@ -26,17 +31,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.job_runners.config.parquet_and_info import ParquetFileItem -from worker.utils import CompleteJobResult - - -class ParquetFileMetadataItem(TypedDict): - dataset: str - config: str - split: str - url: str - filename: str - size: int - num_rows: int - parquet_metadata_subpath: str - - -class ConfigParquetMetadataResponse(TypedDict): - parquet_files_metadata: List[ParquetFileMetadataItem] diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py index e46b9108..2610f994 100644 --- a/services/worker/src/worker/job_runners/config/size.py +++ b/services/worker/src/worker/job_runners/config/size.py @@ -5 +4,0 @@ import logging -from typing import TypedDict @@ -10,0 +10 @@ from libcommon.simple_cache import get_previous_step_or_raise +from worker.dtos import CompleteJobResult, ConfigSize, ConfigSizeResponse, SplitSize @@ -12,30 +11,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import CompleteJobResult - - -class ConfigSize(TypedDict): - dataset: str - config: str - num_bytes_original_files: int - num_bytes_parquet_files: int - num_bytes_memory: int - num_rows: int - num_columns: int - - -class SplitSize(TypedDict): - dataset: str - config: str - split: str - num_bytes_parquet_files: int - num_bytes_memory: int - num_rows: int - num_columns: int - - -class ConfigSizeContent(TypedDict): - config: ConfigSize - splits: list[SplitSize] - - -class ConfigSizeResponse(TypedDict): - size: ConfigSizeContent diff --git a/services/worker/src/worker/job_runners/config/split_names_from_info.py b/services/worker/src/worker/job_runners/config/split_names_from_info.py index 7037de05..6b97ce14 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_info.py @@ -13,0 +14 @@ from libcommon.simple_cache import get_previous_step_or_raise +from worker.dtos import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList @@ -15 +15,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index 6c4ddc10..8c300c26 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -19,0 +20 @@ from libcommon.exceptions import ( +from worker.dtos import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList @@ -21 +21,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas -from worker.utils import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py index faf74109..51b7a4e1 100644 --- a/services/worker/src/worker/job_runners/dataset/config_names.py +++ b/services/worker/src/worker/job_runners/dataset/config_names.py @@ -5 +5 @@ import logging -from typing import List, Optional, TypedDict, Union +from typing import List, Optional, Union @@ -16,0 +17 @@ from libcommon.exceptions import ( +from worker.dtos import CompleteJobResult, ConfigNameItem, DatasetConfigNamesResponse @@ -20,10 +20,0 @@ from worker.job_runners.dataset.dataset_job_runner import ( -from worker.utils import CompleteJobResult - - -class ConfigNameItem(TypedDict): - dataset: str - config: str - - -class DatasetConfigNamesResponse(TypedDict): - config_names: List[ConfigNameItem] diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py index ad0c1408..d5251d9b 100644 --- a/services/worker/src/worker/job_runners/dataset/info.py +++ b/services/worker/src/worker/job_runners/dataset/info.py @@ -6 +6 @@ from http import HTTPStatus -from typing import Any, Dict, List, Tuple, TypedDict +from typing import Any, Dict, Tuple @@ -15,0 +16 @@ from libcommon.simple_cache import ( +from worker.dtos import DatasetInfoResponse, JobResult, PreviousJob @@ -17,7 +17,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult, PreviousJob - - -class DatasetInfoResponse(TypedDict): - dataset_info: Dict[str, Any] - pending: List[PreviousJob] - failed: List[PreviousJob] diff --git a/services/worker/src/worker/job_runners/dataset/is_valid.py b/services/worker/src/worker/job_runners/dataset/is_valid.py index 9023a1d9..cb45bf8f 100644 --- a/services/worker/src/worker/job_runners/dataset/is_valid.py +++ b/services/worker/src/worker/job_runners/dataset/is_valid.py @@ -5 +5 @@ import logging -from typing import Tuple, TypedDict +from typing import Tuple @@ -9,0 +10 @@ from libcommon.simple_cache import get_validity_by_kind +from worker.dtos import DatasetIsValidResponse, JobResult @@ -11,6 +11,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult - - -class DatasetIsValidResponse(TypedDict): - valid: bool - diff --git a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py index 530d3145..06533df7 100644 --- a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py @@ -15,0 +16 @@ from libcommon.simple_cache import ( +from worker.dtos import JobResult, OptInOutUrlsCountResponse @@ -17 +17,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult, OptInOutUrlsCountResponse diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index 4a360950..2ebb9bac 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -6 +6 @@ from http import HTTPStatus -from typing import List, Tuple, TypedDict +from typing import Tuple @@ -16,2 +16,7 @@ from libcommon.simple_cache import ( -from worker.job_runners.config.parquet import ConfigParquetResponse -from worker.job_runners.config.parquet_and_info import ParquetFileItem +from worker.dtos import ( + ConfigParquetResponse, + DatasetParquetResponse, + JobResult, + ParquetFileItem, + PreviousJob, +) @@ -19,7 +23,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult, PreviousJob - - -class DatasetParquetResponse(TypedDict): - parquet_files: List[ParquetFileItem] - pending: list[PreviousJob] - failed: list[PreviousJob] diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py index 3b69e948..b48c0f6e 100644 --- a/services/worker/src/worker/job_runners/dataset/size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -6 +6 @@ from http import HTTPStatus -from typing import Tuple, TypedDict +from typing import Tuple @@ -16 +16,9 @@ from libcommon.simple_cache import ( -from worker.job_runners.config.size import ConfigSize, ConfigSizeResponse, SplitSize +from worker.dtos import ( + ConfigSize, + ConfigSizeResponse, + DatasetSize, + DatasetSizeResponse, + JobResult, + PreviousJob, + SplitSize, +) @@ -18,21 +25,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult, PreviousJob - - -class DatasetSize(TypedDict): - dataset: str - num_bytes_original_files: int - num_bytes_parquet_files: int - num_bytes_memory: int - num_rows: int - - -class DatasetSizeContent(TypedDict): - dataset: DatasetSize - configs: list[ConfigSize] - splits: list[SplitSize] - - -class DatasetSizeResponse(TypedDict): - size: DatasetSizeContent - pending: list[PreviousJob] - failed: list[PreviousJob] diff --git a/services/worker/src/worker/job_runners/dataset/split_names.py b/services/worker/src/worker/job_runners/dataset/split_names.py index 259e66d4..06aa9222 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names.py +++ b/services/worker/src/worker/job_runners/dataset/split_names.py @@ -12,2 +12 @@ from libcommon.simple_cache import get_best_response, get_previous_step_or_raise -from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import ( +from worker.dtos import ( @@ -19,0 +19 @@ from worker.utils import ( +from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index 07bdb663..ba42e24e 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -25,2 +25 @@ from worker.config import AppConfig, FirstRowsConfig -from worker.job_runners.split.split_job_runner import SplitJobRunner -from worker.utils import ( +from worker.dtos import ( @@ -32,3 +30,0 @@ from worker.utils import ( - create_truncated_row_items, - get_json_size, - to_features_list, @@ -35,0 +32,2 @@ from worker.utils import ( +from worker.job_runners.split.split_job_runner import SplitJobRunner +from worker.utils import create_truncated_row_items, get_json_size, to_features_list diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 3dc01f41..a9694d04 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -35,0 +36 @@ from worker.config import AppConfig, FirstRowsConfig +from worker.dtos import CompleteJobResult, JobRunnerInfo, Row, SplitFirstRowsResponse @@ -38,4 +38,0 @@ from worker.utils import ( - CompleteJobResult, - JobRunnerInfo, - Row, - SplitFirstRowsResponse, diff --git a/services/worker/src/worker/job_runners/split/image_url_columns.py b/services/worker/src/worker/job_runners/split/image_url_columns.py index c2b19607..84756946 100644 --- a/services/worker/src/worker/job_runners/split/image_url_columns.py +++ b/services/worker/src/worker/job_runners/split/image_url_columns.py @@ -11,2 +11 @@ from libcommon.utils import is_image_url -from worker.job_runners.split.split_job_runner import SplitJobRunner -from worker.utils import ( +from worker.dtos import ( @@ -16,0 +16 @@ from worker.utils import ( +from worker.job_runners.split.split_job_runner import SplitJobRunner diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py index af094266..0b602b84 100644 --- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py @@ -9,0 +10 @@ from libcommon.simple_cache import get_previous_step_or_raise +from worker.dtos import CompleteJobResult, OptInOutUrlsCountResponse @@ -11 +11,0 @@ from worker.job_runners.split.split_job_runner import SplitJobRunner -from worker.utils import CompleteJobResult, OptInOutUrlsCountResponse diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py index dbb20277..2ccb4443 100644 --- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py @@ -24,0 +25 @@ from worker.config import AppConfig, OptInOutUrlsScanConfig +from worker.dtos import CompleteJobResult, OptInOutUrlsScanResponse, OptUrl @@ -26,6 +27 @@ from worker.job_runners.split.split_job_runner import SplitJobRunnerWithDatasets -from worker.utils import ( - CompleteJobResult, - OptInOutUrlsScanResponse, - OptUrl, - get_rows_or_raise, -) +from worker.utils import get_rows_or_raise diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 982da72d..fab6b031 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -9 +8,0 @@ import warnings -from dataclasses import dataclass, field @@ -14 +12,0 @@ from typing import ( - Mapping, @@ -19 +16,0 @@ from typing import ( - TypedDict, @@ -36,0 +34,2 @@ from libcommon.utils import orjson_dumps +from worker.dtos import FeatureItem, Row, RowItem, RowsContent + @@ -41,110 +39,0 @@ MAX_IMAGE_PIXELS = 10_000_000_000 -class JobRunnerInfo(TypedDict): - job_type: str - job_runner_version: int - - -@dataclass -class JobResult: - content: Mapping[str, Any] - progress: float - - def __post_init__(self) -> None: - if self.progress < 0.0 or self.progress > 1.0: - raise ValueError(f"Progress should be between 0 and 1, but got {self.progress}") - - -@dataclass -class CompleteJobResult(JobResult): - content: Mapping[str, Any] - progress: float = field(init=False, default=1.0) - - -class DatasetItem(TypedDict): - dataset: str - - -class ConfigItem(DatasetItem): - config: Optional[str] - - -class SplitItem(ConfigItem): - split: Optional[str] - - -class SplitsList(TypedDict): - splits: List[SplitItem] - - -class FailedConfigItem(ConfigItem): - error: Mapping[str, Any] - - -class DatasetSplitNamesResponse(TypedDict): - splits: List[SplitItem] - pending: List[ConfigItem] - failed: List[FailedConfigItem] - - -class PreviousJob(TypedDict): - kind: str - dataset: str - config: Optional[str] - split: Optional[str] - - -class FeatureItem(TypedDict): - feature_idx: int - name: str - type: Mapping[str, Any] - - -class RowItem(TypedDict): - row_idx: int - row: Mapping[str, Any] - truncated_cells: List[str] - - -class SplitFirstRowsResponse(TypedDict): - dataset: str - config: str - split: str - features: List[FeatureItem] - rows: List[RowItem] - - -class OptUrl(TypedDict): - url: str - row_idx: int - column_name: str - - -class OptInOutUrlsCountResponse(TypedDict): - urls_columns: List[str] - num_opt_in_urls: int - num_opt_out_urls: int - num_urls: int - num_scanned_rows: int - has_urls_columns: bool - full_scan: Optional[bool] - - -class OptInOutUrlsScanResponse(OptInOutUrlsCountResponse): - opt_in_urls: List[OptUrl] - opt_out_urls: List[OptUrl] - - -class ImageUrlColumnsResponse(TypedDict): - columns: List[str] - - -Row = Mapping[str, Any] - - -class RowsContent(TypedDict): - rows: List[Row] - all_fetched: bool - - -# TODO: separate functions from common classes and named dicts otherwise this file will continue growing - - diff --git a/services/worker/tests/job_runners/config/test_config_job_runner.py b/services/worker/tests/job_runners/config/test_config_job_runner.py index 1e8fa750..01a5ce8e 100644 --- a/services/worker/tests/job_runners/config/test_config_job_runner.py +++ b/services/worker/tests/job_runners/config/test_config_job_runner.py @@ -11,0 +12 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -13 +13,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import CompleteJobResult diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py index 4b58e5f7..031fee96 100644 --- a/services/worker/tests/job_runners/config/test_parquet.py +++ b/services/worker/tests/job_runners/config/test_parquet.py @@ -15,5 +15 @@ from worker.config import AppConfig -from worker.job_runners.config.parquet import ( - ConfigParquetJobRunner, - ConfigParquetResponse, -) -from worker.job_runners.config.parquet_and_info import ( +from worker.dtos import ( @@ -20,0 +17 @@ from worker.job_runners.config.parquet_and_info import ( + ConfigParquetResponse, @@ -22,0 +20 @@ from worker.job_runners.config.parquet_and_info import ( +from worker.job_runners.config.parquet import ConfigParquetJobRunner diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index e9631ea8..ec3bc9c2 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -36,0 +37 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -52 +52,0 @@ from worker.resources import LibrariesResource -from worker.utils import CompleteJobResult diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py index fadd6e71..56a3815d 100644 --- a/services/worker/tests/job_runners/config/test_parquet_metadata.py +++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py @@ -22,4 +22 @@ from worker.config import AppConfig -from worker.job_runners.config.parquet import ConfigParquetResponse -from worker.job_runners.config.parquet_and_info import ParquetFileItem -from worker.job_runners.config.parquet_metadata import ( - ConfigParquetMetadataJobRunner, +from worker.dtos import ( @@ -26,0 +24,2 @@ from worker.job_runners.config.parquet_metadata import ( + ConfigParquetResponse, + ParquetFileItem, @@ -28,0 +28 @@ from worker.job_runners.config.parquet_metadata import ( +from worker.job_runners.config.parquet_metadata import ConfigParquetMetadataJobRunner diff --git a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py index 2d8a95d2..49f7fa06 100644 --- a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py +++ b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py @@ -11,0 +12 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -13 +13,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import CompleteJobResult diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py index 94c8d7d4..26aeb369 100644 --- a/services/worker/tests/job_runners/dataset/test_info.py +++ b/services/worker/tests/job_runners/dataset/test_info.py @@ -14,0 +15 @@ from worker.config import AppConfig +from worker.dtos import PreviousJob @@ -16 +16,0 @@ from worker.job_runners.dataset.info import DatasetInfoJobRunner -from worker.utils import PreviousJob diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py index c698848c..61cfbfbc 100644 --- a/services/worker/tests/job_runners/dataset/test_parquet.py +++ b/services/worker/tests/job_runners/dataset/test_parquet.py @@ -15,6 +15,2 @@ from worker.config import AppConfig -from worker.job_runners.config.parquet import ConfigParquetResponse -from worker.job_runners.config.parquet_and_info import ParquetFileItem -from worker.job_runners.dataset.parquet import ( - DatasetParquetJobRunner, - DatasetParquetResponse, -) +from worker.dtos import ConfigParquetResponse, DatasetParquetResponse, ParquetFileItem +from worker.job_runners.dataset.parquet import DatasetParquetJobRunner diff --git a/services/worker/tests/job_runners/split/test_image_url_columns.py b/services/worker/tests/job_runners/split/test_image_url_columns.py index cf13a2d3..68fb6f01 100644 --- a/services/worker/tests/job_runners/split/test_image_url_columns.py +++ b/services/worker/tests/job_runners/split/test_image_url_columns.py @@ -17,0 +18 @@ from worker.config import AppConfig +from worker.dtos import ImageUrlColumnsResponse @@ -19 +19,0 @@ from worker.job_runners.split.image_url_columns import SplitImageUrlColumnsJobRu -from worker.utils import ImageUrlColumnsResponse diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py index 8c8d3d74..2de50d1d 100644 --- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py @@ -23,0 +24 @@ from worker.config import AppConfig +from worker.dtos import ImageUrlColumnsResponse @@ -29 +29,0 @@ from worker.resources import LibrariesResource -from worker.utils import ImageUrlColumnsResponse diff --git a/services/worker/tests/job_runners/split/test_split_job_runner.py b/services/worker/tests/job_runners/split/test_split_job_runner.py index d944a8ea..80d88ba4 100644 --- a/services/worker/tests/job_runners/split/test_split_job_runner.py +++ b/services/worker/tests/job_runners/split/test_split_job_runner.py @@ -11,0 +12 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -13 +13,0 @@ from worker.job_runners.split.split_job_runner import SplitJobRunner -from worker.utils import CompleteJobResult diff --git a/services/worker/tests/job_runners/test__job_runner_with_cache.py b/services/worker/tests/job_runners/test__job_runner_with_cache.py index db580ea6..11aa9fd1 100644 --- a/services/worker/tests/job_runners/test__job_runner_with_cache.py +++ b/services/worker/tests/job_runners/test__job_runner_with_cache.py @@ -13,0 +14 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -16 +16,0 @@ from worker.resources import LibrariesResource -from worker.utils import CompleteJobResult diff --git a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py index 30ea2795..2667da40 100644 --- a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py +++ b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py @@ -13,0 +14 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -18 +18,0 @@ from worker.resources import LibrariesResource -from worker.utils import CompleteJobResult diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py index 93b8df65..e4eb41dd 100644 --- a/services/worker/tests/test_job_manager.py +++ b/services/worker/tests/test_job_manager.py @@ -13,0 +14 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -16 +16,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import CompleteJobResult diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py index 4176c770..35fc1c6a 100644 --- a/services/worker/tests/test_loop.py +++ b/services/worker/tests/test_loop.py @@ -7,0 +8 @@ from worker.config import AppConfig +from worker.dtos import CompleteJobResult @@ -12 +12,0 @@ from worker.resources import LibrariesResource -from worker.utils import CompleteJobResult
120c2547e31780e584d0a37ac4d9305f4b988195
Andrea Francis Soria Jimenez
2023-06-26T11:01:35
Create missing Jobs when /rows cache does not exists yet (#1425)
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py index e0ddad9f..f95a63f8 100644 --- a/libs/libcommon/src/libcommon/parquet_utils.py +++ b/libs/libcommon/src/libcommon/parquet_utils.py @@ -20 +19,0 @@ from libcommon.constants import PARQUET_REVISION -from libcommon.exceptions import UnexpectedError @@ -399,12 +398,8 @@ class RowsIndex: - try: - result = get_previous_step_or_raise( - kinds=cache_kinds, - dataset=self.dataset, - config=self.config, - split=None, - ) - self.revision = result.response["dataset_git_revision"] - content = result.response["content"] - except Exception as e: - raise UnexpectedError("Could not get the list of parquet files to fetch the rows from.") from e - # ^ TODO: improve the error, depending on the case + result = get_previous_step_or_raise( + kinds=cache_kinds, + dataset=self.dataset, + config=self.config, + split=None, + ) + self.revision = result.response["dataset_git_revision"] + content = result.response["content"] diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index 34d68dc4..13479c33 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -113,0 +114 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi + hf_endpoint=app_config.common.hf_endpoint, diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py index 8b8a4dab..4ddb8bc5 100644 --- a/services/api/src/api/routes/endpoint.py +++ b/services/api/src/api/routes/endpoint.py @@ -9,2 +8,0 @@ from typing import List, Mapping, Optional, Tuple, TypedDict -from libcommon.dataset import get_dataset_git_revision -from libcommon.orchestrator import DatasetOrchestrator @@ -18 +15,0 @@ from libcommon.simple_cache import ( -from libcommon.utils import Priority @@ -28,2 +24,0 @@ from api.utils import ( - ResponseNotFoundError, - ResponseNotReadyError, @@ -34,0 +30 @@ from api.utils import ( + try_backfill_dataset, @@ -86,30 +82,8 @@ def get_cache_entry_from_steps( - dataset_orchestrator = DatasetOrchestrator(dataset=dataset, processing_graph=processing_graph) - if not dataset_orchestrator.has_some_cache(): - # We have to check if the dataset exists and is supported - try: - revision = get_dataset_git_revision( - dataset=dataset, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - hf_timeout_seconds=hf_timeout_seconds, - ) - except Exception as e: - # The dataset is not supported - raise ResponseNotFoundError("Not found.") from e - # The dataset is supported, and the revision is known. We set the revision (it will create the jobs) - # and tell the user to retry. - dataset_orchestrator.set_revision(revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[]) - raise ResponseNotReadyError( - "The server is busier than usual and the response is not ready yet. Please retry later." - ) - elif dataset_orchestrator.has_pending_ancestor_jobs( - processing_step_names=[processing_step.name for processing_step in processing_steps] - ): - # some jobs are still in progress, the cache entries could exist in the future - raise ResponseNotReadyError( - "The server is busier than usual and the response is not ready yet. Please retry later." - ) - else: - # no pending job: the cache entry will not be created - raise ResponseNotFoundError("Not found.") - + try_backfill_dataset( + processing_steps=processing_steps, + processing_graph=processing_graph, + dataset=dataset, + hf_endpoint=hf_endpoint, + hf_timeout_seconds=hf_timeout_seconds, + hf_token=hf_token, + ) diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 6f8d2d51..177865f5 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -16,0 +17 @@ from libcommon.prometheus import StepProfiler +from libcommon.simple_cache import CachedArtifactError @@ -34,0 +36 @@ from api.utils import ( + try_backfill_dataset, @@ -254,0 +257 @@ def create_rows_endpoint( + hf_endpoint: str, @@ -309,6 +312,21 @@ def create_rows_endpoint( - rows_index = indexer.get_rows_index( - dataset=dataset, - config=config, - split=split, - ) - revision = rows_index.revision + try: + rows_index = indexer.get_rows_index( + dataset=dataset, + config=config, + split=split, + ) + revision = rows_index.revision + except CachedArtifactError: + config_parquet_processing_steps = processing_graph.get_config_parquet_processing_steps() + config_parquet_metadata_processing_steps = ( + processing_graph.get_config_parquet_metadata_processing_steps() + ) + try_backfill_dataset( + processing_steps=config_parquet_metadata_processing_steps + + config_parquet_processing_steps, + processing_graph=processing_graph, + dataset=dataset, + hf_endpoint=hf_endpoint, + hf_timeout_seconds=hf_timeout_seconds, + hf_token=hf_token, + ) diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py index 99fe6657..079529ee 100644 --- a/services/api/src/api/utils.py +++ b/services/api/src/api/utils.py @@ -7,0 +8 @@ from typing import Any, Callable, Coroutine, List, Literal, Optional +from libcommon.dataset import get_dataset_git_revision @@ -9 +10,3 @@ from libcommon.exceptions import CustomError -from libcommon.utils import orjson_dumps +from libcommon.orchestrator import DatasetOrchestrator +from libcommon.processing_graph import ProcessingGraph, ProcessingStep +from libcommon.utils import Priority, orjson_dumps @@ -168,0 +172,39 @@ def are_valid_parameters(parameters: List[Any]) -> bool: +def try_backfill_dataset( + processing_steps: List[ProcessingStep], + dataset: str, + processing_graph: ProcessingGraph, + hf_endpoint: str, + hf_token: Optional[str] = None, + hf_timeout_seconds: Optional[float] = None, +) -> None: + dataset_orchestrator = DatasetOrchestrator(dataset=dataset, processing_graph=processing_graph) + if not dataset_orchestrator.has_some_cache(): + # We have to check if the dataset exists and is supported + try: + revision = get_dataset_git_revision( + dataset=dataset, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + hf_timeout_seconds=hf_timeout_seconds, + ) + except Exception as e: + # The dataset is not supported + raise ResponseNotFoundError("Not found.") from e + # The dataset is supported, and the revision is known. We set the revision (it will create the jobs) + # and tell the user to retry. + dataset_orchestrator.set_revision(revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[]) + raise ResponseNotReadyError( + "The server is busier than usual and the response is not ready yet. Please retry later." + ) + elif dataset_orchestrator.has_pending_ancestor_jobs( + processing_step_names=[processing_step.name for processing_step in processing_steps] + ): + # some jobs are still in progress, the cache entries could exist in the future + raise ResponseNotReadyError( + "The server is busier than usual and the response is not ready yet. Please retry later." + ) + else: + # no pending job: the cache entry will not be created + raise ResponseNotFoundError("Not found.") + + diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py index ebdd22a4..ab566739 100644 --- a/services/api/tests/routes/test_endpoint.py +++ b/services/api/tests/routes/test_endpoint.py @@ -156 +156 @@ def test_get_cache_entry_from_steps() -> None: - with patch("api.routes.endpoint.get_dataset_git_revision", return_value=revision): + with patch("libcommon.dataset.get_dataset_git_revision", return_value=revision):
07887d39931663f1e4c1a78cd1140c6395a8d356
Quentin Lhoest
2023-06-26T10:40:24
Ignore duckdb files in parquet and info (#1429)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 1e1c4d9b..313166d5 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -798,0 +799,27 @@ def create_commits( +def get_delete_operations( + parquet_operations: List[CommitOperationAdd], all_repo_files: Set[str], config_names: Set[str], config: str +) -> List[CommitOperationDelete]: + # - get files that will be preserved in repo: + # 1. parquet files belonging to any other config (otherwise outdated files might be preserved) + # 2. duckdb files belonging to any config + # 3. .gitattributes + pattern_in_any_config_dir = re.compile(f"^({'|'.join(re.escape(conf) for conf in config_names)})/") + pattern_in_any_other_config_dir = re.compile( + f"^({'|'.join(re.escape(conf) for conf in config_names.difference({config}))})/" + ) + files_to_ignore: Set[str] = { + file + for file in all_repo_files + if (pattern_in_any_other_config_dir.match(file) and file.endswith(".parquet")) + or (pattern_in_any_config_dir.match(file) and file.endswith(".duckdb")) + }.union({".gitattributes"}) + # - get files to be deleted - all files except for: + # - the files to be preserved + # - parquet files obtained for current config at this processing step + files_to_add = [operation.path_in_repo for operation in parquet_operations] + files_to_delete = all_repo_files - set(files_to_add).union(files_to_ignore) + delete_operations = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete] + logging.debug(f"{delete_operations=}") + return delete_operations + + @@ -850 +876,0 @@ def commit_parquet_conversion( - # - get repo parquet files @@ -852,18 +878,3 @@ def commit_parquet_conversion( - repo_parquet_files: Set[str] = {file for file in all_repo_files if file.endswith(".parquet")} - # - get files that will be preserved in repo: files belonging to other configs and .gitattributes - # we exclude files of current config because otherwise outdated files might be preserved - files_to_ignore: Set[str] = { - file - for other_config in config_names.difference({config}) - for file in repo_parquet_files - if file.startswith(f"{other_config}/") - }.union({".gitattributes"}) - # - get files to be deleted - all files except for: - # - parquet files obtained for current config at this processing step, - # - parquet files belonging to other existing configs - # - .gitattributes - files_to_add = [operation.path_in_repo for operation in parquet_operations] - files_to_delete = all_repo_files - set(files_to_add).union(files_to_ignore) - delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete] - logging.debug(f"{delete_operations=}") - + delete_operations = get_delete_operations( + parquet_operations=parquet_operations, all_repo_files=all_repo_files, config_names=config_names, config=config + ) diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 9de4ac77..e9631ea8 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -11 +11 @@ from pathlib import Path -from typing import Any, Callable, Iterator, List, Optional, TypedDict +from typing import Any, Callable, Iterator, List, Optional, Set, TypedDict @@ -40,0 +41 @@ from worker.job_runners.config.parquet_and_info import ( + get_delete_operations, @@ -796,0 +798,31 @@ def test_concurrency( + + [email protected]( + "parquet_files,all_repo_files,config_names,config,deleted_files", + [ + ( + set(), + {"dummy", "c1/dummy", "c1/0.parquet", "c2/0.parquet", "c1/index.duckdb"}, + {"c1", "c2"}, + "c1", + {"dummy", "c1/dummy", "c1/0.parquet"}, + ), + ( + {"c1/0.parquet"}, + {"dummy", "c1/dummy", "c1/0.parquet", "c2/0.parquet", "c1/index.duckdb"}, + {"c1", "c2"}, + "c1", + {"dummy", "c1/dummy"}, + ), + ], +) +def test_get_delete_operations( + parquet_files: Set[str], all_repo_files: Set[str], config_names: Set[str], config: str, deleted_files: Set[str] +) -> None: + parquet_operations = [ + CommitOperationAdd(path_in_repo=path_in_repo, path_or_fileobj=b"") for path_in_repo in parquet_files + ] + delete_operations = get_delete_operations( + parquet_operations=parquet_operations, all_repo_files=all_repo_files, config_names=config_names, config=config + ) + assert set(delete_operation.path_in_repo for delete_operation in delete_operations) == deleted_files
1d668b0ad9ea88fec9f37101a42b1934d61d726f
Quentin Lhoest
2023-06-26T08:21:49
Add auth in rows (#1427)
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 2806e5b0..6f8d2d51 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -271 +271 @@ def create_rows_endpoint( - httpfs=HTTPFileSystem(), + httpfs=HTTPFileSystem(storage_options={"headers": {"authorization": f"Bearer {hf_token}"}}),
ad4351dc9ee2995cf716aa83afb50f8f1fb305f1
Quentin Lhoest
2023-06-23T16:01:52
Fix regression: use parquet metadata when possible (#1424)
diff --git a/e2e/tests/test_11_api.py b/e2e/tests/test_11_api.py index d6abeef3..7732ea1f 100644 --- a/e2e/tests/test_11_api.py +++ b/e2e/tests/test_11_api.py @@ -109 +109 @@ def test_rows_endpoint( - limit = 10 + length = 10 @@ -111 +111 @@ def test_rows_endpoint( - relative_url=f"/rows?dataset={dataset}&config={config}&split={split}&offset={offset}&limit={limit}", + relative_url=f"/rows?dataset={dataset}&config={config}&split={split}&offset={offset}&length={length}", diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py index 3a498888..e0ddad9f 100644 --- a/libs/libcommon/src/libcommon/parquet_utils.py +++ b/libs/libcommon/src/libcommon/parquet_utils.py @@ -7 +7 @@ from os import PathLike -from typing import Callable, List, Literal, Optional, Tuple, TypedDict, Union, cast +from typing import Callable, List, Literal, Optional, Tuple, TypedDict, Union @@ -60,22 +59,0 @@ class ParquetFileMetadataItem(TypedDict): -class HTTPFileSystemSession(object): - _singleton: Optional["HTTPFileSystemSession"] = None - - @classmethod - def get_instance(cls) -> "HTTPFileSystemSession": - if cls._singleton is None: - HTTPFileSystemSession() - return cast("HTTPFileSystemSession", HTTPFileSystemSession._singleton) - - def __new__(cls) -> "HTTPFileSystemSession": - if HTTPFileSystemSession._singleton is not None: - raise RuntimeError( - "cannot initialize another instance of HTTPFileSystemSession, use .get_instance() instead" - ) - return super(HTTPFileSystemSession, cls).__new__(cls) - - def __init__(self) -> None: - HTTPFileSystemSession._singleton = self - self.httpfs = HTTPFileSystem() - self.session = asyncio.run(self.httpfs.set_session()) - - @@ -236,0 +215 @@ class ParquetIndexWithMetadata: + httpfs: HTTPFileSystem @@ -238,0 +218,6 @@ class ParquetIndexWithMetadata: + def __post_init__(self) -> None: + if self.httpfs._session is None: + self.httpfs_session = asyncio.run(self.httpfs.set_session()) + else: + self.httpfs_session = self.httpfs._session + @@ -273,3 +257,0 @@ class ParquetIndexWithMetadata: - httpFileSystemSession = HTTPFileSystemSession.get_instance() - session = httpFileSystemSession.session - httpfs = httpFileSystemSession.httpfs @@ -278 +260,8 @@ class ParquetIndexWithMetadata: - HTTPFile(httpfs, url, session=session, size=size, loop=httpfs.loop, cache_type=None), + HTTPFile( + self.httpfs, + url, + session=self.httpfs_session, + size=size, + loop=self.httpfs.loop, + cache_type=None, + ), @@ -319,0 +309 @@ class ParquetIndexWithMetadata: + httpfs: HTTPFileSystem, @@ -359,0 +350 @@ class ParquetIndexWithMetadata: + httpfs=httpfs, @@ -370,0 +362 @@ class RowsIndex: + httpfs: HfFileSystem, @@ -379,0 +372 @@ class RowsIndex: + self.httpfs = httpfs @@ -403,2 +396,2 @@ class RowsIndex: - cache_kinds = [step.cache_kind for step in config_parquet_processing_steps] - cache_kinds.extend([step.cache_kind for step in config_parquet_metadata_processing_steps]) + cache_kinds = [step.cache_kind for step in config_parquet_metadata_processing_steps] + cache_kinds.extend([step.cache_kind for step in config_parquet_processing_steps]) @@ -438,0 +432 @@ class RowsIndex: + httpfs=self.httpfs, @@ -465,0 +460 @@ class Indexer: + httpfs: HTTPFileSystem, @@ -471,0 +467 @@ class Indexer: + self.httpfs = httpfs @@ -492,0 +489 @@ class Indexer: + httpfs=self.httpfs, diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 618a6e02..2806e5b0 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -12,0 +13 @@ from datasets import Features +from fsspec.implementations.http import HTTPFileSystem @@ -269,0 +271 @@ def create_rows_endpoint( + httpfs=HTTPFileSystem(), @@ -274,0 +277 @@ def create_rows_endpoint( + await indexer.httpfs.set_session() diff --git a/services/api/tests/routes/test_rows.py b/services/api/tests/routes/test_rows.py index 9dbad3ed..7613d52a 100644 --- a/services/api/tests/routes/test_rows.py +++ b/services/api/tests/routes/test_rows.py @@ -14,0 +15 @@ from fsspec import AbstractFileSystem +from fsspec.implementations.http import HTTPFileSystem @@ -247 +248,5 @@ def dataset_image_with_config_parquet() -> dict[str, Any]: -def indexer(app_config: AppConfig, processing_graph: ProcessingGraph, parquet_metadata_directory: StrPath) -> Indexer: +def indexer( + app_config: AppConfig, + processing_graph: ProcessingGraph, + parquet_metadata_directory: StrPath, +) -> Indexer: @@ -251,0 +257 @@ def indexer(app_config: AppConfig, processing_graph: ProcessingGraph, parquet_me + httpfs=HTTPFileSystem(), diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index a4f64ece..07bdb663 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -7,0 +8 @@ from datasets import Audio, Features, Image +from fsspec.implementations.http import HTTPFileSystem @@ -198,0 +200 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner): + httpfs=HTTPFileSystem(),
bddd6c15995b6f72a93a8b1cd692aca58cc2a4c8
Quentin Lhoest
2023-06-23T12:02:30
update to datasets 2.13.1 (#1422)
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock index d43c651e..0de28c38 100644 --- a/front/admin_ui/poetry.lock +++ b/front/admin_ui/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -529 +529 @@ name = "datasets" -version = "2.13.0" +version = "2.13.1" @@ -535,2 +535,2 @@ files = [ - {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, - {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, @@ -1248 +1248 @@ appdirs = "^1.4.4" -datasets = {version = "^2.13.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} @@ -2754 +2753,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock index 6b1a8a05..acc68895 100644 --- a/jobs/cache_maintenance/poetry.lock +++ b/jobs/cache_maintenance/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -579 +579 @@ name = "datasets" -version = "2.13.0" +version = "2.13.1" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, - {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, @@ -1026 +1026 @@ appdirs = "^1.4.4" -datasets = {version = "^2.13.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} @@ -2566 +2565,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock index 6b1a8a05..acc68895 100644 --- a/jobs/mongodb_migration/poetry.lock +++ b/jobs/mongodb_migration/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -579 +579 @@ name = "datasets" -version = "2.13.0" +version = "2.13.1" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, - {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, @@ -1026 +1026 @@ appdirs = "^1.4.4" -datasets = {version = "^2.13.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} @@ -2566 +2565,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock index 99b16a6a..aa8af056 100644 --- a/libs/libcommon/poetry.lock +++ b/libs/libcommon/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -579 +579 @@ name = "datasets" -version = "2.13.0" +version = "2.13.1" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, - {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, @@ -2547 +2546,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -3015 +3014 @@ python-versions = "3.9.15" -content-hash = "bfdd77727f041d88e26a4da8c23a6b8434f20c6747f068f1ca3c11d841e38aa1" +content-hash = "c796cce59e2755d2a0b49d39a532262156b903ef92d51cfd769f3e66eceb932c" diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml index 8dcfc6a6..c4234875 100644 --- a/libs/libcommon/pyproject.toml +++ b/libs/libcommon/pyproject.toml @@ -10 +10 @@ appdirs = "^1.4.4" -datasets = {version = "^2.13.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock index e138e9d3..bea23aef 100644 --- a/services/admin/poetry.lock +++ b/services/admin/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -579 +579 @@ name = "datasets" -version = "2.13.0" +version = "2.13.1" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, - {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, @@ -1084 +1084 @@ appdirs = "^1.4.4" -datasets = {version = "^2.13.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} @@ -2661 +2660,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, diff --git a/services/api/poetry.lock b/services/api/poetry.lock index 77c2de7d..d688e743 100644 --- a/services/api/poetry.lock +++ b/services/api/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -621 +621 @@ name = "datasets" -version = "2.13.0" +version = "2.13.1" @@ -627,2 +627,2 @@ files = [ - {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, - {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, @@ -1146 +1146 @@ appdirs = "^1.4.4" -datasets = {version = "^2.13.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} @@ -2852 +2851,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 5f59e6b5..0ece7030 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -910 +910 @@ name = "datasets" -version = "2.13.0" +version = "2.13.1" @@ -916,2 +916,2 @@ files = [ - {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, - {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, + {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"}, + {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"}, @@ -1780 +1780 @@ appdirs = "^1.4.4" -datasets = {version = "^2.13.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.1", extras = ["audio", "vision"]} @@ -2865 +2864,0 @@ files = [ - {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -4401 +4399,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -4694,0 +4693,2 @@ files = [ + {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:9c9b14fbb73ec4cb0f209722a1489020fd8614c92ae22589f2309c48cefdf21f"}, + {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_x86_64.whl", hash = "sha256:6a54539bd076746f69ae8bef7282f981674fe4dbf59c3a84c4af86ae6bae9d5c"},
5584b9cf9e72e4b6b6b22fa436a5f0936a4b3d96
Quentin Lhoest
2023-06-22T12:51:40
keep image and audio untruncated (#1417)
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index 64112c28..a4f64ece 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -7 +7 @@ from typing import List -from datasets import Features +from datasets import Audio, Features, Image @@ -142,0 +143 @@ def compute_first_rows_response( + columns_to_keep_untruncated = [col for col, feature in features.items() if isinstance(feature, (Image, Audio))] @@ -147,0 +149 @@ def compute_first_rows_response( + columns_to_keep_untruncated=columns_to_keep_untruncated, diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index a49cc8ea..3dc01f41 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -8 +8,8 @@ from typing import List, Optional, Union -from datasets import Features, IterableDataset, get_dataset_config_info, load_dataset +from datasets import ( + Audio, + Features, + Image, + IterableDataset, + get_dataset_config_info, + load_dataset, +) @@ -246,0 +254 @@ def compute_first_rows_response( + columns_to_keep_untruncated = [col for col, feature in features.items() if isinstance(feature, (Image, Audio))] @@ -251,0 +260 @@ def compute_first_rows_response( + columns_to_keep_untruncated=columns_to_keep_untruncated, diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index b4869d67..982da72d 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -201 +201 @@ def utf8_byte_truncate(text: str, max_bytes: int) -> str: -def truncate_row_item(row_item: RowItem, min_cell_bytes: int) -> RowItem: +def truncate_row_item(row_item: RowItem, min_cell_bytes: int, columns_to_keep_untruncated: List[str]) -> RowItem: @@ -208 +208 @@ def truncate_row_item(row_item: RowItem, min_cell_bytes: int) -> RowItem: - if len(cell_json) <= min_cell_bytes: + if len(cell_json) <= min_cell_bytes or column_name in columns_to_keep_untruncated: @@ -224 +224,3 @@ COMMA_SIZE = 1 # the comma "," is encoded with one byte in utf-8 -def truncate_row_items(row_items: List[RowItem], min_cell_bytes: int, rows_max_bytes: int) -> List[RowItem]: +def truncate_row_items( + row_items: List[RowItem], min_cell_bytes: int, rows_max_bytes: int, columns_to_keep_untruncated: List[str] +) -> List[RowItem]: @@ -233 +235,3 @@ def truncate_row_items(row_items: List[RowItem], min_cell_bytes: int, rows_max_b - row_item = truncate_row_item(row_item=row_item, min_cell_bytes=min_cell_bytes) + row_item = truncate_row_item( + row_item=row_item, min_cell_bytes=min_cell_bytes, columns_to_keep_untruncated=columns_to_keep_untruncated + ) @@ -251,0 +256 @@ def create_truncated_row_items( + columns_to_keep_untruncated: List[str], @@ -277 +282,6 @@ def create_truncated_row_items( - return truncate_row_items(row_items=row_items, min_cell_bytes=min_cell_bytes, rows_max_bytes=rows_max_bytes) + return truncate_row_items( + row_items=row_items, + min_cell_bytes=min_cell_bytes, + rows_max_bytes=rows_max_bytes, + columns_to_keep_untruncated=columns_to_keep_untruncated, + )
8763a497aa8f40970cc9008c9e68072f2c964794
Rémy
2023-06-22T07:51:06
feat: use external-secrets to read secrets from AWS (#1419)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index e7420e4e..96acf9f7 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -51,0 +52,10 @@ secrets: + externalSecret: + enabled: true + secretName: "datasets-server-prod-secrets" + secretStoreName: "datasets-server-prod-secretstore" + parameters: + MONGO_URL: "hub-prod-datasets-server-mongo-url" + HF_TOKEN: "hub-prod-datasets-server-hf-token" + PARQUET_CONVERTER_HF_TOKEN: "hub-prod-datasets-server-parquet-converter-hf-token" + WEBHOOK_SECRET: "hub-prod-datasets-server-webhook-secret" + SPAWNING_TOKEN: "hub-prod-datasets-server-spawning-token" @@ -54,2 +64 @@ secrets: - secretName: "mongo-url" - value: mongo:// + secretName: "datasets-server-prod-secrets" @@ -58 +67 @@ secrets: - secretName: "hf-token" + secretName: "datasets-server-prod-secrets" @@ -61 +70 @@ secrets: - secretName: "parquet-converter-hf-token" + secretName: "datasets-server-prod-secrets" @@ -64 +73 @@ secrets: - secretName: "webhook-secret" + secretName: "datasets-server-prod-secrets" @@ -67 +76 @@ secrets: - secretName: "spawning-token" + secretName: "datasets-server-prod-secrets" diff --git a/chart/env/staging.yaml b/chart/env/staging.yaml index 54263cb8..3014a6ef 100644 --- a/chart/env/staging.yaml +++ b/chart/env/staging.yaml @@ -49,0 +50,10 @@ secrets: + externalSecret: + enabled: true + secretName: "datasets-server-staging-secrets" + secretStoreName: "datasets-server-ephemeral-secretstore" + parameters: + MONGO_URL: "hub-ephemeral-datasets-server-mongo-url" + HF_TOKEN: "hub-ephemeral-datasets-server-hf-token" + PARQUET_CONVERTER_HF_TOKEN: "hub-ephemeral-datasets-server-parquet-converter-hf-token" + WEBHOOK_SECRET: "hub-ephemeral-datasets-server-webhook-secret" + SPAWNING_TOKEN: "hub-ephemeral-datasets-server-spawning-token" @@ -52,2 +62 @@ secrets: - secretName: "mongo-url" - value: mongo:// + secretName: "datasets-server-staging-secrets" @@ -56 +65 @@ secrets: - secretName: "datasets-server-hf-token" + secretName: "datasets-server-staging-secrets" @@ -59 +68 @@ secrets: - secretName: "parquet-converter-hf-token" + secretName: "datasets-server-staging-secrets" @@ -62,2 +71 @@ secrets: - secretName: "webhook-secret" - value: "" + secretName: "datasets-server-staging-secrets" @@ -66 +74 @@ secrets: - secretName: "spawning-token" + secretName: "datasets-server-staging-secrets" diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index 3d9765fc..6f2d1411 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -49 +49 @@ - key: HF_TOKEN + key: PARQUET_CONVERTER_HF_TOKEN diff --git a/chart/templates/secrets.yaml b/chart/templates/secrets.yaml new file mode 100644 index 00000000..6e371747 --- /dev/null +++ b/chart/templates/secrets.yaml @@ -0,0 +1,20 @@ +{{- if .Values.secrets.externalSecret.enabled }} +apiVersion: "external-secrets.io/v1beta1" +kind: ExternalSecret +metadata: + name: {{ include "name" $ }}-external-secret + namespace: {{ $.Release.Namespace }} +spec: + refreshInterval: 1h + secretStoreRef: + name: {{ .Values.secrets.externalSecret.secretStoreName }} + kind: SecretStore + target: + name: {{ .Values.secrets.externalSecret.secretName }} + data: + {{- range $key, $value := .Values.secrets.externalSecret.parameters }} + - secretKey: {{ $key | quote }} + remoteRef: + key: {{ $value | quote }} + {{- end }} +{{- end }} \ No newline at end of file diff --git a/chart/values.yaml b/chart/values.yaml index 0f99fb5d..4a57fc7a 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -68,0 +69,5 @@ secrets: + externalSecret: + enabled: false + secretName: "" + secretStoreName: "" + parameters: { }
31424f197038003216e23ebae9d68de91a3b1a3b
Sylvain Lesage
2023-06-21T15:06:14
Test concurrency in parquet and info (#1415)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index a0c0f528..1e1c4d9b 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -1005,6 +1004,0 @@ def compute_config_parquet_and_info_response( - # check if the repo exists and get the list of refs - try: - refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE) - except RepositoryNotFoundError as err: - raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err - @@ -1021,0 +1016,2 @@ def compute_config_parquet_and_info_response( + except FileNotFoundError as err: + raise DatasetNotFoundError("The dataset, or the revision, does not exist on the Hub.") from err @@ -1041,11 +1036,0 @@ def compute_config_parquet_and_info_response( - # create the target revision if we managed to get the parquet files and it does not exist yet - # (clone from initial commit to avoid cloning all repo's files) - try: - if all(ref.ref != target_revision for ref in refs.converts): - initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id - committer_hf_api.create_branch( - repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit, exist_ok=True - ) - except RepositoryNotFoundError as err: - raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err - @@ -1053 +1038,2 @@ def compute_config_parquet_and_info_response( - sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300] + sleeps = [1, 1, 1, 1, 1, 10, 10, 10, 10, 100] * 3 + # ^ timeouts after ~7 minutes @@ -1054,0 +1041,15 @@ def compute_config_parquet_and_info_response( + # create the target revision if we managed to get the parquet files and it does not exist yet + # (clone from initial commit to avoid cloning all repo's files) + refs = retry(on=[requests.exceptions.ConnectionError], sleeps=[1, 1, 1, 10, 10])(hf_api.list_repo_refs)( + repo_id=dataset, repo_type=DATASET_TYPE + ) + if all(ref.ref != target_revision for ref in refs.converts): + initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id + committer_hf_api.create_branch( + repo_id=dataset, + branch=target_revision, + repo_type=DATASET_TYPE, + revision=initial_commit, + exist_ok=True, + ) + # commit the parquet files @@ -1064,0 +1066,2 @@ def compute_config_parquet_and_info_response( + # call the API again to get the list of parquet files + target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True) @@ -1066,0 +1070,2 @@ def compute_config_parquet_and_info_response( + except RepositoryNotFoundError as err: + raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err @@ -1068,2 +1072,0 @@ def compute_config_parquet_and_info_response( - # call the API again to get the list of parquet files - target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True) diff --git a/services/worker/tests/fixtures/files.py b/services/worker/tests/fixtures/files.py index f2530413..1bc2a1f3 100644 --- a/services/worker/tests/fixtures/files.py +++ b/services/worker/tests/fixtures/files.py @@ -151,0 +152,34 @@ def dataset_script_with_two_configs_path(tmp_path_factory: pytest.TempPathFactor +# N = 15 +DATASET_SCRIPT_WITH_N_CONFIGS = """ +import os + +import datasets +from datasets import DatasetInfo, BuilderConfig, Features, Split, SplitGenerator, Value + + +class DummyDataset(datasets.GeneratorBasedBuilder): + + BUILDER_CONFIGS = [BuilderConfig(name="config"+str(i)) for i in range(15)] + + def _info(self) -> DatasetInfo: + return DatasetInfo(features=Features({"text": Value("string")})) + + def _split_generators(self, dl_manager): + return [ + SplitGenerator(Split.TRAIN, gen_kwargs={"text": self.config.name}), + ] + + def _generate_examples(self, text, **kwargs): + for i in range(1000): + yield i, {"text": text} +""" + + [email protected](scope="session") +def dataset_script_with_n_configs_path(tmp_path_factory: pytest.TempPathFactory) -> str: + path = str(tmp_path_factory.mktemp("data") / "{dataset_name}.py") + with open(path, "w", newline="") as f: + f.write(DATASET_SCRIPT_WITH_N_CONFIGS) + return path + + diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 34941810..f34ca587 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -262,0 +263,7 @@ def hub_public_legacy_configs(dataset_script_with_two_configs_path: str) -> Iter [email protected](scope="session") +def hub_public_n_configs(dataset_script_with_n_configs_path: str) -> Iterator[str]: + repo_id = create_hub_dataset_repo(prefix="n_configs", file_paths=[dataset_script_with_n_configs_path]) + yield repo_id + delete_hub_dataset_repo(repo_id=repo_id) + + diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 41533372..9de4ac77 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -9 +9,3 @@ from http import HTTPStatus -from typing import Any, Callable, Iterator, List, Optional +from multiprocessing import Pool +from pathlib import Path +from typing import Any, Callable, Iterator, List, Optional, TypedDict @@ -28 +30,2 @@ from libcommon.exceptions import ( -from libcommon.processing_graph import ProcessingGraph +from libcommon.processing_graph import ProcessingGraph, ProcessingStep +from libcommon.queue import Queue @@ -31 +34 @@ from libcommon.simple_cache import CachedArtifactError, upsert_response -from libcommon.utils import Priority +from libcommon.utils import JobInfo, JobParams, Priority @@ -33,0 +37 @@ from worker.config import AppConfig +from worker.job_manager import JobManager @@ -44,0 +49 @@ from worker.job_runners.config.parquet_and_info import ( +from worker.job_runners.dataset.config_names import DatasetConfigNamesJobRunner @@ -45,0 +51 @@ from worker.resources import LibrariesResource +from worker.utils import CompleteJobResult @@ -651,0 +658,139 @@ def test_create_commits( + + +GetDatasetConfigNamesJobRunner = Callable[[str, AppConfig], DatasetConfigNamesJobRunner] + + [email protected] +def get_dataset_config_names_job_runner( + libraries_resource: LibrariesResource, + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetDatasetConfigNamesJobRunner: + def _get_job_runner( + dataset: str, + app_config: AppConfig, + ) -> DatasetConfigNamesJobRunner: + processing_step_name = DatasetConfigNamesJobRunner.get_job_type() + processing_graph = ProcessingGraph( + { + processing_step_name: { + "input_type": "dataset", + "job_runner_version": DatasetConfigNamesJobRunner.get_job_runner_version(), + } + } + ) + return DatasetConfigNamesJobRunner( + job_info={ + "type": DatasetConfigNamesJobRunner.get_job_type(), + "params": { + "dataset": dataset, + "revision": "revision", + "config": None, + "split": None, + }, + "job_id": "job_id", + "priority": Priority.NORMAL, + }, + app_config=app_config, + processing_step=processing_graph.get_processing_step(processing_step_name), + hf_datasets_cache=libraries_resource.hf_datasets_cache, + ) + + return _get_job_runner + + +class JobRunnerArgs(TypedDict): + dataset: str + revision: str + config: str + app_config: AppConfig + tmp_path: Path + + +def launch_job_runner(job_runner_args: JobRunnerArgs) -> CompleteJobResult: + config = job_runner_args["config"] + dataset = job_runner_args["dataset"] + revision = job_runner_args["revision"] + app_config = job_runner_args["app_config"] + tmp_path = job_runner_args["tmp_path"] + job_runner = ConfigParquetAndInfoJobRunner( + job_info=JobInfo( + job_id=f"job_{config}", + type="config-parquet-and-info", + params=JobParams(dataset=dataset, revision=revision, config=config, split=None), + priority=Priority.NORMAL, + ), + app_config=app_config, + processing_step=ProcessingStep( + name="config-parquet-and-info", + input_type="config", + job_runner_version=ConfigParquetAndInfoJobRunner.get_job_runner_version(), + ), + hf_datasets_cache=tmp_path, + ) + return job_runner.compute() + + +def test_concurrency( + hub_public_n_configs: str, + app_config: AppConfig, + tmp_path: Path, + get_dataset_config_names_job_runner: GetDatasetConfigNamesJobRunner, + queue_mongo_resource: QueueMongoResource, + cache_mongo_resource: CacheMongoResource, +) -> None: + """ + Test that multiple job runners (to compute config-parquet-and-info) can run in parallel, + without having conflicts when sending commits to the Hub. + For this test, we need a lot of configs for the same dataset (say 20) and one job runner for each. + Ideally we would try for both quick and slow jobs. + """ + repo_id = hub_public_n_configs + hf_api = HfApi(endpoint=CI_HUB_ENDPOINT, token=CI_USER_TOKEN) + revision = hf_api.dataset_info(repo_id=repo_id, files_metadata=False).sha + if revision is None: + raise ValueError(f"Could not find revision for dataset {repo_id}") + + # fill the cache for the step dataset-config-names, required by the job_runner + # it's a lot of code 😅 + job_info = JobInfo( + job_id="not_used", + type="dataset-config-names", + params=JobParams(dataset=repo_id, revision=revision, config=None, split=None), + priority=Priority.NORMAL, + ) + queue = Queue() + queue.create_jobs([job_info]) + job_info = queue.start_job(job_types_only=["dataset-config-names"]) + job_manager = JobManager( + job_info=job_info, + app_config=app_config, + processing_graph=ProcessingGraph( + { + "dataset-config-names": { + "input_type": "dataset", + "provides_dataset_config_names": True, + "job_runner_version": DatasetConfigNamesJobRunner.get_job_runner_version(), + } + } + ), + job_runner=get_dataset_config_names_job_runner(repo_id, app_config), + ) + job_result = job_manager.run_job() + job_manager.finish(job_result=job_result) + if not job_result["output"]: + raise ValueError("Could not get config names") + configs = [str(config_name["config"]) for config_name in job_result["output"]["content"]["config_names"]] + + # launch the job runners + NUM_JOB_RUNNERS = 10 + with Pool(NUM_JOB_RUNNERS) as pool: + pool.map( + launch_job_runner, + [ + JobRunnerArgs( + dataset=repo_id, revision=revision, config=config, app_config=app_config, tmp_path=tmp_path + ) + for config in configs + ], + )
58c55b64c093929e7f08c527951c50b88879d47b
Sylvain Lesage
2023-06-21T14:25:15
Fix the lock (#1414)
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 1371ee6a..550b1a52 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -254,6 +253,0 @@ class lock(contextlib.AbstractContextManager["lock"]): - try: - Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save( - write_concern={"w": "majority", "fsync": True} - ) - except NotUniqueError: - pass @@ -261 +255 @@ class lock(contextlib.AbstractContextManager["lock"]): - acquired = ( + try: @@ -263 +257,5 @@ class lock(contextlib.AbstractContextManager["lock"]): - job_id=self.job_id, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True} + upsert=True, + write_concern={"w": "majority", "fsync": True}, + read_concern={"level": "majority"}, + job_id=self.job_id, + updated_at=get_datetime(), @@ -265,3 +262,0 @@ class lock(contextlib.AbstractContextManager["lock"]): - > 0 - ) - if acquired: @@ -269,2 +264,3 @@ class lock(contextlib.AbstractContextManager["lock"]): - logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'") - time.sleep(sleep) + except NotUniqueError: + logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'") + time.sleep(sleep) @@ -275 +271,5 @@ class lock(contextlib.AbstractContextManager["lock"]): - job_id=None, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True} + upsert=True, + write_concern={"w": "majority", "fsync": True}, + read_concern={"level": "majority"}, + job_id=None, + updated_at=get_datetime(), diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 18f3e259..82b930fb 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -5,0 +6,2 @@ import os +import random +import time @@ -391,0 +394,5 @@ def test_has_ttl_index_on_finished_at_field() -> None: +def random_sleep() -> None: + MAX_SLEEP_MS = 40 + time.sleep(MAX_SLEEP_MS / 1000 * random.random()) + + @@ -392,0 +400 @@ def increment(tmp_file: Path) -> None: + random_sleep() @@ -394,0 +403 @@ def increment(tmp_file: Path) -> None: + random_sleep() @@ -396,0 +406 @@ def increment(tmp_file: Path) -> None: + random_sleep() @@ -400 +410,2 @@ def locked_increment(tmp_file: Path) -> None: - with lock(key="test_lock", job_id=str(os.getpid())): + sleeps = [0.05, 0.05, 0.05, 1, 1, 1, 1, 1, 1, 5, 5, 5, 5] + with lock(key="test_lock", job_id=str(os.getpid()), sleeps=sleeps): @@ -420 +431 @@ def git_branch_locked_increment(tmp_file: Path) -> None: - sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300] + sleeps = [0.05, 0.05, 0.05, 1, 1, 1, 1, 1, 1, 5, 5, 5, 5]
2848ab4660367d9a74a01ae2f4566c0e243d3009
Albert Villanova del Moral
2023-06-21T14:22:44
Remove unused code from /rows API endpoint (#1394)
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 702fea88..618a6e02 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -4 +3,0 @@ -import asyncio @@ -14 +12,0 @@ from datasets import Features -from fsspec.implementations.http import HTTPFileSystem @@ -41,4 +38,0 @@ logger = logging.getLogger(__name__) -httpfs = HTTPFileSystem() -session = asyncio.run(httpfs.set_session()) - -
178ed7b0ff6199d85a5503bd8804e70cbf02c1f2
Quentin Lhoest
2023-06-21T14:21:56
unblock DFKI-SLT/few-nerd (#1412)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 06b172c5..e7420e4e 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -91 +91 @@ parquetAndInfo: - blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member" + blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
1bb3b274b750e551fb943a729706755d5519467f
Quentin Lhoest
2023-06-21T09:54:11
Retry get_parquet_file_and_size (#1404)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 73dfb0e8..a0c0f528 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -16,0 +17 @@ import numpy as np +import pyarrow as pa @@ -618,0 +620,6 @@ def copy_parquet_files(builder: DatasetBuilder) -> List[CommitOperationCopy]: +class NotAParquetFileError(ValueError): + """When a remote parquet file can't be parsed""" + + pass + + @@ -624,0 +632,14 @@ def get_parquet_file_and_size(url: str, fs: HTTPFileSystem, hf_token: Optional[s +def retry_get_parquet_file_and_size( + url: str, fs: HTTPFileSystem, hf_token: Optional[str] +) -> Tuple[pq.ParquetFile, int]: + try: + sleeps = [1, 1, 1, 10, 10, 10] + pf, size = retry(on=[pa.ArrowInvalid], sleeps=sleeps)(get_parquet_file_and_size)(url, fs, hf_token) + return pf, size + except RuntimeError as err: + if err.__cause__ and isinstance(err.__cause__, pa.ArrowInvalid): + raise NotAParquetFileError(f"Not a parquet file: '{url}'") from err.__cause__ + else: + raise err + + @@ -639 +660 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None: - partial(get_parquet_file_and_size, fs=fs, hf_token=hf_token), + partial(retry_get_parquet_file_and_size, fs=fs, hf_token=hf_token),
dd8f9fc58c9f652578e796d6bd1ffaa887dfc527
Sylvain Lesage
2023-06-20T21:26:59
fix: 🐛 try to ensure only one job can get the lock (#1410)
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 27128ccb..1371ee6a 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -255 +255,3 @@ class lock(contextlib.AbstractContextManager["lock"]): - Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save() + Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save( + write_concern={"w": "majority", "fsync": True} + ) @@ -261 +263 @@ class lock(contextlib.AbstractContextManager["lock"]): - job_id=self.job_id, updated_at=get_datetime() + job_id=self.job_id, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True} @@ -272 +274,3 @@ class lock(contextlib.AbstractContextManager["lock"]): - Lock.objects(key=self.key, job_id=self.job_id).update(job_id=None, updated_at=get_datetime()) + Lock.objects(key=self.key, job_id=self.job_id).update( + job_id=None, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True} + )
86bd831913706c4beecd197b54015e4de2ddb0ab
Sylvain Lesage
2023-06-20T20:35:11
test: 💍 add a test on lock.git_branch (#1409)
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 02b9578e..18f3e259 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -3,0 +4 @@ +import json @@ -415,0 +417,26 @@ def test_lock(tmp_path_factory: pytest.TempPathFactory, queue_mongo_resource: Qu + + +def git_branch_locked_increment(tmp_file: Path) -> None: + sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300] + dataset = "dataset" + branch = "refs/convert/parquet" + with lock.git_branch(dataset=dataset, branch=branch, job_id=str(os.getpid()), sleeps=sleeps): + increment(tmp_file) + + +def test_lock_git_branch(tmp_path_factory: pytest.TempPathFactory, queue_mongo_resource: QueueMongoResource) -> None: + tmp_file = Path(tmp_path_factory.mktemp("test_lock") / "tmp.txt") + tmp_file.touch() + max_parallel_jobs = 5 + num_jobs = 43 + + with Pool(max_parallel_jobs, initializer=queue_mongo_resource.allocate) as pool: + pool.map(git_branch_locked_increment, [tmp_file] * num_jobs) + + expected = num_jobs + with open(tmp_file, "r") as f: + assert int(f.read()) == expected + assert Lock.objects().count() == 1 + assert Lock.objects().get().key == json.dumps({"dataset": "dataset", "branch": "refs/convert/parquet"}) + assert Lock.objects().get().job_id is None + Lock.objects().delete()
f38b99d2d49a0e5deb08014705c61ace13282f88
Sylvain Lesage
2023-06-20T19:08:39
Add tests on create_commits (#1408)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 2769bdd8..73dfb0e8 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -731 +731 @@ def create_commits( - The ma number of operations per commit, to avoid time out errors from the Hub. Defaults to 500. + The max number of operations per commit, to avoid time out errors from the Hub. Defaults to 500. diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index da77c3fc..34941810 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -621 +621 @@ def hub_reponses_empty(hub_public_empty: str) -> HubDatasetTest: -def hub_reponses_public(hub_public_csv: str) -> HubDatasetTest: +def hub_responses_public(hub_public_csv: str) -> HubDatasetTest: diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 2121d53d..41533372 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -17 +17 @@ from datasets import Audio, Features, Image, Value, load_dataset_builder -from huggingface_hub.hf_api import HfApi +from huggingface_hub.hf_api import CommitOperationAdd, HfApi @@ -35,0 +36 @@ from worker.job_runners.config.parquet_and_info import ( + create_commits, @@ -118 +119 @@ def test_compute( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -120,2 +121,2 @@ def test_compute( - dataset = hub_reponses_public["name"] - config = hub_reponses_public["config_names_response"]["config_names"][0]["config"] + dataset = hub_responses_public["name"] + config = hub_responses_public["config_names_response"]["config_names"][0]["config"] @@ -126 +127 @@ def test_compute( - content=hub_reponses_public["config_names_response"], + content=hub_responses_public["config_names_response"], @@ -134 +135 @@ def test_compute( - assert_content_is_equal(content, hub_reponses_public["parquet_and_info_response"]) + assert_content_is_equal(content, hub_responses_public["parquet_and_info_response"]) @@ -446 +447 @@ def test_compute_splits_response_simple_csv_ok( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -454 +455 @@ def test_compute_splits_response_simple_csv_ok( - hub_datasets = {"public": hub_reponses_public, "audio": hub_reponses_audio, "gated": hub_reponses_gated} + hub_datasets = {"public": hub_responses_public, "audio": hub_reponses_audio, "gated": hub_reponses_gated} @@ -531 +532 @@ def test_compute_splits_response_simple_csv_error_2( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -538,2 +539,2 @@ def test_compute_splits_response_simple_csv_error_2( - dataset = hub_reponses_public["name"] - config_names_response = hub_reponses_public["config_names_response"] + dataset = hub_responses_public["name"] + config_names_response = hub_responses_public["config_names_response"] @@ -559 +560 @@ def test_previous_step_error( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -562,2 +563,2 @@ def test_previous_step_error( - dataset = hub_reponses_public["name"] - config = hub_reponses_public["config_names_response"]["config_names"][0]["config"] + dataset = hub_responses_public["name"] + config = hub_responses_public["config_names_response"]["config_names"][0]["config"] @@ -615,0 +617,35 @@ def test_get_writer_batch_size(ds_info: datasets.info.DatasetInfo, has_big_chunk + + [email protected]( + "max_operations_per_commit,use_parent_commit,expected_num_commits", + [(2, False, 1), (1, False, 2), (2, True, 1), (1, True, 2)], +) +def test_create_commits( + hub_public_legacy_configs: str, max_operations_per_commit: int, use_parent_commit: bool, expected_num_commits: int +) -> None: + NUM_FILES = 2 + repo_id = hub_public_legacy_configs + hf_api = HfApi(endpoint=CI_HUB_ENDPOINT, token=CI_USER_TOKEN) + if use_parent_commit: + target_dataset_info = hf_api.dataset_info(repo_id=repo_id, files_metadata=False) + parent_commit = target_dataset_info.sha + else: + parent_commit = None + directory = f".test_create_commits_{max_operations_per_commit}_{use_parent_commit}" + operations: List[CommitOperationAdd] = [ + CommitOperationAdd(path_in_repo=f"{directory}/file{i}.txt", path_or_fileobj=f"content{i}".encode("UTF-8")) + for i in range(NUM_FILES) + ] + commit_infos = create_commits( + hf_api=hf_api, + repo_id=repo_id, + operations=operations, + commit_message="test", + max_operations_per_commit=max_operations_per_commit, + parent_commit=parent_commit, + ) + assert len(commit_infos) == expected_num_commits + # check that the files were created + filenames = hf_api.list_repo_files(repo_id=repo_id, repo_type="dataset") + for i in range(NUM_FILES): + assert f"{directory}/file{i}.txt" in filenames diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index 56e28c7e..2e5f3215 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -91 +91 @@ def test_compute_split_names_from_streaming_response( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -105 +105 @@ def test_compute_split_names_from_streaming_response( - "public": hub_reponses_public, + "public": hub_responses_public, diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py index a94427f4..dc6aafe2 100644 --- a/services/worker/tests/job_runners/dataset/test_config_names.py +++ b/services/worker/tests/job_runners/dataset/test_config_names.py @@ -113 +113 @@ def test_compute_splits_response_simple_csv( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -127 +127 @@ def test_compute_splits_response_simple_csv( - "public": hub_reponses_public, + "public": hub_responses_public, diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index fd2dd64b..9f255de8 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -116 +116 @@ def test_number_rows( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -140 +140 @@ def test_number_rows( - "public": hub_reponses_public, + "public": hub_responses_public, diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py index e3fa31ac..8c8d3d74 100644 --- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py @@ -176 +176 @@ def test_compute( - hub_reponses_public: HubDatasetTest, + hub_responses_public: HubDatasetTest, @@ -185 +185 @@ def test_compute( - hub_datasets = {"public": hub_reponses_public, "spawning_opt_in_out": hub_reponses_spawning_opt_in_out} + hub_datasets = {"public": hub_responses_public, "spawning_opt_in_out": hub_reponses_spawning_opt_in_out}
ca55a35ebd4c9b26af7ea19508bf773e3f953739
Andrea Francis Soria Jimenez
2023-06-20T18:43:07
Rename parent job runners (#1398)
diff --git a/services/worker/src/worker/job_runners/_cached_directory_job_runner.py b/services/worker/src/worker/job_runners/_job_runner_with_cache.py similarity index 98% rename from services/worker/src/worker/job_runners/_cached_directory_job_runner.py rename to services/worker/src/worker/job_runners/_job_runner_with_cache.py index 1ff02d16..1eca7d3b 100644 --- a/services/worker/src/worker/job_runners/_cached_directory_job_runner.py +++ b/services/worker/src/worker/job_runners/_job_runner_with_cache.py @@ -19 +19 @@ from worker.job_runner import JobRunner -class CacheDirectoryJobRunner(JobRunner): +class JobRunnerWithCache(JobRunner): diff --git a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py b/services/worker/src/worker/job_runners/_job_runner_with_datasets_cache.py similarity index 91% rename from services/worker/src/worker/job_runners/_datasets_based_job_runner.py rename to services/worker/src/worker/job_runners/_job_runner_with_datasets_cache.py index f07c0da9..dc007a02 100644 --- a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py +++ b/services/worker/src/worker/job_runners/_job_runner_with_datasets_cache.py @@ -13 +13 @@ from worker.config import AppConfig -from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner +from worker.job_runners._job_runner_with_cache import JobRunnerWithCache @@ -16 +16 @@ from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRun -class DatasetsBasedJobRunner(CacheDirectoryJobRunner): +class JobRunnerWithDatasetsCache(JobRunnerWithCache): diff --git a/services/worker/src/worker/job_runners/config/config_job_runner.py b/services/worker/src/worker/job_runners/config/config_job_runner.py index 75723d36..b18d09c2 100644 --- a/services/worker/src/worker/job_runners/config/config_job_runner.py +++ b/services/worker/src/worker/job_runners/config/config_job_runner.py @@ -11 +11,3 @@ from worker.config import AppConfig -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.job_runners._job_runner_with_datasets_cache import ( + JobRunnerWithDatasetsCache, +) @@ -30 +32 @@ class ConfigJobRunner(DatasetJobRunner): -class ConfigCachedJobRunner(DatasetsBasedJobRunner, ConfigJobRunner): +class ConfigJobRunnerWithDatasetsCache(JobRunnerWithDatasetsCache, ConfigJobRunner): @@ -38 +40 @@ class ConfigCachedJobRunner(DatasetsBasedJobRunner, ConfigJobRunner): - DatasetsBasedJobRunner.__init__( + JobRunnerWithDatasetsCache.__init__( diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index d75e9e55..2769bdd8 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -77 +77 @@ from worker.config import AppConfig, ParquetAndInfoConfig -from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner +from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatasetsCache @@ -1074 +1074 @@ def compute_config_parquet_and_info_response( -class ConfigParquetAndInfoJobRunner(ConfigCachedJobRunner): +class ConfigParquetAndInfoJobRunner(ConfigJobRunnerWithDatasetsCache): diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py index 8208dece..6c4ddc10 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py @@ -20 +20 @@ from libcommon.exceptions import ( -from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner +from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatasetsCache @@ -80 +80 @@ def compute_split_names_from_streaming_response( -class ConfigSplitNamesFromStreamingJobRunner(ConfigCachedJobRunner): +class ConfigSplitNamesFromStreamingJobRunner(ConfigJobRunnerWithDatasetsCache): diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py index a3a80354..faf74109 100644 --- a/services/worker/src/worker/job_runners/dataset/config_names.py +++ b/services/worker/src/worker/job_runners/dataset/config_names.py @@ -17 +17,3 @@ from libcommon.exceptions import ( -from worker.job_runners.dataset.dataset_job_runner import DatasetCachedJobRunner +from worker.job_runners.dataset.dataset_job_runner import ( + DatasetJobRunnerWithDatasetsCache, +) @@ -83 +85 @@ def compute_config_names_response( -class DatasetConfigNamesJobRunner(DatasetCachedJobRunner): +class DatasetConfigNamesJobRunner(DatasetJobRunnerWithDatasetsCache): diff --git a/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py b/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py index a21f9e14..693e01a3 100644 --- a/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py +++ b/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py @@ -12 +12,3 @@ from worker.job_runner import JobRunner -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.job_runners._job_runner_with_datasets_cache import ( + JobRunnerWithDatasetsCache, +) @@ -30 +32 @@ class DatasetJobRunner(JobRunner): -class DatasetCachedJobRunner(DatasetsBasedJobRunner, DatasetJobRunner): +class DatasetJobRunnerWithDatasetsCache(JobRunnerWithDatasetsCache, DatasetJobRunner): @@ -38 +40 @@ class DatasetCachedJobRunner(DatasetsBasedJobRunner, DatasetJobRunner): - DatasetsBasedJobRunner.__init__( + JobRunnerWithDatasetsCache.__init__( diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index 43221906..a49cc8ea 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -29 +29 @@ from worker.config import AppConfig, FirstRowsConfig -from worker.job_runners.split.split_job_runner import SplitCachedJobRunner +from worker.job_runners.split.split_job_runner import SplitJobRunnerWithDatasetsCache @@ -261 +261 @@ def compute_first_rows_response( -class SplitFirstRowsFromStreamingJobRunner(SplitCachedJobRunner): +class SplitFirstRowsFromStreamingJobRunner(SplitJobRunnerWithDatasetsCache): diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py index 7544df77..dbb20277 100644 --- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py @@ -25 +25 @@ from worker.config import AppConfig, OptInOutUrlsScanConfig -from worker.job_runners.split.split_job_runner import SplitCachedJobRunner +from worker.job_runners.split.split_job_runner import SplitJobRunnerWithDatasetsCache @@ -266 +266 @@ def compute_opt_in_out_urls_scan_response( -class SplitOptInOutUrlsScanJobRunner(SplitCachedJobRunner): +class SplitOptInOutUrlsScanJobRunner(SplitJobRunnerWithDatasetsCache): diff --git a/services/worker/src/worker/job_runners/split/split_job_runner.py b/services/worker/src/worker/job_runners/split/split_job_runner.py index 55a87b28..cca736bd 100644 --- a/services/worker/src/worker/job_runners/split/split_job_runner.py +++ b/services/worker/src/worker/job_runners/split/split_job_runner.py @@ -11 +11,4 @@ from worker.config import AppConfig -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.job_runners._job_runner_with_cache import JobRunnerWithCache +from worker.job_runners._job_runner_with_datasets_cache import ( + JobRunnerWithDatasetsCache, +) @@ -30 +33 @@ class SplitJobRunner(ConfigJobRunner): -class SplitCachedJobRunner(DatasetsBasedJobRunner, SplitJobRunner): +class SplitJobRunnerWithDatasetsCache(JobRunnerWithDatasetsCache, SplitJobRunner): @@ -38 +41 @@ class SplitCachedJobRunner(DatasetsBasedJobRunner, SplitJobRunner): - DatasetsBasedJobRunner.__init__( + JobRunnerWithDatasetsCache.__init__( @@ -50,0 +54,23 @@ class SplitCachedJobRunner(DatasetsBasedJobRunner, SplitJobRunner): + + +class SplitJobRunnerWithCache(JobRunnerWithCache, SplitJobRunner): + def __init__( + self, + job_info: JobInfo, + app_config: AppConfig, + processing_step: ProcessingStep, + cache_directory: Path, + ) -> None: + JobRunnerWithCache.__init__( + self, + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + cache_directory=cache_directory, + ) + SplitJobRunner.__init__( + self, + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + ) diff --git a/services/worker/tests/job_runners/test__cached_directory_job_runner.py b/services/worker/tests/job_runners/test__job_runner_with_cache.py similarity index 97% rename from services/worker/tests/job_runners/test__cached_directory_job_runner.py rename to services/worker/tests/job_runners/test__job_runner_with_cache.py index a8a83a9b..db580ea6 100644 --- a/services/worker/tests/job_runners/test__cached_directory_job_runner.py +++ b/services/worker/tests/job_runners/test__job_runner_with_cache.py @@ -14 +14 @@ from worker.config import AppConfig -from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner +from worker.job_runners._job_runner_with_cache import JobRunnerWithCache @@ -21 +21 @@ from ..fixtures.hub import get_default_config_split -class DummyJobRunner(CacheDirectoryJobRunner): +class DummyJobRunner(JobRunnerWithCache): diff --git a/services/worker/tests/job_runners/test__datasets_based_job_runner.py b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py similarity index 96% rename from services/worker/tests/job_runners/test__datasets_based_job_runner.py rename to services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py index 207200d8..30ea2795 100644 --- a/services/worker/tests/job_runners/test__datasets_based_job_runner.py +++ b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py @@ -14 +14,3 @@ from worker.config import AppConfig -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.job_runners._job_runner_with_datasets_cache import ( + JobRunnerWithDatasetsCache, +) @@ -21 +23 @@ from ..fixtures.hub import get_default_config_split -class DummyJobRunner(DatasetsBasedJobRunner): +class DummyJobRunner(JobRunnerWithDatasetsCache):
9643c233e12765ed83373e8cf56a64336674196b
Andrea Francis Soria Jimenez
2023-06-20T15:19:59
Modify TTL index condition (#1405)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 06412c36..4b6dba13 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -224,0 +225,7 @@ class MigrationsCollector: + MigrationQueueDeleteTTLIndex( + version="202306201100", + description=( + "delete the TTL index on the 'finished_at' field in the queue database to update its TTL condition" + ), + field_name="finished_at", + ), diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index 2c1d062d..09cb6301 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -4 +3,0 @@ -import pytest @@ -120 +118,0 @@ def test_metrics_deletion_migration(mongo_host: str) -> None: [email protected](reason="temporaly disabled because of index removal") diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index b1c5e80c..27128ccb 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -26,0 +27 @@ from libcommon.constants import ( + QUEUE_TTL_SECONDS, @@ -137,0 +139,5 @@ class Job(Document): + { + "fields": ["finished_at"], + "expireAfterSeconds": QUEUE_TTL_SECONDS, + "partialFilterExpression": {"status": {"$in": [Status.SUCCESS, Status.ERROR, Status.CANCELLED]}}, + }, diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 5262acd1..02b9578e 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -379 +378,0 @@ def test_queue_get_zombies() -> None: [email protected](reason="temporaly disabled because of index removal")
82f3b7393b0b57bd3cf5fc7599e1dad9141632e6
Andrea Francis Soria Jimenez
2023-06-20T14:59:07
Temporaly remove TTL index (#1403)
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index 09cb6301..2c1d062d 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -3,0 +4 @@ +import pytest @@ -118,0 +120 @@ def test_metrics_deletion_migration(mongo_host: str) -> None: [email protected](reason="temporaly disabled because of index removal") diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 6c233cab..b1c5e80c 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -27 +26,0 @@ from libcommon.constants import ( - QUEUE_TTL_SECONDS, @@ -139 +137,0 @@ class Job(Document): - {"fields": ["finished_at"], "expireAfterSeconds": QUEUE_TTL_SECONDS}, diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 02b9578e..5262acd1 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -378,0 +379 @@ def test_queue_get_zombies() -> None: [email protected](reason="temporaly disabled because of index removal")
bf7cc0b3b0c4315d299603217683b9fb1050de5e
Quentin Lhoest
2023-06-20T13:52:22
increase max job duration (#1395)
diff --git a/chart/values.yaml b/chart/values.yaml index 5294d83c..0f99fb5d 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -131 +131 @@ worker: - maxJobDurationSeconds: 1200 + maxJobDurationSeconds: 2400
d017e7c79a4196535d38743b163293d2d296aa1c
Sylvain Lesage
2023-06-20T13:28:26
fix: 🐛 split the default value to get a list of strings (#1401)
diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py index e1717543..1bbab28c 100644 --- a/libs/libcommon/src/libcommon/orchestrator.py +++ b/libs/libcommon/src/libcommon/orchestrator.py @@ -330 +330 @@ class DatasetBackfillPlan(Plan): - self.error_codes_to_retry = list(ERROR_CODES_TO_RETRY) + self.error_codes_to_retry = ERROR_CODES_TO_RETRY.split(",")
7626cd0dfff2d5d70618893636a2e63e85b4ffa4
Sylvain Lesage
2023-06-20T12:36:44
Raise retryable error on hfhubhttperror (#1393)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 163a431d..06b172c5 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -135 +135 @@ backfill: - error_codes_to_retry: "" + error_codes_to_retry: "CreateCommitError,LockedDatasetTimeoutError" diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index fc0dd91e..cd41126f 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -43,0 +44,2 @@ PARQUET_REVISION = "refs/convert/parquet" + +ERROR_CODES_TO_RETRY = "CreateCommitError,LockedDatasetTimeoutError" diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py index 672111e1..f9de8f3f 100644 --- a/libs/libcommon/src/libcommon/exceptions.py +++ b/libs/libcommon/src/libcommon/exceptions.py @@ -76,0 +77 @@ CacheableErrorCode = Literal[ + "CreateCommitError", @@ -144,0 +146,7 @@ class ConfigNamesError(CacheableError): +class CreateCommitError(CacheableError): + """Raised when a commit could not be created on the Hub.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "CreateCommitError", cause, False) + + diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py index fb8518bd..e1717543 100644 --- a/libs/libcommon/src/libcommon/orchestrator.py +++ b/libs/libcommon/src/libcommon/orchestrator.py @@ -10,0 +11 @@ import pandas as pd +from libcommon.constants import ERROR_CODES_TO_RETRY @@ -327,0 +329,2 @@ class DatasetBackfillPlan(Plan): + if self.error_codes_to_retry is None: + self.error_codes_to_retry = list(ERROR_CODES_TO_RETRY) diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 22460dc6..d75e9e55 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -50,0 +51 @@ from libcommon.exceptions import ( + CreateCommitError, @@ -745,2 +746,2 @@ def create_commits( - [`huggingface_hub.utils.HfHubHTTPError`]: - If another commit happened after `parent_commit` or in between the commits. + [`libcommon.exceptions.CreateCommitError`]: + If one of the commits could not be created on the Hub. @@ -748,0 +750 @@ def create_commits( + sleeps = [1, 1, 1, 10, 10, 10] @@ -752,3 +754,3 @@ def create_commits( - retry_create_commit = retry(on=[HfHubHTTPError], sleeps=[1, 1, 1, 10, 10, 10])(hf_api.create_commit) - commit_infos.append( - retry_create_commit( + retry_create_commit = retry(on=[HfHubHTTPError], sleeps=sleeps)(hf_api.create_commit) + try: + commit_info = retry_create_commit( @@ -760 +762 @@ def create_commits( - parent_commit=parent_commit if not commit_infos else commit_infos[-1].oid, + parent_commit=commit_infos[-1].oid if commit_infos else parent_commit, @@ -762 +764,11 @@ def create_commits( - ) + except RuntimeError as e: + if e.__cause__ and isinstance(e.__cause__, HfHubHTTPError): + raise CreateCommitError( + message=( + f"Commit {commit_idx}/{len(offsets)} could not be created on the Hub (after" + f" {len(sleeps)} attempts)." + ), + cause=e.__cause__, + ) from e.__cause__ + raise e + commit_infos.append(commit_info) @@ -813,4 +825,2 @@ def commit_parquet_conversion( - [`RuntimeError`]: If it fails after several retries. - The retry fails can come from: - [`huggingface_hub.utils.HfHubHTTPError`]: - If another commit happened after `parent_commit` or in between the commits. + [`libcommon.exceptions.CreateCommitError`]: + If one of the commits could not be created on the Hub. @@ -911,0 +922,2 @@ def compute_config_parquet_and_info_response( + - [`libcommon.exceptions.CreateCommitError`]: + If one of the commits could not be created on the Hub. diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 7472e39b..b4869d67 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -322 +322 @@ class retry: - raise RuntimeError(f"Give up after {attempt} attempts with {type(last_err)}") from last_err + raise RuntimeError(f"Give up after {attempt} attempts. The last one raised {type(last_err)}") from last_err
8d2f9ab14c3771e64d8341b879f36ef0ec96b7e3
Andrea Francis Soria Jimenez
2023-06-20T12:21:31
New parent job runner for cached data (#1388)
diff --git a/services/worker/src/worker/job_runners/_cached_directory_job_runner.py b/services/worker/src/worker/job_runners/_cached_directory_job_runner.py new file mode 100644 index 00000000..1ff02d16 --- /dev/null +++ b/services/worker/src/worker/job_runners/_cached_directory_job_runner.py @@ -0,0 +1,63 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2023 The HuggingFace Authors. + +import json +import random +import re +from hashlib import sha1 +from pathlib import Path +from typing import Optional + +from libcommon.processing_graph import ProcessingStep +from libcommon.storage import init_dir, remove_dir +from libcommon.utils import JobInfo + +from worker.config import AppConfig +from worker.job_runner import JobRunner + + +class CacheDirectoryJobRunner(JobRunner): + """Base class for job runners that use a temporary cache directory.""" + + base_cache_directory: Path + cache_subdirectory: Optional[Path] = None + + def __init__( + self, + job_info: JobInfo, + app_config: AppConfig, + processing_step: ProcessingStep, + cache_directory: Path, + ) -> None: + super().__init__( + job_info=job_info, + app_config=app_config, + processing_step=processing_step, + ) + self.base_cache_directory = cache_directory + + def get_cache_subdirectory(self, digits: int = 14) -> str: + random_str = f"{random.randrange(10**(digits - 1), 10**digits)}" # nosec B311 + # TODO: Refactor, need a way to generate payload based only on provided params + payload = ( + random_str, + self.get_job_type(), + self.job_info["params"]["dataset"], + self.job_info["params"]["config"], + self.job_info["params"]["split"], + ) + hash_suffix = sha1(json.dumps(payload, sort_keys=True).encode(), usedforsecurity=False).hexdigest()[:8] + prefix = f"{random_str}-{self.get_job_type()}-{self.job_info['params']['dataset']}"[:64] + subdirectory = f"{prefix}-{hash_suffix}" + return "".join([c if re.match(r"[\w-]", c) else "-" for c in subdirectory]) + + def pre_compute(self) -> None: + new_directory = self.base_cache_directory / self.get_cache_subdirectory() + self.cache_subdirectory = Path(init_dir(new_directory)) + + def post_compute(self) -> None: + # empty the cache after the job to save storage space + previous_cache = self.cache_subdirectory + if previous_cache is not None: + remove_dir(previous_cache) + self.cache_subdirectory = None diff --git a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py index 310b4bde..f07c0da9 100644 --- a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py +++ b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py @@ -4 +3,0 @@ -import json @@ -6,3 +4,0 @@ import logging -import random -import re -from hashlib import sha1 @@ -14 +9,0 @@ from libcommon.processing_graph import ProcessingStep -from libcommon.storage import init_dir, remove_dir @@ -17,2 +12,2 @@ from libcommon.utils import JobInfo -from worker.config import AppConfig, DatasetsBasedConfig -from worker.job_runner import JobRunner +from worker.config import AppConfig +from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner @@ -21 +16 @@ from worker.job_runner import JobRunner -class DatasetsBasedJobRunner(JobRunner): +class DatasetsBasedJobRunner(CacheDirectoryJobRunner): @@ -24,8 +18,0 @@ class DatasetsBasedJobRunner(JobRunner): - datasets_based_config: DatasetsBasedConfig - base_datasets_cache: Path - - # the datasets library cache directories (for data, downloads, extraction, NOT for modules) - # the job runner should have only one running job at the same time, then it should - # be safe to use a global variable (and to set the datasets cache globally) - datasets_cache: Optional[Path] = None - @@ -42,0 +30 @@ class DatasetsBasedJobRunner(JobRunner): + cache_directory=hf_datasets_cache, @@ -44,17 +31,0 @@ class DatasetsBasedJobRunner(JobRunner): - self.datasets_based_config = app_config.datasets_based - self.base_datasets_cache = hf_datasets_cache - - def get_cache_subdirectory(self, digits: int = 14) -> str: - random_str = f"{random.randrange(10**(digits - 1), 10**digits)}" # nosec B311 - # TODO: Refactor, need a way to generate payload based only on provided params - payload = ( - random_str, - self.get_job_type(), - self.job_info["params"]["dataset"], - self.job_info["params"]["config"], - self.job_info["params"]["split"], - ) - hash_suffix = sha1(json.dumps(payload, sort_keys=True).encode(), usedforsecurity=False).hexdigest()[:8] - prefix = f"{random_str}-{self.get_job_type()}-{self.job_info['params']['dataset']}"[:64] - subdirectory = f"{prefix}-{hash_suffix}" - return "".join([c if re.match(r"[\w-]", c) else "-" for c in subdirectory]) @@ -62,3 +33,2 @@ class DatasetsBasedJobRunner(JobRunner): - def set_datasets_cache(self, datasets_cache: Path) -> None: - self.datasets_cache = Path(init_dir(datasets_cache)) - datasets.config.HF_DATASETS_CACHE = self.datasets_cache + def set_datasets_cache(self, cache_subdirectory: Optional[Path]) -> None: + datasets.config.HF_DATASETS_CACHE = cache_subdirectory @@ -73,15 +42,0 @@ class DatasetsBasedJobRunner(JobRunner): - def unset_datasets_cache(self) -> None: - previous_datasets_cache = self.datasets_cache - self.set_datasets_cache(self.base_datasets_cache) - if previous_datasets_cache is not None and self.datasets_cache != previous_datasets_cache: - remove_dir(previous_datasets_cache) - logging.debug(f"temporary datasets data cache deleted: {previous_datasets_cache}") - self.datasets_cache = None - - def set_cache(self) -> None: - cache_subdirectory = self.get_cache_subdirectory() - self.set_datasets_cache(self.base_datasets_cache / cache_subdirectory) - - def unset_cache(self) -> None: - self.unset_datasets_cache() - @@ -89 +44,2 @@ class DatasetsBasedJobRunner(JobRunner): - self.set_cache() + super().pre_compute() + self.set_datasets_cache(self.cache_subdirectory) @@ -92,2 +48,2 @@ class DatasetsBasedJobRunner(JobRunner): - # empty the cache after the job to save storage space - self.unset_cache() + super().post_compute() + self.set_datasets_cache(self.base_cache_directory) diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__cached_directory_job_runner.py similarity index 68% rename from services/worker/tests/job_runners/test__datasets_based_worker.py rename to services/worker/tests/job_runners/test__cached_directory_job_runner.py index a57c7578..a8a83a9b 100644 --- a/services/worker/tests/job_runners/test__datasets_based_worker.py +++ b/services/worker/tests/job_runners/test__cached_directory_job_runner.py @@ -2 +2 @@ -# Copyright 2022 The HuggingFace Authors. +# Copyright 2023 The HuggingFace Authors. @@ -8 +7,0 @@ from typing import Callable, Optional -import datasets.config @@ -15 +14 @@ from worker.config import AppConfig -from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner @@ -22 +21 @@ from ..fixtures.hub import get_default_config_split -class DummyJobRunner(DatasetsBasedJobRunner): +class DummyJobRunner(CacheDirectoryJobRunner): @@ -26,2 +24,0 @@ class DummyJobRunner(DatasetsBasedJobRunner): - # ^ borrowing the type, so that the processing step exists and the job runner can be initialized - # refactoring libcommon.processing_graph might help avoiding this @@ -75 +72 @@ def get_job_runner( - hf_datasets_cache=libraries_resource.hf_datasets_cache, + cache_directory=libraries_resource.hf_datasets_cache, @@ -112,12 +109 @@ def test_get_cache_subdirectory( -def test_set_and_unset_datasets_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: - dataset, config, split = get_default_config_split("dataset") - job_runner = get_job_runner(dataset, config, split, app_config) - base_path = job_runner.base_datasets_cache - dummy_path = base_path / "dummy" - job_runner.set_datasets_cache(dummy_path) - assert_datasets_cache_path(path=dummy_path, exists=True) - job_runner.unset_datasets_cache() - assert_datasets_cache_path(path=base_path, exists=True) - - -def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: +def test_pre_compute_post_compute(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: @@ -126,5 +112,5 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner - datasets_base_path = job_runner.base_datasets_cache - job_runner.set_cache() - assert str(datasets.config.HF_DATASETS_CACHE).startswith(str(datasets_base_path)) - assert "dummy-job-runner-user-dataset" in str(datasets.config.HF_DATASETS_CACHE) - job_runner.unset_cache() + datasets_base_path = job_runner.base_cache_directory + job_runner.pre_compute() + datasets_cache_subdirectory = job_runner.cache_subdirectory + assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=True) + job_runner.post_compute() @@ -131,0 +118 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner + assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=False) @@ -134 +121,2 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner -def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) -> None: +def assert_datasets_cache_path(path: Optional[Path], exists: bool) -> None: + assert path is not None @@ -136,3 +123,0 @@ def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) -> - assert (datasets.config.HF_DATASETS_CACHE == path) is equals - assert (datasets.config.DOWNLOADED_DATASETS_PATH == path / datasets.config.DOWNLOADED_DATASETS_DIR) is equals - assert (datasets.config.EXTRACTED_DATASETS_PATH == path / datasets.config.EXTRACTED_DATASETS_DIR) is equals diff --git a/services/worker/tests/job_runners/test__datasets_based_job_runner.py b/services/worker/tests/job_runners/test__datasets_based_job_runner.py new file mode 100644 index 00000000..207200d8 --- /dev/null +++ b/services/worker/tests/job_runners/test__datasets_based_job_runner.py @@ -0,0 +1,108 @@ +# SPDX-License-Identifier: Apache-2.0 +# Copyright 2022 The HuggingFace Authors. + +from pathlib import Path +from typing import Callable, Optional + +import datasets.config +import pytest +from libcommon.processing_graph import ProcessingGraph +from libcommon.resources import CacheMongoResource, QueueMongoResource +from libcommon.utils import Priority + +from worker.config import AppConfig +from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner +from worker.resources import LibrariesResource +from worker.utils import CompleteJobResult + +from ..fixtures.hub import get_default_config_split + + +class DummyJobRunner(DatasetsBasedJobRunner): + @staticmethod + def get_job_type() -> str: + return "dummy-job-runner" + # ^ borrowing the type, so that the processing step exists and the job runner can be initialized + # refactoring libcommon.processing_graph might help avoiding this + + @staticmethod + def get_job_runner_version() -> int: + return 1 + + def compute(self) -> CompleteJobResult: + return CompleteJobResult({"col1": "a" * 200}) + + +GetJobRunner = Callable[[str, Optional[str], Optional[str], AppConfig], DummyJobRunner] + + [email protected] +def get_job_runner( + libraries_resource: LibrariesResource, + cache_mongo_resource: CacheMongoResource, + queue_mongo_resource: QueueMongoResource, +) -> GetJobRunner: + def _get_job_runner( + dataset: str, + config: Optional[str], + split: Optional[str], + app_config: AppConfig, + ) -> DummyJobRunner: + processing_step_name = DummyJobRunner.get_job_type() + processing_graph = ProcessingGraph( + { + processing_step_name: { + "input_type": "dataset", + "job_runner_version": DummyJobRunner.get_job_runner_version(), + } + } + ) + return DummyJobRunner( + job_info={ + "type": DummyJobRunner.get_job_type(), + "params": { + "dataset": dataset, + "revision": "revision", + "config": config, + "split": split, + }, + "job_id": "job_id", + "priority": Priority.NORMAL, + }, + app_config=app_config, + processing_step=processing_graph.get_processing_step(processing_step_name), + hf_datasets_cache=libraries_resource.hf_datasets_cache, + ) + + return _get_job_runner + + +def test_set_datasets_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset, config, split = get_default_config_split("dataset") + job_runner = get_job_runner(dataset, config, split, app_config) + base_path = job_runner.base_cache_directory + dummy_path = base_path / "dummy" + job_runner.set_datasets_cache(dummy_path) + assert str(datasets.config.HF_DATASETS_CACHE).startswith(str(dummy_path)) + + +def test_pre_compute_post_compute(app_config: AppConfig, get_job_runner: GetJobRunner) -> None: + dataset, config, split = get_default_config_split("user/dataset") + job_runner = get_job_runner(dataset, config, split, app_config) + datasets_base_path = job_runner.base_cache_directory + job_runner.pre_compute() + datasets_cache_subdirectory = job_runner.cache_subdirectory + assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=True) + assert str(datasets.config.HF_DATASETS_CACHE).startswith(str(datasets_base_path)) + assert "dummy-job-runner-user-dataset" in str(datasets.config.HF_DATASETS_CACHE) + job_runner.post_compute() + assert_datasets_cache_path(path=datasets_base_path, exists=True) + assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=False, equals=False) + + +def assert_datasets_cache_path(path: Optional[Path], exists: bool, equals: bool = True) -> None: + assert path is not None + assert path.exists() is exists + assert (datasets.config.HF_DATASETS_CACHE == path) is equals + assert (datasets.config.DOWNLOADED_DATASETS_PATH == path / datasets.config.DOWNLOADED_DATASETS_DIR) is equals + assert (datasets.config.EXTRACTED_DATASETS_PATH == path / datasets.config.EXTRACTED_DATASETS_DIR) is equals
b7f13c8ee5a1b774c75874095b368a79309a226f
Sylvain Lesage
2023-06-19T12:35:59
feat: 🎸 10x the size of supported images (#1392)
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index bc187b6f..7472e39b 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -37 +37 @@ from libcommon.utils import orjson_dumps -MAX_IMAGE_PIXELS = 1_000_000_000 +MAX_IMAGE_PIXELS = 10_000_000_000
cb563d9451eb6898c7be87a34bd2f0262fd4330a
Sylvain Lesage
2023-06-19T12:12:18
Rename dev to staging, and use staging mongodb cluster (#1383)
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml index 878e3e3b..d1394601 100644 --- a/.github/workflows/cd.yml +++ b/.github/workflows/cd.yml @@ -79,2 +79,2 @@ jobs: - - name: Lint chart with dev values - run: helm lint --values env/dev.yaml + - name: Lint chart with staging values + run: helm lint --values env/staging.yaml @@ -86 +86 @@ jobs: - deploy-dev-and-prod: + deploy-staging-and-prod: diff --git a/.github/workflows/chart-pr.yml b/.github/workflows/chart-pr.yml index cd5789bc..93880b23 100644 --- a/.github/workflows/chart-pr.yml +++ b/.github/workflows/chart-pr.yml @@ -22,2 +22,2 @@ jobs: - - name: Lint chart with dev values - run: helm lint --values env/dev.yaml + - name: Lint chart with staging values + run: helm lint --values env/staging.yaml diff --git a/chart/Makefile b/chart/Makefile index 555a89f3..57716f53 100644 --- a/chart/Makefile +++ b/chart/Makefile @@ -1,2 +0,0 @@ -K8S_NAMESPACE := datasets-server - @@ -7,36 +4,0 @@ init: -.PHONY: uninstall -uninstall: - helm uninstall $(ENV) -n $(K8S_NAMESPACE) - -.PHONY: diff -diff: - helm diff upgrade --install $(ENV) . --values env/$(ENV).yaml -n $(K8S_NAMESPACE) - -.PHONY: upgrade -upgrade: - helm upgrade --install $(ENV) . --values env/$(ENV).yaml -n $(K8S_NAMESPACE) - -.PHONY: diff-dev -diff-dev: - @make diff ENV=dev - -.PHONY: uninstall-dev -uninstall-dev: - @make uninstall ENV=dev - -.PHONY: upgrade-dev -upgrade-dev: - @make upgrade ENV=dev - -.PHONY: diff-prod -diff-prod: - @make diff ENV=prod - -.PHONY: uninstall-prod -uninstall-prod: - @make uninstall ENV=prod - -.PHONY: upgrade-prod -upgrade-prod: - @make upgrade ENV=prod - diff --git a/chart/README.md b/chart/README.md index 26865e85..aa378210 100644 --- a/chart/README.md +++ b/chart/README.md @@ -16,18 +16 @@ Note that this Helm chart is used to manage the deployment of the `datasets-serv -To deploy to the `hub-ephemeral` Kubernetes cluster, ensure to first: - -- install the tools (aws, kubectl, helm) -- authenticate with AWS -- select the `hub-ephemeral` cluster - -Dry run: - -```shell -make init -make diff-dev -``` - -Deploy: - -```shell -make upgrade-dev -``` +To deploy, go to https://cd.internal.huggingface.tech/applications. diff --git a/chart/env/dev.yaml b/chart/env/staging.yaml similarity index 91% rename from chart/env/dev.yaml rename to chart/env/staging.yaml index b66d3f88..54263cb8 100644 --- a/chart/env/dev.yaml +++ b/chart/env/staging.yaml @@ -7,4 +6,0 @@ global: - imageRegistry: "" - imagePullSecrets: [] - privateHub: - enabled: false @@ -12,0 +9 @@ global: + # ^ the domain contains "dev", not "staging". We don't change for now. @@ -54 +51 @@ secrets: - fromSecret: false + fromSecret: true @@ -78 +75 @@ mongodb: - enabled: true + enabled: false @@ -87,0 +85,3 @@ log: +firstRows: + maxBytes: "200_000" + @@ -104,0 +105,4 @@ cacheMaintenance: + log: + level: "debug" + backfill: + error_codes_to_retry: "" @@ -117,2 +121,2 @@ metricsCollector: - schedule: "*/5 * * * *" - # every five minutes + schedule: "*/2 * * * *" + # every two minutes @@ -122 +126 @@ metricsCollector: - cpu: 0 + cpu: 1 @@ -124 +128,2 @@ metricsCollector: - cpu: 0 + cpu: 1 + memory: "512Mi" @@ -163,2 +168,2 @@ ingress: - alb.ingress.kubernetes.io/load-balancer-name: "hub-datasets-server-dev" - alb.ingress.kubernetes.io/tags: "Env=dev,Project=datasets-server,Terraform=true" + alb.ingress.kubernetes.io/load-balancer-name: "hub-datasets-server-staging" + alb.ingress.kubernetes.io/tags: "Env=staging,Project=datasets-server,Terraform=true"
170b1c6be8bbfc0347056ac5b89e9e6b816ecc6c
Sylvain Lesage
2023-06-19T11:21:39
fix: 🐛 support bigger images (#1387)
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 8d817352..5f59e6b5 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -4695,2 +4694,0 @@ files = [ - {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:9c9b14fbb73ec4cb0f209722a1489020fd8614c92ae22589f2309c48cefdf21f"}, - {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_x86_64.whl", hash = "sha256:6a54539bd076746f69ae8bef7282f981674fe4dbf59c3a84c4af86ae6bae9d5c"}, @@ -5012,0 +5011,12 @@ test = ["black (>=22.3.0,<23.0.0)", "coverage (>=5.2,<6.0)", "isort (>=5.0.6,<6. +[[package]] +name = "types-pillow" +version = "9.5.0.4" +description = "Typing stubs for Pillow" +category = "dev" +optional = false +python-versions = "*" +files = [ + {file = "types-Pillow-9.5.0.4.tar.gz", hash = "sha256:f1b6af47abd151847ee25911ffeba784899bc7dc7f9eba8ca6a5aac522b012ef"}, + {file = "types_Pillow-9.5.0.4-py3-none-any.whl", hash = "sha256:69427d9fa4320ff6e30f00fb9c0dd71185dc0a16de4757774220104759483466"}, +] + @@ -5584 +5594 @@ python-versions = "3.9.15" -content-hash = "ad3c5c34e9ea75e4cb4394930bb35c5afdff65fe03c5f51b8cbebbea37f62f1d" +content-hash = "109af95b92d54671ee4da9ddd322e3b03b5f9882ec38c9ace064afcbe37b5a2b" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index 9fabc531..f5feb938 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -22,0 +23 @@ lxml = "^4.9.2" +mirakuru = "^2.4.2" @@ -28 +28,0 @@ pdf2image = "^1.16.2" -pyarrow = "^11.0.0" @@ -29,0 +30 @@ py7zr = "^0.20.4" +pyarrow = "^11.0.0" @@ -47 +47,0 @@ wget = "^3.2" -mirakuru = "^2.4.2" @@ -59,0 +60 @@ pytest-cov = "^2.12.1" +types-pillow = "^9.5.0.4" diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 14d209a2..bc187b6f 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -24,0 +25 @@ from typing import ( +import PIL @@ -35,0 +37,3 @@ from libcommon.utils import orjson_dumps +MAX_IMAGE_PIXELS = 1_000_000_000 +# ^ see https://pillow.readthedocs.io/en/stable/reference/Image.html#PIL.Image.MAX_IMAGE_PIXELS + @@ -333,0 +338 @@ def get_rows( + PIL.Image.MAX_IMAGE_PIXELS = MAX_IMAGE_PIXELS
aa4a6c6442ce9f5f40ed664a0afb1f0c401be72d
Bas Krahmer
2023-06-19T10:39:35
Add docker internal to extra_hosts (#1390)
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml index 90a1e9cb..b9ca8e9e 100644 --- a/tools/docker-compose-dev-base.yml +++ b/tools/docker-compose-dev-base.yml @@ -27,0 +28,2 @@ services: + extra_hosts: + - "host.docker.internal:host-gateway"
de349b5df053b8ea1dcc1ed5e626b8de3a223f07
Albert Villanova del Moral
2023-06-19T08:55:12
Fix typo in erro rmessage (#1391)
diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py index b1045c31..a3a80354 100644 --- a/services/worker/src/worker/job_runners/dataset/config_names.py +++ b/services/worker/src/worker/job_runners/dataset/config_names.py @@ -77 +77 @@ def compute_config_names_response( - f"The maximun number of configs allowed is {max_number}, dataset has {number_of_configs} configs." + f"The maximum number of configs allowed is {max_number}, dataset has {number_of_configs} configs."
03d633edfb5f5852c5b2f70e685860509ca9657d
Bas Krahmer
2023-06-19T08:51:24
Fix closing brackets and GH action link (#1389)
diff --git a/DEVELOPER_GUIDE.md b/DEVELOPER_GUIDE.md index d446e6d0..e2101dc5 100644 --- a/DEVELOPER_GUIDE.md +++ b/DEVELOPER_GUIDE.md @@ -54 +54 @@ If you use VSCode, it might be useful to use the ["monorepo" workspace](./.vscod -The repository is structured as a monorepo, with Python libraries and applications in [jobs](./jobs)), [libs](./libs) and [services](./services): +The repository is structured as a monorepo, with Python libraries and applications in [jobs](./jobs), [libs](./libs) and [services](./services): @@ -100 +100 @@ The following environments contain all the modules: reverse proxy, API server, a -The CI checks the quality of the code through a [GitHub action](./.github/workflows/quality.yml). To manually format the code of a job, library, service or worker: +The CI checks the quality of the code through a [GitHub action](./.github/workflows/_quality-python.yml). To manually format the code of a job, library, service or worker: diff --git a/services/reverse-proxy/README.md b/services/reverse-proxy/README.md index 2801bf44..d322fda2 100644 --- a/services/reverse-proxy/README.md +++ b/services/reverse-proxy/README.md @@ -11 +11 @@ The reverse proxy uses nginx: -- it serves the static assets directly (the API also serves them if required, but it's unnecessary to go through starlette for this, and it generates errors in Safari, see [1](https://github.com/encode/starlette/issues/950) and [2](https://developer.apple.com/library/archive/documentation/AppleApplications/Reference/SafariWebContent/CreatingVideoforSafarioniPhone/CreatingVideoforSafarioniPhone.html#//apple_ref/doc/uid/TP40006514-SW6) +- it serves the static assets directly (the API also serves them if required, but it's unnecessary to go through starlette for this, and it generates errors in Safari, see [1](https://github.com/encode/starlette/issues/950) and [2](https://developer.apple.com/library/archive/documentation/AppleApplications/Reference/SafariWebContent/CreatingVideoforSafarioniPhone/CreatingVideoforSafarioniPhone.html#//apple_ref/doc/uid/TP40006514-SW6))
dbd15643a377d9747c21000c364e231cfd6492df
Steven Liu
2023-06-16T16:10:04
[docs] Improvements (#1376)
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml index 0767f5ab..b4dbd129 100644 --- a/docs/source/_toctree.yml +++ b/docs/source/_toctree.yml @@ -6,0 +7,2 @@ + - local: analyze_data + title: Analyze a dataset on the Hub diff --git a/docs/source/analyze_data.mdx b/docs/source/analyze_data.mdx new file mode 100644 index 00000000..e6c1a673 --- /dev/null +++ b/docs/source/analyze_data.mdx @@ -0,0 +1,63 @@ +# Analyze a dataset on the Hub + +[[open-in-colab]] + +In the Quickstart, you were introduced to various endpoints for interacting with datasets on the Hub. One of the most useful ones is the `/parquet` endpoint, which allows you to get a dataset stored on the Hub and analyze it. This is a great way to explore the dataset, and get a better understanding of it's contents. + +To demonstrate, this guide will show you an end-to-end example of how to retrieve a dataset from the Hub and do some basic data analysis with the Pandas library. + +## Get a dataset + +The [Hub](https://huggingface.co/datasets) is home to more than 40,000 datasets across a wide variety of tasks, sizes, and languages. For this example, you'll use the [`codeparrot/codecomplex`](https://huggingface.co/datasets/codeparrot/codecomplex) dataset, but feel free to explore and find another dataset that interests you! The dataset contains Java code from programming competitions, and the time complexity of the code is labeled by a group of algorithm experts. + +Let's say you're interested in the average length of the submitted code as it relates to the time complexity. Here's how you can get started. + +Use the `/parquet` endpoint to convert the dataset to a Parquet file and return the URL to it: + +```py +import requests +API_URL = "https://datasets-server.huggingface.co/parquet?dataset=codeparrot/codecomplex" +def query(): + response = requests.get(API_URL) + return response.json() +data = query() +print(data) +{'parquet_files': + [ + {'dataset': 'codeparrot/codecomplex', 'config': 'codeparrot--codecomplex', 'split': 'train', 'url': 'https://huggingface.co/datasets/codeparrot/codecomplex/resolve/refs%2Fconvert%2Fparquet/codeparrot--codecomplex/json-train.parquet', 'filename': 'json-train.parquet', 'size': 4115908} + ], + 'pending': [], 'failed': [] +} +``` + +## Read dataset with Pandas + +With the URL, you can read the Parquet file into a Pandas DataFrame: + +```py +import pandas as pd + +url = "https://huggingface.co/datasets/codeparrot/codecomplex/resolve/refs%2Fconvert%2Fparquet/codeparrot--codecomplex/json-train.parquet" +df = pd.read_parquet(url) +df.head(5) +``` + +| src | complexity | problem | from | +|--------------------------------------------------:|-----------:|--------------------------------:|-----------:| +| import java.io.*;\nimport java.math.BigInteger... | quadratic | 1179_B. Tolik and His Uncle | CODEFORCES | +| import java.util.Scanner;\n \npublic class pil... | linear | 1197_B. Pillars | CODEFORCES | +| import java.io.BufferedReader;\nimport java.io... | linear | 1059_C. Sequence Transformation | CODEFORCES | +| import java.util.*;\n\nimport java.io.*;\npubl... | linear | 1011_A. Stages | CODEFORCES | +| import java.io.OutputStream;\nimport java.io.I... | linear | 1190_C. Tokitsukaze and Duel | CODEFORCES | + +## Calculate mean code length by time complexity + +Pandas is a powerful library for data analysis; group the dataset by time complexity, apply a function to calculate the average length of the code snippet, and plot the results: + +```py +df.groupby('complexity')['src'].apply(lambda x: x.str.len().mean()).sort_values(ascending=False).plot.barh(color="orange") +``` + +<div class="flex justify-center"> + <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets-server/codecomplex.png"/> +</div> \ No newline at end of file diff --git a/docs/source/quick_start.mdx b/docs/source/quick_start.mdx index d784bb20..b798b8e8 100644 --- a/docs/source/quick_start.mdx +++ b/docs/source/quick_start.mdx @@ -2,0 +3,2 @@ +[[open-in-colab]] + @@ -89,0 +92,7 @@ curl https://datasets-server.huggingface.co/is-valid?dataset=rotten_tomatoes \ +You'll see the following error if you're trying to access a gated dataset without providing your user token: + +```py +print(data) +{'error': 'The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication.'} +``` + @@ -130,0 +140,16 @@ curl https://datasets-server.huggingface.co/valid \ +This returns a list of all the datasets that load without an error: + +```py +print(data) +{ + "valid": [ + "0n1xus/codexglue", + "0n1xus/pytorrent-standalone", + "0x7194633/rupile", + "51la5/keyword-extraction", + ..., + ..., + ] +} +``` + @@ -169,0 +195,7 @@ curl https://datasets-server.huggingface.co/is-valid?dataset=rotten_tomatoes \ +This returns whether the `valid` key is `true` or `false`: + +```py +print(data) +{'valid': True} +``` + @@ -210,0 +243,15 @@ curl https://datasets-server.huggingface.co/splits?dataset=rotten_tomatoes \ +This returns the available configuration and splits in the dataset: + +```py +print(data) +{'splits': + [ + {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'train'}, + {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'validation'}, + {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'test'} + ], + 'pending': [], + 'failed': [] +} +``` + @@ -251,0 +299,20 @@ curl https://datasets-server.huggingface.co/first-rows?dataset=rotten_tomatoes&c +This returns the first 100 rows of the dataset: + +```py +print(data) +{'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'train', + 'features': + [ + {'feature_idx': 0, 'name': 'text', 'type': {'dtype': 'string', '_type': 'Value'}}, + {'feature_idx': 1, 'name': 'label', 'type': {'names': ['neg', 'pos'], '_type': 'ClassLabel'}} + ], + 'rows': + [ + {'row_idx': 0, 'row': {'text': 'the rock is destined to be the 21st century\'s new " conan " and that he\'s going to make a splash even greater than arnold schwarzenegger , jean-claud van damme or steven segal .', 'label': 1}, 'truncated_cells': []}, + {'row_idx': 1, 'row': {'text': 'the gorgeously elaborate continuation of " the lord of the rings " trilogy is so huge that a column of words cannot adequately describe co-writer/director peter jackson\'s expanded vision of j . r . r . tolkien\'s middle-earth .', 'label': 1}, 'truncated_cells': []} + ..., + ..., + ], +} +``` + @@ -296,0 +364,21 @@ You can download slices of 100 rows maximum at a time. +The response looks like: + +```py +print(data) +{'features': + [ + {'feature_idx': 0, 'name': 'text', 'type': {'dtype': 'string', '_type': 'Value'}}, + {'feature_idx': 1, 'name': 'label', 'type': {'names': ['neg', 'pos'], '_type': 'ClassLabel'}}], + 'rows': + [ + {'row_idx': 150, 'row': {'text': 'enormously likable , partly because it is aware of its own grasp of the absurd .', 'label': 1}, 'truncated_cells': []}, + {'row_idx': 151, 'row': {'text': "here's a british flick gleefully unconcerned with plausibility , yet just as determined to entertain you .", 'label': 1}, 'truncated_cells': []}, + {'row_idx': 152, 'row': {'text': "it's an old story , but a lively script , sharp acting and partially animated interludes make just a kiss seem minty fresh .", 'label': 1}, 'truncated_cells': []}, + {'row_idx': 153, 'row': {'text': 'must be seen to be believed .', 'label': 1}, 'truncated_cells': []}, + {'row_idx': 154, 'row': {'text': "ray liotta and jason patric do some of their best work in their underwritten roles , but don't be fooled : nobody deserves any prizes here .", 'label': 1}, 'truncated_cells': []}, + ..., + ..., + ] +} +``` + @@ -336,0 +425,15 @@ curl https://datasets-server.huggingface.co/parquet?dataset=rotten_tomatoes \ + +This returns a URL to the Parquet file for each split: + +```py +print(data) +{'parquet_files': + [ + {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'test', 'url': 'https://huggingface.co/datasets/rotten_tomatoes/resolve/refs%2Fconvert%2Fparquet/default/rotten_tomatoes-test.parquet', 'filename': 'rotten_tomatoes-test.parquet', 'size': 92206}, + {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'train', 'url': 'https://huggingface.co/datasets/rotten_tomatoes/resolve/refs%2Fconvert%2Fparquet/default/rotten_tomatoes-train.parquet', 'filename': 'rotten_tomatoes-train.parquet', 'size': 698845}, + {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'validation', 'url': 'https://huggingface.co/datasets/rotten_tomatoes/resolve/refs%2Fconvert%2Fparquet/default/rotten_tomatoes-validation.parquet', 'filename': 'rotten_tomatoes-validation.parquet', 'size': 90001} + ], + 'pending': [], + 'failed': [] +} +``` \ No newline at end of file
a3da0bfc18e7990d36ddb6b0820868857303cdc9
Quentin Lhoest
2023-06-16T12:59:01
Multi commit Parquet copies (#1355)
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py index f7048d5a..672111e1 100644 --- a/libs/libcommon/src/libcommon/exceptions.py +++ b/libs/libcommon/src/libcommon/exceptions.py @@ -101,0 +102 @@ CacheableErrorCode = Literal[ + "LockedDatasetTimeoutError", @@ -239,0 +241,7 @@ class DatasetWithTooManyParquetFilesError(CacheableError): +class LockedDatasetTimeoutError(CacheableError): + """Raised when a dataset is locked by another job.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "LockedDatasetTimeoutError", cause, True) + + diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 14f54073..6c233cab 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -4,0 +5 @@ import contextlib +import json @@ -216 +217 @@ class Lock(Document): -class lock: +class lock(contextlib.AbstractContextManager["lock"]): @@ -241 +242,3 @@ class lock: - def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = (0.05, 0.05, 0.05, 1, 1, 1, 5)) -> None: + _default_sleeps = (0.05, 0.05, 0.05, 1, 1, 1, 5) + + def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> None: @@ -276,0 +280,14 @@ class lock: + @classmethod + def git_branch(cls, dataset: str, branch: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> "lock": + """ + Lock a git branch of a dataset on the hub for read/write + + Args: + dataset (`str`): the dataset repository + branch (`str`): the branch to lock + job_id (`str`): the current job id that holds the lock + sleeps (`Sequence[float]`): the time in seconds to sleep between each attempt to acquire the lock + """ + key = json.dumps({"dataset": dataset, "branch": branch}) + return cls(key=key, job_id=job_id, sleeps=sleeps) + diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index d9d10395..22460dc6 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -40,2 +40,2 @@ from huggingface_hub._commit_api import ( -from huggingface_hub.hf_api import DatasetInfo, HfApi, RepoFile -from huggingface_hub.utils._errors import RepositoryNotFoundError +from huggingface_hub.hf_api import CommitInfo, DatasetInfo, HfApi, RepoFile +from huggingface_hub.utils._errors import HfHubHTTPError, RepositoryNotFoundError @@ -64,0 +65 @@ from libcommon.exceptions import ( + LockedDatasetTimeoutError, @@ -68,0 +70 @@ from libcommon.processing_graph import ProcessingStep +from libcommon.queue import lock @@ -75 +77 @@ from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner -from worker.utils import CompleteJobResult +from worker.utils import CompleteJobResult, retry @@ -93,0 +96 @@ MAX_FILES_PER_DIRECTORY = 10_000 # hf hub limitation +MAX_OPERATIONS_PER_COMMIT = 500 @@ -680,0 +684,166 @@ def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]: +def create_commits( + hf_api: HfApi, + repo_id: str, + operations: List[CommitOperation], + *, + commit_message: str, + revision: Optional[str] = None, + parent_commit: Optional[str] = None, + max_operations_per_commit: int = MAX_OPERATIONS_PER_COMMIT, +) -> List[CommitInfo]: + """ + Creates one or several commits in the given dataset repo, deleting & uploading files as needed. + + Args: + hf_api (`huggingface_hub.HfApi`): + The HfApi to use to commit the operations. + repo_id (`str`): + The repository in which the commit will be created, for example: + `"username/my_dataset"` + operations (`Iterable` of [`huggingface_hub.hf_api.CommitOperation`]): + An iterable of operations to include in the commit, either: + + - [`huggingface_hub.hf_api.CommitOperationAdd`] to upload a file + - [`huggingface_hub.hf_api.CommitOperationDelete`] to delete a file + - [`huggingface_hub.hf_api.CommitOperationCopy`] to copy a file + commit_message (`str`): + The summary (first line) of the commit that will be created. + commit_description (`str`, *optional*): + The description of the commit that will be created + token (`str`, *optional*): + Authentication token, obtained with `HfApi.login` method. Will + default to the stored token. + repo_type (`str`, *optional*): + Set to `"dataset"` or `"space"` if uploading to a dataset or + space, `None` or `"model"` if uploading to a model. Default is + `None`. + revision (`str`, *optional*): + The git revision to commit from. Defaults to the head of the `"main"` branch. + parent_commit (`str`, *optional*): + The OID / SHA of the parent commit, as a hexadecimal string. + Shorthands (7 first characters) are also supported. If specified and `create_pr` is `False`, + the commit will fail if `revision` does not point to `parent_commit`. If specified and `create_pr` + is `True`, the pull request will be created from `parent_commit`. Specifying `parent_commit` + ensures the repo has not changed before committing the changes, and can be especially useful + if the repo is updated / committed to concurrently. + max_operations_per_commit (`int`, *optional*): + The ma number of operations per commit, to avoid time out errors from the Hub. Defaults to 500. + Returns: + [`List[huggingface_hub.CommitInfo]`]: + List of [`CommitInfo`] containing information about the newly created commit (commit hash, commit + url, pr url, commit message,...). + Raises: + [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) + If commit message is empty. + [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) + If parent commit is not a valid commit OID. + [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) + If the Hub API returns an HTTP 400 error (bad request) + [`huggingface_hub.utils.RepositoryNotFoundError`]: + If repository is not found (error 404): wrong repo_id/repo_type, private + but not authenticated or repo does not exist. + [`huggingface_hub.utils.HfHubHTTPError`]: + If another commit happened after `parent_commit` or in between the commits. + """ + commit_infos: List[CommitInfo] = [] + offsets = range(0, len(operations), max_operations_per_commit) + for commit_idx, offset in enumerate(offsets): + batch_msg = f" (step {commit_idx + 1} of {len(offsets)})" if len(offsets) > 1 else "" + retry_create_commit = retry(on=[HfHubHTTPError], sleeps=[1, 1, 1, 10, 10, 10])(hf_api.create_commit) + commit_infos.append( + retry_create_commit( + repo_id=repo_id, + repo_type=DATASET_TYPE, + revision=revision, + operations=operations[offset : offset + max_operations_per_commit], # noqa: E203 + commit_message=commit_message + batch_msg, + parent_commit=parent_commit if not commit_infos else commit_infos[-1].oid, + ) + ) + return commit_infos + + +def commit_parquet_conversion( + hf_api: HfApi, + committer_hf_api: HfApi, + dataset: str, + config: str, + config_names: Set[str], + parquet_operations: List[CommitOperation], + commit_message: str, + target_revision: Optional[str], +) -> List[CommitInfo]: + """ + Creates one or several commits in the given dataset repo, deleting & uploading files as needed. + + Args: + hf_api (`huggingface_hub.HfApi`): + The HfApi to get the dataset info. + committer_hf_api (`huggingface_hub.HfApi`): + The HfApi to use to commit the operations. + dataset (`str`): + The dataset in which the commit will be created, for example: + `"username/my_dataset"` + config (`str`): + The dataset configuration. + config_names (`List[str]`): + The list of all the configurations of this dataset. This is used to clean + the other fiels and directories in the repo, if any. + parquet_operations (`List[huggingface_hub.hf_api.CommitOperation]`): + List of commit operation for the parquet conversion. It could be + file additions or file copies for example. + commit_message (`str`): + The summary (first line) of the commit that will be created. + target_revision (`str`, *optional*): + The git revision to commit from. Defaults to the head of the `"main"` branch. + Returns: + [`List[huggingface_hub.CommitInfo]`]: + List of [`CommitInfo`] containing information about the newly created commit (commit hash, commit + url, pr url, commit message,...). + Raises: + [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) + If commit message is empty. + [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) + If parent commit is not a valid commit OID. + [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError) + If the Hub API returns an HTTP 400 error (bad request) + [`huggingface_hub.utils.RepositoryNotFoundError`]: + If repository is not found (error 404): wrong repo_id/repo_type, private + but not authenticated or repo does not exist. + [`RuntimeError`]: If it fails after several retries. + The retry fails can come from: + [`huggingface_hub.utils.HfHubHTTPError`]: + If another commit happened after `parent_commit` or in between the commits. + """ + target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False) + # - get repo parquet files + all_repo_files: Set[str] = {f.rfilename for f in target_dataset_info.siblings} + repo_parquet_files: Set[str] = {file for file in all_repo_files if file.endswith(".parquet")} + # - get files that will be preserved in repo: files belonging to other configs and .gitattributes + # we exclude files of current config because otherwise outdated files might be preserved + files_to_ignore: Set[str] = { + file + for other_config in config_names.difference({config}) + for file in repo_parquet_files + if file.startswith(f"{other_config}/") + }.union({".gitattributes"}) + # - get files to be deleted - all files except for: + # - parquet files obtained for current config at this processing step, + # - parquet files belonging to other existing configs + # - .gitattributes + files_to_add = [operation.path_in_repo for operation in parquet_operations] + files_to_delete = all_repo_files - set(files_to_add).union(files_to_ignore) + delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete] + logging.debug(f"{delete_operations=}") + + operations = delete_operations + parquet_operations + return create_commits( + committer_hf_api, + repo_id=dataset, + revision=target_revision, + operations=operations, + commit_message=commit_message, + parent_commit=target_dataset_info.sha, + ) + + @@ -681,0 +851 @@ def compute_config_parquet_and_info_response( + job_id: str, @@ -699,0 +870,2 @@ def compute_config_parquet_and_info_response( + job_id (`str`): + The id of the current Job. It is used to lock the access to the parquet conversion branch on the Hub. @@ -847,30 +1019,15 @@ def compute_config_parquet_and_info_response( - builder_info = asdict(builder.info) - target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False) - # - get repo parquet files - all_repo_files: Set[str] = {f.rfilename for f in target_dataset_info.siblings} - repo_parquet_files: Set[str] = {file for file in all_repo_files if file.endswith(".parquet")} - # - get files that will be preserved in repo: files belonging to other configs and .gitattributes - # we exclude files of current config because otherwise outdated files might be preserved - files_to_ignore: Set[str] = { - file - for other_config in config_names.difference({config}) - for file in repo_parquet_files - if file.startswith(f"{other_config}/") - }.union({".gitattributes"}) - # - get files to be deleted - all files except for: - # - parquet files obtained for current config at this processing step, - # - parquet files belonging to other existing configs - # - .gitattributes - config_files_to_add = [operation.path_in_repo for operation in parquet_operations] - files_to_delete = all_repo_files - set(config_files_to_add).union(files_to_ignore) - delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete] - logging.debug(f"{delete_operations=}") - - committer_hf_api.create_commit( - repo_id=dataset, - repo_type=DATASET_TYPE, - revision=target_revision, - operations=delete_operations + parquet_operations, - commit_message=commit_message, - parent_commit=target_dataset_info.sha, - ) + try: + sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300] + with lock.git_branch(dataset=dataset, branch=target_revision, job_id=job_id, sleeps=sleeps): + commit_parquet_conversion( + hf_api=hf_api, + committer_hf_api=committer_hf_api, + dataset=dataset, + config=config, + parquet_operations=parquet_operations, + config_names=config_names, + target_revision=target_revision, + commit_message=commit_message, + ) + except TimeoutError as err: + raise LockedDatasetTimeoutError("the dataset is currently locked, please try again later.") from err @@ -901 +1058 @@ def compute_config_parquet_and_info_response( - dataset_info=builder_info, + dataset_info=asdict(builder.info), @@ -933,0 +1091 @@ class ConfigParquetAndInfoJobRunner(ConfigCachedJobRunner): + job_id=self.job_info["job_id"], diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 4399bd6e..14d209a2 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -15,0 +16,3 @@ from typing import ( + Sequence, + Tuple, + Type, @@ -287,0 +291,2 @@ FuncT = TypeVar("FuncT", bound=Callable[..., Any]) +RETRY_SLEEPS = (1, 1, 1, 10, 10, 10, 60, 60, 60, 10 * 60) +RETRY_ON: Tuple[Type[Exception]] = (Exception,) @@ -290 +295 @@ FuncT = TypeVar("FuncT", bound=Callable[..., Any]) -def retry(func: FuncT) -> FuncT: +class retry: @@ -292,24 +297,27 @@ def retry(func: FuncT) -> FuncT: - SLEEPS = [1, 7, 70, 7 * 60, 70 * 60] - MAX_ATTEMPTS = len(SLEEPS) - - @functools.wraps(func) - def decorator(*args: Any, **kwargs: Any) -> Any: - attempt = 0 - last_err = None - while attempt < MAX_ATTEMPTS: - try: - """always sleep before calling the function. It will prevent rate limiting in the first place""" - duration = SLEEPS[attempt] - logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.") - time.sleep(duration) - return func(*args, **kwargs) - except ConnectionError as err: - logging.info("Got a ConnectionError, possibly due to rate limiting. Let's retry.") - last_err = err - attempt += 1 - raise RuntimeError(f"Give up after {attempt} attempts with ConnectionError") from last_err - - return cast(FuncT, decorator) - - -@retry + + def __init__(self, sleeps: Sequence[int] = RETRY_SLEEPS, on: Sequence[Type[Exception]] = RETRY_ON) -> None: + self.sleeps = sleeps + self.on = on + + def __call__(self, func: FuncT) -> FuncT: + @functools.wraps(func) + def decorator(*args: Any, **kwargs: Any) -> Any: + attempt = 0 + last_err = None + while attempt < len(self.sleeps): + try: + """always sleep before calling the function. It will prevent rate limiting in the first place""" + duration = self.sleeps[attempt] + logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.") + time.sleep(duration) + return func(*args, **kwargs) + except tuple(self.on) as err: + logging.info(f"Got a {type(err)}. Let's retry.") + last_err = err + attempt += 1 + raise RuntimeError(f"Give up after {attempt} attempts with {type(last_err)}") from last_err + + return cast(FuncT, decorator) + + +@retry(on=[ConnectionError])
031daa5a5beee4e44573fb487f094d1429f3292d
Sylvain Lesage
2023-06-16T07:29:54
Revert "Delete ttl index from queue.py code (#1378)" (#1380)
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index 2c1d062d..09cb6301 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -4 +3,0 @@ -import pytest @@ -120 +118,0 @@ def test_metrics_deletion_migration(mongo_host: str) -> None: [email protected](reason="temporaly disabled because of index removal") diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index bec930c1..14f54073 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -25,0 +26 @@ from libcommon.constants import ( + QUEUE_TTL_SECONDS, @@ -136,0 +138 @@ class Job(Document): + {"fields": ["finished_at"], "expireAfterSeconds": QUEUE_TTL_SECONDS}, diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 5262acd1..02b9578e 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -379 +378,0 @@ def test_queue_get_zombies() -> None: [email protected](reason="temporaly disabled because of index removal")
c96497924667ce699ef4cb77fc04377f276337d0
Andrea Francis Soria Jimenez
2023-06-15T23:20:24
Rollback index (#1379)
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index 09cb6301..2c1d062d 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -3,0 +4 @@ +import pytest @@ -118,0 +120 @@ def test_metrics_deletion_migration(mongo_host: str) -> None: [email protected](reason="temporaly disabled because of index removal") diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index e226b886..bec930c1 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -26 +25,0 @@ from libcommon.constants import ( - QUEUE_TTL_SECONDS, @@ -138,5 +136,0 @@ class Job(Document): - { - "fields": ["finished_at"], - "expireAfterSeconds": QUEUE_TTL_SECONDS, - "partialFilterExpression": {"status": {"$in": [Status.SUCCESS, Status.ERROR, Status.CANCELLED]}}, - }, diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 02b9578e..5262acd1 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -378,0 +379 @@ def test_queue_get_zombies() -> None: [email protected](reason="temporaly disabled because of index removal")
335754dd4284b19ccfa3b75e0f96171858ecfd2a
Andrea Francis Soria Jimenez
2023-06-15T22:32:21
Adding condition to Jobs Collection - ttl index (#1325)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py index 88e61e94..06412c36 100644 --- a/jobs/mongodb_migration/src/mongodb_migration/collector.py +++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py @@ -217,0 +218,7 @@ class MigrationsCollector: + MigrationQueueDeleteTTLIndex( + version="20230607154800", + description=( + "delete the TTL index on the 'finished_at' field in the queue database to update its TTL condition" + ), + field_name="finished_at", + ), diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index 2c1d062d..09cb6301 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -4 +3,0 @@ -import pytest @@ -120 +118,0 @@ def test_metrics_deletion_migration(mongo_host: str) -> None: [email protected](reason="temporaly disabled because of index removal") diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index bec930c1..e226b886 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -25,0 +26 @@ from libcommon.constants import ( + QUEUE_TTL_SECONDS, @@ -136,0 +138,5 @@ class Job(Document): + { + "fields": ["finished_at"], + "expireAfterSeconds": QUEUE_TTL_SECONDS, + "partialFilterExpression": {"status": {"$in": [Status.SUCCESS, Status.ERROR, Status.CANCELLED]}}, + }, diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 5262acd1..02b9578e 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -379 +378,0 @@ def test_queue_get_zombies() -> None: [email protected](reason="temporaly disabled because of index removal")
47ea65b2567db4482579cd7000393cf0a15b412e
Andrea Francis Soria Jimenez
2023-06-15T22:08:06
Delete ttl index from queue.py code (#1378)
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py index 09cb6301..2c1d062d 100644 --- a/jobs/mongodb_migration/tests/test_deletion_migrations.py +++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py @@ -3,0 +4 @@ +import pytest @@ -118,0 +120 @@ def test_metrics_deletion_migration(mongo_host: str) -> None: [email protected](reason="temporaly disabled because of index removal") diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 14f54073..bec930c1 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -26 +25,0 @@ from libcommon.constants import ( - QUEUE_TTL_SECONDS, @@ -138 +136,0 @@ class Job(Document): - {"fields": ["finished_at"], "expireAfterSeconds": QUEUE_TTL_SECONDS}, diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 02b9578e..5262acd1 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -378,0 +379 @@ def test_queue_get_zombies() -> None: [email protected](reason="temporaly disabled because of index removal")
32c833873068ac89428df7ac0a8802de6ee33414
Quentin Lhoest
2023-06-15T20:53:42
Refac hub_datasets fixture (#1373)
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 110dba75..da77c3fc 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -576,16 +576,2 @@ SPAWNING_OPT_IN_OUT_rows = ["http://testurl.test/test_image.jpg", "http://testur [email protected](scope="session") -def hub_datasets( - hub_public_empty: str, - hub_public_csv: str, - hub_private_csv: str, - hub_gated_csv: str, - hub_public_jsonl: str, - hub_public_audio: str, - hub_public_image: str, - hub_public_images_list: str, - hub_public_big: str, - hub_public_big_no_info: str, - hub_public_big_csv: str, - hub_public_external_files: str, - hub_public_spawning_opt_in_out: str, -) -> HubDatasets: [email protected] +def hub_reponses_does_not_exist() -> HubDatasetTest: @@ -593,124 +579,180 @@ def hub_datasets( - "does_not_exist": { - "name": "does_not_exist", - "config_names_response": None, - "splits_response": None, - "first_rows_response": None, - "parquet_and_info_response": None, - }, - "does_not_exist_config": { - "name": "does_not_exist_config", - "config_names_response": None, - "splits_response": None, - "first_rows_response": None, - "parquet_and_info_response": None, - }, - "does_not_exist_split": { - "name": "does_not_exist_split", - "config_names_response": None, - "splits_response": None, - "first_rows_response": None, - "parquet_and_info_response": None, - }, - "empty": { - "name": hub_public_empty, - "config_names_response": None, - "splits_response": None, - "first_rows_response": None, - "parquet_and_info_response": None, - }, - "public": { - "name": hub_public_csv, - "config_names_response": create_config_names_response(hub_public_csv), - "splits_response": create_splits_response(hub_public_csv), - "first_rows_response": create_first_rows_response(hub_public_csv, DATA_cols, DATA_rows), - "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_csv, data_type="csv"), - }, - "private": { - "name": hub_private_csv, - "config_names_response": create_config_names_response(hub_private_csv), - "splits_response": create_splits_response(hub_private_csv), - "first_rows_response": create_first_rows_response(hub_private_csv, DATA_cols, DATA_rows), - "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_private_csv, data_type="csv"), - }, - "gated": { - "name": hub_gated_csv, - "config_names_response": create_config_names_response(hub_gated_csv), - "splits_response": create_splits_response(hub_gated_csv), - "first_rows_response": create_first_rows_response(hub_gated_csv, DATA_cols, DATA_rows), - "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_gated_csv, data_type="csv"), - }, - "jsonl": { - "name": hub_public_jsonl, - "config_names_response": create_config_names_response(hub_public_jsonl), - "splits_response": create_splits_response(hub_public_jsonl), - "first_rows_response": create_first_rows_response(hub_public_jsonl, JSONL_cols, JSONL_rows), - "parquet_and_info_response": None, - }, - "audio": { - "name": hub_public_audio, - "config_names_response": create_config_names_response(hub_public_audio), - "splits_response": create_splits_response(hub_public_audio), - "first_rows_response": create_first_rows_response( - hub_public_audio, AUDIO_cols, get_AUDIO_rows(hub_public_audio) - ), - "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_audio, data_type="audio"), - }, - "image": { - "name": hub_public_image, - "config_names_response": create_config_names_response(hub_public_image), - "splits_response": create_splits_response(hub_public_image), - "first_rows_response": create_first_rows_response( - hub_public_image, IMAGE_cols, get_IMAGE_rows(hub_public_image) - ), - "parquet_and_info_response": None, - }, - "images_list": { - "name": hub_public_images_list, - "config_names_response": create_config_names_response(hub_public_images_list), - "splits_response": create_splits_response(hub_public_images_list), - "first_rows_response": create_first_rows_response( - hub_public_images_list, IMAGES_LIST_cols, get_IMAGES_LIST_rows(hub_public_images_list) - ), - "parquet_and_info_response": None, - }, - "big": { - "name": hub_public_big, - "config_names_response": create_config_names_response(hub_public_big), - "splits_response": create_splits_response(hub_public_big), - "first_rows_response": create_first_rows_response(hub_public_big, BIG_cols, BIG_rows), - "parquet_and_info_response": create_parquet_and_info_response( - dataset=hub_public_big, data_type="big_parquet" - ), - }, - "big-no-info": { - "name": hub_public_big_no_info, - "config_names_response": create_config_names_response(hub_public_big_no_info), - "splits_response": create_splits_response(hub_public_big_no_info), - "first_rows_response": create_first_rows_response(hub_public_big_no_info, BIG_cols, BIG_rows), - "parquet_and_info_response": create_parquet_and_info_response( - dataset=hub_public_big_no_info, data_type="big_parquet_no_info" - ), - }, - "big-csv": { - "name": hub_public_big_csv, - "config_names_response": create_config_names_response(hub_public_big_csv), - "splits_response": create_splits_response(hub_public_big_csv), - "first_rows_response": create_first_rows_response(hub_public_big_csv, BIG_cols, BIG_rows), - "parquet_and_info_response": None, - }, - "external_files": { - "name": hub_public_external_files, - "config_names_response": create_config_names_response(hub_public_external_files), - "splits_response": create_splits_response(hub_public_external_files), - "first_rows_response": create_first_rows_response(hub_public_external_files, TEXT_cols, TEXT_rows), - "parquet_and_info_response": None, - }, - "spawning_opt_in_out": { - "name": hub_public_spawning_opt_in_out, - "config_names_response": create_config_names_response(hub_public_spawning_opt_in_out), - "splits_response": create_splits_response(hub_public_spawning_opt_in_out), - "first_rows_response": create_first_rows_response( - hub_public_spawning_opt_in_out, SPAWNING_OPT_IN_OUT_cols, SPAWNING_OPT_IN_OUT_rows - ), - "parquet_and_info_response": None, - }, + "name": "does_not_exist", + "config_names_response": None, + "splits_response": None, + "first_rows_response": None, + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_does_not_exist_config() -> HubDatasetTest: + return { + "name": "does_not_exist_config", + "config_names_response": None, + "splits_response": None, + "first_rows_response": None, + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_does_not_exist_split() -> HubDatasetTest: + return { + "name": "does_not_exist_split", + "config_names_response": None, + "splits_response": None, + "first_rows_response": None, + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_empty(hub_public_empty: str) -> HubDatasetTest: + return { + "name": hub_public_empty, + "config_names_response": None, + "splits_response": None, + "first_rows_response": None, + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_public(hub_public_csv: str) -> HubDatasetTest: + return { + "name": hub_public_csv, + "config_names_response": create_config_names_response(hub_public_csv), + "splits_response": create_splits_response(hub_public_csv), + "first_rows_response": create_first_rows_response(hub_public_csv, DATA_cols, DATA_rows), + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_csv, data_type="csv"), + } + + [email protected] +def hub_reponses_private(hub_private_csv: str) -> HubDatasetTest: + return { + "name": hub_private_csv, + "config_names_response": create_config_names_response(hub_private_csv), + "splits_response": create_splits_response(hub_private_csv), + "first_rows_response": create_first_rows_response(hub_private_csv, DATA_cols, DATA_rows), + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_private_csv, data_type="csv"), + } + + [email protected] +def hub_reponses_gated(hub_gated_csv: str) -> HubDatasetTest: + return { + "name": hub_gated_csv, + "config_names_response": create_config_names_response(hub_gated_csv), + "splits_response": create_splits_response(hub_gated_csv), + "first_rows_response": create_first_rows_response(hub_gated_csv, DATA_cols, DATA_rows), + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_gated_csv, data_type="csv"), + } + + [email protected] +def hub_reponses_jsonl(hub_public_jsonl: str) -> HubDatasetTest: + return { + "name": hub_public_jsonl, + "config_names_response": create_config_names_response(hub_public_jsonl), + "splits_response": create_splits_response(hub_public_jsonl), + "first_rows_response": create_first_rows_response(hub_public_jsonl, JSONL_cols, JSONL_rows), + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_audio(hub_public_audio: str) -> HubDatasetTest: + return { + "name": hub_public_audio, + "config_names_response": create_config_names_response(hub_public_audio), + "splits_response": create_splits_response(hub_public_audio), + "first_rows_response": create_first_rows_response( + hub_public_audio, AUDIO_cols, get_AUDIO_rows(hub_public_audio) + ), + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_audio, data_type="audio"), + } + + [email protected] +def hub_reponses_image(hub_public_image: str) -> HubDatasetTest: + return { + "name": hub_public_image, + "config_names_response": create_config_names_response(hub_public_image), + "splits_response": create_splits_response(hub_public_image), + "first_rows_response": create_first_rows_response( + hub_public_image, IMAGE_cols, get_IMAGE_rows(hub_public_image) + ), + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_images_list(hub_public_images_list: str) -> HubDatasetTest: + return { + "name": hub_public_images_list, + "config_names_response": create_config_names_response(hub_public_images_list), + "splits_response": create_splits_response(hub_public_images_list), + "first_rows_response": create_first_rows_response( + hub_public_images_list, IMAGES_LIST_cols, get_IMAGES_LIST_rows(hub_public_images_list) + ), + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_big(hub_public_big: str) -> HubDatasetTest: + return { + "name": hub_public_big, + "config_names_response": create_config_names_response(hub_public_big), + "splits_response": create_splits_response(hub_public_big), + "first_rows_response": create_first_rows_response(hub_public_big, BIG_cols, BIG_rows), + "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_big, data_type="big_parquet"), + } + + [email protected] +def hub_reponses_big_no_info(hub_public_big_no_info: str) -> HubDatasetTest: + return { + "name": hub_public_big_no_info, + "config_names_response": create_config_names_response(hub_public_big_no_info), + "splits_response": create_splits_response(hub_public_big_no_info), + "first_rows_response": create_first_rows_response(hub_public_big_no_info, BIG_cols, BIG_rows), + "parquet_and_info_response": create_parquet_and_info_response( + dataset=hub_public_big_no_info, data_type="big_parquet_no_info" + ), + } + + [email protected] +def hub_reponses_big_csv(hub_public_big_csv: str) -> HubDatasetTest: + return { + "name": hub_public_big_csv, + "config_names_response": create_config_names_response(hub_public_big_csv), + "splits_response": create_splits_response(hub_public_big_csv), + "first_rows_response": create_first_rows_response(hub_public_big_csv, BIG_cols, BIG_rows), + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_external_files(hub_public_external_files: str) -> HubDatasetTest: + return { + "name": hub_public_external_files, + "config_names_response": create_config_names_response(hub_public_external_files), + "splits_response": create_splits_response(hub_public_external_files), + "first_rows_response": create_first_rows_response(hub_public_external_files, TEXT_cols, TEXT_rows), + "parquet_and_info_response": None, + } + + [email protected] +def hub_reponses_spawning_opt_in_out(hub_public_spawning_opt_in_out: str) -> HubDatasetTest: + return { + "name": hub_public_spawning_opt_in_out, + "config_names_response": create_config_names_response(hub_public_spawning_opt_in_out), + "splits_response": create_splits_response(hub_public_spawning_opt_in_out), + "first_rows_response": create_first_rows_response( + hub_public_spawning_opt_in_out, SPAWNING_OPT_IN_OUT_cols, SPAWNING_OPT_IN_OUT_rows + ), + "parquet_and_info_response": None, diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 66944344..2121d53d 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -4,0 +5 @@ import io +from contextlib import contextmanager @@ -46 +47 @@ from ...constants import CI_HUB_ENDPOINT, CI_USER_TOKEN -from ...fixtures.hub import HubDatasets +from ...fixtures.hub import HubDatasetTest @@ -49,14 +50,5 @@ from ...fixtures.hub import HubDatasets -# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200 [email protected](scope="module", autouse=True) -def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyPatch]: - mp = pytest.MonkeyPatch() - mp.setenv( - "PARQUET_AND_INFO_BLOCKED_DATASETS", - ",".join(value["name"] for value in hub_datasets.values() if "jsonl" in value["name"]), - ) - mp.setenv( - "PARQUET_AND_INFO_SUPPORTED_DATASETS", - ",".join(value["name"] for value in hub_datasets.values() if "big" not in value["name"]), - ) - yield mp - mp.undo() +@contextmanager +def blocked(app_config: AppConfig, repo_id: str) -> Iterator[None]: + app_config.parquet_and_info.blocked_datasets.append(repo_id) + yield + app_config.parquet_and_info.blocked_datasets.remove(repo_id) @@ -126 +118 @@ def test_compute( - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, @@ -128,2 +120,2 @@ def test_compute( - dataset = hub_datasets["public"]["name"] - config = hub_datasets["public"]["config_names_response"]["config_names"][0]["config"] + dataset = hub_reponses_public["name"] + config = hub_reponses_public["config_names_response"]["config_names"][0]["config"] @@ -134 +126 @@ def test_compute( - content=hub_datasets["public"]["config_names_response"], + content=hub_reponses_public["config_names_response"], @@ -142 +134 @@ def test_compute( - assert_content_is_equal(content, hub_datasets["public"]["parquet_and_info_response"]) + assert_content_is_equal(content, hub_reponses_public["parquet_and_info_response"]) @@ -245 +237,2 @@ def test_raise_if_too_big_from_hub( - hub_datasets: HubDatasets, + hub_public_csv: str, + hub_public_big: str, @@ -250 +243 @@ def test_raise_if_too_big_from_hub( - dataset = hub_datasets[name]["name"] + dataset = hub_public_csv if name == "public" else hub_public_big @@ -274 +267,2 @@ def test_raise_if_too_big_from_datasets( - hub_datasets: HubDatasets, + hub_public_csv: str, + hub_public_big: str, @@ -279 +273 @@ def test_raise_if_too_big_from_datasets( - dataset = hub_datasets[name]["name"] + dataset = hub_public_csv if name == "public" else hub_public_big @@ -363 +357 @@ def test_supported_if_big_parquet( - hub_datasets: HubDatasets, + hub_reponses_big: HubDatasetTest, @@ -368,2 +362,2 @@ def test_supported_if_big_parquet( - dataset = hub_datasets["big"]["name"] - config = hub_datasets["big"]["config_names_response"]["config_names"][0]["config"] + dataset = hub_reponses_big["name"] + config = hub_reponses_big["config_names_response"]["config_names"][0]["config"] @@ -374 +368 @@ def test_supported_if_big_parquet( - content=hub_datasets["big"]["config_names_response"], + content=hub_reponses_big["config_names_response"], @@ -382 +376 @@ def test_supported_if_big_parquet( - assert_content_is_equal(content, hub_datasets["big"]["parquet_and_info_response"]) + assert_content_is_equal(content, hub_reponses_big["parquet_and_info_response"]) @@ -388 +382 @@ def test_not_supported_if_big_non_parquet( - hub_datasets: HubDatasets, + hub_reponses_big_csv: HubDatasetTest, @@ -392,2 +386,2 @@ def test_not_supported_if_big_non_parquet( - dataset = hub_datasets["big-csv"]["name"] - config = hub_datasets["big-csv"]["config_names_response"]["config_names"][0]["config"] + dataset = hub_reponses_big_csv["name"] + config = hub_reponses_big_csv["config_names_response"]["config_names"][0]["config"] @@ -398 +392 @@ def test_not_supported_if_big_non_parquet( - content=hub_datasets["big-csv"]["config_names_response"], + content=hub_reponses_big_csv["config_names_response"], @@ -409 +403 @@ def test_supported_if_gated( - hub_datasets: HubDatasets, + hub_reponses_gated: HubDatasetTest, @@ -412,2 +406,2 @@ def test_supported_if_gated( - dataset = hub_datasets["gated"]["name"] - config = hub_datasets["gated"]["config_names_response"]["config_names"][0]["config"] + dataset = hub_reponses_gated["name"] + config = hub_reponses_gated["config_names_response"]["config_names"][0]["config"] @@ -418 +412 @@ def test_supported_if_gated( - content=hub_datasets["gated"]["config_names_response"], + content=hub_reponses_gated["config_names_response"], @@ -429 +423 @@ def test_blocked( - hub_datasets: HubDatasets, + hub_reponses_jsonl: HubDatasetTest, @@ -432,12 +426,13 @@ def test_blocked( - dataset = hub_datasets["jsonl"]["name"] - config = hub_datasets["jsonl"]["config_names_response"]["config_names"][0]["config"] - upsert_response( - kind="dataset-config-names", - dataset=dataset, - http_status=HTTPStatus.OK, - content=hub_datasets["jsonl"]["config_names_response"], - ) - job_runner = get_job_runner(dataset, config, app_config) - with pytest.raises(CustomError) as e: - job_runner.compute() - assert e.typename == "DatasetInBlockListError" + with blocked(app_config, repo_id=hub_reponses_jsonl["name"]): + dataset = hub_reponses_jsonl["name"] + config = hub_reponses_jsonl["config_names_response"]["config_names"][0]["config"] + upsert_response( + kind="dataset-config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_reponses_jsonl["config_names_response"], + ) + job_runner = get_job_runner(dataset, config, app_config) + with pytest.raises(CustomError) as e: + job_runner.compute() + assert e.typename == "DatasetInBlockListError" @@ -451 +446,3 @@ def test_compute_splits_response_simple_csv_ok( - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, + hub_reponses_audio: HubDatasetTest, + hub_reponses_gated: HubDatasetTest, @@ -456,0 +454 @@ def test_compute_splits_response_simple_csv_ok( + hub_datasets = {"public": hub_reponses_public, "audio": hub_reponses_audio, "gated": hub_reponses_gated} @@ -495 +493 @@ def test_compute_splits_response_simple_csv_error( - hub_datasets: HubDatasets, + hub_reponses_private: HubDatasetTest, @@ -502,2 +500,2 @@ def test_compute_splits_response_simple_csv_error( - dataset = hub_datasets[name]["name"] - config_names_response = hub_datasets[name]["config_names_response"] + dataset = hub_reponses_private["name"] + config_names_response = hub_reponses_private["config_names_response"] @@ -509 +507 @@ def test_compute_splits_response_simple_csv_error( - content=hub_datasets[name]["config_names_response"], + content=hub_reponses_private["config_names_response"], @@ -533 +531 @@ def test_compute_splits_response_simple_csv_error_2( - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, @@ -540,2 +538,2 @@ def test_compute_splits_response_simple_csv_error_2( - dataset = hub_datasets[name]["name"] - config_names_response = hub_datasets[name]["config_names_response"] + dataset = hub_reponses_public["name"] + config_names_response = hub_reponses_public["config_names_response"] @@ -561,2 +559 @@ def test_previous_step_error( - hub_public_csv: str, - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, @@ -565,2 +562,2 @@ def test_previous_step_error( - dataset = hub_datasets["public"]["name"] - config = hub_datasets["public"]["config_names_response"]["config_names"][0]["config"] + dataset = hub_reponses_public["name"] + config = hub_reponses_public["config_names_response"]["config_names"][0]["config"] diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py index 2d60370b..56e28c7e 100644 --- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py +++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py @@ -20 +20 @@ from worker.resources import LibrariesResource -from ...fixtures.hub import HubDatasets, get_default_config_split +from ...fixtures.hub import HubDatasetTest, get_default_config_split @@ -91 +91,6 @@ def test_compute_split_names_from_streaming_response( - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, + hub_reponses_audio: HubDatasetTest, + hub_reponses_gated: HubDatasetTest, + hub_reponses_private: HubDatasetTest, + hub_reponses_empty: HubDatasetTest, + hub_reponses_does_not_exist: HubDatasetTest, @@ -98,0 +104,8 @@ def test_compute_split_names_from_streaming_response( + hub_datasets = { + "public": hub_reponses_public, + "audio": hub_reponses_audio, + "gated": hub_reponses_gated, + "private": hub_reponses_private, + "empty": hub_reponses_empty, + "does_not_exist": hub_reponses_does_not_exist, + } diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py index 3f0ed076..a94427f4 100644 --- a/services/worker/tests/job_runners/dataset/test_config_names.py +++ b/services/worker/tests/job_runners/dataset/test_config_names.py @@ -18 +18 @@ from worker.resources import LibrariesResource -from ...fixtures.hub import HubDatasets +from ...fixtures.hub import HubDatasetTest @@ -113 +113,6 @@ def test_compute_splits_response_simple_csv( - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, + hub_reponses_audio: HubDatasetTest, + hub_reponses_gated: HubDatasetTest, + hub_reponses_private: HubDatasetTest, + hub_reponses_empty: HubDatasetTest, + hub_reponses_does_not_exist: HubDatasetTest, @@ -120,0 +126,8 @@ def test_compute_splits_response_simple_csv( + hub_datasets = { + "public": hub_reponses_public, + "audio": hub_reponses_audio, + "gated": hub_reponses_gated, + "private": hub_reponses_private, + "empty": hub_reponses_empty, + "does_not_exist": hub_reponses_does_not_exist, + } diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py index e5ed7cfa..fd2dd64b 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py @@ -24 +24 @@ from worker.utils import get_json_size -from ...fixtures.hub import HubDatasets, get_default_config_split +from ...fixtures.hub import HubDatasetTest, get_default_config_split @@ -116 +116,10 @@ def test_number_rows( - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, + hub_reponses_audio: HubDatasetTest, + hub_reponses_image: HubDatasetTest, + hub_reponses_images_list: HubDatasetTest, + hub_reponses_jsonl: HubDatasetTest, + hub_reponses_gated: HubDatasetTest, + hub_reponses_private: HubDatasetTest, + hub_reponses_empty: HubDatasetTest, + hub_reponses_does_not_exist_config: HubDatasetTest, + hub_reponses_does_not_exist_split: HubDatasetTest, @@ -129,0 +139,12 @@ def test_number_rows( + hub_datasets = { + "public": hub_reponses_public, + "audio": hub_reponses_audio, + "image": hub_reponses_image, + "images_list": hub_reponses_images_list, + "jsonl": hub_reponses_jsonl, + "gated": hub_reponses_gated, + "private": hub_reponses_private, + "empty": hub_reponses_empty, + "does_not_exist_config": hub_reponses_does_not_exist_config, + "does_not_exist_split": hub_reponses_does_not_exist_split, + } @@ -187 +208,2 @@ def test_truncation( - hub_datasets: HubDatasets, + hub_public_csv: str, + hub_public_big: str, @@ -195 +217,2 @@ def test_truncation( - dataset, config, split = get_default_config_split(hub_datasets[name]["name"]) + dataset = hub_public_csv if name == "public" else hub_public_big + dataset, config, split = get_default_config_split(dataset) diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py index f56e08b4..e3fa31ac 100644 --- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py @@ -32 +32 @@ from ...constants import CI_SPAWNING_TOKEN -from ...fixtures.hub import HubDatasets, get_default_config_split +from ...fixtures.hub import HubDatasetTest, get_default_config_split @@ -176 +176,2 @@ def test_compute( - hub_datasets: HubDatasets, + hub_reponses_public: HubDatasetTest, + hub_reponses_spawning_opt_in_out: HubDatasetTest, @@ -183,0 +185 @@ def test_compute( + hub_datasets = {"public": hub_reponses_public, "spawning_opt_in_out": hub_reponses_spawning_opt_in_out} @@ -238 +240 @@ def test_compute_failed( - hub_datasets: HubDatasets, + hub_reponses_spawning_opt_in_out: HubDatasetTest, @@ -247 +249 @@ def test_compute_failed( - dataset = hub_datasets["spawning_opt_in_out"]["name"] + dataset = hub_reponses_spawning_opt_in_out["name"]
ae6ffbe68ec50aa144e2b8986a7cd85c161d1faa
Quentin Lhoest
2023-06-15T20:37:34
fix fill_builder_info (#1375)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 8f2f0f10..d9d10395 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -631,0 +632 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None: + split = str(split) # in case it's a NamedSplit
0ce4b5b508ac660eba658ac77a8a90883541516c
Steven Liu
2023-06-15T20:22:50
add build notebook (#1377)
diff --git a/.github/workflows/doc-build.yml b/.github/workflows/doc-build.yml index 8dc7ac99..6238f7c7 100644 --- a/.github/workflows/doc-build.yml +++ b/.github/workflows/doc-build.yml @@ -19,0 +20 @@ jobs: + notebook_folder: datasets-server_doc
6865d8fab8c9476ab1efccd713479c0b432bd22d
Albert Villanova del Moral
2023-06-15T15:59:03
Update datasets dependency to 2.13.0 version (#1372)
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock index cd3fcb36..d43c651e 100644 --- a/front/admin_ui/poetry.lock +++ b/front/admin_ui/poetry.lock @@ -529 +529 @@ name = "datasets" -version = "2.12.0" +version = "2.13.0" @@ -535,2 +535,2 @@ files = [ - {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"}, - {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"}, + {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, + {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, @@ -553 +552,0 @@ requests = ">=2.19.0" -responses = "<0.19" @@ -562 +561 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -565 +564 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -570 +569 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -1249 +1248 @@ appdirs = "^1.4.4" -datasets = {version = "^2.12.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.0", extras = ["audio", "vision"]} @@ -2605,19 +2603,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "responses" -version = "0.18.0" -description = "A utility library for mocking out the `requests` Python library." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, - {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, -] - -[package.dependencies] -requests = ">=2.0,<3.0" -urllib3 = ">=1.25.10" - -[package.extras] -tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] - diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock index bd34f45d..6b1a8a05 100644 --- a/jobs/cache_maintenance/poetry.lock +++ b/jobs/cache_maintenance/poetry.lock @@ -579 +579 @@ name = "datasets" -version = "2.12.0" +version = "2.13.0" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"}, - {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"}, + {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, + {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, @@ -603 +602,0 @@ requests = ">=2.19.0" -responses = "<0.19" @@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -1027 +1026 @@ appdirs = "^1.4.4" -datasets = {version = "^2.12.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.0", extras = ["audio", "vision"]} @@ -2390,19 +2388,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "responses" -version = "0.18.0" -description = "A utility library for mocking out the `requests` Python library." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, - {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, -] - -[package.dependencies] -requests = ">=2.0,<3.0" -urllib3 = ">=1.25.10" - -[package.extras] -tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] - diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock index bd34f45d..6b1a8a05 100644 --- a/jobs/mongodb_migration/poetry.lock +++ b/jobs/mongodb_migration/poetry.lock @@ -579 +579 @@ name = "datasets" -version = "2.12.0" +version = "2.13.0" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"}, - {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"}, + {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, + {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, @@ -603 +602,0 @@ requests = ">=2.19.0" -responses = "<0.19" @@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -1027 +1026 @@ appdirs = "^1.4.4" -datasets = {version = "^2.12.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.0", extras = ["audio", "vision"]} @@ -2390,19 +2388,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "responses" -version = "0.18.0" -description = "A utility library for mocking out the `requests` Python library." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, - {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, -] - -[package.dependencies] -requests = ">=2.0,<3.0" -urllib3 = ">=1.25.10" - -[package.extras] -tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] - diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock index 1ddc15bc..99b16a6a 100644 --- a/libs/libcommon/poetry.lock +++ b/libs/libcommon/poetry.lock @@ -579 +579 @@ name = "datasets" -version = "2.12.0" +version = "2.13.0" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"}, - {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"}, + {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, + {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, @@ -603 +602,0 @@ requests = ">=2.19.0" -responses = "<0.19" @@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -2371,19 +2369,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "responses" -version = "0.18.0" -description = "A utility library for mocking out the `requests` Python library." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, - {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, -] - -[package.dependencies] -requests = ">=2.0,<3.0" -urllib3 = ">=1.25.10" - -[package.extras] -tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] - @@ -3035 +3015 @@ python-versions = "3.9.15" -content-hash = "3eab4cefa7dd3f05dcf9657d8a1e81f7d7df494f07b6eb0f85fbc97e8a6aa56e" +content-hash = "bfdd77727f041d88e26a4da8c23a6b8434f20c6747f068f1ca3c11d841e38aa1" diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml index 1bc1d006..8dcfc6a6 100644 --- a/libs/libcommon/pyproject.toml +++ b/libs/libcommon/pyproject.toml @@ -10 +10 @@ appdirs = "^1.4.4" -datasets = { extras = ["audio", "vision"], version = "^2.12.0" } +datasets = {version = "^2.13.0", extras = ["audio", "vision"]} diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock index 928e56ca..e138e9d3 100644 --- a/services/admin/poetry.lock +++ b/services/admin/poetry.lock @@ -579 +579 @@ name = "datasets" -version = "2.12.0" +version = "2.13.0" @@ -585,2 +585,2 @@ files = [ - {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"}, - {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"}, + {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, + {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, @@ -603 +602,0 @@ requests = ">=2.19.0" -responses = "<0.19" @@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -1085 +1084 @@ appdirs = "^1.4.4" -datasets = {version = "^2.12.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.0", extras = ["audio", "vision"]} @@ -2452 +2451 @@ description = "A utility library for mocking out the `requests` Python library." -category = "main" +category = "dev" diff --git a/services/api/poetry.lock b/services/api/poetry.lock index 70a86b1c..77c2de7d 100644 --- a/services/api/poetry.lock +++ b/services/api/poetry.lock @@ -621 +621 @@ name = "datasets" -version = "2.12.0" +version = "2.13.0" @@ -627,2 +627,2 @@ files = [ - {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"}, - {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"}, + {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, + {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, @@ -645 +644,0 @@ requests = ">=2.19.0" -responses = "<0.19" @@ -654 +653 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -657 +656 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -662 +661 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -1147 +1146 @@ appdirs = "^1.4.4" -datasets = {version = "^2.12.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.0", extras = ["audio", "vision"]} @@ -2658,19 +2656,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"] -[[package]] -name = "responses" -version = "0.18.0" -description = "A utility library for mocking out the `requests` Python library." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, - {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, -] - -[package.dependencies] -requests = ">=2.0,<3.0" -urllib3 = ">=1.25.10" - -[package.extras] -tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] - @@ -3433 +3413 @@ python-versions = "3.9.15" -content-hash = "8f47c3a3309578c035ed9ed97d3b5b965b9efaabca2c187e5e501303ab3310aa" +content-hash = "781ea6fdf22b868dd68ea1f37735c621a9064e2ee616b3e2538fa8924697b1b2" diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml index d1055331..7fc589ed 100644 --- a/services/api/pyproject.toml +++ b/services/api/pyproject.toml @@ -10 +9,0 @@ cryptography = "^41.0.1" -datasets = { extras = ["audio", "vision"], version = "^2.11.0" } diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 6ae2e954..8d817352 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -910 +910 @@ name = "datasets" -version = "2.12.0" +version = "2.13.0" @@ -916,2 +916,2 @@ files = [ - {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"}, - {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"}, + {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"}, + {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"}, @@ -934 +933,0 @@ requests = ">=2.19.0" -responses = "<0.19" @@ -943 +942 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", -dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -946 +945 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"] -metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] +metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"] @@ -951 +950 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"] -tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] +tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"] @@ -1781 +1780 @@ appdirs = "^1.4.4" -datasets = {version = "^2.12.0", extras = ["audio", "vision"]} +datasets = {version = "^2.13.0", extras = ["audio", "vision"]} @@ -4149,19 +4147,0 @@ rsa = ["oauthlib[signedtoken] (>=3.0.0)"] -[[package]] -name = "responses" -version = "0.18.0" -description = "A utility library for mocking out the `requests` Python library." -category = "main" -optional = false -python-versions = ">=3.7" -files = [ - {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"}, - {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"}, -] - -[package.dependencies] -requests = ">=2.0,<3.0" -urllib3 = ">=1.25.10" - -[package.extras] -tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"] -
8396a0d518d446b24334c0e4daf4841327fdf56f
Andrea Francis Soria Jimenez
2023-06-15T14:58:42
Adding limit for number of configs (#1371)
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl index ffbec5ae..3d9765fc 100644 --- a/chart/templates/_envWorker.tpl +++ b/chart/templates/_envWorker.tpl @@ -86,0 +87,2 @@ +- name: CONFIG_NAMES_MAX_NUMBER + value: {{ .Values.configNames.maxNumber | quote }} diff --git a/chart/values.yaml b/chart/values.yaml index aaae2c89..5294d83c 100644 --- a/chart/values.yaml +++ b/chart/values.yaml @@ -182,0 +183,4 @@ optInOutUrlsScan: +configNames: + # the max number of configs per dataset + maxNumber: 3_000 + diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py index eaf8cc95..f7048d5a 100644 --- a/libs/libcommon/src/libcommon/exceptions.py +++ b/libs/libcommon/src/libcommon/exceptions.py @@ -87,0 +88 @@ CacheableErrorCode = Literal[ + "DatasetWithTooManyConfigsError", @@ -481,0 +483,7 @@ class UnsupportedExternalFilesError(CacheableError): + + +class DatasetWithTooManyConfigsError(CacheableError): + """Raised when the number of configs of a dataset exceeded the limit.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyConfigsError", cause, True) diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index 3419e9aa..bc3fd397 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -219,0 +220,16 @@ class NumbaConfig: +CONFIG_NAMES_MAX_NUMBER = 3_000 + + +@dataclass(frozen=True) +class ConfigNamesConfig: + max_number: int = CONFIG_NAMES_MAX_NUMBER + + @classmethod + def from_env(cls) -> "ConfigNamesConfig": + env = Env(expand_vars=True) + with env.prefixed("CONFIG_NAMES_"): + return cls( + max_number=env.int(name="MAX_NUMBER", default=CONFIG_NAMES_MAX_NUMBER), + ) + + @@ -224,0 +241 @@ class AppConfig: + config_names: ConfigNamesConfig = field(default_factory=ConfigNamesConfig) @@ -240,0 +258 @@ class AppConfig: + config_names=ConfigNamesConfig.from_env(), diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py index b9c447be..b1045c31 100644 --- a/services/worker/src/worker/job_runners/dataset/config_names.py +++ b/services/worker/src/worker/job_runners/dataset/config_names.py @@ -12,0 +13 @@ from libcommon.exceptions import ( + DatasetWithTooManyConfigsError, @@ -30,0 +32 @@ def compute_config_names_response( + max_number: int, @@ -70,0 +73,7 @@ def compute_config_names_response( + + number_of_configs = len(config_name_items) + if number_of_configs > max_number: + raise DatasetWithTooManyConfigsError( + f"The maximun number of configs allowed is {max_number}, dataset has {number_of_configs} configs." + ) + @@ -85 +94,5 @@ class DatasetConfigNamesJobRunner(DatasetCachedJobRunner): - compute_config_names_response(dataset=self.dataset, hf_token=self.app_config.common.hf_token) + compute_config_names_response( + dataset=self.dataset, + hf_token=self.app_config.common.hf_token, + max_number=self.app_config.config_names.max_number, + ) diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py index 9116bb5d..3f0ed076 100644 --- a/services/worker/tests/job_runners/dataset/test_config_names.py +++ b/services/worker/tests/job_runners/dataset/test_config_names.py @@ -5,0 +6 @@ from typing import Callable +from unittest.mock import patch @@ -68,0 +70,27 @@ def test_compute(app_config: AppConfig, hub_public_csv: str, get_job_runner: Get [email protected]( + "max_number_of_configs,error_code", + [ + (1, "DatasetWithTooManyConfigsError"), + (2, None), + (3, None), + ], +) +def test_compute_too_many_configs( + app_config: AppConfig, get_job_runner: GetJobRunner, max_number_of_configs: int, error_code: str +) -> None: + dataset = "dataset" + configs = ["config_1", "config_2"] + job_runner = get_job_runner( + dataset, + replace(app_config, config_names=replace(app_config.config_names, max_number=max_number_of_configs)), + ) + + with patch("worker.job_runners.dataset.config_names.get_dataset_config_names", return_value=configs): + if error_code: + with pytest.raises(CustomError) as exc_info: + job_runner.compute() + assert exc_info.value.code == error_code + else: + assert job_runner.compute() is not None + + diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml index 4f1cf859..d6b2bd92 100644 --- a/tools/docker-compose-datasets-server.yml +++ b/tools/docker-compose-datasets-server.yml @@ -99,0 +100 @@ services: + CONFIG_NAMES_MAX_NUMBER: ${CONFIG_NAMES_MAX_NUMBER-3_000} diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml index 6c7e39af..c50d781f 100644 --- a/tools/docker-compose-dev-datasets-server.yml +++ b/tools/docker-compose-dev-datasets-server.yml @@ -103,0 +104 @@ services: + CONFIG_NAMES_MAX_NUMBER: ${CONFIG_NAMES_MAX_NUMBER-3_000}
364900464d575c160bd74b80e85857b3a84bb03b
Sylvain Lesage
2023-06-14T15:25:09
feat: 🎸 reduce the resources (#1368)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 4d70bcc0..163a431d 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -263 +263 @@ workers: - replicas: 160 + replicas: 20 @@ -278 +278 @@ workers: - replicas: 12 + replicas: 4
3dad30355148f2bf0a886014bafbff87a12719c1
Quentin Lhoest
2023-06-14T14:53:45
Add lock for queue based on mongodb (#1357)
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py index 0a3a5494..fc0dd91e 100644 --- a/libs/libcommon/src/libcommon/constants.py +++ b/libs/libcommon/src/libcommon/constants.py @@ -12,0 +13 @@ QUEUE_COLLECTION_JOBS = "jobsBlue" +QUEUE_COLLECTION_LOCKS = "locks" diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index 9b636995..14f54073 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -5,0 +6 @@ import logging +import time @@ -11 +12,2 @@ from operator import itemgetter -from typing import Generic, List, Optional, Type, TypedDict, TypeVar +from types import TracebackType +from typing import Generic, List, Literal, Optional, Sequence, Type, TypedDict, TypeVar @@ -15,0 +18 @@ from mongoengine import Document, DoesNotExist +from mongoengine.errors import NotUniqueError @@ -20,0 +24 @@ from libcommon.constants import ( + QUEUE_COLLECTION_LOCKS, @@ -200,0 +205,73 @@ class Job(Document): +class Lock(Document): + meta = {"collection": QUEUE_COLLECTION_LOCKS, "db_alias": QUEUE_MONGOENGINE_ALIAS, "indexes": [("key", "job_id")]} + key = StringField(primary_key=True) + job_id = StringField() + + created_at = DateTimeField(required=True) + updated_at = DateTimeField() + + objects = QuerySetManager["Lock"]() + + +class lock: + """ + Provides a simple way of inter-worker communication using a MongoDB lock. + A lock is used to indicate another worker of your application that a resource + or working directory is currently used in a job. + + Example of usage: + + ```python + key = json.dumps({"type": job.type, "dataset": job.dataset}) + with lock(key=key, job_id=job.pk): + ... + ``` + + Or using a try/except: + + ```python + try: + key = json.dumps({"type": job.type, "dataset": job.dataset}) + lock(key=key, job_id=job.pk).acquire() + except TimeoutError: + ... + ``` + """ + + def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = (0.05, 0.05, 0.05, 1, 1, 1, 5)) -> None: + self.key = key + self.job_id = job_id + self.sleeps = sleeps + + def acquire(self) -> None: + try: + Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save() + except NotUniqueError: + pass + for sleep in self.sleeps: + acquired = ( + Lock.objects(key=self.key, job_id__in=[None, self.job_id]).update( + job_id=self.job_id, updated_at=get_datetime() + ) + > 0 + ) + if acquired: + return + logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'") + time.sleep(sleep) + raise TimeoutError("lock couldn't be acquired") + + def release(self) -> None: + Lock.objects(key=self.key, job_id=self.job_id).update(job_id=None, updated_at=get_datetime()) + + def __enter__(self) -> "lock": + self.acquire() + return self + + def __exit__( + self, exctype: Optional[Type[BaseException]], excinst: Optional[BaseException], exctb: Optional[TracebackType] + ) -> Literal[False]: + self.release() + return False + + @@ -817,0 +895 @@ def _clean_queue_database() -> None: + Lock.drop_collection() # type: ignore diff --git a/libs/libcommon/src/libcommon/resources.py b/libs/libcommon/src/libcommon/resources.py index e89121ce..b91e851d 100644 --- a/libs/libcommon/src/libcommon/resources.py +++ b/libs/libcommon/src/libcommon/resources.py @@ -6 +6 @@ from types import TracebackType -from typing import Optional, Type, TypeVar +from typing import Any, Optional, Type, TypeVar @@ -106,0 +107,5 @@ class MongoResource(Resource): + def __reduce__(self) -> tuple[Any, ...]: + # Needed to be able to use the resource in subprocesses in tests (e.g. tests/test_queue.py::test_lock). + # This is because the _client in not picklable. + return (MongoResource, (self.database, self.host, self.mongoengine_alias, self.server_selection_timeout_ms)) + diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py index 36226b22..02b9578e 100644 --- a/libs/libcommon/tests/test_queue.py +++ b/libs/libcommon/tests/test_queue.py @@ -3,0 +4 @@ +import os @@ -4,0 +6,2 @@ from datetime import datetime, timedelta +from multiprocessing import Pool +from pathlib import Path @@ -12 +15 @@ from libcommon.constants import QUEUE_TTL_SECONDS -from libcommon.queue import EmptyQueueError, Job, Queue +from libcommon.queue import EmptyQueueError, Job, Lock, Queue, lock @@ -385,0 +389,27 @@ def test_has_ttl_index_on_finished_at_field() -> None: + + +def increment(tmp_file: Path) -> None: + with open(tmp_file, "r") as f: + current = int(f.read() or 0) + with open(tmp_file, "w") as f: + f.write(str(current + 1)) + + +def locked_increment(tmp_file: Path) -> None: + with lock(key="test_lock", job_id=str(os.getpid())): + increment(tmp_file) + + +def test_lock(tmp_path_factory: pytest.TempPathFactory, queue_mongo_resource: QueueMongoResource) -> None: + tmp_file = Path(tmp_path_factory.mktemp("test_lock") / "tmp.txt") + tmp_file.touch() + max_parallel_jobs = 4 + num_jobs = 42 + + with Pool(max_parallel_jobs, initializer=queue_mongo_resource.allocate) as pool: + pool.map(locked_increment, [tmp_file] * num_jobs) + + expected = num_jobs + with open(tmp_file, "r") as f: + assert int(f.read()) == expected + Lock.objects(key="test_lock").delete()
ca1d155af54b1aa372d9677115d12ce94fd1021b
Sylvain Lesage
2023-06-14T11:49:04
feat: 🎸 run jobs killing routines with some random delay (#1364)
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py index af6bcc0d..9b636995 100644 --- a/libs/libcommon/src/libcommon/queue.py +++ b/libs/libcommon/src/libcommon/queue.py @@ -786 +786 @@ class Queue: - def get_zombies(self, max_seconds_without_heartbeat: int) -> List[JobInfo]: + def get_zombies(self, max_seconds_without_heartbeat: float) -> List[JobInfo]: diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py index b2c2ecf9..3419e9aa 100644 --- a/services/worker/src/worker/config.py +++ b/services/worker/src/worker/config.py @@ -39 +39 @@ class WorkerConfig: - heartbeat_interval_seconds: int = WORKER_HEARTBEAT_INTERVAL_SECONDS + heartbeat_interval_seconds: float = WORKER_HEARTBEAT_INTERVAL_SECONDS @@ -42,2 +42,2 @@ class WorkerConfig: - kill_long_job_interval_seconds: int = WORKER_KILL_LONG_JOB_INTERVAL_SECONDS - kill_zombies_interval_seconds: int = WORKER_KILL_ZOMBIES_INTERVAL_SECONDS + kill_long_job_interval_seconds: float = WORKER_KILL_LONG_JOB_INTERVAL_SECONDS + kill_zombies_interval_seconds: float = WORKER_KILL_ZOMBIES_INTERVAL_SECONDS @@ -45 +45 @@ class WorkerConfig: - max_job_duration_seconds: int = WORKER_MAX_JOB_DURATION_SECONDS + max_job_duration_seconds: float = WORKER_MAX_JOB_DURATION_SECONDS @@ -49 +49 @@ class WorkerConfig: - sleep_seconds: int = WORKER_SLEEP_SECONDS + sleep_seconds: float = WORKER_SLEEP_SECONDS @@ -59 +59 @@ class WorkerConfig: - heartbeat_interval_seconds=env.int( + heartbeat_interval_seconds=env.float( @@ -64 +64 @@ class WorkerConfig: - kill_long_job_interval_seconds=env.int( + kill_long_job_interval_seconds=env.float( @@ -67 +67 @@ class WorkerConfig: - kill_zombies_interval_seconds=env.int( + kill_zombies_interval_seconds=env.float( @@ -71 +71 @@ class WorkerConfig: - max_job_duration_seconds=env.int( + max_job_duration_seconds=env.float( @@ -77 +77 @@ class WorkerConfig: - sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_SLEEP_SECONDS), + sleep_seconds=env.float(name="SLEEP_SECONDS", default=WORKER_SLEEP_SECONDS), diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py index a0acf747..d663a775 100644 --- a/services/worker/src/worker/executor.py +++ b/services/worker/src/worker/executor.py @@ -8 +8,2 @@ from datetime import datetime, timedelta -from typing import Any, Callable, Optional +from random import random +from typing import Any, Callable, Optional, Tuple, Union @@ -27 +28,5 @@ async def every( - func: Callable[..., Optional[Any]], *args: Any, seconds: int, stop_on: Optional[Any] = None, **kwargs: Any + func: Callable[..., Optional[Any]], + *args: Any, + seconds: Union[float, Tuple[float, float]], + stop_on: Optional[Any] = None, + **kwargs: Any, @@ -33 +38,4 @@ async def every( - await asyncio.sleep(seconds) + delay = ( + seconds[0] + (seconds[1] - seconds[0]) * random() if isinstance(seconds, tuple) else seconds # nosec B311 + ) + await asyncio.sleep(delay) @@ -81 +89,9 @@ class WorkerExecutor: - loop.create_task(every(self.kill_zombies, seconds=self.app_config.worker.kill_zombies_interval_seconds)) + loop.create_task( + every( + self.kill_zombies, + seconds=( + self.app_config.worker.kill_zombies_interval_seconds * 0.5, + self.app_config.worker.kill_zombies_interval_seconds * 1.5, + ), + ) + ) @@ -86 +102,4 @@ class WorkerExecutor: - seconds=self.app_config.worker.kill_long_job_interval_seconds, + seconds=( + self.app_config.worker.kill_long_job_interval_seconds * 0.5, + self.app_config.worker.kill_long_job_interval_seconds * 1.5, + ), @@ -90 +109 @@ class WorkerExecutor: - every(self.is_worker_alive, worker_loop_executor=worker_loop_executor, seconds=1, stop_on=False) + every(self.is_worker_alive, worker_loop_executor=worker_loop_executor, seconds=1.0, stop_on=False)
7bebee49ebf214eefb6410241fd544dbb4454aa4
Sylvain Lesage
2023-06-14T09:40:52
ci: 🎡 don't run doc CI if no doc has been changed (#1363)
diff --git a/.github/workflows/doc-pr-build.yml b/.github/workflows/doc-pr-build.yml index 9c920c4f..6ee6a305 100644 --- a/.github/workflows/doc-pr-build.yml +++ b/.github/workflows/doc-pr-build.yml @@ -7,0 +8,3 @@ on: + paths: + - "docs/**" + - ".github/workflows/doc-pr-build.yml" diff --git a/.github/workflows/doc-pr-delete-trigger.yml b/.github/workflows/doc-pr-delete-trigger.yml index ebc0d0ad..89037f89 100644 --- a/.github/workflows/doc-pr-delete-trigger.yml +++ b/.github/workflows/doc-pr-delete-trigger.yml @@ -8,0 +9,3 @@ on: + paths: + - "docs/**" + - ".github/workflows/doc-pr-delete-trigger.yml"
7b266d24ee81c2f5b86fbd352466e97b34d46c99
Sylvain Lesage
2023-06-14T09:25:58
style: 💄 fix style for github action files (#1362)
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml index 3e0a01f5..0126945c 100644 --- a/.github/workflows/_e2e_tests.yml +++ b/.github/workflows/_e2e_tests.yml @@ -21,2 +21 @@ jobs: - - - name: Checkout + - name: Checkout @@ -24,2 +23 @@ jobs: - - - name: Build and launch the services (no cache) + - name: Build and launch the services (no cache) @@ -48,2 +46 @@ jobs: - - - name: Install poetry + - name: Install poetry @@ -51,2 +48 @@ jobs: - - - name: Use Python + - name: Use Python @@ -56 +52 @@ jobs: - cache: 'poetry' + cache: "poetry" @@ -59,2 +55 @@ jobs: - - - name: Install dependencies + - name: Install dependencies @@ -64,2 +59 @@ jobs: - - - name: End-to-end tests + - name: End-to-end tests @@ -85 +79 @@ jobs: - WORKER_SLEEP_SECONDS : "1" + WORKER_SLEEP_SECONDS: "1" diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml index 9f9638e0..c285df42 100644 --- a/.github/workflows/_quality-python.yml +++ b/.github/workflows/_quality-python.yml @@ -5 +5 @@ name: Check Python code quality -on: +on: @@ -31 +31 @@ jobs: - cache: 'poetry' + cache: "poetry" diff --git a/.github/workflows/_unit-tests-python.yml b/.github/workflows/_unit-tests-python.yml index 78d2daef..763748bf 100644 --- a/.github/workflows/_unit-tests-python.yml +++ b/.github/workflows/_unit-tests-python.yml @@ -5 +5 @@ name: Launch Python unit tests -on: +on: @@ -32 +32 @@ jobs: - cache: 'poetry' + cache: "poetry" @@ -48 +48 @@ jobs: - METRICS_MONGO_URL: mongodb://localhost:${{ env.mongo-port }} + METRICS_MONGO_URL: mongodb://localhost:${{ env.mongo-port }} diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml index d938a7f6..878e3e3b 100644 --- a/.github/workflows/cd.yml +++ b/.github/workflows/cd.yml @@ -89 +89 @@ jobs: - needs: [ build-and-push-images ] + needs: [build-and-push-images] diff --git a/.github/workflows/chart-pr.yml b/.github/workflows/chart-pr.yml index 6e33e252..cd5789bc 100644 --- a/.github/workflows/chart-pr.yml +++ b/.github/workflows/chart-pr.yml @@ -8,2 +8,2 @@ on: - - 'chart/**' - - '.github/workflows/chart-pr.yml' + - "chart/**" + - ".github/workflows/chart-pr.yml" diff --git a/.github/workflows/doc-build.yml b/.github/workflows/doc-build.yml index 7378ed14..8dc7ac99 100644 --- a/.github/workflows/doc-build.yml +++ b/.github/workflows/doc-build.yml @@ -11,2 +11,2 @@ on: - - 'docs/**' - - '.github/workflows/doc-build.yml' + - "docs/**" + - ".github/workflows/doc-build.yml" @@ -15 +15 @@ jobs: - build: + build: diff --git a/.github/workflows/doc-pr-build.yml b/.github/workflows/doc-pr-build.yml index 7dc41015..9c920c4f 100644 --- a/.github/workflows/doc-pr-build.yml +++ b/.github/workflows/doc-pr-build.yml @@ -8 +7,0 @@ on: - diff --git a/.github/workflows/doc-pr-delete-trigger.yml b/.github/workflows/doc-pr-delete-trigger.yml index b23e2591..ebc0d0ad 100644 --- a/.github/workflows/doc-pr-delete-trigger.yml +++ b/.github/workflows/doc-pr-delete-trigger.yml @@ -8,2 +8 @@ on: - types: [ closed ] - + types: [closed] diff --git a/.github/workflows/doc-pr-upload.yml b/.github/workflows/doc-pr-upload.yml index 517a69ca..99c3a4cf 100644 --- a/.github/workflows/doc-pr-upload.yml +++ b/.github/workflows/doc-pr-upload.yml @@ -19 +19 @@ jobs: - comment_bot_token: ${{ secrets.COMMENT_BOT_TOKEN }} \ No newline at end of file + comment_bot_token: ${{ secrets.COMMENT_BOT_TOKEN }} diff --git a/.github/workflows/e2e.yml b/.github/workflows/e2e.yml index 81a6ad18..3fbd1662 100644 --- a/.github/workflows/e2e.yml +++ b/.github/workflows/e2e.yml @@ -11,9 +11,9 @@ on: - - 'e2e/**' - - 'libs/**' - - 'services/**' - - 'chart/static-files/openapi.json' - - '.github/workflows/_e2e_tests.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/e2e.yml' - - 'tools/Python.mk' - - 'tools/docker-compose-datasets-server.yml' + - "e2e/**" + - "libs/**" + - "services/**" + - "chart/static-files/openapi.json" + - ".github/workflows/_e2e_tests.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/e2e.yml" + - "tools/Python.mk" + - "tools/docker-compose-datasets-server.yml" @@ -22,9 +22,9 @@ on: - - 'e2e/**' - - 'libs/**' - - 'services/**' - - 'chart/static-files/openapi.json' - - '.github/workflows/_e2e_tests.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/e2e.yml' - - 'tools/Python.mk' - - 'tools/docker-compose-datasets-server.yml' + - "e2e/**" + - "libs/**" + - "services/**" + - "chart/static-files/openapi.json" + - ".github/workflows/_e2e_tests.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/e2e.yml" + - "tools/Python.mk" + - "tools/docker-compose-datasets-server.yml" diff --git a/.github/workflows/j-cache-maintenance.yml b/.github/workflows/j-cache-maintenance.yml index 74052dcf..f0ad6474 100644 --- a/.github/workflows/j-cache-maintenance.yml +++ b/.github/workflows/j-cache-maintenance.yml @@ -11,6 +11,6 @@ on: - - 'jobs/cache_maintenance/**' - - 'libs/libcommon/**' - - '.github/workflows/j-cache-maintenance.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "jobs/cache_maintenance/**" + - "libs/libcommon/**" + - ".github/workflows/j-cache-maintenance.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" @@ -19,6 +19,6 @@ on: - - 'jobs/cache_maintenance/**' - - 'libs/libcommon/**' - - '.github/workflows/j-cache-maintenance.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "jobs/cache_maintenance/**" + - "libs/libcommon/**" + - ".github/workflows/j-cache-maintenance.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" diff --git a/.github/workflows/j-migration.yml b/.github/workflows/j-migration.yml index 201eaea8..40ffa4ed 100644 --- a/.github/workflows/j-migration.yml +++ b/.github/workflows/j-migration.yml @@ -11,6 +11,6 @@ on: - - 'jobs/mongodb_migration/**' - - 'libs/libcommon/**' - - '.github/workflows/j-migration.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "jobs/mongodb_migration/**" + - "libs/libcommon/**" + - ".github/workflows/j-migration.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" @@ -19,6 +19,6 @@ on: - - 'jobs/mongodb_migration/**' - - 'libs/libcommon/**' - - '.github/workflows/j-migration.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "jobs/mongodb_migration/**" + - "libs/libcommon/**" + - ".github/workflows/j-migration.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" diff --git a/.github/workflows/l-libcommon.yml b/.github/workflows/l-libcommon.yml index 3ba6a069..96a8c1ae 100644 --- a/.github/workflows/l-libcommon.yml +++ b/.github/workflows/l-libcommon.yml @@ -11,5 +11,5 @@ on: - - 'libs/libcommon/**' - - '.github/workflows/l-libcommon.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "libs/libcommon/**" + - ".github/workflows/l-libcommon.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" @@ -18,5 +18,5 @@ on: - - 'libs/libcommon/**' - - '.github/workflows/l-libcommon.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "libs/libcommon/**" + - ".github/workflows/l-libcommon.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" diff --git a/.github/workflows/openapi-spec.yml b/.github/workflows/openapi-spec.yml index a8062cfc..bfae8dcf 100644 --- a/.github/workflows/openapi-spec.yml +++ b/.github/workflows/openapi-spec.yml @@ -11,2 +11,2 @@ on: - - 'chart/static-files/opanapi.json' - - '.github/workflows/openapi.yml' + - "chart/static-files/opanapi.json" + - ".github/workflows/openapi.yml" @@ -15,2 +15,2 @@ on: - - 'chart/static-files/opanapi.json' - - '.github/workflows/openapi.yml' + - "chart/static-files/opanapi.json" + - ".github/workflows/openapi.yml" @@ -38 +38 @@ jobs: - cache: 'poetry' + cache: "poetry" diff --git a/.github/workflows/publish-helm.yml b/.github/workflows/publish-helm.yml index fe9b8997..0c176e80 100644 --- a/.github/workflows/publish-helm.yml +++ b/.github/workflows/publish-helm.yml @@ -9 +9 @@ on: - - 'chart/**' + - "chart/**" diff --git a/.github/workflows/s-admin.yml b/.github/workflows/s-admin.yml index 7dd4c411..0be67500 100644 --- a/.github/workflows/s-admin.yml +++ b/.github/workflows/s-admin.yml @@ -11,6 +11,6 @@ on: - - 'libs/libcommon/**' - - 'services/admin/**' - - '.github/workflows/s-admin.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "libs/libcommon/**" + - "services/admin/**" + - ".github/workflows/s-admin.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" @@ -19,6 +19,6 @@ on: - - 'libs/libcommon/**' - - 'services/admin/**' - - '.github/workflows/s-admin.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "libs/libcommon/**" + - "services/admin/**" + - ".github/workflows/s-admin.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" diff --git a/.github/workflows/s-api.yml b/.github/workflows/s-api.yml index 972d6639..63c440cc 100644 --- a/.github/workflows/s-api.yml +++ b/.github/workflows/s-api.yml @@ -11,6 +11,6 @@ on: - - 'libs/libcommon/**' - - 'services/api/**' - - '.github/workflows/s-api.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "libs/libcommon/**" + - "services/api/**" + - ".github/workflows/s-api.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" @@ -19,6 +19,6 @@ on: - - 'libs/libcommon/**' - - 'services/api/**' - - '.github/workflows/s-api.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' + - "libs/libcommon/**" + - "services/api/**" + - ".github/workflows/s-api.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" diff --git a/.github/workflows/s-worker.yml b/.github/workflows/s-worker.yml index ffb3be37..146a6bcf 100644 --- a/.github/workflows/s-worker.yml +++ b/.github/workflows/s-worker.yml @@ -11,7 +11,7 @@ on: - - 'libs/libcommon/**' - - 'services/worker/**' - - '.github/workflows/s-worker.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' - - 'vendors/' + - "libs/libcommon/**" + - "services/worker/**" + - ".github/workflows/s-worker.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" + - "vendors/" @@ -20,7 +20,7 @@ on: - - 'libs/libcommon/**' - - 'services/worker/**' - - '.github/workflows/s-worker.yml' - - '.github/workflows/_quality-python.yml' - - '.github/workflows/_unit-tests-python.yml' - - 'tools/docker-compose-mongo.yml' - - 'vendors/' + - "libs/libcommon/**" + - "services/worker/**" + - ".github/workflows/s-worker.yml" + - ".github/workflows/_quality-python.yml" + - ".github/workflows/_unit-tests-python.yml" + - "tools/docker-compose-mongo.yml" + - "vendors/" diff --git a/.github/workflows/stale.yml b/.github/workflows/stale.yml index ede171dc..9586e941 100644 --- a/.github/workflows/stale.yml +++ b/.github/workflows/stale.yml @@ -18 +18 @@ jobs: - - uses: actions/checkout@v3 + - uses: actions/checkout@v3 @@ -20,4 +20,4 @@ jobs: - - name: Setup Python - uses: actions/setup-python@v4 - with: - python-version: 3.7 + - name: Setup Python + uses: actions/setup-python@v4 + with: + python-version: 3.7 @@ -25,6 +25,6 @@ jobs: - - name: Install requirements - run: | - pip install PyGithub - - name: Close stale issues - run: | - python tools/stale.py + - name: Install requirements + run: | + pip install PyGithub + - name: Close stale issues + run: | + python tools/stale.py
7b4b78f04473d8f00e7c2226381982a9a84ab48c
Sylvain Lesage
2023-06-13T22:27:15
feat: 🎸 add resources for temporary load (#1360)
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 163a431d..4d70bcc0 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -263 +263 @@ workers: - replicas: 20 + replicas: 160 @@ -278 +278 @@ workers: - replicas: 4 + replicas: 12
2990bbc93b6622d536bb8bb50ea7fdfa72b17112
Steven Liu
2023-06-12T18:40:42
separate page for each library (#1354)
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml index dcd3ef45..0767f5ab 100644 --- a/docs/source/_toctree.yml +++ b/docs/source/_toctree.yml @@ -19,2 +19,10 @@ - - local: parquet_process - title: Query datasets from Datasets Server + - title: Query datasets from Datasets Server + sections: + - local: parquet_process + title: Overview + - local: duckdb + title: DuckDB + - local: polars + title: Polars + - local: pandas + title: Pandas diff --git a/docs/source/duckdb.mdx b/docs/source/duckdb.mdx new file mode 100644 index 00000000..b55fcd6c --- /dev/null +++ b/docs/source/duckdb.mdx @@ -0,0 +1,90 @@ +# DuckDB + +[DuckDB](https://duckdb.org/docs/) is a database that supports reading and querying Parquet files really fast. Begin by creating a connection to DuckDB, and then install and load the [`httpfs`](https://duckdb.org/docs/extensions/httpfs.html) extension to read and write remote files: + +<inferencesnippet> +<python> +```py +import duckdb + +url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet" + +con = duckdb.connect() +con.execute("INSTALL httpfs;") +con.execute("LOAD httpfs;") +``` +</python> +<js> +```js +var duckdb = require('duckdb'); +var db = new duckdb.Database(':memory:'); +var con = db.connect(); +con.exec('INSTALL httpfs'); +con.exec('LOAD httpfs'); + +const url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet" +``` +</js> +</inferencesnippet> + +Now you can write and execute your SQL query on the Parquet file: + +<inferencesnippet> +<python> +```py +con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '{url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)") +┌───────────┬──────────────┬────────────────────┐ +│ horoscope │ count_star() │ avg_blog_length │ +│ varchar │ int64 │ double │ +├───────────┼──────────────┼────────────────────┤ +│ Aquarius │ 34062 │ 1129.218836239798 │ +│ Cancer │ 41509 │ 1098.366812016671 │ +│ Capricorn │ 33961 │ 1073.2002002296751 │ +│ Libra │ 40302 │ 1072.0718326633914 │ +│ Leo │ 40587 │ 1064.0536871412028 │ +└───────────┴──────────────┴────────────────────┘ +``` +</python> +<js> +```js +con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '${url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) { + if (err) { + throw err; + } + console.log(res) +}); +``` +</js> +</inferencesnippet> + +To query multiple files - for example, if the dataset is sharded: + +<inferencesnippet> +<python> +```py +con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet({urls[:2]}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)") +┌─────────────┬──────────────┬────────────────────┐ +│ horoscope │ count_star() │ avg_blog_length │ +│ varchar │ int64 │ double │ +├─────────────┼──────────────┼────────────────────┤ +│ Aquarius │ 49568 │ 1125.8306770497095 │ +│ Cancer │ 63512 │ 1097.95608703867 │ +│ Libra │ 60304 │ 1060.6110539931017 │ +│ Capricorn │ 49402 │ 1059.5552609206104 │ +│ Sagittarius │ 50431 │ 1057.4589835616982 │ +└─────────────┴──────────────┴────────────────────┘ +``` +</python> +<js> +```js +con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet(${JSON.stringify(urls)}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) { + if (err) { + throw err; + } + console.log(res) +}); +``` +</js> +</inferencesnippet> + +[DuckDB-Wasm](https://duckdb.org/docs/api/wasm), a package powered by [WebAssembly](https://webassembly.org/), is also available for running DuckDB in any browser. This could be useful, for instance, if you want to create a web app to query Parquet files from the browser! diff --git a/docs/source/pandas.mdx b/docs/source/pandas.mdx new file mode 100644 index 00000000..f15c4d29 --- /dev/null +++ b/docs/source/pandas.mdx @@ -0,0 +1,29 @@ +# Pandas + +[Pandas](https://pandas.pydata.org/docs/index.html) is a popular DataFrame library for data analysis. + +To read from a single Parquet file, use the [`read_parquet`](https://pandas.pydata.org/docs/reference/api/pandas.read_parquet.html) function to read it into a DataFrame: + +```py +import pandas as pd + +df = ( + pd.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") + .groupby('horoscope')['text'] + .apply(lambda x: x.str.len().mean()) + .sort_values(ascending=False) + .head(5) +) +``` + +To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pandas.pydata.org/docs/reference/api/pandas.concat.html) function to concatenate the files into a single DataFrame: + +```py +df = ( + pd.concat([pd.read_parquet(url) for url in urls]) + .groupby('horoscope')['text'] + .apply(lambda x: x.str.len().mean()) + .sort_values(ascending=False) + .head(5) +) +``` \ No newline at end of file diff --git a/docs/source/parquet_process.mdx b/docs/source/parquet_process.mdx index 40798731..1ace2523 100644 --- a/docs/source/parquet_process.mdx +++ b/docs/source/parquet_process.mdx @@ -1 +1 @@ -# Query datasets from Datasets Server +# Overview @@ -3 +3 @@ -Datasets Server automatically converts and publishes datasets on the Hub as Parquet files. [Parquet](https://parquet.apache.org/docs/) files are column-based and they shine when you're working with big data. There are several ways you can work with Parquet files, and this guide will show you how to: +Datasets Server automatically converts and publishes public datasets less than 5GB on the Hub as Parquet files. [Parquet](https://parquet.apache.org/docs/) files are column-based and they shine when you're working with big data. There are several different libraries you can use to work with the published Parquet files: @@ -5,231 +5,3 @@ Datasets Server automatically converts and publishes datasets on the Hub as Parq -- read and query Parquet files with Pandas and Polars -- connect, read and query Parquet files with DuckDB and DuckDB-Wasm - -## Polars - -[Polars](https://pola-rs.github.io/polars-book/user-guide/introduction.html) is a fast DataFrame library written in Rust with Arrow as its foundation. - -<Tip> - -💡 Learn more about how to get the dataset URLs in the [List Parquet files](parquet) guide. - -</Tip> - -Let's start by grabbing the URLs to the `train` split of the [`blog_authorship_corpus`](https://huggingface.co/datasets/blog_authorship_corpus) dataset from Datasets Server: - -```py -r = requests.get("https://datasets-server.huggingface.co/parquet?dataset=blog_authorship_corpus") -j = r.json() -urls = [f['url'] for f in j['parquet_files'] if f['split'] == 'train'] -urls -['https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet', - 'https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00001-of-00002.parquet'] -``` - -To read from a single Parquet file, use the [`read_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.read_parquet.html) function to read it into a DataFrame and then execute your query: - -```py -import polars as pl - -df = ( - pl.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") - .groupby("horoscope") - .agg( - [ - pl.count(), - pl.col("text").str.n_chars().mean().alias("avg_blog_length") - ] - ) - .sort("avg_blog_length", descending=True) - .limit(5) -) -print(df) -shape: (5, 3) -┌───────────┬───────┬─────────────────┐ -│ horoscope ┆ count ┆ avg_blog_length │ -│ --- ┆ --- ┆ --- │ -│ str ┆ u32 ┆ f64 │ -╞═══════════╪═══════╪═════════════════╡ -│ Aquarius ┆ 34062 ┆ 1129.218836 │ -│ Cancer ┆ 41509 ┆ 1098.366812 │ -│ Capricorn ┆ 33961 ┆ 1073.2002 │ -│ Libra ┆ 40302 ┆ 1072.071833 │ -│ Leo ┆ 40587 ┆ 1064.053687 │ -└───────────┴───────┴─────────────────┘ -``` - -To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.concat.html) function to concatenate the files into a single DataFrame: - -```py -import polars as pl -df = ( - pl.concat([pl.read_parquet(url) for url in urls]) - .groupby("horoscope") - .agg( - [ - pl.count(), - pl.col("text").str.n_chars().mean().alias("avg_blog_length") - ] - ) - .sort("avg_blog_length", descending=True) - .limit(5) -) -print(df) -shape: (5, 3) -┌─────────────┬───────┬─────────────────┐ -│ horoscope ┆ count ┆ avg_blog_length │ -│ --- ┆ --- ┆ --- │ -│ str ┆ u32 ┆ f64 │ -╞═════════════╪═══════╪═════════════════╡ -│ Aquarius ┆ 49568 ┆ 1125.830677 │ -│ Cancer ┆ 63512 ┆ 1097.956087 │ -│ Libra ┆ 60304 ┆ 1060.611054 │ -│ Capricorn ┆ 49402 ┆ 1059.555261 │ -│ Sagittarius ┆ 50431 ┆ 1057.458984 │ -└─────────────┴───────┴─────────────────┘ -``` - -### Lazy API - -Polars offers a [lazy API](https://pola-rs.github.io/polars-book/user-guide/lazy-api/intro.html) that is more performant and memory-efficient for large Parquet files. The LazyFrame API keeps track of what you want to do, and it'll only execute the entire query when you're ready. This way, the lazy API doesn't load everything into RAM beforehand, and it allows you to work with datasets larger than your available RAM. - -To lazily read a Parquet file, use the [`scan_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.scan_parquet.html) function instead. Then, execute the entire query with the [`collect`](https://pola-rs.github.io/polars/py-polars/html/reference/lazyframe/api/polars.LazyFrame.collect.html) function: - -```py -import polars as pl - -q = ( - pl.scan_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") - .groupby("horoscope") - .agg( - [ - pl.count(), - pl.col("text").str.n_chars().mean().alias("avg_blog_length") - ] - ) - .sort("avg_blog_length", descending=True) - .limit(5) -) -df = q.collect() -``` - -## Pandas - -You can also use the popular Pandas DataFrame library to read Parquet files. - -To read from a single Parquet file, use the [`read_parquet`](https://pandas.pydata.org/docs/reference/api/pandas.read_parquet.html) function to read it into a DataFrame: - -```py -import pandas as pd - -df = ( - pd.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") - .groupby('horoscope')['text'] - .apply(lambda x: x.str.len().mean()) - .sort_values(ascending=False) - .head(5) -) -``` - -To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pandas.pydata.org/docs/reference/api/pandas.concat.html) function to concatenate the files into a single DataFrame: - -```py -df = ( - pd.concat([pd.read_parquet(url) for url in urls]) - .groupby('horoscope')['text'] - .apply(lambda x: x.str.len().mean()) - .sort_values(ascending=False) - .head(5) -) -``` - -## DuckDB - -[DuckDB](https://duckdb.org/docs/) is a database that supports reading and querying Parquet files really fast. Begin by creating a connection to DuckDB, and then install and load the [`httpfs`](https://duckdb.org/docs/extensions/httpfs.html) extension to read and write remote files: - -<inferencesnippet> -<python> -```py -import duckdb - -url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet" - -con = duckdb.connect() -con.execute("INSTALL httpfs;") -con.execute("LOAD httpfs;") -``` -</python> -<js> -```js -var duckdb = require('duckdb'); -var db = new duckdb.Database(':memory:'); -var con = db.connect(); -con.exec('INSTALL httpfs'); -con.exec('LOAD httpfs'); - -const url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet" -``` -</js> -</inferencesnippet> - -Now you can write and execute your SQL query on the Parquet file: - -<inferencesnippet> -<python> -```py -con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '{url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)") -┌───────────┬──────────────┬────────────────────┐ -│ horoscope │ count_star() │ avg_blog_length │ -│ varchar │ int64 │ double │ -├───────────┼──────────────┼────────────────────┤ -│ Aquarius │ 34062 │ 1129.218836239798 │ -│ Cancer │ 41509 │ 1098.366812016671 │ -│ Capricorn │ 33961 │ 1073.2002002296751 │ -│ Libra │ 40302 │ 1072.0718326633914 │ -│ Leo │ 40587 │ 1064.0536871412028 │ -└───────────┴──────────────┴────────────────────┘ -``` -</python> -<js> -```js -con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '${url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) { - if (err) { - throw err; - } - console.log(res) -}); -``` -</js> -</inferencesnippet> - -To query multiple files - for example, if the dataset is sharded: - -<inferencesnippet> -<python> -```py -con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet({urls[:2]}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)") -┌─────────────┬──────────────┬────────────────────┐ -│ horoscope │ count_star() │ avg_blog_length │ -│ varchar │ int64 │ double │ -├─────────────┼──────────────┼────────────────────┤ -│ Aquarius │ 49568 │ 1125.8306770497095 │ -│ Cancer │ 63512 │ 1097.95608703867 │ -│ Libra │ 60304 │ 1060.6110539931017 │ -│ Capricorn │ 49402 │ 1059.5552609206104 │ -│ Sagittarius │ 50431 │ 1057.4589835616982 │ -└─────────────┴──────────────┴────────────────────┘ -``` -</python> -<js> -```js -con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet(${JSON.stringify(urls)}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) { - if (err) { - throw err; - } - console.log(res) -}); -``` -</js> -</inferencesnippet> - -[DuckDB-Wasm](https://duckdb.org/docs/api/wasm), a package powered by [WebAssembly](https://webassembly.org/), is also available for running DuckDB in any browser. This could be useful, for instance, if you want to create a web app to query Parquet files from the browser! +- [Polars](https://pola-rs.github.io/polars-book/user-guide/), a Rust based DataFrame library +- [Pandas](https://pandas.pydata.org/docs/index.html), a data analysis tool for working with data structures +- [DuckDB](https://duckdb.org/docs/), a high-performance SQL database for analytical queries \ No newline at end of file diff --git a/docs/source/polars.mdx b/docs/source/polars.mdx new file mode 100644 index 00000000..a02c2167 --- /dev/null +++ b/docs/source/polars.mdx @@ -0,0 +1,107 @@ +# Polars + +[Polars](https://pola-rs.github.io/polars-book/user-guide/) is a fast DataFrame library written in Rust with Arrow as its foundation. + +<Tip> + +💡 Learn more about how to get the dataset URLs in the [List Parquet files](parquet) guide. + +</Tip> + +Let's start by grabbing the URLs to the `train` split of the [`blog_authorship_corpus`](https://huggingface.co/datasets/blog_authorship_corpus) dataset from Datasets Server: + +```py +r = requests.get("https://datasets-server.huggingface.co/parquet?dataset=blog_authorship_corpus") +j = r.json() +urls = [f['url'] for f in j['parquet_files'] if f['split'] == 'train'] +urls +['https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet', + 'https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00001-of-00002.parquet'] +``` + +To read from a single Parquet file, use the [`read_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.read_parquet.html) function to read it into a DataFrame and then execute your query: + +```py +import polars as pl + +df = ( + pl.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") + .groupby("horoscope") + .agg( + [ + pl.count(), + pl.col("text").str.n_chars().mean().alias("avg_blog_length") + ] + ) + .sort("avg_blog_length", descending=True) + .limit(5) +) +print(df) +shape: (5, 3) +┌───────────┬───────┬─────────────────┐ +│ horoscope ┆ count ┆ avg_blog_length │ +│ --- ┆ --- ┆ --- │ +│ str ┆ u32 ┆ f64 │ +╞═══════════╪═══════╪═════════════════╡ +│ Aquarius ┆ 34062 ┆ 1129.218836 │ +│ Cancer ┆ 41509 ┆ 1098.366812 │ +│ Capricorn ┆ 33961 ┆ 1073.2002 │ +│ Libra ┆ 40302 ┆ 1072.071833 │ +│ Leo ┆ 40587 ┆ 1064.053687 │ +└───────────┴───────┴─────────────────┘ +``` + +To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.concat.html) function to concatenate the files into a single DataFrame: + +```py +import polars as pl +df = ( + pl.concat([pl.read_parquet(url) for url in urls]) + .groupby("horoscope") + .agg( + [ + pl.count(), + pl.col("text").str.n_chars().mean().alias("avg_blog_length") + ] + ) + .sort("avg_blog_length", descending=True) + .limit(5) +) +print(df) +shape: (5, 3) +┌─────────────┬───────┬─────────────────┐ +│ horoscope ┆ count ┆ avg_blog_length │ +│ --- ┆ --- ┆ --- │ +│ str ┆ u32 ┆ f64 │ +╞═════════════╪═══════╪═════════════════╡ +│ Aquarius ┆ 49568 ┆ 1125.830677 │ +│ Cancer ┆ 63512 ┆ 1097.956087 │ +│ Libra ┆ 60304 ┆ 1060.611054 │ +│ Capricorn ┆ 49402 ┆ 1059.555261 │ +│ Sagittarius ┆ 50431 ┆ 1057.458984 │ +└─────────────┴───────┴─────────────────┘ +``` + +## Lazy API + +Polars offers a [lazy API](https://pola-rs.github.io/polars-book/user-guide/lazy/using/) that is more performant and memory-efficient for large Parquet files. The LazyFrame API keeps track of what you want to do, and it'll only execute the entire query when you're ready. This way, the lazy API doesn't load everything into RAM beforehand, and it allows you to work with datasets larger than your available RAM. + +To lazily read a Parquet file, use the [`scan_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.scan_parquet.html) function instead. Then, execute the entire query with the [`collect`](https://pola-rs.github.io/polars/py-polars/html/reference/lazyframe/api/polars.LazyFrame.collect.html) function: + +```py +import polars as pl + +q = ( + pl.scan_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet") + .groupby("horoscope") + .agg( + [ + pl.count(), + pl.col("text").str.n_chars().mean().alias("avg_blog_length") + ] + ) + .sort("avg_blog_length", descending=True) + .limit(5) +) +df = q.collect() +``` \ No newline at end of file
8c53732d93921822a96c9b256fd26f449519b069
Sylvain Lesage
2023-06-12T16:29:05
refactor: 💡 rename required_by_dataset_viewer: enables_preview (#1353)
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json index 296ba170..11b6b1fe 100644 --- a/chart/static-files/openapi.json +++ b/chart/static-files/openapi.json @@ -293,4 +293 @@ - "enum": [ - "RowsPostProcessingError", - "UnexpectedError" - ] + "enum": ["RowsPostProcessingError", "UnexpectedError"] @@ -899 +896 @@ - "required": ["valid"], + "required": ["preview", "viewer", "valid"], @@ -901 +898,5 @@ - "valid": { + "preview": { + "type": "array", + "items": { "type": "string" } + }, + "viewer": { @@ -903,0 +905,5 @@ + }, + "valid": { + "type": "array", + "items": { "type": "string" }, + "deprecated": true @@ -2216 +2222 @@ - "maximum": 100, + "maximum": 100, @@ -3149 +3155 @@ - "description": "The list of the Hub datasets that work without an error (for /splits and /first-rows).", + "description": "The lists of the Hub datasets that work without an error, by type. It lists the datasets that have a Dataset Viewer (i.e. have been converted to Parquet format, and can be paginated) and the datasets that only have the Dataset Preview (the first 100 rows).", @@ -3151 +3157 @@ - "description": "See Valid datasets (Hub docs)", + "description": "See Valid datasets (Hub docs)", @@ -3175,0 +3182,9 @@ + "preview": [ + "0n1xus/codexglue", + "0x7194633/rupile", + "AHussain0418/day2_data" + ], + "viewer": [ + "0n1xus/pytorrent-standalone", + "51la5/keyword-extraction" + ], diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py index 8a074b41..e4d63e70 100644 --- a/libs/libcommon/src/libcommon/config.py +++ b/libs/libcommon/src/libcommon/config.py @@ -217 +217 @@ class ProcessingGraphConfig: - "required_by_dataset_viewer": True, + "enables_preview": True, @@ -240 +240 @@ class ProcessingGraphConfig: - "required_by_dataset_viewer": True, + "enables_preview": True, @@ -266,0 +267 @@ class ProcessingGraphConfig: + "enables_viewer": True, diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py index 1b50b779..f6f1e636 100644 --- a/libs/libcommon/src/libcommon/processing_graph.py +++ b/libs/libcommon/src/libcommon/processing_graph.py @@ -50 +50,2 @@ class ProcessingStepSpecification(TypedDict, total=False): - required_by_dataset_viewer: Literal[True] + enables_preview: Literal[True] + enables_viewer: Literal[True] @@ -137 +138,2 @@ class ProcessingGraph: - _processing_steps_required_by_dataset_viewer: List[ProcessingStep] = field(init=False) + _processing_steps_enables_preview: List[ProcessingStep] = field(init=False) + _processing_steps_enables_viewer: List[ProcessingStep] = field(init=False) @@ -182 +184,2 @@ class ProcessingGraph: - required_by_dataset_viewer=specification.get("required_by_dataset_viewer", False), + enables_preview=specification.get("enables_preview", False), + enables_viewer=specification.get("enables_viewer", False), @@ -216 +219 @@ class ProcessingGraph: - self._processing_steps_required_by_dataset_viewer = [ + self._processing_steps_enables_preview = [ @@ -218 +221,6 @@ class ProcessingGraph: - for (processing_step_name, required) in _nx_graph.nodes(data="required_by_dataset_viewer") + for (processing_step_name, required) in _nx_graph.nodes(data="enables_preview") + if required + ] + self._processing_steps_enables_viewer = [ + self._processing_steps[processing_step_name] + for (processing_step_name, required) in _nx_graph.nodes(data="enables_viewer") @@ -373 +381,13 @@ class ProcessingGraph: - def get_processing_steps_required_by_dataset_viewer(self) -> List[ProcessingStep]: + def get_processing_steps_enables_preview(self) -> List[ProcessingStep]: + """ + Get the processing steps that enable the dataset preview (first rows). + + The returned processing steps are copies of the original ones, so that they can be modified without affecting + the original ones. + + Returns: + List[ProcessingStep]: The list of processing steps that enable the dataset preview + """ + return copy_processing_steps_list(self._processing_steps_enables_preview) + + def get_processing_steps_enables_viewer(self) -> List[ProcessingStep]: @@ -375 +395 @@ class ProcessingGraph: - Get the processing steps required by the dataset viewer. + Get the processing steps that enable the dataset viewer (all rows). @@ -381 +401 @@ class ProcessingGraph: - List[ProcessingStep]: The list of processing steps required by the dataset viewer + List[ProcessingStep]: The list of processing steps that enable the dataset viewer @@ -383 +403 @@ class ProcessingGraph: - return copy_processing_steps_list(self._processing_steps_required_by_dataset_viewer) + return copy_processing_steps_list(self._processing_steps_enables_viewer) diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py index a4861594..49bd37fb 100644 --- a/libs/libcommon/tests/test_processing_graph.py +++ b/libs/libcommon/tests/test_processing_graph.py @@ -303,3 +303,8 @@ def test_default_graph_first_steps(graph: ProcessingGraph) -> None: -def test_default_graph_required_by_dataset_viewer(graph: ProcessingGraph) -> None: - required_by_dataset_viewer = ["split-first-rows-from-streaming", "split-first-rows-from-parquet"] - assert_lists_are_equal(graph.get_processing_steps_required_by_dataset_viewer(), required_by_dataset_viewer) +def test_default_graph_enables_preview(graph: ProcessingGraph) -> None: + enables_preview = ["split-first-rows-from-streaming", "split-first-rows-from-parquet"] + assert_lists_are_equal(graph.get_processing_steps_enables_preview(), enables_preview) + + +def test_default_graph_enables_viewer(graph: ProcessingGraph) -> None: + enables_viewer = ["config-size"] + assert_lists_are_equal(graph.get_processing_steps_enables_viewer(), enables_viewer) diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py index 4cf69d0b..606353e4 100644 --- a/services/api/src/api/routes/valid.py +++ b/services/api/src/api/routes/valid.py @@ -5 +5,2 @@ import logging -from typing import List +from dataclasses import dataclass, field +from typing import List, Set, TypedDict @@ -7 +8 @@ from typing import List -from libcommon.processing_graph import ProcessingGraph +from libcommon.processing_graph import ProcessingGraph, ProcessingStep @@ -21,11 +22,42 @@ from api.utils import ( -def get_valid(processing_graph: ProcessingGraph) -> List[str]: - # a dataset is considered valid if at least one response of any of the - # "required_by_dataset_viewer" steps is valid. - processing_steps = processing_graph.get_processing_steps_required_by_dataset_viewer() - if not processing_steps: - return [] - datasets = set.union( - *[get_valid_datasets(kind=processing_step.cache_kind) for processing_step in processing_steps] - ) - # note that the list is sorted alphabetically for consistency - return sorted(datasets) +class ValidContent(TypedDict): + valid: List[str] + preview: List[str] + viewer: List[str] + + +@dataclass +class ValidDatasets: + processing_graph: ProcessingGraph + content: ValidContent = field(init=False) + + def __post_init__(self) -> None: + _viewer_set: Set[str] = self._get_valid_set( + processing_steps=self.processing_graph.get_processing_steps_enables_viewer() + ) + _preview_set: Set[str] = self._get_valid_set( + processing_steps=self.processing_graph.get_processing_steps_enables_preview() + ).difference(_viewer_set) + _valid_set = set.union(_viewer_set, _preview_set) + self.content = ValidContent( + valid=sorted(_valid_set), + preview=sorted(_preview_set), + viewer=sorted(_viewer_set), + ) + + def _get_valid_set(self, processing_steps: List[ProcessingStep]) -> Set[str]: + """Returns the list of the valid datasets for the list of steps + + A dataset is considered valid if at least one response of any of the artifacts for any of the + steps is valid. + + Args: + processing_steps (List[ProcessingStep]): The list of processing steps + + Returns: + List[str]: The list of valid datasets for the steps + """ + if not processing_steps: + return set() + return set.union( + *[get_valid_datasets(kind=processing_step.cache_kind) for processing_step in processing_steps] + ) @@ -45 +77 @@ def create_valid_endpoint( - content = {"valid": get_valid(processing_graph=processing_graph)} + content = ValidDatasets(processing_graph=processing_graph).content diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py index 317b0946..38fb758f 100644 --- a/services/api/tests/routes/test_valid.py +++ b/services/api/tests/routes/test_valid.py @@ -9 +9 @@ from api.config import AppConfig -from api.routes.valid import get_valid +from api.routes.valid import ValidDatasets @@ -29 +29 @@ def clean_mongo_databases(app_config: AppConfig) -> None: - {step_1: {"required_by_dataset_viewer": True}}, + {step_1: {"enables_preview": True}}, @@ -34 +34,2 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) -> - assert get_valid(processing_graph=processing_graph) == [] + valid_datasets = ValidDatasets(processing_graph=processing_graph) + assert valid_datasets.content == {"valid": [], "preview": [], "viewer": []} @@ -38 +39 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) -> - "processing_graph_specification,expected_valid", + "processing_graph_specification,expected_preview,expected_viewer,expected_valid", @@ -40,4 +41,6 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) -> - ({step_1: {}}, []), - ({step_1: {"required_by_dataset_viewer": True}}, ["dataset"]), - ({step_1: {}, step_2: {"required_by_dataset_viewer": True}}, []), - ({step_1: {"required_by_dataset_viewer": True}, step_2: {"required_by_dataset_viewer": True}}, ["dataset"]), + ({step_1: {}, step_2: {}}, [], [], []), + ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset"], [], ["dataset"]), + ({step_1: {}, step_2: {"enables_preview": True}}, [], [], []), + ({step_1: {"enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]), + ({step_1: {}, step_2: {"enables_viewer": True}}, [], [], []), + ({step_1: {"enables_preview": True, "enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]), @@ -46 +49,6 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) -> -def test_one_dataset(processing_graph_specification: ProcessingGraphSpecification, expected_valid: List[str]) -> None: +def test_one_dataset( + processing_graph_specification: ProcessingGraphSpecification, + expected_preview: List[str], + expected_viewer: List[str], + expected_valid: List[str], +) -> None: @@ -51 +59,6 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio - assert get_valid(processing_graph=processing_graph) == expected_valid + valid_datasets = ValidDatasets(processing_graph=processing_graph) + assert valid_datasets.content == { + "valid": expected_valid, + "preview": expected_preview, + "viewer": expected_viewer, + } @@ -55 +68 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio - "processing_graph_specification,expected_valid", + "processing_graph_specification,expected_preview,expected_viewer,expected_valid", @@ -57,3 +70,3 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio - ({step_1: {}, step_2: {}}, []), - ({step_1: {"required_by_dataset_viewer": True}, step_2: {}}, ["dataset1"]), - ({step_1: {}, step_2: {"required_by_dataset_viewer": True}}, ["dataset2"]), + ({step_1: {}, step_2: {}}, [], [], []), + ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset1"], [], ["dataset1"]), + ({step_1: {}, step_2: {"enables_preview": True}}, ["dataset2"], [], ["dataset2"]), @@ -61 +74,9 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio - {step_1: {"required_by_dataset_viewer": True}, step_2: {"required_by_dataset_viewer": True}}, + {step_1: {"enables_preview": True}, step_2: {"enables_preview": True}}, + ["dataset1", "dataset2"], + [], + ["dataset1", "dataset2"], + ), + ( + {step_1: {"enables_preview": True}, step_2: {"enables_viewer": True}}, + ["dataset1"], + ["dataset2"], @@ -66 +87,6 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio -def test_two_datasets(processing_graph_specification: ProcessingGraphSpecification, expected_valid: List[str]) -> None: +def test_two_datasets( + processing_graph_specification: ProcessingGraphSpecification, + expected_preview: List[str], + expected_viewer: List[str], + expected_valid: List[str], +) -> None: @@ -80 +106,6 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - assert get_valid(processing_graph=processing_graph) == expected_valid + valid_datasets = ValidDatasets(processing_graph=processing_graph) + assert valid_datasets.content == { + "valid": expected_valid, + "preview": expected_preview, + "viewer": expected_viewer, + } @@ -84 +115 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - "processing_graph_specification,expected_valid", + "processing_graph_specification,expected_preview,expected_viewer,expected_valid", @@ -92,0 +124,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati + [], + [], @@ -96 +129 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - dataset_step: {"required_by_dataset_viewer": True}, + dataset_step: {"enables_preview": True}, @@ -100,0 +134,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati + [], + ["dataset"], @@ -108 +143 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - "required_by_dataset_viewer": True, + "enables_preview": True, @@ -112,0 +148,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati + [], + ["dataset"], @@ -118 +155 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - split_step: {"input_type": "split", "triggered_by": config_step, "required_by_dataset_viewer": True}, + split_step: {"input_type": "split", "triggered_by": config_step, "enables_preview": True}, @@ -120,0 +158,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati + [], + ["dataset"], @@ -124 +163 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - dataset_step: {"required_by_dataset_viewer": True}, + dataset_step: {"enables_preview": True}, @@ -128 +167 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - "required_by_dataset_viewer": True, + "enables_preview": True, @@ -130 +169 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati - split_step: {"input_type": "split", "triggered_by": config_step, "required_by_dataset_viewer": True}, + split_step: {"input_type": "split", "triggered_by": config_step, "enables_preview": True}, @@ -132,0 +172,16 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati + [], + ["dataset"], + ), + ( + { + dataset_step: {}, + config_step: { + "input_type": "config", + "triggered_by": dataset_step, + "enables_viewer": True, + }, + split_step: {"input_type": "split", "triggered_by": config_step, "enables_preview": True}, + }, + [], + ["dataset"], + ["dataset"], @@ -136 +191,6 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati -def test_three_steps(processing_graph_specification: ProcessingGraphSpecification, expected_valid: List[str]) -> None: +def test_three_steps( + processing_graph_specification: ProcessingGraphSpecification, + expected_preview: List[str], + expected_viewer: List[str], + expected_valid: List[str], +) -> None: @@ -162 +222,6 @@ def test_three_steps(processing_graph_specification: ProcessingGraphSpecificatio - assert get_valid(processing_graph=processing_graph) == expected_valid + valid_datasets = ValidDatasets(processing_graph=processing_graph) + assert valid_datasets.content == { + "valid": expected_valid, + "preview": expected_preview, + "viewer": expected_viewer, + } @@ -166 +231 @@ def test_errors() -> None: - processing_graph = ProcessingGraph({dataset_step: {"required_by_dataset_viewer": True}}) + processing_graph = ProcessingGraph({dataset_step: {"enables_preview": True}}) @@ -174 +239,12 @@ def test_errors() -> None: - assert get_valid(processing_graph=processing_graph) == [dataset_a, dataset_b] + valid_datasets = ValidDatasets(processing_graph=processing_graph) + assert valid_datasets.content == { + "valid": [ + dataset_a, + dataset_b, + ], + "preview": [ + dataset_a, + dataset_b, + ], + "viewer": [], + } diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py index 32c83a97..a883d39b 100644 --- a/services/api/tests/test_app.py +++ b/services/api/tests/test_app.py @@ -51,0 +52,2 @@ def test_get_valid_datasets(client: TestClient) -> None: + assert "preview" in response.json() + assert "viewer" in response.json()
db7913058dc3a940c862dc4649d260c5bde33147
Sylvain Lesage
2023-06-12T12:56:29
feat: 🎸 upgrade duckdb and gradio (#1351)
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock index 19956d9d..cd3fcb36 100644 --- a/front/admin_ui/poetry.lock +++ b/front/admin_ui/poetry.lock @@ -619 +619 @@ name = "duckdb" -version = "0.6.1" +version = "0.8.0" @@ -625,47 +625,47 @@ files = [ - {file = "duckdb-0.6.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e566514f9327f89264e98ac14ee7a84fbd9857328028258422c3e8375ee19d25"}, - {file = "duckdb-0.6.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b31c2883de5b19591a2852165e6b3f9821f77af649835f27bc146b26e4aa30cb"}, - {file = "duckdb-0.6.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:998165b2fb1f1d2b0ad742096015ea70878f7d40304643c7424c3ed3ddf07bfc"}, - {file = "duckdb-0.6.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3941b3a1e8a1cdb7b90ab3917b87af816e71f9692e5ada7f19b6b60969f731e5"}, - {file = "duckdb-0.6.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:143611bd1b7c13343f087d4d423a7a8a4f33a114c5326171e867febf3f0fcfe1"}, - {file = "duckdb-0.6.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:125ba45e8b08f28858f918ec9cbd3a19975e5d8d9e8275ef4ad924028a616e14"}, - {file = "duckdb-0.6.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e609a65b31c92f2f7166831f74b56f5ed54b33d8c2c4b4c3974c26fdc50464c5"}, - {file = "duckdb-0.6.1-cp310-cp310-win32.whl", hash = "sha256:b39045074fb9a3f068496475a5d627ad4fa572fa3b4980e3b479c11d0b706f2d"}, - {file = "duckdb-0.6.1-cp310-cp310-win_amd64.whl", hash = "sha256:16fa96ffaa3d842a9355a633fb8bc092d119be08d4bc02013946d8594417bc14"}, - {file = "duckdb-0.6.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4bbe2f6c1b109c626f9318eee80934ad2a5b81a51409c6b5083c6c5f9bdb125"}, - {file = "duckdb-0.6.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cfea36b58928ce778d17280d4fb3bf0a2d7cff407667baedd69c5b41463ac0fd"}, - {file = "duckdb-0.6.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:0b64eb53d0d0695814bf1b65c0f91ab7ed66b515f89c88038f65ad5e0762571c"}, - {file = "duckdb-0.6.1-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:35b01bc724e1933293f4c34f410d2833bfbb56d5743b515d805bbfed0651476e"}, - {file = "duckdb-0.6.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fec2c2466654ce786843bda2bfba71e0e4719106b41d36b17ceb1901e130aa71"}, - {file = "duckdb-0.6.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:82cd30f5cf368658ef879b1c60276bc8650cf67cfe3dc3e3009438ba39251333"}, - {file = "duckdb-0.6.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:a782bbfb7f5e97d4a9c834c9e78f023fb8b3f6687c22ca99841e6ed944b724da"}, - {file = "duckdb-0.6.1-cp311-cp311-win32.whl", hash = "sha256:e3702d4a9ade54c6403f6615a98bbec2020a76a60f5db7fcf085df1bd270e66e"}, - {file = "duckdb-0.6.1-cp311-cp311-win_amd64.whl", hash = "sha256:93b074f473d68c944b0eeb2edcafd91ad11da8432b484836efaaab4e26351d48"}, - {file = "duckdb-0.6.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:adae183924d6d479202c39072e37d440b511326e84525bcb7432bca85f86caba"}, - {file = "duckdb-0.6.1-cp36-cp36m-win32.whl", hash = "sha256:546a1cd17595bd1dd009daf6f36705aa6f95337154360ce44932157d353dcd80"}, - {file = "duckdb-0.6.1-cp36-cp36m-win_amd64.whl", hash = "sha256:87b0d00eb9d1a7ebe437276203e0cdc93b4a2154ba9688c65e8d2a8735839ec6"}, - {file = "duckdb-0.6.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8442e074de6e1969c3d2b24363a5a6d7f866d5ac3f4e358e357495b389eff6c1"}, - {file = "duckdb-0.6.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0a6bf2ae7bec803352dade14561cb0b461b2422e70f75d9f09b36ba2dad2613b"}, - {file = "duckdb-0.6.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5054792f22733f89d9cbbced2bafd8772d72d0fe77f159310221cefcf981c680"}, - {file = "duckdb-0.6.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:21cc503dffc2c68bb825e4eb3098e82f40e910b3d09e1b3b7f090d39ad53fbea"}, - {file = "duckdb-0.6.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:54b3da77ad893e99c073087ff7f75a8c98154ac5139d317149f12b74367211db"}, - {file = "duckdb-0.6.1-cp37-cp37m-win32.whl", hash = "sha256:f1d709aa6a26172a3eab804b57763d5cdc1a4b785ac1fc2b09568578e52032ee"}, - {file = "duckdb-0.6.1-cp37-cp37m-win_amd64.whl", hash = "sha256:f4edcaa471d791393e37f63e3c7c728fa6324e3ac7e768b9dc2ea49065cd37cc"}, - {file = "duckdb-0.6.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:d218c2dd3bda51fb79e622b7b2266183ac9493834b55010aa01273fa5b7a7105"}, - {file = "duckdb-0.6.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0c7155cb93ab432eca44b651256c359281d26d927ff43badaf1d2276dd770832"}, - {file = "duckdb-0.6.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0925778200090d3d5d8b6bb42b4d05d24db1e8912484ba3b7e7b7f8569f17dcb"}, - {file = "duckdb-0.6.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8b544dd04bb851d08bc68b317a7683cec6091547ae75555d075f8c8a7edb626e"}, - {file = "duckdb-0.6.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f2c37d5a0391cf3a3a66e63215968ffb78e6b84f659529fa4bd10478f6203071"}, - {file = "duckdb-0.6.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ce376966260eb5c351fcc6af627a979dbbcae3efeb2e70f85b23aa45a21e289d"}, - {file = "duckdb-0.6.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:73c974b09dd08dff5e8bdedba11c7d0aa0fc46ca93954ee7d19e1e18c9883ac1"}, - {file = "duckdb-0.6.1-cp38-cp38-win32.whl", hash = "sha256:bfe39ed3a03e8b1ed764f58f513b37b24afe110d245803a41655d16d391ad9f1"}, - {file = "duckdb-0.6.1-cp38-cp38-win_amd64.whl", hash = "sha256:afa97d982dbe6b125631a17e222142e79bee88f7a13fc4cee92d09285e31ec83"}, - {file = "duckdb-0.6.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:c35ff4b1117096ef72d101524df0079da36c3735d52fcf1d907ccffa63bd6202"}, - {file = "duckdb-0.6.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5c54910fbb6de0f21d562e18a5c91540c19876db61b862fc9ffc8e31be8b3f03"}, - {file = "duckdb-0.6.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:99a7172563a3ae67d867572ce27cf3962f58e76f491cb7f602f08c2af39213b3"}, - {file = "duckdb-0.6.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7363ffe857d00216b659116647fbf1e925cb3895699015d4a4e50b746de13041"}, - {file = "duckdb-0.6.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:06c1cef25f896b2284ba048108f645c72fab5c54aa5a6f62f95663f44ff8a79b"}, - {file = "duckdb-0.6.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:e92dd6aad7e8c29d002947376b6f5ce28cae29eb3b6b58a64a46cdbfc5cb7943"}, - {file = "duckdb-0.6.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4b280b2d8a01ecd4fe2feab041df70233c534fafbe33a38565b52c1e017529c7"}, - {file = "duckdb-0.6.1-cp39-cp39-win32.whl", hash = "sha256:d9212d76e90b8469743924a4d22bef845be310d0d193d54ae17d9ef1f753cfa7"}, - {file = "duckdb-0.6.1-cp39-cp39-win_amd64.whl", hash = "sha256:00b7be8f67ec1a8edaa8844f521267baa1a795f4c482bfad56c72c26e1862ab2"}, - {file = "duckdb-0.6.1.tar.gz", hash = "sha256:6d26e9f1afcb924a6057785e506810d48332d4764ddc4a5b414d0f2bf0cacfb4"}, + {file = "duckdb-0.8.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:6455aee00af30770c20f4a8c5e4347918cf59b578f49ee996a13807b12911871"}, + {file = "duckdb-0.8.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b8cf0622ae7f86d4ce72791f8928af4357a46824aadf1b6879c7936b3db65344"}, + {file = "duckdb-0.8.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:6132e8183ca3ae08a593e43c97cb189794077dedd48546e27ce43bd6a51a9c33"}, + {file = "duckdb-0.8.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe29e5343fa2a95f2cde4519a4f4533f4fd551a48d2d9a8ab5220d40ebf53610"}, + {file = "duckdb-0.8.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:945165987ca87c097dc0e578dcf47a100cad77e1c29f5dd8443d53ce159dc22e"}, + {file = "duckdb-0.8.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:673c60daf7ada1d9a8518286a6893ec45efabb64602954af5f3d98f42912fda6"}, + {file = "duckdb-0.8.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d5075fe1ff97ae62331ca5c61e3597e6e9f7682a6fdd418c23ba5c4873ed5cd1"}, + {file = "duckdb-0.8.0-cp310-cp310-win32.whl", hash = "sha256:001f5102f45d3d67f389fa8520046c8f55a99e2c6d43b8e68b38ea93261c5395"}, + {file = "duckdb-0.8.0-cp310-cp310-win_amd64.whl", hash = "sha256:cb00800f2e1e865584b13221e0121fce9341bb3a39a93e569d563eaed281f528"}, + {file = "duckdb-0.8.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b2707096d6df4321044fcde2c9f04da632d11a8be60957fd09d49a42fae71a29"}, + {file = "duckdb-0.8.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b27df1b70ae74d2c88efb5ffca8490954fdc678099509a9c4404ca30acc53426"}, + {file = "duckdb-0.8.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75a97c800271b52dd0f37696d074c50576dcb4b2750b6115932a98696a268070"}, + {file = "duckdb-0.8.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:804cac261a5e016506a6d67838a65d19b06a237f7949f1704f0e800eb708286a"}, + {file = "duckdb-0.8.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c6b9abca7fa6713e1d031c18485343b4de99742c7e1b85c10718aa2f31a4e2c6"}, + {file = "duckdb-0.8.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:51aa6d606d49072abcfeb3be209eb559ac94c1b5e70f58ac3adbb94aca9cd69f"}, + {file = "duckdb-0.8.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:7c8dc769aaf2be0a1c57995ca657e5b92c1c56fc8437edb720ca6cab571adf14"}, + {file = "duckdb-0.8.0-cp311-cp311-win32.whl", hash = "sha256:c4207d18b42387c4a035846d8878eb967070198be8ac26fd77797ce320d1a400"}, + {file = "duckdb-0.8.0-cp311-cp311-win_amd64.whl", hash = "sha256:0c392257547c20794c3072fcbca99a49ef0a49974005d755e93893e2b4875267"}, + {file = "duckdb-0.8.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:2832379e122020814dbe869af7b9ddf3c9f21474cf345531145b099c63ffe17e"}, + {file = "duckdb-0.8.0-cp36-cp36m-win32.whl", hash = "sha256:914896526f7caba86b170f2c4f17f11fd06540325deeb0000cb4fb24ec732966"}, + {file = "duckdb-0.8.0-cp36-cp36m-win_amd64.whl", hash = "sha256:022ebda86d0e3204cdc206e4af45aa9f0ae0668b34c2c68cf88e08355af4a372"}, + {file = "duckdb-0.8.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:96a31c0f3f4ccbf0f5b18f94319f37691205d82f80aae48c6fe04860d743eb2c"}, + {file = "duckdb-0.8.0-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a07c73c6e6a8cf4ce1a634625e0d1b17e5b817242a8a530d26ed84508dfbdc26"}, + {file = "duckdb-0.8.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:424acbd6e857531b06448d757d7c2557938dbddbff0632092090efbf413b4699"}, + {file = "duckdb-0.8.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c83cfd2a868f1acb0692b9c3fd5ef1d7da8faa1348c6eabf421fbf5d8c2f3eb8"}, + {file = "duckdb-0.8.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5c6f6b2d8db56936f662c649539df81856b5a8cb769a31f9544edf18af2a11ff"}, + {file = "duckdb-0.8.0-cp37-cp37m-win32.whl", hash = "sha256:0bd6376b40a512172eaf4aa816813b1b9d68994292ca436ce626ccd5f77f8184"}, + {file = "duckdb-0.8.0-cp37-cp37m-win_amd64.whl", hash = "sha256:931221885bcf1e7dfce2400f11fd048a7beef566b775f1453bb1db89b828e810"}, + {file = "duckdb-0.8.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:42e7853d963d68e72403ea208bcf806b0f28c7b44db0aa85ce49bb124d56c133"}, + {file = "duckdb-0.8.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:fcc338399175be3d43366576600aef7d72e82114d415992a7a95aded98a0f3fd"}, + {file = "duckdb-0.8.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:03dd08a4624d6b581a59f9f9dbfd34902416398d16795ad19f92361cf21fd9b5"}, + {file = "duckdb-0.8.0-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0c7c24ea0c9d8563dbd5ad49ccb54b7a9a3c7b8c2833d35e5d32a08549cacea5"}, + {file = "duckdb-0.8.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cb58f6505cc0f34b4e976154302d26563d2e5d16b206758daaa04b65e55d9dd8"}, + {file = "duckdb-0.8.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ef37ac7880100c4b3f913c8483a29a13f8289313b9a07df019fadfa8e7427544"}, + {file = "duckdb-0.8.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:c2a4f5ee913ca8a6a069c78f8944b9934ffdbc71fd935f9576fdcea2a6f476f1"}, + {file = "duckdb-0.8.0-cp38-cp38-win32.whl", hash = "sha256:73831c6d7aefcb5f4072cd677b9efebecbf6c578946d21710791e10a1fc41b9a"}, + {file = "duckdb-0.8.0-cp38-cp38-win_amd64.whl", hash = "sha256:faa36d2854734364d234f37d7ef4f3d763b73cd6b0f799cbc2a0e3b7e2575450"}, + {file = "duckdb-0.8.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:50a31ec237ed619e50f9ab79eb0ec5111eb9697d4475da6e0ab22c08495ce26b"}, + {file = "duckdb-0.8.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:351abb4cc2d229d043920c4bc2a4c29ca31a79fef7d7ef8f6011cf4331f297bf"}, + {file = "duckdb-0.8.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:568550a163aca6a787bef8313e358590254de3f4019025a8d68c3a61253fedc1"}, + {file = "duckdb-0.8.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2b82617f0e7f9fc080eda217090d82b42d4fad083bc9f6d58dfda9cecb7e3b29"}, + {file = "duckdb-0.8.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d01c9be34d272532b75e8faedda0ff77fa76d1034cde60b8f5768ae85680d6d3"}, + {file = "duckdb-0.8.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:8549d6a6bf5f00c012b6916f605416226507e733a3ffc57451682afd6e674d1b"}, + {file = "duckdb-0.8.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:8d145c6d51e55743c3ed1a74cffa109d9e72f82b07e203b436cfa453c925313a"}, + {file = "duckdb-0.8.0-cp39-cp39-win32.whl", hash = "sha256:f8610dfd21e90d7b04e8598b244bf3ad68599fd6ba0daad3428c03cbfd74dced"}, + {file = "duckdb-0.8.0-cp39-cp39-win_amd64.whl", hash = "sha256:d0f0f104d30418808bafbe9bccdcd238588a07bd246b3cff13842d60bfd8e8ba"}, + {file = "duckdb-0.8.0.tar.gz", hash = "sha256:c68da35bab5072a64ada2646a5b343da620ddc75a7a6e84aa4a1e0628a7ec18f"}, @@ -674,3 +673,0 @@ files = [ -[package.dependencies] -numpy = ">=1.14" - @@ -915 +912 @@ name = "gradio" -version = "3.28.1" +version = "3.34.0" @@ -921,2 +918,2 @@ files = [ - {file = "gradio-3.28.1-py3-none-any.whl", hash = "sha256:0811f04be88ee789d921fee0089c69b7da3b12dd7c937f383e4c0998663b4f21"}, - {file = "gradio-3.28.1.tar.gz", hash = "sha256:15d959e73f77bb6980334a159ff177867f6636a2a1007a7154d55a1d6e6d8648"}, + {file = "gradio-3.34.0-py3-none-any.whl", hash = "sha256:1cd8b25b598d983561d64f0a039af819382f1376c676aa9f84972c46b6875741"}, + {file = "gradio-3.34.0.tar.gz", hash = "sha256:fd7fa7257ffc749f9dc7c297eba554eaa1e5acd1a5f9c973250b2080932d6a41"}, @@ -931 +928 @@ ffmpy = "*" -gradio-client = ">=0.1.3" +gradio-client = ">=0.2.6" @@ -933 +930 @@ httpx = "*" -huggingface-hub = ">=0.13.0" +huggingface-hub = ">=0.14.0" @@ -944,0 +942 @@ pydub = "*" +pygments = ">=2.12.0" @@ -950 +948 @@ typing-extensions = "*" -uvicorn = "*" +uvicorn = ">=0.14.0" @@ -955 +953 @@ name = "gradio-client" -version = "0.2.5" +version = "0.2.6" @@ -961,2 +959,2 @@ files = [ - {file = "gradio_client-0.2.5-py3-none-any.whl", hash = "sha256:922a5188c93797adce023b4caa655318b9c09834095d31763443c1e7a707e301"}, - {file = "gradio_client-0.2.5.tar.gz", hash = "sha256:1a24dd7a09976db70fd32cdb5512b6f7b15c1a084f310b5f6b3021c9794d90a4"}, + {file = "gradio_client-0.2.6-py3-none-any.whl", hash = "sha256:daa3e2f7697cc07821ccefa7486b923e71c1e459d5b3c6e35318f343b566d789"}, + {file = "gradio_client-0.2.6.tar.gz", hash = "sha256:a5d5c5799ce33ae3107e1d30992c27050f50506f15dd70481a39b13ac47e2613"}, @@ -2267,0 +2266,15 @@ files = [ +[[package]] +name = "pygments" +version = "2.15.1" +description = "Pygments is a syntax highlighting package written in Python." +category = "main" +optional = false +python-versions = ">=3.7" +files = [ + {file = "Pygments-2.15.1-py3-none-any.whl", hash = "sha256:db2db3deb4b4179f399a09054b023b6a586b76499d36965813c71aa8ed7b5fd1"}, + {file = "Pygments-2.15.1.tar.gz", hash = "sha256:8ace4d3c1dd481894b2005f560ead0f9f19ee64fe983366be1a21e171d12775c"}, +] + +[package.extras] +plugins = ["importlib-metadata"] + @@ -3270 +3283 @@ python-versions = "3.9.15" -content-hash = "9e29354c885f6c0d8aba1c5e79d6087a8b1419b69418e4db09caa28e32d771ae" +content-hash = "4c403bae2a09145ce4fb1217296ad92a9a01b184bf2d913aceb98a777b55c133" diff --git a/front/admin_ui/pyproject.toml b/front/admin_ui/pyproject.toml index 8947fe80..cf4bebc5 100644 --- a/front/admin_ui/pyproject.toml +++ b/front/admin_ui/pyproject.toml @@ -8 +8 @@ authors = ["Quentin Lhoest <[email protected]>"] -gradio = "3.28.1" +gradio = "^3.34.0" @@ -14 +14 @@ huggingface-hub = "^0.15.1" -duckdb = "^0.6.1" +duckdb = "^0.8.0" diff --git a/front/admin_ui/requirements.txt b/front/admin_ui/requirements.txt index e178a30b..3f903ace 100644 --- a/front/admin_ui/requirements.txt +++ b/front/admin_ui/requirements.txt @@ -1 +1 @@ -gradio==3.28.1 +gradio==3.34.0 @@ -7 +7 @@ huggingface-hub~=0.15.1 -duckdb~=0.6.1 \ No newline at end of file +duckdb~=0.8.0
ba115a9145e2c24f01f6d58dabdd6cf78fc8b4da
Sylvain Lesage
2023-06-12T11:19:00
Fix admin app (#1348)
diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py index 770ff8f3..a39a8e20 100644 --- a/front/admin_ui/app.py +++ b/front/admin_ui/app.py @@ -27,0 +28,2 @@ DSS_ENDPOINT = DEV_DSS_ENDPOINT if DEV else PROD_DSS_ENDPOINT +PROCESSING_GRAPH = ProcessingGraph(ProcessingGraphConfig().specification) + @@ -43,2 +45 @@ def draw_graph(width, height): - config = ProcessingGraphConfig() - graph = ProcessingGraph(config.specification)._nx_graph + graph = PROCESSING_GRAPH._nx_graph @@ -81 +82,2 @@ with gr.Blocks() as demo: - refresh_type = gr.Textbox(label="Processing step type", placeholder="split-first-rows-from-streaming") + job_types = [processing_step.job_type for processing_step in PROCESSING_GRAPH.get_topologically_ordered_processing_steps()] + refresh_type = gr.Dropdown(job_types, multiselect=False, label="job type", value=job_types[0]) @@ -171 +173 @@ with gr.Blocks() as demo: - "type": job_type, + "kind": cached_response["kind"], @@ -187 +189 @@ with gr.Blocks() as demo: - "type": job_type, + "type": job["type"], @@ -188,0 +191 @@ with gr.Blocks() as demo: + "revision": job["revision"],
deba0fd4c5a538922f15c723838b6c0bd056d7a1
Quentin Lhoest
2023-06-12T10:10:51
fix parquet-and-info download_size (#1347)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 91de381e..8f2f0f10 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -628,0 +629 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None: + builder.info.download_size = 0
42b364b24e03912d9e70148057c7468eb3f8a047
Albert Villanova del Moral
2023-06-12T09:25:24
Update locked cachecontrol version (#1344)
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock index 54ed9fd8..19956d9d 100644 --- a/front/admin_ui/poetry.lock +++ b/front/admin_ui/poetry.lock @@ -1266,0 +1267 @@ starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock index b67c295d..bd34f45d 100644 --- a/jobs/cache_maintenance/poetry.lock +++ b/jobs/cache_maintenance/poetry.lock @@ -265 +265 @@ name = "cachecontrol" -version = "0.13.0" +version = "0.13.1" @@ -269 +269 @@ optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" @@ -271,2 +271,2 @@ files = [ - {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"}, - {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"}, + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, @@ -280,0 +281 @@ requests = ">=2.16.0" +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] @@ -1041,0 +1043 @@ starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock index b67c295d..bd34f45d 100644 --- a/jobs/mongodb_migration/poetry.lock +++ b/jobs/mongodb_migration/poetry.lock @@ -265 +265 @@ name = "cachecontrol" -version = "0.13.0" +version = "0.13.1" @@ -269 +269 @@ optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" @@ -271,2 +271,2 @@ files = [ - {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"}, - {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"}, + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, @@ -280,0 +281 @@ requests = ">=2.16.0" +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] @@ -1041,0 +1043 @@ starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock index 534f3756..1ddc15bc 100644 --- a/libs/libcommon/poetry.lock +++ b/libs/libcommon/poetry.lock @@ -265 +265 @@ name = "cachecontrol" -version = "0.13.0" +version = "0.13.1" @@ -269 +269 @@ optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" @@ -271,2 +271,2 @@ files = [ - {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"}, - {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"}, + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, @@ -280,0 +281 @@ requests = ">=2.16.0" +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] @@ -3034 +3035 @@ python-versions = "3.9.15" -content-hash = "859a5062a386f1bc1cec980e4006f3f7f6c7c5b418902232d16994cb9ffb6947" +content-hash = "3eab4cefa7dd3f05dcf9657d8a1e81f7d7df494f07b6eb0f85fbc97e8a6aa56e" diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock index 8ff004b9..928e56ca 100644 --- a/services/admin/poetry.lock +++ b/services/admin/poetry.lock @@ -265 +265 @@ name = "cachecontrol" -version = "0.13.0" +version = "0.13.1" @@ -269 +269 @@ optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" @@ -271,2 +271,2 @@ files = [ - {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"}, - {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"}, + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, @@ -280,0 +281 @@ requests = ">=2.16.0" +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] @@ -1099,0 +1101 @@ starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" diff --git a/services/api/poetry.lock b/services/api/poetry.lock index a4b0eaf9..70a86b1c 100644 --- a/services/api/poetry.lock +++ b/services/api/poetry.lock @@ -265 +265 @@ name = "cachecontrol" -version = "0.13.0" +version = "0.13.1" @@ -269 +269 @@ optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" @@ -271,2 +271,2 @@ files = [ - {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"}, - {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"}, + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, @@ -280,0 +281 @@ requests = ">=2.16.0" +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] @@ -1161,0 +1163 @@ starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index 92064110..6ae2e954 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. +# This file is automatically @generated by Poetry and should not be changed by hand. @@ -473 +473 @@ name = "cachecontrol" -version = "0.13.0" +version = "0.13.1" @@ -477 +477 @@ optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" @@ -479,2 +479,2 @@ files = [ - {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"}, - {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"}, + {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"}, + {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"}, @@ -488,0 +489 @@ requests = ">=2.16.0" +dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"] @@ -1795,0 +1797 @@ starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" @@ -2863,0 +2866 @@ files = [ + {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -4417,0 +4421 @@ files = [ + {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
e0eed9c9ad09d57a857e6498c44b76d6bd71ff76
Sylvain Lesage
2023-06-10T11:54:23
feat: 🎸 give 10x more time for the jobs for cerebras/SlimPajama (#1342)
diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py index e369028d..a0acf747 100644 --- a/services/worker/src/worker/executor.py +++ b/services/worker/src/worker/executor.py @@ -135 +135,5 @@ class WorkerExecutor: - if last_updated + timedelta(seconds=self.app_config.worker.max_job_duration_seconds) <= get_datetime(): + coefficient = 10 if long_job["params"]["dataset"] == "cerebras/SlimPajama-627B" else 1 + if ( + last_updated + timedelta(seconds=coefficient * self.app_config.worker.max_job_duration_seconds) + <= get_datetime() + ):
da29c537bc969d2dc0973e010bb94222384b2f62
Andrea Francis Soria Jimenez
2023-06-09T18:05:29
refactor: first-rows-from-parquet use same code as in /rows (#1287)
diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml index 24ca6c2a..1bc1d006 100644 --- a/libs/libcommon/pyproject.toml +++ b/libs/libcommon/pyproject.toml @@ -26,0 +27 @@ starlette-prometheus = "^0.9.0" +tqdm = "^4.65.0" @@ -71,0 +73,2 @@ module = [ + "tqdm.*", + "fsspec.*", diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py new file mode 100644 index 00000000..3a498888 --- /dev/null +++ b/libs/libcommon/src/libcommon/parquet_utils.py @@ -0,0 +1,496 @@ +import asyncio +import logging +import os +from dataclasses import dataclass +from functools import lru_cache, partial +from os import PathLike +from typing import Callable, List, Literal, Optional, Tuple, TypedDict, Union, cast + +import numpy as np +import numpy.typing as npt +import pyarrow as pa +import pyarrow.parquet as pq +from datasets import Features +from fsspec.implementations.http import HTTPFile, HTTPFileSystem +from huggingface_hub import HfFileSystem +from huggingface_hub.hf_file_system import safe_quote +from tqdm.contrib.concurrent import thread_map + +from libcommon.constants import PARQUET_REVISION +from libcommon.exceptions import UnexpectedError +from libcommon.processing_graph import ProcessingGraph +from libcommon.prometheus import StepProfiler +from libcommon.simple_cache import get_previous_step_or_raise + +StrPath = Union[str, PathLike[str]] + + +class ParquetResponseEmptyError(Exception): + pass + + +class ParquetResponseFormatError(Exception): + pass + + +class FileSystemError(Exception): + pass + + +class ParquetFileItem(TypedDict): + dataset: str + config: str + split: str + url: str + filename: str + size: int + + +class ParquetFileMetadataItem(TypedDict): + dataset: str + config: str + split: str + url: str + filename: str + size: int + num_rows: int + parquet_metadata_subpath: str + + +class HTTPFileSystemSession(object): + _singleton: Optional["HTTPFileSystemSession"] = None + + @classmethod + def get_instance(cls) -> "HTTPFileSystemSession": + if cls._singleton is None: + HTTPFileSystemSession() + return cast("HTTPFileSystemSession", HTTPFileSystemSession._singleton) + + def __new__(cls) -> "HTTPFileSystemSession": + if HTTPFileSystemSession._singleton is not None: + raise RuntimeError( + "cannot initialize another instance of HTTPFileSystemSession, use .get_instance() instead" + ) + return super(HTTPFileSystemSession, cls).__new__(cls) + + def __init__(self) -> None: + HTTPFileSystemSession._singleton = self + self.httpfs = HTTPFileSystem() + self.session = asyncio.run(self.httpfs.set_session()) + + +def get_supported_unsupported_columns( + features: Features, + unsupported_features_magic_strings: List[str] = [], +) -> Tuple[List[str], List[str]]: + supported_columns, unsupported_columns = [], [] + + for column, feature in features.items(): + str_feature = str(feature) + str_column = str(column) + if any(magic_string in str_feature for magic_string in unsupported_features_magic_strings): + unsupported_columns.append(str_column) + else: + supported_columns.append(str_column) + return supported_columns, unsupported_columns + + +# TODO: how to invalidate the cache when the parquet branch is created or deleted? +@lru_cache(maxsize=128) +def get_hf_fs(hf_token: Optional[str]) -> HfFileSystem: + """Get the Hugging Face filesystem. + + Args: + hf_token (Optional[str]): The token to access the filesystem. + Returns: + HfFileSystem: The Hugging Face filesystem. + """ + return HfFileSystem(token=hf_token) + + +def get_hf_parquet_uris(paths: List[str], dataset: str) -> List[str]: + """Get the Hugging Face URIs from the Parquet branch of the dataset repository (see PARQUET_REVISION). + + Args: + paths (List[str]): List of paths. + dataset (str): The dataset name. + Returns: + List[str]: List of Parquet URIs. + """ + return [f"hf://datasets/{dataset}@{safe_quote(PARQUET_REVISION)}/{path}" for path in paths] + + +@dataclass +class ParquetIndexWithoutMetadata: + features: Features + supported_columns: List[str] + unsupported_columns: List[str] + row_group_offsets: npt.NDArray[np.int64] + row_group_readers: List[Callable[[], pa.Table]] + + def query(self, offset: int, length: int) -> pa.Table: + """Query the parquet files + + Note that this implementation will always read at least one row group, to get the list of columns and always + have the same schema, even if the requested rows are invalid (out of range). + + Args: + offset (int): The first row to read. + length (int): The number of rows to read. + + Returns: + pa.Table: The requested rows. + """ + if (len(self.row_group_offsets) == 0) or (len(self.row_group_readers) == 0): + raise ParquetResponseEmptyError("No parquet files found.") + last_row_in_parquet = self.row_group_offsets[-1] - 1 + first_row = min(offset, last_row_in_parquet) + last_row = min(offset + length - 1, last_row_in_parquet) + first_row_group_id, last_row_group_id = np.searchsorted( + self.row_group_offsets, [first_row, last_row], side="right" + ) + pa_table = pa.concat_tables( + [self.row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)] + ) + first_row_in_pa_table = self.row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0 + return pa_table.slice(offset - first_row_in_pa_table, length) + + @staticmethod + def from_parquet_file_items( + parquet_file_items: List[ParquetFileItem], + dataset: str, + config: str, + split: str, + hf_token: Optional[str], + unsupported_features_magic_strings: List[str] = [], + ) -> "ParquetIndexWithoutMetadata": + try: + sources = sorted(f"{config}/{parquet_file['filename']}" for parquet_file in parquet_file_items) + except Exception as e: + raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e + logging.debug( + f"Found {len(sources)} parquet files for dataset={dataset}, config={config}, split={split}: {sources}" + ) + if not sources: + raise ParquetResponseEmptyError("No parquet files found.") + with StepProfiler( + method="parquet_index_without_metadata.from_parquet_file_items", step="get the Hub's dataset filesystem" + ): + fs = get_hf_fs(hf_token=hf_token) + with StepProfiler(method="parquet_index_without_metadata.from_parquet_file_items", step="get the source URIs"): + source_uris = get_hf_parquet_uris(sources, dataset=dataset) + with StepProfiler( + method="parquet_index_without_metadata.from_parquet_file_items", + step="get one parquet reader per parquet file", + ): + desc = f"{dataset}/{config}/{split}" + try: + parquet_files: List[pq.ParquetFile] = thread_map( + partial(pq.ParquetFile, filesystem=fs), source_uris, desc=desc, unit="pq", disable=True + ) + except Exception as e: + raise FileSystemError(f"Could not read the parquet files: {e}") from e + with StepProfiler( + method="parquet_index_without_metadata.from_parquet_file_items", step="get the dataset's features" + ): + features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema()) + supported_columns, unsupported_columns = get_supported_unsupported_columns( + features, unsupported_features_magic_strings=unsupported_features_magic_strings + ) + + with StepProfiler( + method="parquet_index_without_metadata.from_parquet_file_items", step="create the row group offsets" + ): + row_group_offsets = np.cumsum( + [ + parquet_file.metadata.row_group(group_id).num_rows + for parquet_file in parquet_files + for group_id in range(parquet_file.metadata.num_row_groups) + ] + ) + with StepProfiler( + method="parquet_index_without_metadata.from_parquet_file_items", step="create the row group readers" + ): + row_group_readers: List[Callable[[], pa.Table]] = [ + partial(parquet_file.read_row_group, i=group_id, columns=supported_columns) + for parquet_file in parquet_files + for group_id in range(parquet_file.metadata.num_row_groups) + ] + return ParquetIndexWithoutMetadata( + features=features, + supported_columns=supported_columns, + unsupported_columns=unsupported_columns, + row_group_offsets=row_group_offsets, + row_group_readers=row_group_readers, + ) + + +@dataclass +class ParquetIndexWithMetadata: + features: Features + supported_columns: List[str] + unsupported_columns: List[str] + parquet_files_urls: List[str] + metadata_paths: List[str] + num_bytes: List[int] + num_rows: List[int] + hf_token: Optional[str] + + def query(self, offset: int, length: int) -> pa.Table: + """Query the parquet files + + Note that this implementation will always read at least one row group, to get the list of columns and always + have the same schema, even if the requested rows are invalid (out of range). + + Args: + offset (int): The first row to read. + length (int): The number of rows to read. + + Returns: + pa.Table: The requested rows. + """ + with StepProfiler( + method="parquet_index_with_metadata.query", step="get the parquet files than contain the requested rows" + ): + parquet_file_offsets = np.cumsum(self.num_rows) + + last_row_in_parquet = parquet_file_offsets[-1] - 1 + first_row = min(offset, last_row_in_parquet) + last_row = min(offset + length - 1, last_row_in_parquet) + first_parquet_file_id, last_parquet_file_id = np.searchsorted( + parquet_file_offsets, [first_row, last_row], side="right" + ) + parquet_offset = ( + offset - parquet_file_offsets[first_parquet_file_id - 1] if first_parquet_file_id > 0 else offset + ) + urls = self.parquet_files_urls[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203 + metadata_paths = self.metadata_paths[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203 + num_bytes = self.num_bytes[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203 + + with StepProfiler( + method="parquet_index_with_metadata.query", step="load the remote parquet files using metadata from disk" + ): + httpFileSystemSession = HTTPFileSystemSession.get_instance() + session = httpFileSystemSession.session + httpfs = httpFileSystemSession.httpfs + parquet_files = [ + pq.ParquetFile( + HTTPFile(httpfs, url, session=session, size=size, loop=httpfs.loop, cache_type=None), + metadata=pq.read_metadata(metadata_path), + pre_buffer=True, + ) + for url, metadata_path, size in zip(urls, metadata_paths, num_bytes) + ] + + with StepProfiler( + method="parquet_index_with_metadata.query", step="get the row groups than contain the requested rows" + ): + row_group_offsets = np.cumsum( + [ + parquet_file.metadata.row_group(group_id).num_rows + for parquet_file in parquet_files + for group_id in range(parquet_file.metadata.num_row_groups) + ] + ) + row_group_readers: List[Callable[[], pa.Table]] = [ + partial(parquet_file.read_row_group, i=group_id, columns=self.supported_columns) + for parquet_file in parquet_files + for group_id in range(parquet_file.metadata.num_row_groups) + ] + + last_row_in_parquet = row_group_offsets[-1] - 1 + first_row = min(parquet_offset, last_row_in_parquet) + last_row = min(parquet_offset + length - 1, last_row_in_parquet) + + first_row_group_id, last_row_group_id = np.searchsorted( + row_group_offsets, [first_row, last_row], side="right" + ) + + with StepProfiler(method="parquet_index_with_metadata.query", step="read the row groups"): + pa_table = pa.concat_tables( + [row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)] + ) + first_row_in_pa_table = row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0 + return pa_table.slice(parquet_offset - first_row_in_pa_table, length) + + @staticmethod + def from_parquet_metadata_items( + parquet_file_metadata_items: List[ParquetFileMetadataItem], + parquet_metadata_directory: StrPath, + hf_token: Optional[str], + unsupported_features_magic_strings: List[str] = [], + ) -> "ParquetIndexWithMetadata": + if not parquet_file_metadata_items: + raise ParquetResponseEmptyError("No parquet files found.") + + with StepProfiler( + method="parquet_index_with_metadata.from_parquet_metadata_items", + step="get the index from parquet metadata", + ): + try: + parquet_files_metadata = sorted( + parquet_file_metadata_items, key=lambda parquet_file_metadata: parquet_file_metadata["filename"] + ) + parquet_files_urls = [parquet_file_metadata["url"] for parquet_file_metadata in parquet_files_metadata] + metadata_paths = [ + os.path.join(parquet_metadata_directory, parquet_file_metadata["parquet_metadata_subpath"]) + for parquet_file_metadata in parquet_files_metadata + ] + num_bytes = [parquet_file_metadata["size"] for parquet_file_metadata in parquet_files_metadata] + num_rows = [parquet_file_metadata["num_rows"] for parquet_file_metadata in parquet_files_metadata] + except Exception as e: + raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e + + with StepProfiler( + method="parquet_index_with_metadata.from_parquet_metadata_items", step="get the dataset's features" + ): + features = Features.from_arrow_schema(pq.read_schema(metadata_paths[0])) + supported_columns, unsupported_columns = get_supported_unsupported_columns( + features, + unsupported_features_magic_strings=unsupported_features_magic_strings, + ) + return ParquetIndexWithMetadata( + features=features, + supported_columns=supported_columns, + unsupported_columns=unsupported_columns, + parquet_files_urls=parquet_files_urls, + metadata_paths=metadata_paths, + num_bytes=num_bytes, + num_rows=num_rows, + hf_token=hf_token, + ) + + +class RowsIndex: + def __init__( + self, + dataset: str, + config: str, + split: str, + processing_graph: ProcessingGraph, + hf_token: Optional[str], + parquet_metadata_directory: StrPath, + unsupported_features_magic_strings: List[str] = [], + ): + self.dataset = dataset + self.revision: Optional[str] = None + self.config = config + self.split = split + self.processing_graph = processing_graph + self.parquet_index = self._init_parquet_index( + hf_token=hf_token, + parquet_metadata_directory=parquet_metadata_directory, + unsupported_features_magic_strings=unsupported_features_magic_strings, + ) + + def _init_parquet_index( + self, + hf_token: Optional[str], + parquet_metadata_directory: StrPath, + unsupported_features_magic_strings: List[str] = [], + ) -> Union[ParquetIndexWithMetadata, ParquetIndexWithoutMetadata]: + with StepProfiler(method="rows_index._init_parquet_index", step="all"): + # get the list of parquet files + with StepProfiler(method="rows_index._init_parquet_index", step="get list of parquet files for split"): + config_parquet_processing_steps = self.processing_graph.get_config_parquet_processing_steps() + if not config_parquet_processing_steps: + raise RuntimeError("No processing steps are configured to provide a config's parquet response.") + + config_parquet_metadata_processing_steps = ( + self.processing_graph.get_config_parquet_metadata_processing_steps() + ) + + cache_kinds = [step.cache_kind for step in config_parquet_processing_steps] + cache_kinds.extend([step.cache_kind for step in config_parquet_metadata_processing_steps]) + + try: + result = get_previous_step_or_raise( + kinds=cache_kinds, + dataset=self.dataset, + config=self.config, + split=None, + ) + self.revision = result.response["dataset_git_revision"] + content = result.response["content"] + except Exception as e: + raise UnexpectedError("Could not get the list of parquet files to fetch the rows from.") from e + # ^ TODO: improve the error, depending on the case + if content and "parquet_files" in content: + return ParquetIndexWithoutMetadata.from_parquet_file_items( + [ + parquet_item + for parquet_item in content["parquet_files"] + if parquet_item["split"] == self.split and parquet_item["config"] == self.config + ], + dataset=self.dataset, + config=self.config, + split=self.split, + hf_token=hf_token, + unsupported_features_magic_strings=unsupported_features_magic_strings, + ) + else: + return ParquetIndexWithMetadata.from_parquet_metadata_items( + [ + parquet_item + for parquet_item in content["parquet_files_metadata"] + if parquet_item["split"] == self.split and parquet_item["config"] == self.config + ], + parquet_metadata_directory=parquet_metadata_directory, + hf_token=hf_token, + unsupported_features_magic_strings=unsupported_features_magic_strings, + ) + + # note that this cache size is global for the class, not per instance + @lru_cache(maxsize=1024) + def query(self, offset: int, length: int) -> pa.Table: + """Query the parquet files + + Note that this implementation will always read at least one row group, to get the list of columns and always + have the same schema, even if the requested rows are invalid (out of range). + + Args: + offset (int): The first row to read. + length (int): The number of rows to read. + + Returns: + pa.Table: The requested rows. + """ + return self.parquet_index.query(offset=offset, length=length) + + +class Indexer: + def __init__( + self, + processing_graph: ProcessingGraph, + parquet_metadata_directory: StrPath, + unsupported_features_magic_strings: List[str] = [], + all_columns_supported_datasets_allow_list: Union[Literal["all"], List[str]] = "all", + hf_token: Optional[str] = None, + ): + self.processing_graph = processing_graph + self.parquet_metadata_directory = parquet_metadata_directory + self.hf_token = hf_token + self.unsupported_features_magic_strings = unsupported_features_magic_strings + self.all_columns_supported_datasets_allow_list = all_columns_supported_datasets_allow_list + + @lru_cache(maxsize=128) + def get_rows_index( + self, + dataset: str, + config: str, + split: str, + ) -> RowsIndex: + filter_magic_strings = ( + self.all_columns_supported_datasets_allow_list != "all" + and dataset not in self.all_columns_supported_datasets_allow_list + ) + unsupported_features_magic_strings = self.unsupported_features_magic_strings if filter_magic_strings else [] + return RowsIndex( + dataset=dataset, + config=config, + split=split, + processing_graph=self.processing_graph, + hf_token=self.hf_token, + parquet_metadata_directory=self.parquet_metadata_directory, + unsupported_features_magic_strings=unsupported_features_magic_strings, + ) diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py index d70a1112..f537eb28 100644 --- a/libs/libcommon/src/libcommon/simple_cache.py +++ b/libs/libcommon/src/libcommon/simple_cache.py @@ -402,0 +403,17 @@ def get_best_response( +def get_previous_step_or_raise( + kinds: List[str], dataset: str, config: Optional[str] = None, split: Optional[str] = None +) -> BestResponse: + """Get the previous step from the cache, or raise an exception if it failed.""" + best_response = get_best_response(kinds=kinds, dataset=dataset, config=config, split=split) + if best_response.response["http_status"] != HTTPStatus.OK: + raise CachedArtifactError( + message="The previous step failed.", + kind=best_response.kind, + dataset=dataset, + config=config, + split=split, + cache_entry_with_details=best_response.response, + ) + return best_response + + diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py index 13479c33..34d68dc4 100644 --- a/services/api/src/api/app.py +++ b/services/api/src/api/app.py @@ -114 +113,0 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi - hf_endpoint=app_config.common.hf_endpoint, diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py index 7fca6036..702fea88 100644 --- a/services/api/src/api/routes/rows.py +++ b/services/api/src/api/routes/rows.py @@ -9,2 +8,0 @@ import shutil -from dataclasses import dataclass -from functools import lru_cache, partial @@ -12,12 +10 @@ from itertools import islice -from os import PathLike -from typing import ( - Any, - Callable, - List, - Literal, - Mapping, - Optional, - Tuple, - TypedDict, - Union, -) +from typing import Any, List, Literal, Mapping, Optional, TypedDict, Union @@ -25,2 +11,0 @@ from typing import ( -import numpy as np -import numpy.typing as npt @@ -28 +12,0 @@ import pyarrow as pa -import pyarrow.parquet as pq @@ -30,3 +14,2 @@ from datasets import Features -from fsspec.implementations.http import HTTPFile, HTTPFileSystem -from huggingface_hub import HfFileSystem -from huggingface_hub.hf_file_system import safe_quote +from fsspec.implementations.http import HTTPFileSystem +from libcommon.parquet_utils import Indexer, StrPath @@ -42 +24,0 @@ from starlette.responses import Response -from tqdm.contrib.concurrent import thread_map @@ -45 +26,0 @@ from api.authentication import auth_check -from api.routes.endpoint import get_cache_entry_from_steps @@ -66,17 +46,0 @@ MAX_ROWS = 100 -PARQUET_REVISION = "refs/convert/parquet" - - -StrPath = Union[str, PathLike[str]] - - -class FileSystemError(Exception): - pass - - -class ParquetResponseFormatError(Exception): - pass - - -class ParquetResponseEmptyError(Exception): - pass - @@ -88,45 +51,0 @@ class ParquetDataProcessingError(Exception): -class ParquetFileItem(TypedDict): - dataset: str - config: str - split: str - url: str - filename: str - size: int - - -class ParquetFileMetadataItem(TypedDict): - dataset: str - config: str - split: str - url: str - filename: str - size: int - num_rows: int - parquet_metadata_subpath: str - - -# TODO: how to invalidate the cache when the parquet branch is created or deleted? -@lru_cache(maxsize=128) -def get_hf_fs(hf_token: Optional[str]) -> HfFileSystem: - """Get the Hugging Face filesystem. - - Args: - hf_token (Optional[str]): The token to access the filesystem. - Returns: - HfFileSystem: The Hugging Face filesystem. - """ - return HfFileSystem(token=hf_token) - - -def get_hf_parquet_uris(paths: List[str], dataset: str) -> List[str]: - """Get the Hugging Face URIs from the Parquet branch of the dataset repository (see PARQUET_REVISION). - - Args: - paths (List[str]): List of paths. - dataset (str): The dataset name. - Returns: - List[str]: List of Parquet URIs. - """ - return [f"hf://datasets/{dataset}@{safe_quote(PARQUET_REVISION)}/{path}" for path in paths] - - @@ -138,359 +56,0 @@ UNSUPPORTED_FEATURES_MAGIC_STRINGS = ["'binary'", "Audio("] - -def get_supported_unsupported_columns(features: Features, dataset_name: str) -> Tuple[List[str], List[str]]: - supported_columns, unsupported_columns = [], [] - unsupported_features_magic_strings = ( - UNSUPPORTED_FEATURES_MAGIC_STRINGS - if ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST != "all" - and dataset_name not in ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST - else [] - ) - for column, feature in features.items(): - str_feature = str(feature) - str_column = str(column) - if any(magic_string in str_feature for magic_string in unsupported_features_magic_strings): - unsupported_columns.append(str_column) - else: - supported_columns.append(str_column) - return supported_columns, unsupported_columns - - -@dataclass -class ParquetIndexWithoutMetadata: - features: Features - supported_columns: List[str] - unsupported_columns: List[str] - row_group_offsets: npt.NDArray[np.int64] - row_group_readers: List[Callable[[], pa.Table]] - - def query(self, offset: int, length: int) -> pa.Table: - """Query the parquet files - - Note that this implementation will always read at least one row group, to get the list of columns and always - have the same schema, even if the requested rows are invalid (out of range). - - Args: - offset (int): The first row to read. - length (int): The number of rows to read. - - Returns: - pa.Table: The requested rows. - """ - if (len(self.row_group_offsets) == 0) or (len(self.row_group_readers) == 0): - raise ParquetResponseEmptyError("No parquet files found.") - last_row_in_parquet = self.row_group_offsets[-1] - 1 - first_row = min(offset, last_row_in_parquet) - last_row = min(offset + length - 1, last_row_in_parquet) - first_row_group_id, last_row_group_id = np.searchsorted( - self.row_group_offsets, [first_row, last_row], side="right" - ) - pa_table = pa.concat_tables( - [self.row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)] - ) - first_row_in_pa_table = self.row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0 - return pa_table.slice(offset - first_row_in_pa_table, length) - - @staticmethod - def from_parquet_file_items( - parquet_file_items: List[ParquetFileItem], dataset: str, config: str, split: str, hf_token: Optional[str] - ) -> "ParquetIndexWithoutMetadata": - try: - sources = sorted(f"{config}/{parquet_file['filename']}" for parquet_file in parquet_file_items) - except Exception as e: - raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e - logging.debug( - f"Found {len(sources)} parquet files for dataset={dataset}, config={config}, split={split}: {sources}" - ) - if not sources: - raise ParquetResponseEmptyError("No parquet files found.") - with StepProfiler(method="rows.index.without_metadata", step="get the Hub's dataset filesystem"): - fs = get_hf_fs(hf_token=hf_token) - with StepProfiler(method="rows.index.without_metadata", step="get the source URIs"): - source_uris = get_hf_parquet_uris(sources, dataset=dataset) - with StepProfiler(method="rows.index.without_metadata", step="get one parquet reader per parquet file"): - desc = f"{dataset}/{config}/{split}" - try: - parquet_files: List[pq.ParquetFile] = thread_map( - partial(pq.ParquetFile, filesystem=fs), source_uris, desc=desc, unit="pq", disable=True - ) - except Exception as e: - raise FileSystemError(f"Could not read the parquet files: {e}") from e - with StepProfiler(method="rows.index.without_metadata", step="get the dataset's features"): - features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema()) - supported_columns, unsupported_columns = get_supported_unsupported_columns(features, dataset_name=dataset) - - with StepProfiler(method="rows.index.without_metadata", step="create the row group offsets"): - row_group_offsets = np.cumsum( - [ - parquet_file.metadata.row_group(group_id).num_rows - for parquet_file in parquet_files - for group_id in range(parquet_file.metadata.num_row_groups) - ] - ) - with StepProfiler(method="rows.index.without_metadata", step="create the row group readers"): - row_group_readers: List[Callable[[], pa.Table]] = [ - partial(parquet_file.read_row_group, i=group_id, columns=supported_columns) - for parquet_file in parquet_files - for group_id in range(parquet_file.metadata.num_row_groups) - ] - return ParquetIndexWithoutMetadata( - features=features, - supported_columns=supported_columns, - unsupported_columns=unsupported_columns, - row_group_offsets=row_group_offsets, - row_group_readers=row_group_readers, - ) - - -@dataclass -class ParquetIndexWithMetadata: - features: Features - supported_columns: List[str] - unsupported_columns: List[str] - parquet_files_urls: List[str] - metadata_paths: List[str] - num_bytes: List[int] - num_rows: List[int] - hf_token: Optional[str] - - def query(self, offset: int, length: int) -> pa.Table: - """Query the parquet files - - Note that this implementation will always read at least one row group, to get the list of columns and always - have the same schema, even if the requested rows are invalid (out of range). - - Args: - offset (int): The first row to read. - length (int): The number of rows to read. - - Returns: - pa.Table: The requested rows. - """ - with StepProfiler( - method="rows.query.with_metadata", step="get the parquet files than contain the requested rows" - ): - parquet_file_offsets = np.cumsum(self.num_rows) - - last_row_in_parquet = parquet_file_offsets[-1] - 1 - first_row = min(offset, last_row_in_parquet) - last_row = min(offset + length - 1, last_row_in_parquet) - first_parquet_file_id, last_parquet_file_id = np.searchsorted( - parquet_file_offsets, [first_row, last_row], side="right" - ) - parquet_offset = ( - offset - parquet_file_offsets[first_parquet_file_id - 1] if first_parquet_file_id > 0 else offset - ) - urls = self.parquet_files_urls[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203 - metadata_paths = self.metadata_paths[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203 - num_bytes = self.num_bytes[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203 - - with StepProfiler( - method="rows.query.with_metadata", step="load the remote parquet files using metadata from disk" - ): - parquet_files = [ - pq.ParquetFile( - HTTPFile(httpfs, url, session=session, size=size, loop=httpfs.loop, cache_type=None), - metadata=pq.read_metadata(metadata_path), - pre_buffer=True, - ) - for url, metadata_path, size in zip(urls, metadata_paths, num_bytes) - ] - - with StepProfiler( - method="rows.query.with_metadata", step="get the row groups than contain the requested rows" - ): - row_group_offsets = np.cumsum( - [ - parquet_file.metadata.row_group(group_id).num_rows - for parquet_file in parquet_files - for group_id in range(parquet_file.metadata.num_row_groups) - ] - ) - row_group_readers: List[Callable[[], pa.Table]] = [ - partial(parquet_file.read_row_group, i=group_id, columns=self.supported_columns) - for parquet_file in parquet_files - for group_id in range(parquet_file.metadata.num_row_groups) - ] - - last_row_in_parquet = row_group_offsets[-1] - 1 - first_row = min(parquet_offset, last_row_in_parquet) - last_row = min(parquet_offset + length - 1, last_row_in_parquet) - - first_row_group_id, last_row_group_id = np.searchsorted( - row_group_offsets, [first_row, last_row], side="right" - ) - - with StepProfiler(method="rows.query.with_metadata", step="read the row groups"): - pa_table = pa.concat_tables( - [row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)] - ) - first_row_in_pa_table = row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0 - return pa_table.slice(parquet_offset - first_row_in_pa_table, length) - - @staticmethod - def from_parquet_metadata_items( - parquet_file_metadata_items: List[ParquetFileMetadataItem], - parquet_metadata_directory: StrPath, - hf_token: Optional[str], - ) -> "ParquetIndexWithMetadata": - if not parquet_file_metadata_items: - raise ParquetResponseEmptyError("No parquet files found.") - - with StepProfiler(method="rows.index", step="get the index from parquet metadata"): - try: - parquet_files_metadata = sorted( - parquet_file_metadata_items, key=lambda parquet_file_metadata: parquet_file_metadata["filename"] - ) - parquet_files_urls = [parquet_file_metadata["url"] for parquet_file_metadata in parquet_files_metadata] - metadata_paths = [ - os.path.join(parquet_metadata_directory, parquet_file_metadata["parquet_metadata_subpath"]) - for parquet_file_metadata in parquet_files_metadata - ] - num_bytes = [parquet_file_metadata["size"] for parquet_file_metadata in parquet_files_metadata] - num_rows = [parquet_file_metadata["num_rows"] for parquet_file_metadata in parquet_files_metadata] - dataset_name = parquet_files_metadata[0]["dataset"] - except Exception as e: - raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e - - with StepProfiler(method="rows.index.with_metadata", step="get the dataset's features"): - features = Features.from_arrow_schema(pq.read_schema(metadata_paths[0])) - supported_columns, unsupported_columns = get_supported_unsupported_columns( - features, dataset_name=dataset_name - ) - return ParquetIndexWithMetadata( - features=features, - supported_columns=supported_columns, - unsupported_columns=unsupported_columns, - parquet_files_urls=parquet_files_urls, - metadata_paths=metadata_paths, - num_bytes=num_bytes, - num_rows=num_rows, - hf_token=hf_token, - ) - - -class RowsIndex: - def __init__( - self, - dataset: str, - config: str, - split: str, - processing_graph: ProcessingGraph, - hf_endpoint: str, - hf_token: Optional[str], - parquet_metadata_directory: StrPath, - ): - self.dataset = dataset - self.revision: Optional[str] = None - self.config = config - self.split = split - self.processing_graph = processing_graph - self.parquet_index = self._init_parquet_index( - hf_endpoint=hf_endpoint, - hf_token=hf_token, - parquet_metadata_directory=parquet_metadata_directory, - ) - - def _init_parquet_index( - self, - hf_endpoint: str, - hf_token: Optional[str], - parquet_metadata_directory: StrPath, - ) -> Union[ParquetIndexWithMetadata, ParquetIndexWithoutMetadata]: - with StepProfiler(method="rows.index", step="all"): - # get the list of parquet files - with StepProfiler(method="rows.index", step="get list of parquet files for split"): - config_parquet_processing_steps = self.processing_graph.get_config_parquet_processing_steps() - config_parquet_metadata_processing_steps = ( - self.processing_graph.get_config_parquet_metadata_processing_steps() - ) - if not config_parquet_processing_steps: - raise RuntimeError("No processing steps are configured to provide a config's parquet response.") - try: - result = get_cache_entry_from_steps( - processing_steps=config_parquet_metadata_processing_steps + config_parquet_processing_steps, - dataset=self.dataset, - config=self.config, - split=None, - processing_graph=self.processing_graph, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - ) - self.revision = result["dataset_git_revision"] - content = result["content"] - except ApiCustomError as e: - raise e - except Exception as e: - raise UnexpectedError("Could not get the list of parquet files to fetch the rows from.") from e - # ^ TODO: improve the error, depending on the case - if content and "parquet_files" in content: - return ParquetIndexWithoutMetadata.from_parquet_file_items( - [ - parquet_item - for parquet_item in content["parquet_files"] - if parquet_item["split"] == self.split and parquet_item["config"] == self.config - ], - dataset=self.dataset, - config=self.config, - split=self.split, - hf_token=hf_token, - ) - else: - return ParquetIndexWithMetadata.from_parquet_metadata_items( - [ - parquet_item - for parquet_item in content["parquet_files_metadata"] - if parquet_item["split"] == self.split and parquet_item["config"] == self.config - ], - parquet_metadata_directory=parquet_metadata_directory, - hf_token=hf_token, - ) - - # note that this cache size is global for the class, not per instance - @lru_cache(maxsize=1024) - def query(self, offset: int, length: int) -> pa.Table: - """Query the parquet files - - Note that this implementation will always read at least one row group, to get the list of columns and always - have the same schema, even if the requested rows are invalid (out of range). - - Args: - offset (int): The first row to read. - length (int): The number of rows to read. - - Returns: - pa.Table: The requested rows. - """ - return self.parquet_index.query(offset=offset, length=length) - - -class Indexer: - def __init__( - self, - processing_graph: ProcessingGraph, - parquet_metadata_directory: StrPath, - hf_endpoint: str, - hf_token: Optional[str] = None, - ): - self.processing_graph = processing_graph - self.parquet_metadata_directory = parquet_metadata_directory - self.hf_endpoint = hf_endpoint - self.hf_token = hf_token - - @lru_cache(maxsize=128) - def get_rows_index( - self, - dataset: str, - config: str, - split: str, - ) -> RowsIndex: - return RowsIndex( - dataset=dataset, - config=config, - split=split, - processing_graph=self.processing_graph, - hf_endpoint=self.hf_endpoint, - hf_token=self.hf_token, - parquet_metadata_directory=self.parquet_metadata_directory, - ) - - @@ -700 +259,0 @@ def create_rows_endpoint( - hf_endpoint: str, @@ -715 +273,0 @@ def create_rows_endpoint( - hf_endpoint=hf_endpoint, @@ -717,0 +276,2 @@ def create_rows_endpoint( + unsupported_features_magic_strings=UNSUPPORTED_FEATURES_MAGIC_STRINGS, + all_columns_supported_datasets_allow_list=ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST, @@ -752 +312,5 @@ def create_rows_endpoint( - rows_index = indexer.get_rows_index(dataset=dataset, config=config, split=split) + rows_index = indexer.get_rows_index( + dataset=dataset, + config=config, + split=split, + ) diff --git a/services/api/tests/routes/test_rows.py b/services/api/tests/routes/test_rows.py index 4b3f7f86..9dbad3ed 100644 --- a/services/api/tests/routes/test_rows.py +++ b/services/api/tests/routes/test_rows.py @@ -14,0 +15,6 @@ from fsspec import AbstractFileSystem +from libcommon.parquet_utils import ( + Indexer, + ParquetIndexWithMetadata, + ParquetIndexWithoutMetadata, + RowsIndex, +) @@ -21,8 +27 @@ from api.config import AppConfig -from api.routes.rows import ( - Indexer, - ParquetIndexWithMetadata, - ParquetIndexWithoutMetadata, - RowsIndex, - clean_cached_assets, - create_response, -) +from api.routes.rows import clean_cached_assets, create_response @@ -251 +249,0 @@ def indexer(app_config: AppConfig, processing_graph: ProcessingGraph, parquet_me - hf_endpoint=app_config.common.hf_endpoint, @@ -268,2 +266,2 @@ def rows_index( - with patch("api.routes.rows.get_hf_fs", return_value=ds_sharded_fs): - with patch("api.routes.rows.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris): + with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_sharded_fs): + with patch("libcommon.parquet_utils.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris): @@ -276,2 +274,2 @@ def test_indexer_get_rows_index( - with patch("api.routes.rows.get_hf_fs", return_value=ds_fs): - with patch("api.routes.rows.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris): + with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_fs): + with patch("libcommon.parquet_utils.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris): @@ -295,2 +293,2 @@ def test_indexer_get_rows_index_sharded( - with patch("api.routes.rows.get_hf_fs", return_value=ds_sharded_fs): - with patch("api.routes.rows.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris): + with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_sharded_fs): + with patch("libcommon.parquet_utils.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris): @@ -325 +323 @@ def rows_index_with_parquet_metadata( - with patch("api.routes.rows.HTTPFile", return_value=f): + with patch("libcommon.parquet_utils.HTTPFile", return_value=f): @@ -333 +331 @@ def test_indexer_get_rows_index_with_parquet_metadata( - with patch("api.routes.rows.HTTPFile", return_value=f): + with patch("libcommon.parquet_utils.HTTPFile", return_value=f): @@ -354 +352 @@ def test_indexer_get_rows_index_sharded_with_parquet_metadata( - with patch("api.routes.rows.HTTPFile", return_value=f): + with patch("libcommon.parquet_utils.HTTPFile", return_value=f): diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py index 0e4bfe82..2352944c 100644 --- a/services/worker/src/worker/job_runner_factory.py +++ b/services/worker/src/worker/job_runner_factory.py @@ -174,0 +175 @@ class JobRunnerFactory(BaseJobRunnerFactory): + processing_graph=self.processing_graph, @@ -175,0 +177 @@ class JobRunnerFactory(BaseJobRunnerFactory): + parquet_metadata_directory=self.parquet_metadata_directory, diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py index 265ef38b..ae0e2760 100644 --- a/services/worker/src/worker/job_runners/config/info.py +++ b/services/worker/src/worker/job_runners/config/info.py @@ -5,0 +6 @@ from libcommon.exceptions import PreviousStepFormatError +from libcommon.simple_cache import get_previous_step_or_raise @@ -8 +9 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import CompleteJobResult, get_previous_step_or_raise +from worker.utils import CompleteJobResult diff --git a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py index 54b18b0b..9c9306c4 100644 --- a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py @@ -10,6 +10,2 @@ from libcommon.exceptions import PreviousStepFormatError -from libcommon.simple_cache import DoesNotExist, get_response - -from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import ( - JobResult, - OptInOutUrlsCountResponse, +from libcommon.simple_cache import ( + DoesNotExist, @@ -16,0 +13 @@ from worker.utils import ( + get_response, @@ -18,0 +16,3 @@ from worker.utils import ( +from worker.job_runners.config.config_job_runner import ConfigJobRunner +from worker.utils import JobResult, OptInOutUrlsCountResponse + diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py index ee88039b..eae1d780 100644 --- a/services/worker/src/worker/job_runners/config/parquet.py +++ b/services/worker/src/worker/job_runners/config/parquet.py @@ -8,0 +9 @@ from libcommon.exceptions import PreviousStepFormatError +from libcommon.simple_cache import get_previous_step_or_raise @@ -12 +13 @@ from worker.job_runners.config.parquet_and_info import ParquetFileItem -from worker.utils import CompleteJobResult, get_previous_step_or_raise +from worker.utils import CompleteJobResult diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index e6c4b1b2..91de381e 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -68,0 +69 @@ from libcommon.processing_graph import ProcessingStep +from libcommon.simple_cache import get_previous_step_or_raise @@ -74 +75 @@ from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner -from worker.utils import CompleteJobResult, get_previous_step_or_raise +from worker.utils import CompleteJobResult diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py index 97600419..e1cb6ef5 100644 --- a/services/worker/src/worker/job_runners/config/parquet_metadata.py +++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py @@ -16,0 +17 @@ from libcommon.processing_graph import ProcessingStep +from libcommon.simple_cache import get_previous_step_or_raise @@ -26 +27 @@ from worker.job_runners.config.parquet_and_info import ParquetFileItem -from worker.utils import CompleteJobResult, get_previous_step_or_raise +from worker.utils import CompleteJobResult diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py index 4b2dfca4..e46b9108 100644 --- a/services/worker/src/worker/job_runners/config/size.py +++ b/services/worker/src/worker/job_runners/config/size.py @@ -8,0 +9 @@ from libcommon.exceptions import PreviousStepFormatError +from libcommon.simple_cache import get_previous_step_or_raise @@ -11 +12 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import CompleteJobResult, get_previous_step_or_raise +from worker.utils import CompleteJobResult diff --git a/services/worker/src/worker/job_runners/config/split_names_from_info.py b/services/worker/src/worker/job_runners/config/split_names_from_info.py index 7d6d9fb1..7037de05 100644 --- a/services/worker/src/worker/job_runners/config/split_names_from_info.py +++ b/services/worker/src/worker/job_runners/config/split_names_from_info.py @@ -11,0 +12 @@ from libcommon.exceptions import PreviousStepFormatError +from libcommon.simple_cache import get_previous_step_or_raise @@ -14,7 +15 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner -from worker.utils import ( - CompleteJobResult, - JobRunnerInfo, - SplitItem, - SplitsList, - get_previous_step_or_raise, -) +from worker.utils import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py index 2cedb82a..ad0c1408 100644 --- a/services/worker/src/worker/job_runners/dataset/info.py +++ b/services/worker/src/worker/job_runners/dataset/info.py @@ -10 +10,5 @@ from libcommon.exceptions import PreviousStepFormatError -from libcommon.simple_cache import DoesNotExist, get_response +from libcommon.simple_cache import ( + DoesNotExist, + get_previous_step_or_raise, + get_response, +) @@ -13 +17 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult, PreviousJob, get_previous_step_or_raise +from worker.utils import JobResult, PreviousJob diff --git a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py index 8a36fbae..530d3145 100644 --- a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py @@ -10,6 +10,2 @@ from libcommon.exceptions import PreviousStepFormatError -from libcommon.simple_cache import DoesNotExist, get_response - -from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import ( - JobResult, - OptInOutUrlsCountResponse, +from libcommon.simple_cache import ( + DoesNotExist, @@ -16,0 +13 @@ from worker.utils import ( + get_response, @@ -18,0 +16,3 @@ from worker.utils import ( +from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner +from worker.utils import JobResult, OptInOutUrlsCountResponse + diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py index 6d12167f..4a360950 100644 --- a/services/worker/src/worker/job_runners/dataset/parquet.py +++ b/services/worker/src/worker/job_runners/dataset/parquet.py @@ -10 +10,5 @@ from libcommon.exceptions import PreviousStepFormatError -from libcommon.simple_cache import DoesNotExist, get_response +from libcommon.simple_cache import ( + DoesNotExist, + get_previous_step_or_raise, + get_response, +) @@ -15 +19 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult, PreviousJob, get_previous_step_or_raise +from worker.utils import JobResult, PreviousJob diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py index 84db9f3c..3b69e948 100644 --- a/services/worker/src/worker/job_runners/dataset/size.py +++ b/services/worker/src/worker/job_runners/dataset/size.py @@ -10 +10,5 @@ from libcommon.exceptions import PreviousStepFormatError -from libcommon.simple_cache import DoesNotExist, get_response +from libcommon.simple_cache import ( + DoesNotExist, + get_previous_step_or_raise, + get_response, +) @@ -14 +18 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner -from worker.utils import JobResult, PreviousJob, get_previous_step_or_raise +from worker.utils import JobResult, PreviousJob diff --git a/services/worker/src/worker/job_runners/dataset/split_names.py b/services/worker/src/worker/job_runners/dataset/split_names.py index eea5a259..259e66d4 100644 --- a/services/worker/src/worker/job_runners/dataset/split_names.py +++ b/services/worker/src/worker/job_runners/dataset/split_names.py @@ -10 +10 @@ from libcommon.exceptions import PreviousStepFormatError -from libcommon.simple_cache import get_best_response +from libcommon.simple_cache import get_best_response, get_previous_step_or_raise @@ -19 +18,0 @@ from worker.utils import ( - get_previous_step_or_raise, diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py index 0d6862b0..64112c28 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py @@ -5,2 +5 @@ import logging -from functools import lru_cache, partial -from typing import List, Optional +from typing import List @@ -8 +6,0 @@ from typing import List, Optional -import pyarrow as pa @@ -10,2 +7,0 @@ from datasets import Features -from huggingface_hub import HfFileSystem -from huggingface_hub.hf_file_system import safe_quote @@ -13 +8,0 @@ from libcommon.constants import ( - PARQUET_REVISION, @@ -18,3 +12,0 @@ from libcommon.exceptions import ( - FileSystemError, - ParquetResponseEmptyError, - PreviousStepFormatError, @@ -25 +17,2 @@ from libcommon.exceptions import ( -from libcommon.processing_graph import ProcessingStep +from libcommon.parquet_utils import Indexer +from libcommon.processing_graph import ProcessingGraph, ProcessingStep @@ -29,2 +21,0 @@ from libcommon.viewer_utils.features import get_cell_value -from pyarrow.parquet import ParquetFile -from tqdm.contrib.concurrent import thread_map @@ -42 +32,0 @@ from worker.utils import ( - get_previous_step_or_raise, @@ -75,24 +64,0 @@ def transform_rows( -@lru_cache(maxsize=128) -def get_hf_fs(hf_token: Optional[str]) -> HfFileSystem: - """Get the Hugging Face filesystem. - - Args: - hf_token (Optional[str]): The token to access the filesystem. - Returns: - HfFileSystem: The Hugging Face filesystem. - """ - return HfFileSystem(token=hf_token) - - -def get_hf_parquet_uris(paths: List[str], dataset: str) -> List[str]: - """Get the Hugging Face URIs from the Parquet branch of the dataset repository (see PARQUET_REVISION). - - Args: - paths (List[str]): List of paths. - dataset (str): The dataset name. - Returns: - List[str]: List of Parquet URIs. - """ - return [f"hf://datasets/{dataset}@{safe_quote(PARQUET_REVISION)}/{path}" for path in paths] - - @@ -104 +69,0 @@ def compute_first_rows_response( - hf_token: Optional[str], @@ -110,0 +76 @@ def compute_first_rows_response( + indexer: Indexer, @@ -114,29 +80,5 @@ def compute_first_rows_response( - # first ensure the tuple (dataset, config, split) exists on the Hub - - config_parquet_best_response = get_previous_step_or_raise(kinds=["config-parquet"], dataset=dataset, config=config) - try: - parquet_files_content = config_parquet_best_response.response["content"]["parquet_files"] - sources = sorted( - f"{config}/{parquet_file['filename']}" - for parquet_file in parquet_files_content - if parquet_file["split"] == split and parquet_file["config"] == config - ) - if not sources: - raise ParquetResponseEmptyError("No parquet files found.") - except Exception as e: - raise PreviousStepFormatError("Previous step did not return the expected content.") from e - - logging.debug(f"Found {len(sources)} parquet files for {dataset=}, {config=}, {split=}: {sources}") - - fs = get_hf_fs(hf_token=hf_token) - source_uris = get_hf_parquet_uris(sources, dataset=dataset) - desc = f"{dataset}/{config}/{split}" - try: - parquet_files: List[ParquetFile] = thread_map( - partial(ParquetFile, filesystem=fs), source_uris, desc=desc, unit="pq", disable=True - ) - except Exception as e: - raise FileSystemError(f"Could not read the parquet files: {e}") from e - - # get the features - features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema()) + rows_index = indexer.get_rows_index( + dataset=dataset, + config=config, + split=split, + ) @@ -143,0 +86,2 @@ def compute_first_rows_response( + # validate the features + features = rows_index.parquet_index.features @@ -169,22 +113 @@ def compute_first_rows_response( - num_rows = 0 - last_row_group_id = 0 - row_group_readers = [] - for parquet_file in parquet_files: - for group_id in range(parquet_file.metadata.num_row_groups): - last_row_group_id = group_id - row_group_readers.append(partial(parquet_file.read_row_group, i=group_id)) - if num_rows + parquet_file.metadata.row_group(group_id).num_rows >= rows_max_number: - num_rows = rows_max_number - break - else: - num_rows += parquet_file.metadata.row_group(group_id).num_rows - else: - continue - break - - if len(row_group_readers) == 0: - raise ParquetResponseEmptyError("No parquet files found.") - - pa_table = pa.concat_tables([row_group_readers[i]() for i in range(last_row_group_id + 1)]) - result = pa_table.slice(0, num_rows) - + pa_table = rows_index.query(offset=0, length=rows_max_number) @@ -199 +122 @@ def compute_first_rows_response( - for idx, row in enumerate(result.to_pylist()) + for idx, row in enumerate(pa_table.to_pylist()) @@ -234,0 +158 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner): + indexed: Indexer @@ -255,0 +180 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner): + processing_graph: ProcessingGraph, @@ -256,0 +182 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner): + parquet_metadata_directory: StrPath, @@ -265,0 +192,8 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner): + self.parquet_metadata_directory = parquet_metadata_directory + self.indexer = Indexer( + processing_graph=processing_graph, + hf_token=self.app_config.common.hf_token, + parquet_metadata_directory=parquet_metadata_directory, + unsupported_features_magic_strings=[], + all_columns_supported_datasets_allow_list="all", + ) @@ -275 +208,0 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner): - hf_token=self.app_config.common.hf_token, @@ -280,0 +214 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner): + indexer=self.indexer, diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py index d3a617cb..43221906 100644 --- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py @@ -22,0 +23 @@ from libcommon.processing_graph import ProcessingStep +from libcommon.simple_cache import get_previous_step_or_raise @@ -36 +36,0 @@ from worker.utils import ( - get_previous_step_or_raise, diff --git a/services/worker/src/worker/job_runners/split/image_url_columns.py b/services/worker/src/worker/job_runners/split/image_url_columns.py index 1686e800..c2b19607 100644 --- a/services/worker/src/worker/job_runners/split/image_url_columns.py +++ b/services/worker/src/worker/job_runners/split/image_url_columns.py @@ -7,0 +8 @@ from libcommon.exceptions import PreviousStepFormatError +from libcommon.simple_cache import get_previous_step_or_raise @@ -15 +15,0 @@ from worker.utils import ( - get_previous_step_or_raise, diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py index cf83ec1f..af094266 100644 --- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py @@ -7,0 +8 @@ from libcommon.exceptions import PreviousStepFormatError +from libcommon.simple_cache import get_previous_step_or_raise @@ -10,5 +11 @@ from worker.job_runners.split.split_job_runner import SplitJobRunner -from worker.utils import ( - CompleteJobResult, - OptInOutUrlsCountResponse, - get_previous_step_or_raise, -) +from worker.utils import CompleteJobResult, OptInOutUrlsCountResponse diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py index 28fd6f0e..7544df77 100644 --- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py +++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py @@ -20,0 +21 @@ from libcommon.processing_graph import ProcessingStep +from libcommon.simple_cache import get_previous_step_or_raise @@ -29 +29,0 @@ from worker.utils import ( - get_previous_step_or_raise, diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py index 17ccd75b..4399bd6e 100644 --- a/services/worker/src/worker/utils.py +++ b/services/worker/src/worker/utils.py @@ -10 +9,0 @@ from dataclasses import dataclass, field -from http import HTTPStatus @@ -32 +30,0 @@ from libcommon.exceptions import NormalRowsError, StreamingRowsError -from libcommon.simple_cache import BestResponse, CachedArtifactError, get_best_response @@ -403,17 +400,0 @@ def get_rows_or_raise( - - -def get_previous_step_or_raise( - kinds: List[str], dataset: str, config: Optional[str] = None, split: Optional[str] = None -) -> BestResponse: - """Get the previous step from the cache, or raise an exception if it failed.""" - best_response = get_best_response(kinds=kinds, dataset=dataset, config=config, split=split) - if best_response.response["http_status"] != HTTPStatus.OK: - raise CachedArtifactError( - message="The previous step failed.", - kind=best_response.kind, - dataset=dataset, - config=config, - split=split, - cache_entry_with_details=best_response.response, - ) - return best_response diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py index 43fd1de7..851a99c6 100644 --- a/services/worker/tests/conftest.py +++ b/services/worker/tests/conftest.py @@ -143 +143 @@ def another_processing_step(test_processing_graph: ProcessingGraph) -> Processin -pytest_plugins = ["tests.fixtures.datasets", "tests.fixtures.files", "tests.fixtures.hub"] +pytest_plugins = ["tests.fixtures.datasets", "tests.fixtures.files", "tests.fixtures.hub", "tests.fixtures.fsspec"] diff --git a/services/worker/tests/fixtures/fsspec.py b/services/worker/tests/fixtures/fsspec.py new file mode 100644 index 00000000..848dceb5 --- /dev/null +++ b/services/worker/tests/fixtures/fsspec.py @@ -0,0 +1,119 @@ +# type: ignore +import posixpath +import shutil +from pathlib import Path +from unittest.mock import patch + +import fsspec +import pytest +from fsspec.implementations.local import ( + AbstractFileSystem, + LocalFileSystem, + stringify_path, +) + + +class MockFileSystem(AbstractFileSystem): + protocol = "mock" + + def __init__(self, *args, local_root_dir, **kwargs): + super().__init__() + self._fs = LocalFileSystem(*args, **kwargs) + self.local_root_dir = Path(local_root_dir).resolve().as_posix() + "/" + + def mkdir(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.mkdir(path, *args, **kwargs) + + def makedirs(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.makedirs(path, *args, **kwargs) + + def rmdir(self, path): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.rmdir(path) + + def ls(self, path, detail=True, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + out = self._fs.ls(path, detail=detail, *args, **kwargs) + if detail: + return [{**info, "name": info["name"][len(self.local_root_dir) :]} for info in out] # noqa: E203 + else: + return [name[len(self.local_root_dir) :] for name in out] # noqa: E203 + + def info(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + out = dict(self._fs.info(path, *args, **kwargs)) + out["name"] = out["name"][len(self.local_root_dir) :] # noqa: E203 + return out + + def cp_file(self, path1, path2, *args, **kwargs): + path1 = posixpath.join(self.local_root_dir, self._strip_protocol(path1)) + path2 = posixpath.join(self.local_root_dir, self._strip_protocol(path2)) + return self._fs.cp_file(path1, path2, *args, **kwargs) + + def rm_file(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.rm_file(path, *args, **kwargs) + + def rm(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.rm(path, *args, **kwargs) + + def _open(self, path, *args, **kwargs): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs._open(path, *args, **kwargs) + + def created(self, path): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.created(path) + + def modified(self, path): + path = posixpath.join(self.local_root_dir, self._strip_protocol(path)) + return self._fs.modified(path) + + @classmethod + def _strip_protocol(cls, path): + path = stringify_path(path) + if path.startswith("mock://"): + path = path[7:] + return path + + +class TmpDirFileSystem(MockFileSystem): + protocol = "tmp" + tmp_dir = None + + def __init__(self, *args, **kwargs): + assert self.tmp_dir is not None, "TmpDirFileSystem.tmp_dir is not set" + super().__init__(*args, **kwargs, local_root_dir=self.tmp_dir, auto_mkdir=True) + + @classmethod + def _strip_protocol(cls, path): + path = stringify_path(path) + if path.startswith("tmp://"): + path = path[6:] + return path + + [email protected] +def mock_fsspec(): + original_registry = fsspec.registry.copy() + fsspec.register_implementation("mock", MockFileSystem) + fsspec.register_implementation("tmp", TmpDirFileSystem) + yield + fsspec.registry = original_registry + + [email protected] +def mockfs(tmp_path_factory, mock_fsspec): + local_fs_dir = tmp_path_factory.mktemp("mockfs") + return MockFileSystem(local_root_dir=local_fs_dir, auto_mkdir=True) + + [email protected] +def tmpfs(tmp_path_factory, mock_fsspec): + tmp_fs_dir = tmp_path_factory.mktemp("tmpfs") + with patch.object(TmpDirFileSystem, "tmp_dir", tmp_fs_dir): + yield TmpDirFileSystem() + shutil.rmtree(tmp_fs_dir) diff --git a/services/worker/tests/job_runners/split/config/dataset-split.parquet b/services/worker/tests/job_runners/split/config/dataset-split.parquet deleted file mode 100644 index 672ab11e..00000000 Binary files a/services/worker/tests/job_runners/split/config/dataset-split.parquet and /dev/null differ diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py index 46a92c80..3dcb37c1 100644 --- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py +++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py @@ -4 +3,0 @@ -import os @@ -7 +6 @@ from http import HTTPStatus -from typing import Callable, List +from typing import Callable, Generator, List @@ -10,0 +10,2 @@ import pytest +from datasets import Dataset +from fsspec import AbstractFileSystem @@ -17 +17,0 @@ from libcommon.utils import Priority -from pyarrow.fs import LocalFileSystem @@ -30,0 +31 @@ def get_job_runner( + parquet_metadata_directory: StrPath, @@ -44 +45,5 @@ def get_job_runner( - "config-level": {"input_type": "dataset", "triggered_by": "dataset-level"}, + "config-level": { + "input_type": "config", + "triggered_by": "dataset-level", + "provides_config_parquet": True, + }, @@ -65,0 +71 @@ def get_job_runner( + processing_graph=processing_graph, @@ -66,0 +73 @@ def get_job_runner( + parquet_metadata_directory=parquet_metadata_directory, @@ -71,0 +79,12 @@ def get_job_runner( [email protected] +def ds() -> Dataset: + return Dataset.from_dict({"col1": [1, 2, 3], "col2": ["a", "b", "c"]}) + + [email protected] +def ds_fs(ds: Dataset, tmpfs: AbstractFileSystem) -> Generator[AbstractFileSystem, None, None]: + with tmpfs.open("config/dataset-split.parquet", "wb") as f: + ds.to_parquet(f) + yield tmpfs + + @@ -84,0 +104 @@ def test_compute( + ds_fs: AbstractFileSystem, @@ -91,0 +112,14 @@ def test_compute( + + config_parquet_content = { + "parquet_files": [ + { + "dataset": dataset, + "config": config, + "split": split, + "url": "https://fake.huggingface.co/datasets/dataset/resolve/refs%2Fconvert%2Fparquet/config/dataset-split.parquet", # noqa: E501 + "filename": "dataset-split.parquet", + "size": 128, + } + ] + } + @@ -93 +127 @@ def test_compute( - kind="config-parquet", + kind="config-level", @@ -96,11 +130 @@ def test_compute( - content={ - "parquet_files": [ - { - "dataset": dataset, - "config": config, - "split": split, - "filename": f"{dataset}-{split}.parquet", - "size": 1000, - } - ] - }, + content=config_parquet_content, @@ -110 +134 @@ def test_compute( - with patch("worker.job_runners.split.first_rows_from_parquet.get_hf_fs") as mock_read: + with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_fs): @@ -112 +136 @@ def test_compute( - "worker.job_runners.split.first_rows_from_parquet.get_hf_parquet_uris", + "libcommon.parquet_utils.get_hf_parquet_uris", @@ -115,6 +138,0 @@ def test_compute( - initial_location = os.getcwd() - os.chdir("tests/job_runners/split") - # TODO: Make localsystem by relative path - fs = LocalFileSystem() - mock_read.return_value = fs - # ^ Mocking file system with local file @@ -156 +174 @@ def test_compute( - assert response["features"][0]["type"]["dtype"] == "int32" + assert response["features"][0]["type"]["dtype"] == "int64" @@ -170 +187,0 @@ def test_compute( - os.chdir(initial_location)
e870cb24073f4c214734b155e6195b5a6d295291
Quentin Lhoest
2023-06-09T17:41:12
Copy parquet when possible (#1321)
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml index 9cc6eb2a..9f9638e0 100644 --- a/.github/workflows/_quality-python.yml +++ b/.github/workflows/_quality-python.yml @@ -52 +52 @@ jobs: - run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d')" + run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed 's/^huggingface-hub @ git.*/huggingface-hub==0.15.1 ; python_full_version == \"3.9.15\" --hash=sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445 --hash=sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081/')" diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml index 1bf2afcf..163a431d 100644 --- a/chart/env/prod.yaml +++ b/chart/env/prod.yaml @@ -91 +91 @@ parquetAndInfo: - blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,bigcode/the-stack-dedup,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member" + blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member" diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py index 708f6257..7d5dfc42 100644 --- a/libs/libcommon/src/libcommon/dataset.py +++ b/libs/libcommon/src/libcommon/dataset.py @@ -63,0 +64,2 @@ def get_dataset_info_for_supported_datasets( + revision: Optional[str] = None, + files_metadata: bool = False, @@ -76,0 +79,6 @@ def get_dataset_info_for_supported_datasets( + revision (`str`, *optional*, defaults to None): + The revision of the dataset repository from which to get the + information. + files_metadata (`bool`, *optional*, defaults to False): + Whether or not to retrieve metadata for files in the repository + (size, LFS metadata, etc). Defaults to `False`. @@ -94 +102,5 @@ def get_dataset_info_for_supported_datasets( - repo_id=dataset, token=hf_token, timeout=hf_timeout_seconds + repo_id=dataset, + token=hf_token, + timeout=hf_timeout_seconds, + revision=revision, + files_metadata=files_metadata, diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py index 8ae12363..eaf8cc95 100644 --- a/libs/libcommon/src/libcommon/exceptions.py +++ b/libs/libcommon/src/libcommon/exceptions.py @@ -87,0 +88 @@ CacheableErrorCode = Literal[ + "DatasetWithTooManyParquetFilesError", @@ -225 +226 @@ class DatasetWithTooManyExternalFilesError(CacheableError): - """Raised when the dataset size (sum of config sizes given by the datasets library) is too big.""" + """Raised when the number of external data files of a dataset is too big.""" @@ -230,0 +232,7 @@ class DatasetWithTooManyExternalFilesError(CacheableError): +class DatasetWithTooManyParquetFilesError(CacheableError): + """Raised when the number of parquet files of a dataset is too big.""" + + def __init__(self, message: str, cause: Optional[BaseException] = None): + super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyParquetFilesError", cause, True) + + diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock index d233d5ff..92064110 100644 --- a/services/worker/poetry.lock +++ b/services/worker/poetry.lock @@ -1 +1 @@ -# This file is automatically @generated by Poetry and should not be changed by hand. +# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand. @@ -1457 +1457 @@ name = "huggingface-hub" -version = "0.15.1" +version = "0.16.0.dev0" @@ -1462,4 +1462,2 @@ python-versions = ">=3.7.0" -files = [ - {file = "huggingface_hub-0.15.1-py3-none-any.whl", hash = "sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445"}, - {file = "huggingface_hub-0.15.1.tar.gz", hash = "sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081"}, -] +files = [] +develop = false @@ -1486,0 +1485,6 @@ typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "t +[package.source] +type = "git" +url = "https://github.com/huggingface/huggingface_hub" +reference = "1055a56b2d2723b55ba4fdf1f3296e04cfd8d6db" +resolved_reference = "1055a56b2d2723b55ba4fdf1f3296e04cfd8d6db" + @@ -2860 +2863,0 @@ files = [ - {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"}, @@ -4415 +4417,0 @@ files = [ - {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"}, @@ -4708,0 +4711,2 @@ files = [ + {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:9c9b14fbb73ec4cb0f209722a1489020fd8614c92ae22589f2309c48cefdf21f"}, + {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_x86_64.whl", hash = "sha256:6a54539bd076746f69ae8bef7282f981674fe4dbf59c3a84c4af86ae6bae9d5c"}, @@ -5596 +5600 @@ python-versions = "3.9.15" -content-hash = "f9c56aa59e9d0c6802245fea8dac8ae28a9b85c8c017d69a7085a9689328ca46" +content-hash = "ad3c5c34e9ea75e4cb4394930bb35c5afdff65fe03c5f51b8cbebbea37f62f1d" diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml index ecdc0f65..9fabc531 100644 --- a/services/worker/pyproject.toml +++ b/services/worker/pyproject.toml @@ -17 +17 @@ gdown = "^4.6.3" -huggingface-hub = "^0.15.1" +huggingface-hub = { git = "https://github.com/huggingface/huggingface_hub", rev = "1055a56b2d2723b55ba4fdf1f3296e04cfd8d6db" } diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py index 3a0c75f6..e6c4b1b2 100644 --- a/services/worker/src/worker/job_runners/config/parquet_and_info.py +++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py @@ -4 +3,0 @@ -import contextlib @@ -17,0 +17 @@ import numpy as np +import pyarrow.parquet as pq @@ -19 +19 @@ import requests -from datasets import DownloadConfig, get_dataset_config_info, load_dataset_builder +from datasets import DownloadConfig, Features, load_dataset_builder @@ -21,0 +22 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError +from datasets.data_files import Url @@ -22,0 +24,2 @@ from datasets.download import StreamingDownloadManager +from datasets.packaged_modules.parquet.parquet import Parquet as ParquetBuilder +from datasets.splits import SplitDict, SplitInfo @@ -29,0 +33 @@ from datasets.utils.py_utils import asdict, map_nested +from fsspec.implementations.http import HTTPFileSystem @@ -32,0 +37 @@ from huggingface_hub._commit_api import ( + CommitOperationCopy, @@ -36 +41 @@ from huggingface_hub.hf_api import DatasetInfo, HfApi, RepoFile -from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFoundError +from huggingface_hub.utils._errors import RepositoryNotFoundError @@ -42,0 +48 @@ from libcommon.constants import ( +from libcommon.dataset import get_dataset_info_for_supported_datasets @@ -48 +53,0 @@ from libcommon.exceptions import ( - DatasetRevisionNotFoundError, @@ -52,0 +58 @@ from libcommon.exceptions import ( + DatasetWithTooManyParquetFilesError, @@ -57,0 +64 @@ from libcommon.exceptions import ( + FileSystemError, @@ -62,0 +70 @@ from libcommon.utils import JobInfo +from tqdm.contrib.concurrent import thread_map @@ -83,0 +92 @@ DATASET_TYPE = "dataset" +MAX_FILES_PER_DIRECTORY = 10_000 # hf hub limitation @@ -94 +103,2 @@ class ParquetFile: - def repo_file(self) -> str: + @property + def path_in_repo(self) -> str: @@ -172,40 +182,10 @@ def raise_if_blocked( -def get_dataset_info_or_raise( - dataset: str, - hf_endpoint: str, - hf_token: Optional[str], - revision: str, -) -> DatasetInfo: - """ - Return the dataset info if possible. - Raise an error if the dataset cannot be accessed (does not exist, gated with extra fields, private) - - Args: - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - hf_endpoint (`str`): - The Hub endpoint (for example: "https://huggingface.co") - hf_token (`str`, `optional`): - An app authentication token with read access to all the datasets. - revision (`str`): - The git revision (e.g. "main" or sha) of the dataset - Returns: - `DatasetInfo`: The dataset info - Raises the following errors: - - [`libcommon.exceptions.DatasetNotFoundError`] - If the repository to download from cannot be found. This may be because it doesn't exist, - or because it is set to `private` and you do not have access. - - [`libcommon.exceptions.DatasetRevisionNotFoundError`] - If the revision does not exist or cannot be accessed using the token. - """ - try: - dataset_info = HfApi(endpoint=hf_endpoint, token=hf_token).dataset_info( - repo_id=dataset, revision=revision, files_metadata=True - ) - except RepositoryNotFoundError as err: - raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err - except RevisionNotFoundError as err: - raise DatasetRevisionNotFoundError("The dataset revision does not exist on the Hub.") from err - if dataset_info.private: - raise DatasetNotFoundError("The dataset does not exist on the Hub.") - return dataset_info +def is_parquet_builder_with_hub_files(builder: DatasetBuilder, hf_endpoint: str) -> bool: + if not isinstance(builder, ParquetBuilder) or not builder.config.data_files: + return False + for split in builder.config.data_files: + for data_file in builder.config.data_files[split]: + if not isinstance(data_file, Url): + return False + if not data_file.startswith(hf_endpoint + "/datasets/" + str(builder.repo_id) + "/"): + return False + return True @@ -243,5 +223 @@ def raise_if_too_big_from_datasets( - dataset: str, - config: str, - hf_endpoint: str, - hf_token: Optional[str], - revision: str, + info: datasets.DatasetInfo, @@ -255,11 +231,2 @@ def raise_if_too_big_from_datasets( - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - config (`str`): - Dataset configuration name - hf_endpoint (`str`): - The Hub endpoint (for example: "https://huggingface.co") - hf_token (`str`, `optional`): - An app authentication token with read access to all the datasets. - revision (`str`): - The git revision (e.g. "main" or sha) of the dataset + info (`datasets.DatasetInfo`): + Dataset info from the datasets library @@ -277,9 +244 @@ def raise_if_too_big_from_datasets( - if datasets.config.HF_ENDPOINT != hf_endpoint: - raise ValueError( - f"Invalid datasets.config.HF_ENDPOINT value: '{datasets.config.HF_ENDPOINT}'. Please set it to:" - f" '{hf_endpoint}'." - ) - dataset_size = 0 - with contextlib.suppress(Exception): - info = get_dataset_config_info(path=dataset, config_name=config, revision=revision, use_auth_token=hf_token) - dataset_size = info.dataset_size if info.dataset_size is not None else 0 + dataset_size = info.dataset_size if info.dataset_size is not None else 0 @@ -295,2 +254 @@ def raise_if_requires_manual_download( - dataset: str, - config: str, + builder: DatasetBuilder, @@ -299 +256,0 @@ def raise_if_requires_manual_download( - revision: str, @@ -305,5 +262,2 @@ def raise_if_requires_manual_download( - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - config (`str`): - Dataset configuration name. + builder (`datasets.builder.DatasetBuilder`): + A dataset builder instance to check. @@ -314,2 +267,0 @@ def raise_if_requires_manual_download( - revision (`str`): - The git revision (e.g. "main" or sha) of the dataset. @@ -331,6 +282,0 @@ def raise_if_requires_manual_download( - builder = load_dataset_builder( - dataset, - name=config, - revision=revision, - use_auth_token=hf_token, - ) @@ -344 +290 @@ def raise_if_requires_manual_download( - raise DatasetManualDownloadError(f"{dataset=} requires manual download.", cause=err) from err + raise DatasetManualDownloadError(f"dataset={builder.repo_id} requires manual download.", cause=err) from err @@ -348,2 +294,2 @@ def raise_if_not_supported( - dataset: str, - config: str, + dataset_info: DatasetInfo, + builder: DatasetBuilder, @@ -352,3 +297,0 @@ def raise_if_not_supported( - revision: str, - supported_datasets: List[str], - blocked_datasets: List[str], @@ -355,0 +299 @@ def raise_if_not_supported( + max_external_data_files: int, @@ -364,5 +308,4 @@ def raise_if_not_supported( - dataset (`str`): - A namespace (user or an organization) and a repo name separated - by a `/`. - config (`str`): - Dataset configuration name + dataset_info (`DatasetInfo`): + The dataset info + builder (`datasets.builder.DatasetBuilder`): + A dataset builder instance to check. @@ -375,5 +317,0 @@ def raise_if_not_supported( - supported_datasets (`List[str]`): - The list of supported datasets, saving the blocked datasets. If empty, all datasets are supported - (saving the blocked datasets). - blocked_datasets (`List[str]`): - The list of blocked datasets. If empty, no dataset is blocked. @@ -382,0 +321,3 @@ def raise_if_not_supported( + max_external_data_files (`int`): + The maximum number of external data files (i.e. not hosted on HF). + If the dataset is under the limit (which means that the files can be fetched), it will be allowed. @@ -387,6 +327,0 @@ def raise_if_not_supported( - - [`libcommon.exceptions.DatasetNotFoundError`]: - if the dataset does not exist, or if the token does not give the sufficient access to the dataset, - - ['requests.exceptions.HTTPError'](https://requests.readthedocs.io/en/latest/api/#requests.HTTPError) - any other error when asking access - - [`libcommon.exceptions.DatasetInBlockListError`] - If the dataset is in the list of blocked datasets. @@ -402,0 +338,14 @@ def raise_if_not_supported( + - [`libcommon.exceptions.DatasetWithTooManyExternalFilesError`] + If the dataset has too many external files to be converted to parquet + - [`libcommon.exceptions.DatasetWithTooBigExternalFilesError`] + If the dataset is too big external files be converted to parquet + - [`libcommon.exceptions.UnsupportedExternalFilesError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`libcommon.exceptions.ExternalFilesSizeRequestHTTPError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`libcommon.exceptions.ExternalFilesSizeRequestConnectionError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`libcommon.exceptions.ExternalFilesSizeRequestTimeoutError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet + - [`libcommon.exceptions.ExternalFilesSizeRequestError`] + If we failed to get the external files sizes to make sure we can convert the dataset to parquet @@ -404 +353 @@ def raise_if_not_supported( - If the datasets.config.HF_ENDPOINT is not set to the expected value + If the datasets.config.HF_ENDPOINT is not set to the expected value @@ -406,4 +355,5 @@ def raise_if_not_supported( - raise_if_blocked(dataset=dataset, blocked_datasets=blocked_datasets) - dataset_info = get_dataset_info_or_raise( - dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, revision=revision - ) + if datasets.config.HF_ENDPOINT != hf_endpoint: + raise ValueError( + f"Invalid datasets.config.HF_ENDPOINT value: '{datasets.config.HF_ENDPOINT}'. Please set it to:" + f" '{hf_endpoint}'." + ) @@ -411,2 +361 @@ def raise_if_not_supported( - dataset=dataset, - config=config, + builder=builder, @@ -415 +363,0 @@ def raise_if_not_supported( - revision=revision, @@ -417,6 +365,5 @@ def raise_if_not_supported( - if dataset in supported_datasets: - return - raise_if_too_big_from_datasets( - dataset=dataset, - config=config, - hf_endpoint=hf_endpoint, + raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size) + raise_if_too_big_from_external_data_files( + builder=builder, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, @@ -424 +371,3 @@ def raise_if_not_supported( - revision=revision, + ) + raise_if_too_big_from_datasets( + builder.info, @@ -427 +375,0 @@ def raise_if_not_supported( - raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size) @@ -628,0 +577,101 @@ def get_writer_batch_size(ds_config_info: datasets.info.DatasetInfo) -> Optional +def copy_parquet_files(builder: DatasetBuilder) -> List[CommitOperationCopy]: + """Copy parquet files by copying the git LFS pointer files""" + data_files = builder.config.data_files + if not data_files: + raise EmptyDatasetError("Empty parquet data_files") + parquet_operations = [] + total_num_parquet_files = sum(len(data_files[split]) for split in data_files) + if total_num_parquet_files >= MAX_FILES_PER_DIRECTORY: + raise DatasetWithTooManyParquetFilesError( + f"The dataset has {total_num_parquet_files} parquet files and can't be linked in the parquet directory " + f"because it exceeds the maximum number of files per directory ({MAX_FILES_PER_DIRECTORY})." + ) + for split in data_files: + num_shards = len(data_files[split]) + for shard_idx, data_file in enumerate(data_files[split]): + src_revision, src_path_in_repo = data_file.split("/datasets/" + builder.repo_id + "/resolve/", 1)[1].split( + "/", 1 + ) + + # for forward compatibility with https://github.com/huggingface/datasets/pull/5331 + parquet_name = str(builder.dataset_name) if hasattr(builder, "dataset_name") else builder.name + + if num_shards > 1: + path_in_repo = ( + f"{builder.config.name}/{parquet_name}-{split}-{shard_idx:05d}-of-{num_shards:05d}.parquet" + ) + else: + path_in_repo = f"{builder.config.name}/{parquet_name}-{split}.parquet" + + parquet_operations.append( + CommitOperationCopy( + src_path_in_repo=src_path_in_repo, path_in_repo=path_in_repo, src_revision=src_revision + ) + ) + return parquet_operations + + +def get_parquet_file_and_size(url: str, fs: HTTPFileSystem, hf_token: Optional[str]) -> Tuple[pq.ParquetFile, int]: + headers = get_authentication_headers_for_url(url, use_auth_token=hf_token) + f = fs.open(url, headers=headers) + return pq.ParquetFile(f), f.size + + +def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None: + """Fill the builder DatasetInfo from the copied parquet files""" + data_files = builder.config.data_files + if not data_files: + raise EmptyDatasetError("Empty parquet data_files") + fs = HTTPFileSystem() + if not builder.info.splits or not builder.info.download_size: + builder.info.splits = SplitDict() + builder.info.dataset_size = 0 + for split in data_files: + try: + parquet_files_and_sizes: List[Tuple[pq.ParquetFile, int]] = thread_map( + partial(get_parquet_file_and_size, fs=fs, hf_token=hf_token), + data_files[split], + unit="pq", + disable=True, + ) + parquet_files, sizes = zip(*parquet_files_and_sizes) + except Exception as e: + raise FileSystemError(f"Could not read the parquet files: {e}") from e + if parquet_files: + first_pf = parquet_files[0] + if builder.info.features is None: + builder.info.features = Features.from_arrow_schema(first_pf.schema_arrow) + first_row_group = first_pf.read_row_group(0) + compression_ratio = first_row_group.nbytes / first_row_group.num_rows + num_examples = sum(parquet_file.metadata.num_rows for parquet_file in parquet_files) + approx_num_bytes = int(compression_ratio * num_examples) + builder.info.splits.add(SplitInfo(split, num_bytes=approx_num_bytes, num_examples=num_examples)) + builder.info.download_size += sum(sizes) + builder.info.dataset_size += approx_num_bytes + + +def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]: + """Download and prepare the dataset as parquet files and fills the builder info""" + # prepare the parquet files locally + writer_batch_size = get_writer_batch_size(builder.info) + if writer_batch_size is not None and ( + builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size + ): + builder._writer_batch_size = writer_batch_size + builder.download_and_prepare( + file_format="parquet" + ) # the parquet files are stored in the cache dir and it fills the info + local_parquet_files = [ + ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=builder.config.name) + for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet") + ] + + # send the files to the target revision + parquet_operations: List[CommitOperationAdd] = [ + CommitOperationAdd(path_in_repo=parquet_file.path_in_repo, path_or_fileobj=parquet_file.local_file) + for parquet_file in local_parquet_files + ] + logging.debug(f"{parquet_operations=}") + return parquet_operations + + @@ -725,10 +774 @@ def compute_config_parquet_and_info_response( - raise_if_not_supported( - dataset=dataset, - config=config, - hf_endpoint=hf_endpoint, - hf_token=hf_token, - revision=source_revision, - supported_datasets=supported_datasets, - blocked_datasets=blocked_datasets, - max_dataset_size=max_dataset_size, - ) + raise_if_blocked(dataset=dataset, blocked_datasets=blocked_datasets) @@ -757,2 +797,6 @@ def compute_config_parquet_and_info_response( - # prepare the parquet files locally - local_parquet_files: List[ParquetFile] = [] + # check if the repo exists and get the list of refs + try: + refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE) + except RepositoryNotFoundError as err: + raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err + @@ -770,17 +813,0 @@ def compute_config_parquet_and_info_response( - writer_batch_size = get_writer_batch_size(builder.info) - if writer_batch_size is not None and ( - builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size - ): - builder._writer_batch_size = writer_batch_size - raise_if_too_big_from_external_data_files( - builder=builder, - max_dataset_size=max_dataset_size, - max_external_data_files=max_external_data_files, - hf_token=hf_token, - ) - builder.download_and_prepare(file_format="parquet") # the parquet files are stored in the cache dir - dataset_info = asdict(builder.info) - local_parquet_files.extend( - ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=config) - for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet") - ) @@ -788 +815,20 @@ def compute_config_parquet_and_info_response( - # create the target revision if it does not exist yet (clone from initial commit to avoid cloning all repo's files) + if is_parquet_builder_with_hub_files(builder, hf_endpoint=hf_endpoint): + parquet_operations = copy_parquet_files(builder) + fill_builder_info(builder, hf_token=hf_token) + else: + dataset_info = get_dataset_info_for_supported_datasets( + dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, revision=source_revision, files_metadata=True + ) + if dataset not in supported_datasets: + raise_if_not_supported( + dataset_info=dataset_info, + builder=builder, + hf_endpoint=hf_endpoint, + hf_token=hf_token, + max_dataset_size=max_dataset_size, + max_external_data_files=max_external_data_files, + ) + parquet_operations = convert_to_parquet(builder) + + # create the target revision if we managed to get the parquet files and it does not exist yet + # (clone from initial commit to avoid cloning all repo's files) @@ -790 +835,0 @@ def compute_config_parquet_and_info_response( - refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE) @@ -794 +839 @@ def compute_config_parquet_and_info_response( - repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit + repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit, exist_ok=True @@ -797 +842 @@ def compute_config_parquet_and_info_response( - raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err + raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err @@ -798,0 +844 @@ def compute_config_parquet_and_info_response( + builder_info = asdict(builder.info) @@ -803,4 +848,0 @@ def compute_config_parquet_and_info_response( - # - get parquet files for current config - config_files_to_add: Dict[str, str] = { - parquet_file.repo_file(): parquet_file.local_file for parquet_file in local_parquet_files - } @@ -818,0 +861 @@ def compute_config_parquet_and_info_response( + config_files_to_add = [operation.path_in_repo for operation in parquet_operations] @@ -823,7 +865,0 @@ def compute_config_parquet_and_info_response( - # send the files to the target revision - add_operations: List[CommitOperation] = [ - CommitOperationAdd(path_in_repo=file, path_or_fileobj=local_file) - for file, local_file in config_files_to_add.items() - ] - logging.debug(f"{add_operations=}") - @@ -834 +870 @@ def compute_config_parquet_and_info_response( - operations=delete_operations + add_operations, + operations=delete_operations + parquet_operations, @@ -862 +898 @@ def compute_config_parquet_and_info_response( - dataset_info=dataset_info, + dataset_info=builder_info, diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py index 0c7287d2..110dba75 100644 --- a/services/worker/tests/fixtures/hub.py +++ b/services/worker/tests/fixtures/hub.py @@ -5,0 +6 @@ +import csv @@ -225,0 +227,17 @@ def hub_public_big(datasets: Mapping[str, Dataset]) -> Iterator[str]: [email protected](scope="session") +def hub_public_big_no_info(datasets: Mapping[str, Dataset]) -> Iterator[str]: + repo_id = create_hub_dataset_repo(prefix="big-no-info", dataset=datasets["big"]) + hf_api.delete_file( + "README.md", repo_id=repo_id, repo_type="dataset", commit_message="Delete README.md", token=CI_USER_TOKEN + ) + yield repo_id + delete_hub_dataset_repo(repo_id=repo_id) + + [email protected](scope="session") +def hub_public_big_csv(big_csv_path: str) -> Iterator[str]: + repo_id = create_hub_dataset_repo(prefix="big-csv", file_paths=[big_csv_path]) + yield repo_id + delete_hub_dataset_repo(repo_id=repo_id) + + @@ -342,0 +361,30 @@ def create_dataset_info_response_for_csv(dataset: str, config: str) -> Any: +def create_dataset_info_response_for_big_parquet() -> Any: + return { + "description": "", + "citation": "", + "homepage": "", + "license": "", + "features": BIG_cols, + "splits": { + "train": {"name": "train", "num_bytes": 5653946, "num_examples": len(BIG_rows), "dataset_name": None} + }, + "download_size": BIG_PARQUET_FILE, + "dataset_size": 5653946, + } + + +def create_dataset_info_response_for_big_parquet_no_info() -> Any: + return { + "description": "", + "citation": "", + "homepage": "", + "license": "", + "features": BIG_cols, + "splits": { + "train": {"name": "train", "num_bytes": 12345, "num_examples": len(BIG_rows), "dataset_name": None} + }, + "download_size": BIG_PARQUET_FILE, + "dataset_size": 12345, + } + + @@ -350,7 +398 @@ def create_dataset_info_response_for_audio() -> Any: - "splits": {"train": {"name": "train", "num_bytes": 59, "num_examples": 1, "dataset_name": "parquet"}}, - "download_checksums": { - "SOME_KEY": { - "num_bytes": AUDIO_PARQUET_SIZE, - "checksum": None, - } - }, + "splits": {"train": {"name": "train", "num_bytes": 54, "num_examples": 1, "dataset_name": None}}, @@ -358,2 +400 @@ def create_dataset_info_response_for_audio() -> Any: - "dataset_size": 59, - "size_in_bytes": 1443, + "dataset_size": 54, @@ -363 +404,3 @@ def create_dataset_info_response_for_audio() -> Any: -def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "audio"]) -> Any: +def create_parquet_and_info_response( + dataset: str, data_type: Literal["csv", "audio", "big_parquet", "big_parquet_no_info"] +) -> Any: @@ -367 +410 @@ def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "au - size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE + size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE if data_type == "audio" else BIG_PARQUET_FILE @@ -371,0 +415,4 @@ def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "au + if data_type == "audio" + else create_dataset_info_response_for_big_parquet() + if data_type == "big_parquet" + else create_dataset_info_response_for_big_parquet_no_info() @@ -391,0 +439 @@ AUDIO_PARQUET_SIZE = 1_384 +BIG_PARQUET_FILE = 38_896 @@ -488 +536 @@ BIG_cols = { - "col": [{"_type": "Value", "dtype": "string"}], + "col": {"_type": "Value", "dtype": "string"}, @@ -491 +539,13 @@ BIG_cols = { -BIG_rows = ["a" * 1_234 for _ in range(4_567)] +BIG_rows = [{"col": "a" * 1_234} for _ in range(4_567)] + + [email protected](scope="session") +def big_csv_path(tmp_path_factory: pytest.TempPathFactory) -> str: + path = str(tmp_path_factory.mktemp("data") / "big_dataset.csv") + with open(path, "w", newline="") as f: + writer = csv.DictWriter(f, fieldnames=list(BIG_cols)) + writer.writeheader() + for row in BIG_rows: + writer.writerow(row) + return path + @@ -526,0 +587,2 @@ def hub_datasets( + hub_public_big_no_info: str, + hub_public_big_csv: str, @@ -618,0 +681,18 @@ def hub_datasets( + "parquet_and_info_response": create_parquet_and_info_response( + dataset=hub_public_big, data_type="big_parquet" + ), + }, + "big-no-info": { + "name": hub_public_big_no_info, + "config_names_response": create_config_names_response(hub_public_big_no_info), + "splits_response": create_splits_response(hub_public_big_no_info), + "first_rows_response": create_first_rows_response(hub_public_big_no_info, BIG_cols, BIG_rows), + "parquet_and_info_response": create_parquet_and_info_response( + dataset=hub_public_big_no_info, data_type="big_parquet_no_info" + ), + }, + "big-csv": { + "name": hub_public_big_csv, + "config_names_response": create_config_names_response(hub_public_big_csv), + "splits_response": create_splits_response(hub_public_big_csv), + "first_rows_response": create_first_rows_response(hub_public_big_csv, BIG_cols, BIG_rows), diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py index 5ed9cd60..66944344 100644 --- a/services/worker/tests/job_runners/config/test_parquet_and_info.py +++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py @@ -15 +15 @@ import requests -from datasets import Audio, Features, Image, Value +from datasets import Audio, Features, Image, Value, load_dataset_builder @@ -16,0 +17 @@ from huggingface_hub.hf_api import HfApi +from libcommon.dataset import get_dataset_info_for_supported_datasets @@ -34 +34,0 @@ from worker.job_runners.config.parquet_and_info import ( - get_dataset_info_or_raise, @@ -38 +37,0 @@ from worker.job_runners.config.parquet_and_info import ( - raise_if_not_supported, @@ -122,4 +120,0 @@ def assert_content_is_equal(content: Any, expected: Any) -> None: - assert len(content_value["download_checksums"]) == 1, content - content_checksum = list(content_value["download_checksums"].values())[0] - expected_checksum = list(expected_value["download_checksums"].values())[0] - assert content_checksum == expected_checksum, content @@ -235,0 +231 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_ + builder = load_dataset_builder(hub_public_manual_download) @@ -238,2 +234 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_ - hub_public_manual_download, - "default", + builder=builder, @@ -242 +236,0 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_ - revision="main", @@ -257 +251 @@ def test_raise_if_too_big_from_hub( - dataset_info = get_dataset_info_or_raise( + dataset_info = get_dataset_info_for_supported_datasets( @@ -261,0 +256 @@ def test_raise_if_too_big_from_hub( + files_metadata=True, @@ -285 +280 @@ def test_raise_if_too_big_from_datasets( - config = hub_datasets[name]["config_names_response"]["config_names"][0]["config"] + builder = load_dataset_builder(dataset) @@ -289,5 +284 @@ def test_raise_if_too_big_from_datasets( - dataset=dataset, - config=config, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - revision="main", + info=builder.info, @@ -298,5 +289 @@ def test_raise_if_too_big_from_datasets( - dataset=dataset, - config=config, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - revision="main", + info=builder.info, @@ -373,9 +360 @@ def test_raise_if_too_many_external_files( [email protected]( - "in_list,raises", - [ - (True, False), - (False, True), - ], -) -def test_raise_if_not_supported( - hub_datasets: HubDatasets, +def test_supported_if_big_parquet( @@ -383,2 +362,2 @@ def test_raise_if_not_supported( - in_list: bool, - raises: bool, + get_job_runner: GetJobRunner, + hub_datasets: HubDatasets, @@ -385,0 +365,3 @@ def test_raise_if_not_supported( + # Not in the list of supported datasets and bigger than the maximum size + # but still supported since it's made of parquet files + # dataset = hub_public_big @@ -388,23 +370,13 @@ def test_raise_if_not_supported( - if raises: - with pytest.raises(DatasetTooBigFromDatasetsError): - raise_if_not_supported( - dataset=dataset, - config=config, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - revision="main", - max_dataset_size=app_config.parquet_and_info.max_dataset_size, - supported_datasets=[dataset] if in_list else ["another_dataset"], - blocked_datasets=[], - ) - else: - raise_if_not_supported( - dataset=dataset, - config=config, - hf_endpoint=app_config.common.hf_endpoint, - hf_token=app_config.common.hf_token, - revision="main", - max_dataset_size=app_config.parquet_and_info.max_dataset_size, - supported_datasets=[dataset] if in_list else ["another_dataset"], - blocked_datasets=[], - ) + upsert_response( + kind="dataset-config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets["big"]["config_names_response"], + ) + job_runner = get_job_runner(dataset, config, app_config) + response = job_runner.compute() + assert response + content = response.content + assert content + assert len(content["parquet_files"]) == 1 + assert_content_is_equal(content, hub_datasets["big"]["parquet_and_info_response"]) @@ -413 +385 @@ def test_raise_if_not_supported( -def test_not_supported_if_big( +def test_not_supported_if_big_non_parquet( @@ -419,3 +391,3 @@ def test_not_supported_if_big( - # dataset = hub_public_big - dataset = hub_datasets["big"]["name"] - config = hub_datasets["big"]["config_names_response"]["config_names"][0]["config"] + # dataset = hub_public_big_csv + dataset = hub_datasets["big-csv"]["name"] + config = hub_datasets["big-csv"]["config_names_response"]["config_names"][0]["config"] @@ -426 +398 @@ def test_not_supported_if_big( - content=hub_datasets["big"]["config_names_response"], + content=hub_datasets["big-csv"]["config_names_response"], @@ -431 +403 @@ def test_not_supported_if_big( - assert e.typename == "DatasetTooBigFromDatasetsError" + assert e.typename == "DatasetTooBigFromHubError" @@ -532,0 +505,6 @@ def test_compute_splits_response_simple_csv_error( + upsert_response( + "dataset-config-names", + dataset=dataset, + http_status=HTTPStatus.OK, + content=hub_datasets[name]["config_names_response"], + )