hash
stringlengths 40
40
| authorName
stringclasses 42
values | authorEmail
stringclasses 41
values | date
timestamp[ms]date 2021-07-26 09:52:55
2025-07-18 10:19:56
| subject
stringlengths 11
116
| diff
stringlengths 0
987k
|
---|---|---|---|---|---|
6215b9d5fa0155ad23fca5151e69d2428c43ad53
|
Sylvain Lesage
| 2023-07-05T15:05:59 |
feat: 🎸 add "difficulty" field to JobDocument (#1489)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 24ea7c61..b8e32aef 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.15.0
+version: 1.15.1
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 61bdd330..ab27b0c7 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -294,0 +295,2 @@ workers:
+ workerDifficultyMax: ""
+ workerDifficultyMin: ""
@@ -309,0 +312,2 @@ workers:
+ workerDifficultyMax: ""
+ workerDifficultyMin: ""
diff --git a/chart/env/staging.yaml b/chart/env/staging.yaml
index 5027dfd9..64848c3e 100644
--- a/chart/env/staging.yaml
+++ b/chart/env/staging.yaml
@@ -235,0 +236,2 @@ workers:
+ workerDifficultyMax: ""
+ workerDifficultyMin: ""
@@ -249,0 +252,2 @@ workers:
+ workerDifficultyMax: ""
+ workerDifficultyMin: ""
diff --git a/chart/templates/worker/_container.tpl b/chart/templates/worker/_container.tpl
index f9b86817..2bfc9ab0 100644
--- a/chart/templates/worker/_container.tpl
+++ b/chart/templates/worker/_container.tpl
@@ -18,0 +19,4 @@
+ - name: WORKER_DIFFICULTY_MAX
+ value: {{ .workerValues.workerDifficultyMax | quote }}
+ - name: WORKER_DIFFICULTY_MIN
+ value: {{ .workerValues.workerDifficultyMin | quote }}
diff --git a/chart/values.yaml b/chart/values.yaml
index e8f01d0b..e9736b2e 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -457,0 +458,4 @@ workers:
+ # max difficulty of the jobs that this worker will process
+ workerDifficultyMax: ""
+ # min difficulty of the jobs that this worker will process
+ workerDifficultyMin: ""
diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py
index aed47435..8811b02f 100644
--- a/jobs/cache_maintenance/tests/test_collect_metrics.py
+++ b/jobs/cache_maintenance/tests/test_collect_metrics.py
@@ -28 +28,6 @@ def test_collect_metrics() -> None:
- job_type=processing_step.job_type, dataset="dataset", revision="revision", config="config", split="split"
+ job_type=processing_step.job_type,
+ dataset="dataset",
+ revision="revision",
+ config="config",
+ split="split",
+ difficulty=50,
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index eaf51193..7430144b 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -52,0 +53,3 @@ from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_con
+from mongodb_migration.migrations._20230705160600_queue_job_add_difficulty import (
+ MigrationQueueAddDifficultyToJob,
+)
@@ -243,0 +247 @@ class MigrationsCollector:
+ MigrationQueueAddDifficultyToJob(version="20230705160600", description="add 'difficulty' field to jobs"),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230705160600_queue_job_add_difficulty.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230705160600_queue_job_add_difficulty.py
new file mode 100644
index 00000000..f02ed2fe
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230705160600_queue_job_add_difficulty.py
@@ -0,0 +1,44 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.config import ProcessingGraphConfig
+from libcommon.constants import (
+ DEFAULT_DIFFICULTY,
+ QUEUE_COLLECTION_JOBS,
+ QUEUE_MONGOENGINE_ALIAS,
+)
+from libcommon.queue import JobDocument
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import Migration
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationQueueAddDifficultyToJob(Migration):
+ def up(self) -> None:
+ logging.info("If missing, add the difficulty with a value that depends on the job type, else 50")
+ db = get_db(QUEUE_MONGOENGINE_ALIAS)
+ processing_graph_config = ProcessingGraphConfig()
+ for job_type, spec in processing_graph_config.specification.items():
+ difficulty = spec.get("difficulty", DEFAULT_DIFFICULTY)
+ db[QUEUE_COLLECTION_JOBS].update_many(
+ {"type": job_type, "difficulty": {"$exists": False}},
+ {"$set": {"difficulty": difficulty}},
+ )
+ db[QUEUE_COLLECTION_JOBS].update_many(
+ {"difficulty": {"$exists": False}},
+ {"$set": {"difficulty": DEFAULT_DIFFICULTY}},
+ )
+
+ def down(self) -> None:
+ logging.info("Remove the difficulty field from all the jobs")
+ db = get_db(QUEUE_MONGOENGINE_ALIAS)
+ db[QUEUE_COLLECTION_JOBS].update_many({}, {"$unset": {"difficulty": ""}})
+
+ def validate(self) -> None:
+ logging.info("Ensure that a random selection of jobs have the 'difficulty' field set")
+
+ check_documents(DocCls=JobDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/tests/migrations/__init__.py b/jobs/mongodb_migration/tests/migrations/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/jobs/mongodb_migration/tests/migrations/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py
index 7fced24f..52bfd45c 100644
--- a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py
+++ b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py
@@ -27,0 +28 @@ def test_queue_delete_indexes_with_force(mongo_host: str) -> None:
+ difficulty=50,
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230705160600_queue_job_add_difficulty.py b/jobs/mongodb_migration/tests/migrations/test_20230705160600_queue_job_add_difficulty.py
new file mode 100644
index 00000000..247f9f61
--- /dev/null
+++ b/jobs/mongodb_migration/tests/migrations/test_20230705160600_queue_job_add_difficulty.py
@@ -0,0 +1,55 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS
+from libcommon.resources import MongoResource
+from mongoengine.connection import get_db
+
+from mongodb_migration.migrations._20230705160600_queue_job_add_difficulty import (
+ MigrationQueueAddDifficultyToJob,
+)
+
+
+def test_queue_add_difficulty_to_jobs(mongo_host: str) -> None:
+ with MongoResource(database="test_queue_add_difficulty_to_jobs", host=mongo_host, mongoengine_alias="queue"):
+ db = get_db(QUEUE_MONGOENGINE_ALIAS)
+ db[QUEUE_COLLECTION_JOBS].insert_many(
+ [
+ {
+ "type": "test",
+ "dataset": "test",
+ "revision": "123",
+ "unicity_id": "test",
+ "namespace": "test",
+ "created_at": "2022-01-01T00:00:00.000000Z",
+ },
+ {
+ "type": "dataset-is-valid",
+ "dataset": "test",
+ "revision": "123",
+ "unicity_id": "test",
+ "namespace": "test",
+ "created_at": "2022-01-01T00:00:00.000000Z",
+ },
+ ]
+ )
+
+ migration = MigrationQueueAddDifficultyToJob(
+ version="20230705160600",
+ description="add difficulty field to jobs",
+ )
+ migration.up()
+
+ results = list(db[QUEUE_COLLECTION_JOBS].find({"type": "test"}))
+ assert len(results) == 1
+ assert results[0]["difficulty"] == 50
+ results = list(db[QUEUE_COLLECTION_JOBS].find({"type": "dataset-is-valid"}))
+ assert len(results) == 1
+ assert results[0]["difficulty"] == 20
+
+ migration.down()
+ results = list(db[QUEUE_COLLECTION_JOBS].find({"dataset": "test"}))
+ assert len(results) == 2
+ assert all("difficulty" not in result for result in results)
+
+ db[QUEUE_COLLECTION_JOBS].drop()
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index aa8b7027..ba56516c 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -127,0 +128 @@ def test_queue_delete_ttl_index(mongo_host: str) -> None:
+ difficulty=50,
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 59006344..472fa14f 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -210,0 +211 @@ class ProcessingGraphConfig:
+ "difficulty": 50,
@@ -216,0 +218 @@ class ProcessingGraphConfig:
+ "difficulty": 60,
@@ -222,0 +225 @@ class ProcessingGraphConfig:
+ "difficulty": 70,
@@ -227,0 +231 @@ class ProcessingGraphConfig:
+ "difficulty": 70,
@@ -233,0 +238 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -239,0 +245 @@ class ProcessingGraphConfig:
+ "difficulty": 50,
@@ -245,0 +252 @@ class ProcessingGraphConfig:
+ "difficulty": 40,
@@ -250,0 +258 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -255,0 +264 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -260,0 +270 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -266,0 +277 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -272,0 +284 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -277,0 +290 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -286,0 +300 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -296,0 +311 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -301,0 +317 @@ class ProcessingGraphConfig:
+ "difficulty": 40,
@@ -306,0 +323 @@ class ProcessingGraphConfig:
+ "difficulty": 70,
@@ -311,0 +329 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -320,0 +339 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -325,0 +345 @@ class ProcessingGraphConfig:
+ "difficulty": 20,
@@ -334,0 +355 @@ class ProcessingGraphConfig:
+ "difficulty": 70,
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
index 5bb181a5..19dc97e7 100644
--- a/libs/libcommon/src/libcommon/constants.py
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -17,0 +18 @@ QUEUE_TTL_SECONDS = 600 # 10 minutes
+DEFAULT_DIFFICULTY = 50
diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py
index 1a5669b0..24be4e57 100644
--- a/libs/libcommon/src/libcommon/orchestrator.py
+++ b/libs/libcommon/src/libcommon/orchestrator.py
@@ -296,0 +297 @@ class AfterJobPlan(Plan):
+ "difficulty": next_processing_step.difficulty,
@@ -565,0 +567 @@ class DatasetBackfillPlan(Plan):
+ "difficulty": artifact_state.processing_step.difficulty,
diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py
index f6f1e636..0a04a70c 100644
--- a/libs/libcommon/src/libcommon/processing_graph.py
+++ b/libs/libcommon/src/libcommon/processing_graph.py
@@ -22 +22,5 @@ import networkx as nx
-from libcommon.constants import DEFAULT_INPUT_TYPE, DEFAULT_JOB_RUNNER_VERSION
+from libcommon.constants import (
+ DEFAULT_DIFFICULTY,
+ DEFAULT_INPUT_TYPE,
+ DEFAULT_JOB_RUNNER_VERSION,
+)
@@ -56,0 +61 @@ class ProcessingStepSpecification(TypedDict, total=False):
+ difficulty: int
@@ -82,0 +88 @@ class ProcessingStep:
+ difficulty: int
@@ -100,0 +107 @@ class ProcessingStep:
+ difficulty=self.difficulty,
@@ -194,0 +202 @@ class ProcessingGraph:
+ difficulty=specification.get("difficulty", DEFAULT_DIFFICULTY),
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 57031369..a9af875b 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -20 +20 @@ from mongoengine.errors import DoesNotExist, NotUniqueError
-from mongoengine.fields import DateTimeField, EnumField, StringField
+from mongoengine.fields import DateTimeField, EnumField, IntField, StringField
@@ -71,0 +72 @@ class JobDict(TypedDict):
+ difficulty: int
@@ -110,0 +112,7 @@ class NoWaitingJobError(Exception):
+class JobQueryFilters(TypedDict, total=False):
+ type__nin: List[str]
+ type__in: List[str]
+ difficulty__gte: int
+ difficulty__lte: int
+
+
@@ -130,0 +139 @@ class JobDocument(Document):
+ difficulty (`int`): The difficulty of the job: 0=easy, 100=hard as a convention.
@@ -170,0 +180 @@ class JobDocument(Document):
+ difficulty = IntField(required=True)
@@ -186,0 +197 @@ class JobDocument(Document):
+ "difficulty": self.difficulty,
@@ -206,0 +218 @@ class JobDocument(Document):
+ "difficulty": self.difficulty,
@@ -227,0 +240 @@ class JobDocument(Document):
+ "difficulty": self.difficulty,
@@ -336,2 +349,2 @@ class Queue:
- - a job can be in the queue only once (unicity_id) in the "started" or "waiting" state
- - a job can be in the queue multiple times in the other states (success, error, cancelled)
+ - a job can be in the queue only once (unicity_id) in the "started" state
+ - a job can be in the queue multiple times in the other states
@@ -338,0 +352 @@ class Queue:
+ - a job has a difficulty (from 0: easy to 100: hard, as a convention)
@@ -347,0 +362 @@ class Queue:
+ difficulty: int,
@@ -360,0 +376 @@ class Queue:
+ difficulty (`int`): The difficulty of the job.
@@ -377,0 +394 @@ class Queue:
+ difficulty=difficulty,
@@ -408,0 +426 @@ class Queue:
+ difficulty=job_info["difficulty"],
@@ -437,0 +456,2 @@ class Queue:
+ difficulty_min: Optional[int] = None,
+ difficulty_max: Optional[int] = None,
@@ -449,0 +470,2 @@ class Queue:
+ difficulty_min: if not None, only jobs with a difficulty greater or equal to this value are considered.
+ difficulty_max: if not None, only jobs with a difficulty lower or equal to this value are considered.
@@ -462 +484 @@ class Queue:
- filters = {}
+ filters: JobQueryFilters = {}
@@ -466,0 +489,4 @@ class Queue:
+ if difficulty_min:
+ filters["difficulty__gte"] = difficulty_min
+ if difficulty_max:
+ filters["difficulty__lte"] = difficulty_max
@@ -523 +549,5 @@ class Queue:
- self, job_types_blocked: Optional[list[str]] = None, job_types_only: Optional[list[str]] = None
+ self,
+ difficulty_min: Optional[int] = None,
+ difficulty_max: Optional[int] = None,
+ job_types_blocked: Optional[list[str]] = None,
+ job_types_only: Optional[list[str]] = None,
@@ -533,0 +564,2 @@ class Queue:
+ difficulty_min: if not None, only jobs with a difficulty greater or equal to this value are considered.
+ difficulty_max: if not None, only jobs with a difficulty lower or equal to this value are considered.
@@ -545 +577,5 @@ class Queue:
- priority=priority, job_types_blocked=job_types_blocked, job_types_only=job_types_only
+ priority=priority,
+ job_types_blocked=job_types_blocked,
+ job_types_only=job_types_only,
+ difficulty_min=difficulty_min,
+ difficulty_max=difficulty_max,
@@ -607 +643,5 @@ class Queue:
- self, job_types_blocked: Optional[list[str]] = None, job_types_only: Optional[list[str]] = None
+ self,
+ difficulty_min: Optional[int] = None,
+ difficulty_max: Optional[int] = None,
+ job_types_blocked: Optional[list[str]] = None,
+ job_types_only: Optional[list[str]] = None,
@@ -614,0 +655,2 @@ class Queue:
+ difficulty_min: if not None, only jobs with a difficulty greater or equal to this value are considered.
+ difficulty_max: if not None, only jobs with a difficulty lower or equal to this value are considered.
@@ -629 +671,4 @@ class Queue:
- job_types_blocked=job_types_blocked, job_types_only=job_types_only
+ job_types_blocked=job_types_blocked,
+ job_types_only=job_types_only,
+ difficulty_min=difficulty_min,
+ difficulty_max=difficulty_max,
@@ -767 +812,6 @@ class Queue:
- job_type=job.type, dataset=job.dataset, revision=job.revision, config=job.config, split=job.split
+ job_type=job.type,
+ dataset=job.dataset,
+ revision=job.revision,
+ config=job.config,
+ split=job.split,
+ difficulty=job.difficulty,
diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py
index ac79f831..1fe129e7 100644
--- a/libs/libcommon/src/libcommon/utils.py
+++ b/libs/libcommon/src/libcommon/utils.py
@@ -38,0 +39 @@ class JobInfo(TypedDict):
+ difficulty: int
@@ -49,0 +51 @@ class FlatJobInfo(TypedDict):
+ difficulty: int
diff --git a/libs/libcommon/tests/test_backfill.py b/libs/libcommon/tests/test_backfill.py
index d69c231a..ddb58f60 100644
--- a/libs/libcommon/tests/test_backfill.py
+++ b/libs/libcommon/tests/test_backfill.py
@@ -35,0 +36 @@ from .utils import (
+ DIFFICULTY,
@@ -815 +816,3 @@ def test_delete_jobs(
- job = queue.add_job(job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority)
+ job = queue.add_job(
+ job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority, difficulty=DIFFICULTY
+ )
diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py
index 32a687c4..8563cad5 100644
--- a/libs/libcommon/tests/test_orchestrator.py
+++ b/libs/libcommon/tests/test_orchestrator.py
@@ -27,0 +28 @@ from .utils import (
+ DIFFICULTY,
@@ -153,0 +155 @@ def test_finish_job(
+ difficulty=DIFFICULTY,
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 792905ee..eaf064b3 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -37,0 +38 @@ def test_add_job() -> None:
+ test_difficulty = 50
@@ -41 +42 @@ def test_add_job() -> None:
- job1 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ job1 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty)
@@ -43 +44 @@ def test_add_job() -> None:
- job2 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ job2 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty)
@@ -58 +59 @@ def test_add_job() -> None:
- job3 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ job3 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty)
@@ -92,0 +94 @@ def test_cancel_jobs_by_job_id(
+ test_difficulty = 50
@@ -98 +100 @@ def test_cancel_jobs_by_job_id(
- job = queue.add_job(job_type=test_type, dataset=job_id, revision="test_revision")
+ job = queue.add_job(job_type=test_type, dataset=job_id, revision="test_revision", difficulty=test_difficulty)
@@ -126,0 +129 @@ def test_priority_logic_creation_order() -> None:
+ test_difficulty = 50
@@ -128,2 +131,16 @@ def test_priority_logic_creation_order() -> None:
- queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1")
- queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2")
+ queue.add_job(
+ job_type=test_type,
+ dataset="dataset1",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
+ )
+ queue.add_job(
+ job_type=test_type,
+ dataset="dataset1",
+ revision=test_revision,
+ config="config",
+ split="split2",
+ difficulty=test_difficulty,
+ )
@@ -138,0 +156 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None:
+ test_difficulty = 50
@@ -140,3 +158,24 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None:
- queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1")
- queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2")
- queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1")
+ queue.add_job(
+ job_type=test_type,
+ dataset="dataset1",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
+ )
+ queue.add_job(
+ job_type=test_type,
+ dataset="dataset1",
+ revision=test_revision,
+ config="config",
+ split="split2",
+ difficulty=test_difficulty,
+ )
+ queue.add_job(
+ job_type=test_type,
+ dataset="dataset2",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
+ )
@@ -153,0 +193 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None:
+ test_difficulty = 50
@@ -155,3 +194,0 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None:
- queue.add_job(job_type=test_type, dataset="org1/dataset1", revision=test_revision, config="config", split="split1")
- queue.add_job(job_type=test_type, dataset="org1/dataset2", revision=test_revision, config="config", split="split1")
- queue.add_job(job_type=test_type, dataset="org2/dataset2", revision=test_revision, config="config", split="split1")
@@ -159 +196,30 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None:
- job_type=test_type, dataset="no_org_dataset3", revision=test_revision, config="config", split="split1"
+ job_type=test_type,
+ dataset="org1/dataset1",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
+ )
+ queue.add_job(
+ job_type=test_type,
+ dataset="org1/dataset2",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
+ )
+ queue.add_job(
+ job_type=test_type,
+ dataset="org2/dataset2",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
+ )
+ queue.add_job(
+ job_type=test_type,
+ dataset="no_org_dataset3",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
@@ -172,0 +239 @@ def test_priority_logic_priority_order() -> None:
+ test_difficulty = 50
@@ -179,0 +247 @@ def test_priority_logic_priority_order() -> None:
+ difficulty=test_difficulty,
@@ -187,0 +256 @@ def test_priority_logic_priority_order() -> None:
+ difficulty=test_difficulty,
@@ -197 +266 @@ def test_priority_logic_priority_order() -> None:
- "job_type,job_types_blocked,job_types_only,should_raise",
+ "job_types_blocked,job_types_only,should_raise",
@@ -199,10 +268,10 @@ def test_priority_logic_priority_order() -> None:
- ("test_type", None, None, False),
- ("test_type", None, ["test_type"], False),
- ("test_type", ["other_type"], None, False),
- ("test_type", ["other_type"], ["test_type"], False),
- ("test_type", None, ["other_type"], True),
- ("test_type", ["test_type"], None, True),
- ("test_type", ["test_type"], ["test_type"], True),
- ("test_type", ["other_type", "test_type"], None, True),
- ("test_type", ["other_type"], ["other_type"], True),
- ("test_type", ["other_type", "test_type"], ["other_type", "test_type"], True),
+ (None, None, False),
+ (None, ["test_type"], False),
+ (["other_type"], None, False),
+ (["other_type"], ["test_type"], False),
+ (None, ["other_type"], True),
+ (["test_type"], None, True),
+ (["test_type"], ["test_type"], True),
+ (["other_type", "test_type"], None, True),
+ (["other_type"], ["other_type"], True),
+ (["other_type", "test_type"], ["other_type", "test_type"], True),
@@ -212 +281 @@ def test_job_types_only(
- job_type: str, job_types_blocked: Optional[list[str]], job_types_only: Optional[list[str]], should_raise: bool
+ job_types_blocked: Optional[list[str]], job_types_only: Optional[list[str]], should_raise: bool
@@ -213,0 +283 @@ def test_job_types_only(
+ job_type = "test_type"
@@ -215,0 +286 @@ def test_job_types_only(
+ test_difficulty = 50
@@ -217 +288,8 @@ def test_job_types_only(
- queue.add_job(job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None)
+ queue.add_job(
+ job_type=job_type,
+ dataset=test_dataset,
+ revision=test_revision,
+ config=None,
+ split=None,
+ difficulty=test_difficulty,
+ )
@@ -228,0 +307,39 @@ def test_job_types_only(
[email protected](
+ "difficulty_min,difficulty_max,should_raise",
+ [
+ (None, None, False),
+ (None, 60, False),
+ (40, None, False),
+ (40, 60, False),
+ (50, 50, False),
+ (None, 40, True),
+ (60, None, True),
+ (60, 60, True),
+ (40, 40, True),
+ ],
+)
+def test_difficulty(difficulty_min: Optional[int], difficulty_max: Optional[int], should_raise: bool) -> None:
+ job_type = "test_type"
+ test_dataset = "test_dataset"
+ test_revision = "test_revision"
+ test_difficulty = 50
+ queue = Queue()
+ queue.add_job(
+ job_type=job_type,
+ dataset=test_dataset,
+ revision=test_revision,
+ config=None,
+ split=None,
+ difficulty=test_difficulty,
+ )
+ assert queue.is_job_in_process(
+ job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None
+ )
+ if should_raise:
+ with pytest.raises(EmptyQueueError):
+ queue.start_job(difficulty_max=difficulty_max, difficulty_min=difficulty_min)
+ else:
+ job_info = queue.start_job(difficulty_max=difficulty_max, difficulty_min=difficulty_min)
+ assert job_info["params"]["dataset"] == test_dataset
+
+
@@ -233,0 +351 @@ def test_count_by_status() -> None:
+ test_difficulty = 50
@@ -242 +360 @@ def test_count_by_status() -> None:
- queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty)
@@ -247 +365 @@ def test_count_by_status() -> None:
- queue.add_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision)
+ queue.add_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision, difficulty=test_difficulty)
@@ -255,0 +374 @@ def test_get_dataset_pending_jobs_for_type() -> None:
+ test_difficulty = 50
@@ -266 +385,8 @@ def test_get_dataset_pending_jobs_for_type() -> None:
- queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
+ queue.add_job(
+ job_type=job_type,
+ dataset=dataset,
+ revision=test_revision,
+ config=config,
+ split=None,
+ difficulty=test_difficulty,
+ )
@@ -272 +398,8 @@ def test_get_dataset_pending_jobs_for_type() -> None:
- queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
+ queue.add_job(
+ job_type=job_type,
+ dataset=dataset,
+ revision=test_revision,
+ config=config,
+ split=None,
+ difficulty=test_difficulty,
+ )
@@ -277 +410,8 @@ def test_get_dataset_pending_jobs_for_type() -> None:
- queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
+ queue.add_job(
+ job_type=job_type,
+ dataset=dataset,
+ revision=test_revision,
+ config=config,
+ split=None,
+ difficulty=test_difficulty,
+ )
@@ -287,0 +428 @@ def test_queue_heartbeat() -> None:
+ test_difficulty = 50
@@ -289 +430,8 @@ def test_queue_heartbeat() -> None:
- job = queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1")
+ job = queue.add_job(
+ job_type=job_type,
+ dataset="dataset1",
+ revision="revision",
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
+ )
@@ -300,0 +449 @@ def test_queue_get_zombies() -> None:
+ test_difficulty = 50
@@ -304 +453,6 @@ def test_queue_get_zombies() -> None:
- job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1"
+ job_type=job_type,
+ dataset="dataset1",
+ revision="revision",
+ config="config",
+ split="split1",
+ difficulty=test_difficulty,
@@ -307 +461,8 @@ def test_queue_get_zombies() -> None:
- queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split2")
+ queue.add_job(
+ job_type=job_type,
+ dataset="dataset1",
+ revision="revision",
+ config="config",
+ split="split2",
+ difficulty=test_difficulty,
+ )
diff --git a/libs/libcommon/tests/utils.py b/libs/libcommon/tests/utils.py
index e4addd22..3ee60d34 100644
--- a/libs/libcommon/tests/utils.py
+++ b/libs/libcommon/tests/utils.py
@@ -34,0 +35 @@ JOB_TYPE = "job_type"
+DIFFICULTY = 50
@@ -363,0 +365 @@ def artifact_id_to_job_info(artifact_id: str) -> JobInfo:
+ difficulty=DIFFICULTY,
diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py
index 33855962..df11c6f8 100644
--- a/services/admin/src/admin/app.py
+++ b/services/admin/src/admin/app.py
@@ -131,0 +132 @@ def create_app() -> Starlette:
+ difficulty=processing_step.difficulty,
diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py
index 3f37f937..f54814b1 100644
--- a/services/admin/src/admin/routes/force_refresh.py
+++ b/services/admin/src/admin/routes/force_refresh.py
@@ -28,0 +29 @@ def create_force_refresh_endpoint(
+ difficulty: int,
@@ -64,0 +66 @@ def create_force_refresh_endpoint(
+ difficulty=difficulty,
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index eb5f4886..7ce3d60a 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -164 +164 @@ def test_get_cache_entry_from_steps() -> None:
- queue.add_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config)
+ queue.add_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config, difficulty=50)
diff --git a/services/worker/README.md b/services/worker/README.md
index 9f794786..778a5ffb 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -13,0 +14,2 @@ Set environment variables to configure the worker.
+- `WORKER_DIFFICULTY_MAX`: the maximum difficulty of the jobs to process. Defaults to None.
+- `WORKER_DIFFICULTY_MIN`: the minimum difficulty of the jobs to process. Defaults to None.
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index d39a3c03..25ba41fa 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -18,0 +19,2 @@ WORKER_CONTENT_MAX_BYTES = 10_000_000
+WORKER_DIFFICULTY_MAX = None
+WORKER_DIFFICULTY_MIN = None
@@ -38,0 +41,2 @@ class WorkerConfig:
+ difficulty_max: Optional[int] = WORKER_DIFFICULTY_MAX
+ difficulty_min: Optional[int] = WORKER_DIFFICULTY_MIN
@@ -58,0 +63,2 @@ class WorkerConfig:
+ difficulty_max=env.int(name="DIFFICULTY_MAX", default=WORKER_DIFFICULTY_MAX),
+ difficulty_min=env.int(name="DIFFICULTY_MIN", default=WORKER_DIFFICULTY_MIN),
diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py
index c0303b21..59e3f51b 100644
--- a/services/worker/src/worker/loop.py
+++ b/services/worker/src/worker/loop.py
@@ -129,0 +130,2 @@ class Loop:
+ difficulty_min=self.app_config.worker.difficulty_min,
+ difficulty_max=self.app_config.worker.difficulty_max,
diff --git a/services/worker/tests/job_runners/config/test_config_job_runner.py b/services/worker/tests/job_runners/config/test_config_job_runner.py
index 01a5ce8e..77a9ebec 100644
--- a/services/worker/tests/job_runners/config/test_config_job_runner.py
+++ b/services/worker/tests/job_runners/config/test_config_job_runner.py
@@ -41,0 +42 @@ def test_failed_creation(test_processing_step: ProcessingStep, app_config: AppCo
+ "difficulty": 50,
@@ -62,0 +64 @@ def test_success_creation(test_processing_step: ProcessingStep, app_config: AppC
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py
index 49be6df2..2b9b1983 100644
--- a/services/worker/tests/job_runners/config/test_info.py
+++ b/services/worker/tests/job_runners/config/test_info.py
@@ -165,0 +166 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py
index bde0b947..8a06c6fb 100644
--- a/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py
+++ b/services/worker/tests/job_runners/config/test_opt_in_out_urls_count.py
@@ -59,0 +60 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py
index 1189453f..223f8f69 100644
--- a/services/worker/tests/job_runners/config/test_parquet.py
+++ b/services/worker/tests/job_runners/config/test_parquet.py
@@ -59,0 +60 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index a78ba301..2054d393 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -106,0 +107 @@ def get_job_runner(
+ "difficulty": 50,
@@ -693,0 +695 @@ def get_dataset_config_names_job_runner(
+ "difficulty": 50,
@@ -722,0 +725 @@ def launch_job_runner(job_runner_args: JobRunnerArgs) -> CompleteJobResult:
+ difficulty=50,
@@ -728,0 +732 @@ def launch_job_runner(job_runner_args: JobRunnerArgs) -> CompleteJobResult:
+ difficulty=50,
@@ -761,0 +766 @@ def test_concurrency(
+ difficulty=50,
diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py
index 571a8f8e..9e0079d6 100644
--- a/services/worker/tests/job_runners/config/test_parquet_metadata.py
+++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py
@@ -80,0 +81 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py
index df89b2e2..33692159 100644
--- a/services/worker/tests/job_runners/config/test_size.py
+++ b/services/worker/tests/job_runners/config/test_size.py
@@ -58,0 +59 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/config/test_split_names_from_info.py b/services/worker/tests/job_runners/config/test_split_names_from_info.py
index ee4b7ccb..5a06b413 100644
--- a/services/worker/tests/job_runners/config/test_split_names_from_info.py
+++ b/services/worker/tests/job_runners/config/test_split_names_from_info.py
@@ -53,0 +54 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
index 2e5f3215..c978f96a 100644
--- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
@@ -57,0 +58 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py
index dc6aafe2..3f2fcfd2 100644
--- a/services/worker/tests/job_runners/dataset/test_config_names.py
+++ b/services/worker/tests/job_runners/dataset/test_config_names.py
@@ -52,0 +53 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py
index 49f7fa06..4fe427c9 100644
--- a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py
+++ b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py
@@ -42,0 +43 @@ def test_failed_creation(test_processing_step: ProcessingStep, app_config: AppCo
+ "difficulty": 50,
@@ -63,0 +65 @@ def test_success_creation(test_processing_step: ProcessingStep, app_config: AppC
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py
index 0d6b8158..385b5fbc 100644
--- a/services/worker/tests/job_runners/dataset/test_info.py
+++ b/services/worker/tests/job_runners/dataset/test_info.py
@@ -137,0 +138 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_is_valid.py b/services/worker/tests/job_runners/dataset/test_is_valid.py
index a6c64001..fc9a9d66 100644
--- a/services/worker/tests/job_runners/dataset/test_is_valid.py
+++ b/services/worker/tests/job_runners/dataset/test_is_valid.py
@@ -106,0 +107 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py
index 37f959de..71d37579 100644
--- a/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py
+++ b/services/worker/tests/job_runners/dataset/test_opt_in_out_urls_count.py
@@ -56,0 +57 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py
index 28c04d3f..8d53cad6 100644
--- a/services/worker/tests/job_runners/dataset/test_parquet.py
+++ b/services/worker/tests/job_runners/dataset/test_parquet.py
@@ -58,0 +59 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_size.py b/services/worker/tests/job_runners/dataset/test_size.py
index 76326d1d..3d2e38d7 100644
--- a/services/worker/tests/job_runners/dataset/test_size.py
+++ b/services/worker/tests/job_runners/dataset/test_size.py
@@ -57,0 +58 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/dataset/test_split_names.py b/services/worker/tests/job_runners/dataset/test_split_names.py
index 1afcf32b..6832f6e9 100644
--- a/services/worker/tests/job_runners/dataset/test_split_names.py
+++ b/services/worker/tests/job_runners/dataset/test_split_names.py
@@ -48,0 +49 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/split/test_duckdb_index.py b/services/worker/tests/job_runners/split/test_duckdb_index.py
index a8fe40cf..ead4aed4 100644
--- a/services/worker/tests/job_runners/split/test_duckdb_index.py
+++ b/services/worker/tests/job_runners/split/test_duckdb_index.py
@@ -72,0 +73 @@ def get_job_runner(
+ "difficulty": 50,
@@ -114,0 +116 @@ def get_parquet_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py
index 3dcb37c1..980eb479 100644
--- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py
+++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py
@@ -67,0 +68 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
index 9f255de8..9fdb0bad 100644
--- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
+++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
@@ -64,0 +65 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/split/test_image_url_columns.py b/services/worker/tests/job_runners/split/test_image_url_columns.py
index 68fb6f01..d49d4bd7 100644
--- a/services/worker/tests/job_runners/split/test_image_url_columns.py
+++ b/services/worker/tests/job_runners/split/test_image_url_columns.py
@@ -59,0 +60 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py
index e3898d3c..399b4d8a 100644
--- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py
+++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_count.py
@@ -61,0 +62 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
index 2de50d1d..72678d0e 100644
--- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
+++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
@@ -77,0 +78 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/split/test_split_job_runner.py b/services/worker/tests/job_runners/split/test_split_job_runner.py
index 80d88ba4..2afde623 100644
--- a/services/worker/tests/job_runners/split/test_split_job_runner.py
+++ b/services/worker/tests/job_runners/split/test_split_job_runner.py
@@ -42,0 +43 @@ def test_failed_creation(test_processing_step: ProcessingStep, app_config: AppCo
+ "difficulty": 50,
@@ -63,0 +65 @@ def test_success_creation(test_processing_step: ProcessingStep, app_config: AppC
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/test__job_runner_with_cache.py b/services/worker/tests/job_runners/test__job_runner_with_cache.py
index 11aa9fd1..2ff25f6c 100644
--- a/services/worker/tests/job_runners/test__job_runner_with_cache.py
+++ b/services/worker/tests/job_runners/test__job_runner_with_cache.py
@@ -68,0 +69 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
index 2667da40..afb25020 100644
--- a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
+++ b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
@@ -72,0 +73 @@ def get_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index 3e30ca26..165c8038 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -44,0 +45 @@ def get_job_info(prefix: str = "base") -> JobInfo:
+ difficulty=50,
@@ -128,0 +130 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I
+ difficulty=job_info["difficulty"],
@@ -161,0 +164 @@ def set_long_running_job_in_queue(
+ difficulty=job_info["difficulty"],
@@ -191,0 +195 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato
+ difficulty=job_info["difficulty"],
diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py
index a6e54257..a01f4249 100644
--- a/services/worker/tests/test_job_manager.py
+++ b/services/worker/tests/test_job_manager.py
@@ -73,0 +74 @@ def test_check_type(
+ difficulty=50,
@@ -95,0 +97 @@ def test_check_type(
+ difficulty=50,
@@ -134,0 +137 @@ def test_backfill(priority: Priority, app_config: AppConfig) -> None:
+ difficulty=50,
@@ -204,0 +208 @@ def test_job_runner_set_crashed(
+ difficulty=50,
@@ -262,0 +267 @@ def test_raise_if_parallel_response_exists(
+ difficulty=50,
@@ -293,0 +299 @@ def test_doesnotexist(app_config: AppConfig) -> None:
+ difficulty=50,
diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py
index e10bc8c0..84765d48 100644
--- a/services/worker/tests/test_job_runner_factory.py
+++ b/services/worker/tests/test_job_runner_factory.py
@@ -63,0 +64 @@ def test_create_job_runner(
+ "difficulty": 50,
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 2a766b92..3097e354 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -72 +72,3 @@ def test_process_next_job(
- loop.queue.add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split)
+ loop.queue.add_job(
+ job_type=job_type, dataset=dataset, revision=revision, config=config, split=split, difficulty=50
+ )
|
|
abf206283f7242efebc6d5fc601e75a2b758a5b0
|
Quentin Lhoest
| 2023-07-05T14:51:01 |
Block config-parquet-metadata job for light workers (#1490)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index d50a3bcd..61bdd330 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -311 +311 @@ workers:
- workerJobTypesOnly: "config-parquet,config-parquet-metadata,dataset-parquet,config-info,dataset-info,config-split-names-from-info,config-size,dataset-size,dataset-split-names,dataset-is-valid,split-image-url-columns,split-opt-in-out-urls-count,config-opt-in-out-urls-count,dataset-opt-in-out-urls-count"
+ workerJobTypesOnly: "config-parquet,dataset-parquet,config-info,dataset-info,config-split-names-from-info,config-size,dataset-size,dataset-split-names,dataset-is-valid,split-image-url-columns,split-opt-in-out-urls-count,config-opt-in-out-urls-count,dataset-opt-in-out-urls-count"
|
|
ad0047367c77c8035e2bcc108dc51190f6562bd7
|
Quentin Lhoest
| 2023-07-05T13:00:47 |
More logging for /rows (#1485)
|
diff --git a/libs/libapi/src/libapi/utils.py b/libs/libapi/src/libapi/utils.py
index 2f2480de..65e7f50f 100644
--- a/libs/libapi/src/libapi/utils.py
+++ b/libs/libapi/src/libapi/utils.py
@@ -3,0 +4 @@
+import logging
@@ -100,0 +102 @@ def try_backfill_dataset(
+ logging.info(f"Set orchestrator revision for dataset={dataset}, revision={revision}")
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py
index 351cb65e..b3bfcbbd 100644
--- a/libs/libcommon/src/libcommon/parquet_utils.py
+++ b/libs/libcommon/src/libcommon/parquet_utils.py
@@ -399,0 +400,4 @@ class RowsIndex:
+ logging.info(
+ f"Create ParquetIndexWithoutMetadata for dataset={self.dataset}, config={self.config},"
+ f" split={self.split}"
+ )
@@ -412,0 +417,4 @@ class RowsIndex:
+ logging.info(
+ f"Create ParquetIndexWithMetadata for dataset={self.dataset}, config={self.config},"
+ f" split={self.split}"
+ )
@@ -439,0 +448,4 @@ class RowsIndex:
+ logging.info(
+ f"Query {type(self.parquet_index).__name__} for dataset={self.dataset}, config={self.config},"
+ f" split={self.split}, offset={offset}, length={length}"
+ )
diff --git a/services/rows/src/rows/routes/rows.py b/services/rows/src/rows/routes/rows.py
index 2529260f..efa0a503 100644
--- a/services/rows/src/rows/routes/rows.py
+++ b/services/rows/src/rows/routes/rows.py
@@ -313,2 +313,2 @@ def create_rows_endpoint(
- with StepProfiler(method="rows_endpoint", step="get row groups index"):
- try:
+ try:
+ with StepProfiler(method="rows_endpoint", step="get row groups index"):
@@ -321,5 +321,6 @@ def create_rows_endpoint(
- except CachedArtifactError:
- config_parquet_processing_steps = processing_graph.get_config_parquet_processing_steps()
- config_parquet_metadata_processing_steps = (
- processing_graph.get_config_parquet_metadata_processing_steps()
- )
+ except CachedArtifactError:
+ config_parquet_processing_steps = processing_graph.get_config_parquet_processing_steps()
+ config_parquet_metadata_processing_steps = (
+ processing_graph.get_config_parquet_metadata_processing_steps()
+ )
+ with StepProfiler(method="rows_endpoint", step="try backfill dataset"):
|
|
dee3002c03c7aa78720513bd960568065f5644a7
|
Polina Kazakova
| 2023-07-05T12:45:38 |
Delete `/parquet-and-dataset-info` endpoint (#1488)
|
diff --git a/e2e/tests/test_11_api.py b/e2e/tests/test_11_api.py
index 48e60d68..8dd3e822 100644
--- a/e2e/tests/test_11_api.py
+++ b/e2e/tests/test_11_api.py
@@ -55 +54,0 @@ def test_auth_e2e(
- ("/parquet-and-dataset-info", "config"),
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 28a62fc7..a7a95078 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -71,3 +70,0 @@ class EndpointConfig:
- "/parquet-and-dataset-info": {
- "config": ["config-parquet-and-info"],
- },
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index 37f8f820..eb5f4886 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -51,6 +50,0 @@ def test_endpoints_definition() -> None:
- parquet_and_info = definition["/parquet-and-dataset-info"]
- assert parquet_and_info is not None
- assert sorted(list(parquet_and_info)) == ["config"]
- assert parquet_and_info["config"] is not None
- assert len(parquet_and_info["config"]) == 1 # Only has one processing step
-
@@ -91 +85 @@ def test_endpoints_definition() -> None:
- # assert old endpoints don't exist
+ # assert old deleted endpoints don't exist
@@ -93,0 +88,2 @@ def test_endpoints_definition() -> None:
+ with raises(KeyError):
+ _ = definition["/parquet-and-dataset-info"]
diff --git a/services/worker/README.md b/services/worker/README.md
index b6299326..9f794786 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -3 +3 @@
-> Worker that pre-computes and caches the response to /splits, /first-rows or /parquet-and-dataset-info.
+> Workers that pre-compute and cache the response to /splits, /first-rows, /parquet, /info and /size.
@@ -7 +7 @@
-Use environment variables to configure the worker. The prefix of each environment variable gives its scope.
+Use environment variables to configure the workers. The prefix of each environment variable gives its scope.
@@ -55 +55 @@ If the Hub is not https://huggingface.co (i.e., if you set the `COMMON_HF_ENDPOI
-Set environment variables to configure the first rows worker (`FIRST_ROWS_` prefix):
+Set environment variables to configure the `first-rows` worker (`FIRST_ROWS_` prefix):
@@ -65 +65 @@ Also, set the assets-related configuration for the first-rows worker. See [../..
-### Parquet and dataset info worker
+### Parquet and info worker
@@ -67 +67 @@ Also, set the assets-related configuration for the first-rows worker. See [../..
-Set environment variables to configure the parquet worker (`PARQUET_AND_INFO_` prefix):
+Set environment variables to configure the `parquet-and-info` worker (`PARQUET_AND_INFO_` prefix):
@@ -80 +80 @@ Set environment variables to configure the parquet worker (`PARQUET_AND_INFO_` p
-Set environment variables to configure the duckdb index worker (`DUCKDB_INDEX_` prefix):
+Set environment variables to configure the `duckdb-index` worker (`DUCKDB_INDEX_` prefix):
@@ -92 +92 @@ Set environment variables to configure the duckdb index worker (`DUCKDB_INDEX_`
-The splits worker does not need any additional configuration.
+The `splits` worker does not need any additional configuration.
|
|
4a8a64c9b19ae649214341f79c2c8cae3d637793
|
Sylvain Lesage
| 2023-07-05T11:23:32 |
feat: 🎸 query operation $in is faster then $nin (#1486)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 63eef88a..d50a3bcd 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -310,2 +310,2 @@ workers:
- workerJobTypesBlocked: "dataset-config-names,config-split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan,split-duckdb-index"
- workerJobTypesOnly: ""
+ workerJobTypesBlocked: ""
+ workerJobTypesOnly: "config-parquet,config-parquet-metadata,dataset-parquet,config-info,dataset-info,config-split-names-from-info,config-size,dataset-size,dataset-split-names,dataset-is-valid,split-image-url-columns,split-opt-in-out-urls-count,config-opt-in-out-urls-count,dataset-opt-in-out-urls-count"
|
|
1528ec93f4b61ec30cc6b7cc950050cfa608edfd
|
Quentin Lhoest
| 2023-07-04T20:31:29 |
minor (#1480)
|
diff --git a/libs/libcommon/src/libcommon/viewer_utils/asset.py b/libs/libcommon/src/libcommon/viewer_utils/asset.py
index 28a8322c..f98eff12 100644
--- a/libs/libcommon/src/libcommon/viewer_utils/asset.py
+++ b/libs/libcommon/src/libcommon/viewer_utils/asset.py
@@ -50 +50,4 @@ def update_last_modified_date_of_rows_in_assets_dir(
- (row_dirs_path / str(row_idx) / DATASETS_SERVER_MDATE_FILENAME).unlink()
+ try:
+ (row_dirs_path / str(row_idx) / DATASETS_SERVER_MDATE_FILENAME).unlink()
+ except FileNotFoundError:
+ pass
|
|
bd372054f7bf14f2430309f964d372cbbe8f6f50
|
Sylvain Lesage
| 2023-07-04T17:27:25 |
feat: 🎸 increase RAM for /rows service (#1484)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index f4f22785..63eef88a 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -287 +287 @@ rows:
- memory: "512Mi"
+ memory: "2Gi"
@@ -290 +290 @@ rows:
- memory: "4Gi"
+ memory: "8Gi"
|
|
2619bd836a0a008485fcb62a07bfde325b0cdd9e
|
Sylvain Lesage
| 2023-07-04T16:26:15 |
Move the /rows endpoint to its own service (#1479)
|
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml
index 0126945c..c4272b3e 100644
--- a/.github/workflows/_e2e_tests.yml
+++ b/.github/workflows/_e2e_tests.yml
@@ -34,0 +35,2 @@ jobs:
+ ROWS_UVICORN_NUM_WORKERS: "2"
+ ROWS_UVICORN_PORT: "8082"
@@ -70,0 +73,2 @@ jobs:
+ ROWS_UVICORN_NUM_WORKERS: "2"
+ ROWS_UVICORN_PORT: "8082"
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml
index d1394601..770cf6e9 100644
--- a/.github/workflows/cd.yml
+++ b/.github/workflows/cd.yml
@@ -31,0 +32,2 @@ jobs:
+ - directory: services
+ project: rows
@@ -110,0 +113,2 @@ jobs:
+ rows:
+ tag: sha-${{ steps.vars.outputs.sha_short }}
diff --git a/.github/workflows/s-api.yml b/.github/workflows/s-api.yml
index 63c440cc..070b4c71 100644
--- a/.github/workflows/s-api.yml
+++ b/.github/workflows/s-api.yml
@@ -10,0 +11 @@ on:
+ - "libs/libapi/**"
@@ -18,0 +20 @@ on:
+ - "libs/libapi/**"
diff --git a/.github/workflows/s-rows.yml b/.github/workflows/s-rows.yml
new file mode 100644
index 00000000..d49ac293
--- /dev/null
+++ b/.github/workflows/s-rows.yml
@@ -0,0 +1,35 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+name: services/rows
+on:
+ workflow_dispatch:
+ push:
+ branches:
+ - main
+ paths:
+ - "libs/libapi/**"
+ - "libs/libcommon/**"
+ - "services/rows/**"
+ - ".github/workflows/s-rows.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
+ pull_request:
+ paths:
+ - "libs/libapi/**"
+ - "libs/libcommon/**"
+ - "services/rows/**"
+ - ".github/workflows/s-rows.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
+jobs:
+ quality:
+ uses: ./.github/workflows/_quality-python.yml
+ with:
+ working-directory: services/rows
+ unit-tests:
+ uses: ./.github/workflows/_unit-tests-python.yml
+ with:
+ working-directory: services/rows
diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace
index 597c0f2f..b1be9dde 100644
--- a/.vscode/monorepo.code-workspace
+++ b/.vscode/monorepo.code-workspace
@@ -34,0 +35,4 @@
+ {
+ "name": "services/rows",
+ "path": "../services/rows"
+ },
diff --git a/Makefile b/Makefile
index bc174bb0..71571e13 100644
--- a/Makefile
+++ b/Makefile
@@ -4,0 +5 @@ export PORT_API := 8180
+export PORT_ROWS := 8182
@@ -23 +24 @@ start:
- MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up
+ MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up
@@ -27 +28 @@ stop:
- MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down
+ MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down
@@ -31 +32 @@ dev-start:
- MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up
+ MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) up
@@ -35 +36 @@ dev-stop:
- MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down
+ MONGO_PORT=${MONGO_PORT} ADMIN_UVICORN_PORT=${PORT_ADMIN} API_UVICORN_PORT=${PORT_API} ROWS_UVICORN_PORT=${PORT_ROWS} PORT_REVERSE_PROXY=${PORT_REVERSE_PROXY} DOCKER_COMPOSE=${DOCKER_COMPOSE} $(MAKE) down
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 732dcf41..24ea7c61 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.14.1
+version: 1.15.0
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 953aeb10..f4f22785 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -13,0 +14 @@ global:
+ rows: 31371
@@ -44,0 +46,5 @@ images:
+ rows:
+ registry: huggingface
+ useGlobalRegistry: false
+ repository: datasets-server-services-rows
+ tag: sha-fb3399a
@@ -250,0 +257,21 @@ api:
+
+ nodeSelector:
+ role-datasets-server: "true"
+ replicas: 4
+ service:
+ type: NodePort
+ resources:
+ requests:
+ cpu: 1
+ memory: "512Mi"
+ limits:
+ cpu: 4
+ memory: "4Gi"
+
+rows:
+ # the timeout in seconds for the requests to the Hugging Face Hub.
+ hfTimeoutSeconds: "1.5"
+ # Number of uvicorn workers for running the application
+ # (2 x $num_cores) + 1
+ # https://docs.gunicorn.org/en/stable/design.html#how-many-workers
+ uvicornNumWorkers: "9"
diff --git a/chart/env/staging.yaml b/chart/env/staging.yaml
index 3014a6ef..5027dfd9 100644
--- a/chart/env/staging.yaml
+++ b/chart/env/staging.yaml
@@ -42,0 +43,5 @@ images:
+ rows:
+ registry: huggingface
+ useGlobalRegistry: false
+ repository: datasets-server-services-rows
+ tag: sha-fb3399a
@@ -213,0 +219,14 @@ api:
+rows:
+ uvicornNumWorkers: "1"
+
+ replicas: 1
+ service:
+ type: NodePort
+ resources:
+ requests:
+ cpu: 100m
+ memory: "512Mi"
+ limits:
+ cpu: 1
+ memory: "4Gi"
+
diff --git a/chart/nginx-templates/default.conf.template b/chart/nginx-templates/default.conf.template
index ac2edf9c..7efde8a0 100644
--- a/chart/nginx-templates/default.conf.template
+++ b/chart/nginx-templates/default.conf.template
@@ -49,0 +50,9 @@ server {
+ location /rows {
+ proxy_pass ${URL_ROWS}/rows;
+ proxy_set_header Host $proxy_host;
+ proxy_set_header X-Real-IP $remote_addr;
+ proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
+ proxy_set_header X-Forwarded-Proto $scheme;
+ proxy_http_version 1.1;
+ }
+
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index ac9370e6..cb73f80a 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -48,0 +49,4 @@ Docker image management
+{{- define "services.rows.image" -}}
+{{ include "hf.common.images.image" (dict "imageRoot" .Values.images.services.rows "global" .Values.global.huggingface) }}
+{{- end -}}
+
@@ -100,0 +105,5 @@ app.kubernetes.io/component: "{{ include "name" . }}-api"
+{{- define "labels.rows" -}}
+{{ include "hf.labels.commons" . }}
+app.kubernetes.io/component: "{{ include "name" . }}-rows"
+{{- end -}}
+
@@ -213,0 +223,8 @@ See https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#a-a
+{{/*
+The URL to access the rows service from another container
+See https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#a-aaaa-records
+*/}}
+{{- define "rows.url" -}}
+{{- printf "http://%s-rows.%s.svc.cluster.local:80" ( include "name" . ) ( .Release.Namespace ) }}
+{{- end }}
+
diff --git a/chart/templates/reverse-proxy/_container.tpl b/chart/templates/reverse-proxy/_container.tpl
index 5f1a0952..6ae0be54 100644
--- a/chart/templates/reverse-proxy/_container.tpl
+++ b/chart/templates/reverse-proxy/_container.tpl
@@ -20,0 +21,2 @@
+ - name: URL_ROWS
+ value: {{ include "rows.url" . | quote }}
diff --git a/chart/templates/services/api/_container.tpl b/chart/templates/services/api/_container.tpl
index 923532fc..4c83f00a 100644
--- a/chart/templates/services/api/_container.tpl
+++ b/chart/templates/services/api/_container.tpl
@@ -9 +8,0 @@
- {{ include "envCachedAssets" . | nindent 2 }}
@@ -11 +9,0 @@
- {{ include "envParquetMetadata" . | nindent 2 }}
@@ -49,3 +46,0 @@
- volumeMounts:
- {{ include "volumeMountCachedAssetsRW" . | nindent 2 }}
- {{ include "volumeMountParquetMetadataRO" . | nindent 2 }}
diff --git a/chart/templates/services/api/deployment.yaml b/chart/templates/services/api/deployment.yaml
index a34cef80..e77b45e3 100644
--- a/chart/templates/services/api/deployment.yaml
+++ b/chart/templates/services/api/deployment.yaml
@@ -26,3 +25,0 @@ spec:
- initContainers:
- {{ include "initContainerCachedAssets" . | nindent 8 }}
- {{ include "initContainerParquetMetadata" . | nindent 8 }}
diff --git a/chart/templates/services/rows/_container.tpl b/chart/templates/services/rows/_container.tpl
new file mode 100644
index 00000000..2cb00e86
--- /dev/null
+++ b/chart/templates/services/rows/_container.tpl
@@ -0,0 +1,66 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+{{- define "containerRows" -}}
+- name: "{{ include "name" . }}-rows"
+ image: {{ include "services.rows.image" . }}
+ imagePullPolicy: {{ .Values.images.pullPolicy }}
+ env:
+ {{ include "envCachedAssets" . | nindent 2 }}
+ {{ include "envCache" . | nindent 2 }}
+ {{ include "envParquetMetadata" . | nindent 2 }}
+ {{ include "envQueue" . | nindent 2 }}
+ {{ include "envCommon" . | nindent 2 }}
+ {{ include "envLog" . | nindent 2 }}
+ {{ include "envNumba" . | nindent 2 }}
+ # service
+ - name: API_HF_AUTH_PATH
+ value: {{ .Values.rows.hfAuthPath | quote }}
+ - name: API_HF_JWT_PUBLIC_KEY_URL
+ value: {{ .Values.rows.hfJwtPublicKeyUrl | quote }}
+ - name: API_HF_JWT_ALGORITHM
+ value: {{ .Values.rows.hfJwtAlgorithm | quote }}
+ - name: API_HF_TIMEOUT_SECONDS
+ value: {{ .Values.rows.hfTimeoutSeconds | quote }}
+ - name: API_HF_WEBHOOK_SECRET
+ {{- if .Values.secrets.hfWebhookSecret.fromSecret }}
+ valueFrom:
+ secretKeyRef:
+ name: {{ .Values.secrets.hfWebhookSecret.secretName | quote }}
+ key: WEBHOOK_SECRET
+ optional: false
+ {{- else }}
+ value: {{ .Values.secrets.hfWebhookSecret.value }}
+ {{- end }}
+ - name: API_MAX_AGE_LONG
+ value: {{ .Values.rows.maxAgeLong | quote }}
+ - name: API_MAX_AGE_SHORT
+ value: {{ .Values.rows.maxAgeShort | quote }}
+ # prometheus
+ - name: PROMETHEUS_MULTIPROC_DIR
+ value: {{ .Values.rows.prometheusMultiprocDirectory | quote }}
+ # uvicorn
+ - name: API_UVICORN_HOSTNAME
+ value: {{ .Values.rows.uvicornHostname | quote }}
+ - name: API_UVICORN_NUM_WORKERS
+ value: {{ .Values.rows.uvicornNumWorkers | quote }}
+ - name: API_UVICORN_PORT
+ value: {{ .Values.rows.uvicornPort | quote }}
+ volumeMounts:
+ {{ include "volumeMountCachedAssetsRW" . | nindent 2 }}
+ {{ include "volumeMountParquetMetadataRO" . | nindent 2 }}
+ securityContext:
+ allowPrivilegeEscalation: false
+ readinessProbe:
+ tcpSocket:
+ port: {{ .Values.rows.uvicornPort }}
+ livenessProbe:
+ tcpSocket:
+ port: {{ .Values.rows.uvicornPort }}
+ ports:
+ - containerPort: {{ .Values.rows.uvicornPort }}
+ name: http
+ protocol: TCP
+ resources:
+ {{ toYaml .Values.rows.resources | nindent 4 }}
+{{- end -}}
diff --git a/chart/templates/services/rows/deployment.yaml b/chart/templates/services/rows/deployment.yaml
new file mode 100644
index 00000000..d5edea1c
--- /dev/null
+++ b/chart/templates/services/rows/deployment.yaml
@@ -0,0 +1,33 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+apiVersion: apps/v1
+kind: Deployment
+metadata:
+ labels: {{ include "labels.rows" . | nindent 4 }}
+ name: "{{ include "name" . }}-rows"
+ namespace: {{ .Release.Namespace }}
+spec:
+ progressDeadlineSeconds: 600
+ replicas: {{ .Values.rows.replicas }}
+ revisionHistoryLimit: 10
+ selector:
+ matchLabels: {{ include "labels.rows" . | nindent 6 }}
+ strategy:
+ rollingUpdate:
+ maxSurge: 25%
+ maxUnavailable: 25%
+ type: RollingUpdate
+ template:
+ metadata:
+ labels: {{ include "labels.rows" . | nindent 8 }}
+ spec:
+ {{- include "image.imagePullSecrets" . | nindent 6 }}
+ initContainers:
+ {{ include "initContainerCachedAssets" . | nindent 8 }}
+ {{ include "initContainerParquetMetadata" . | nindent 8 }}
+ containers: {{ include "containerRows" . | nindent 8 }}
+ nodeSelector: {{ toYaml .Values.rows.nodeSelector | nindent 8 }}
+ tolerations: {{ toYaml .Values.rows.tolerations | nindent 8 }}
+ volumes: {{ include "volumeData" . | nindent 8 }}
+ securityContext: {{ include "securityContext" . | nindent 8 }}
diff --git a/chart/templates/services/rows/pdb.yaml b/chart/templates/services/rows/pdb.yaml
new file mode 100644
index 00000000..21c64d80
--- /dev/null
+++ b/chart/templates/services/rows/pdb.yaml
@@ -0,0 +1,13 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+apiVersion: policy/v1
+kind: PodDisruptionBudget
+metadata:
+ labels: {{ include "labels.rows" . | nindent 4 }}
+ name: "{{ include "name" . }}-rows"
+ namespace: {{ .Release.Namespace }}
+spec:
+ maxUnavailable: 1
+ selector:
+ matchLabels: {{ include "labels.rows" . | nindent 6 }}
diff --git a/chart/templates/services/rows/service.yaml b/chart/templates/services/rows/service.yaml
new file mode 100644
index 00000000..3f5bdd23
--- /dev/null
+++ b/chart/templates/services/rows/service.yaml
@@ -0,0 +1,22 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+{{ $serviceType := .Values.reverseProxy.service.type | default .Values.global.huggingface.service.type }}
+apiVersion: v1
+kind: Service
+metadata:
+ name: "{{ include "name" . }}-rows"
+ annotations: {{ toYaml .Values.rows.service.annotations | nindent 4 }}
+ namespace: {{ .Release.Namespace }}
+ labels: {{ include "labels.rows" . | nindent 4 }}
+spec:
+ ports:
+ - name: http
+ port: 80
+ protocol: TCP
+ {{- if eq "NodePort" $serviceType }}
+ nodePort: {{ .Values.global.huggingface.service.ports.datasetsServer.rows }}
+ {{- end }}
+ targetPort: {{ .Values.rows.uvicornPort }}
+ selector: {{ include "labels.rows" . | nindent 4 }}
+ type: {{ $serviceType }}
diff --git a/chart/templates/services/rows/servicemonitor.yaml b/chart/templates/services/rows/servicemonitor.yaml
new file mode 100644
index 00000000..c53863ae
--- /dev/null
+++ b/chart/templates/services/rows/servicemonitor.yaml
@@ -0,0 +1,20 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+{{- if .Values.monitoring.enabled }}
+apiVersion: monitoring.coreos.com/v1
+kind: ServiceMonitor
+metadata:
+ labels: {{ include "labels.rows" . | nindent 4 }}
+ name: "{{ include "name" . }}-rows"
+ namespace: {{ .Release.Namespace }}
+spec:
+ endpoints:
+ - path: /metrics
+ port: http
+ namespaceSelector:
+ matchNames:
+ - {{ .Release.Namespace }}
+ selector:
+ matchLabels: {{ include "labels.rows" . | nindent 6 }}
+{{- end }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 212ccc51..e8f01d0b 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -19,0 +20 @@ global:
+ rows: 30023
@@ -50,0 +52,5 @@ images:
+ rows:
+ registry: huggingface
+ useGlobalRegistry: false
+ repository: datasets-server-services-rows
+ tag: sha-fb3399a
@@ -409,0 +416,38 @@ api:
+rows:
+ # the path of the external authentication service on the hub.
+ # The string must contain `%s` which will be replaced with the dataset name.
+ hfAuthPath: "/api/datasets/%s/auth-check"
+ # the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format.
+ # It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication
+ # check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored.
+ hfJwtPublicKeyUrl: "https://huggingface.co/api/keys/jwt"
+ # the algorithm used to encode the JWT.
+ hfJwtAlgorithm: "EdDSA"
+ # the timeout in seconds for the requests to the Hugging Face Hub.
+ hfTimeoutSeconds: "0.2"
+ # Number of seconds to set in the `max-age` header on data endpoints
+ maxAgeLong: "120"
+ # Number of seconds to set in the `max-age` header on technical endpoints
+ maxAgeShort: "10"
+ # Directory where the uvicorn workers will write the prometheus metrics
+ # see https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn
+ prometheusMultiprocDirectory: "/tmp"
+ # Hostname - it must not be set to localhost to work in Kube!
+ uvicornHostname: "0.0.0.0"
+ # Number of uvicorn workers for running the application
+ uvicornNumWorkers: "1"
+ # Application endpoint port
+ uvicornPort: 8082
+
+ nodeSelector: {}
+ replicas: 1
+ resources:
+ requests:
+ cpu: 0
+ limits:
+ cpu: 0
+ service:
+ type: ""
+ annotations: {}
+ tolerations: []
+
diff --git a/e2e/Makefile b/e2e/Makefile
index dfe55a19..c1ac31e8 100644
--- a/e2e/Makefile
+++ b/e2e/Makefile
@@ -11,0 +12,2 @@ export API_UVICORN_PORT := 9080
+export ROWS_UVICORN_NUM_WORKERS := 2
+export ROWS_UVICORN_PORT := 9082
diff --git a/e2e/tests/test_40_rows_healthcheck.py b/e2e/tests/test_40_rows_healthcheck.py
new file mode 100644
index 00000000..65ac68b1
--- /dev/null
+++ b/e2e/tests/test_40_rows_healthcheck.py
@@ -0,0 +1,11 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from .utils import ROWS_URL, poll
+
+
+def test_healthcheck() -> None:
+ # this tests ensures the /healthcheck and the /metrics endpoints are hidden
+ response = poll("/healthcheck", expected_code=200, url=ROWS_URL)
+ assert response.status_code == 200, f"{response.status_code} - {response.text}"
+ assert "ok" in response.text, response.text
diff --git a/e2e/tests/test_41_rows_metrics.py b/e2e/tests/test_41_rows_metrics.py
new file mode 100644
index 00000000..5ef19930
--- /dev/null
+++ b/e2e/tests/test_41_rows_metrics.py
@@ -0,0 +1,49 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import os
+import re
+from typing import Mapping
+
+from .utils import ROWS_URL, get
+
+
+def has_metric(name: str, labels: Mapping[str, str], metric_names: set[str]) -> bool:
+ label_str = ",".join([f'{k}="{v}"' for k, v in labels.items()])
+ s = name + "{" + label_str + "}"
+ return any(re.match(s, metric_name) is not None for metric_name in metric_names)
+
+
+def test_metrics() -> None:
+ assert "PROMETHEUS_MULTIPROC_DIR" in os.environ
+ response = get("/metrics", url=ROWS_URL)
+ assert response.status_code == 200, f"{response.status_code} - {response.text}"
+ content = response.text
+ lines = content.split("\n")
+ # examples:
+ # starlette_requests_total{method="GET",path_template="/metrics"} 1.0
+ # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05
+ metrics = {
+ parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1))
+ }
+ # see https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn
+ assert "process_start_time_seconds" not in metrics
+
+ # the middleware should have recorded the request
+ name = 'starlette_requests_total{method="GET",path_template="/metrics"}'
+ assert name in metrics, metrics
+ assert metrics[name] > 0, metrics
+
+ metric_names = set(metrics.keys())
+ for endpoint in ["/rows"]:
+ # these metrics are only available in the admin API
+ assert not has_metric(
+ name="queue_jobs_total",
+ labels={"pid": "[0-9]*", "queue": endpoint, "status": "started"},
+ metric_names=metric_names,
+ ), f"queue_jobs_total - endpoint={endpoint} found in {metrics}"
+ assert not has_metric(
+ name="responses_in_cache_total",
+ labels={"error_code": "None", "http_status": "200", "path": endpoint, "pid": "[0-9]*"},
+ metric_names=metric_names,
+ ), f"responses_in_cache_total - endpoint {endpoint} found in {metrics}"
diff --git a/e2e/tests/utils.py b/e2e/tests/utils.py
index 8e3466c7..681aa72a 100644
--- a/e2e/tests/utils.py
+++ b/e2e/tests/utils.py
@@ -15,0 +16 @@ ADMIN_UVICORN_PORT = os.environ.get("ADMIN_UVICORN_PORT", "8081")
+ROWS_UVICORN_PORT = os.environ.get("ROWS_UVICORN_PORT", "8082")
@@ -21,0 +23 @@ API_URL = f"http://localhost:{API_UVICORN_PORT}"
+ROWS_URL = f"http://localhost:{ROWS_UVICORN_PORT}"
@@ -127 +129,3 @@ def log(response: Response, url: str = URL, relative_url: Optional[str] = None,
- return f"{response.status_code} - {response.headers} - {response.text} - {url}{extra}"
+ return (
+ f"{dataset=} - {relative_url=} - {response.status_code} - {response.headers} - {response.text} - {url}{extra}"
+ )
diff --git a/services/admin/Makefile b/services/admin/Makefile
index cd552232..23d6ba6e 100644
--- a/services/admin/Makefile
+++ b/services/admin/Makefile
@@ -30 +30 @@ test:
- PROMETHEUS_MULTIPROC_DIR=/tmp/admin.prometheus poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
+ PROMETHEUS_MULTIPROC_DIR=/tmp/admin.prometheus poetry run python -m pytest -vv -x -k "test_metrics" ${ADDOPTS} $(TEST_PATH)
diff --git a/services/api/Makefile b/services/api/Makefile
index 44aa2fbe..439cb211 100644
--- a/services/api/Makefile
+++ b/services/api/Makefile
@@ -29 +29 @@ test:
- PROMETHEUS_MULTIPROC_DIR=/tmp/api.prometheus poetry run python -m pytest -vv -x -k "test_metrics or test_prometheus" ${ADDOPTS} $(TEST_PATH)
+ PROMETHEUS_MULTIPROC_DIR=/tmp/api.prometheus poetry run python -m pytest -vv -x -k "test_metrics" ${ADDOPTS} $(TEST_PATH)
diff --git a/services/api/README.md b/services/api/README.md
index 45239b60..5cf8431a 100644
--- a/services/api/README.md
+++ b/services/api/README.md
@@ -7 +7 @@
-The worker can be configured using environment variables. They are grouped by scope.
+The service can be configured using environment variables. They are grouped by scope.
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 8f6992b2..85625f38 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -12 +11,0 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource
-from libcommon.storage import exists, init_cached_assets_dir, init_parquet_metadata_dir
@@ -22 +20,0 @@ from api.routes.endpoint import EndpointsDefinition, create_endpoint
-from api.routes.rows import create_rows_endpoint
@@ -36,4 +33,0 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- cached_assets_directory = init_cached_assets_dir(directory=app_config.cached_assets.storage_directory)
- parquet_metadata_directory = init_parquet_metadata_dir(directory=app_config.parquet_metadata.storage_directory)
- if not exists(cached_assets_directory):
- raise RuntimeError("The assets storage directory could not be accessed. Exiting.")
@@ -111,18 +104,0 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- Route(
- "/rows",
- endpoint=create_rows_endpoint(
- processing_graph=processing_graph,
- cached_assets_base_url=app_config.cached_assets.base_url,
- cached_assets_directory=cached_assets_directory,
- parquet_metadata_directory=parquet_metadata_directory,
- hf_endpoint=app_config.common.hf_endpoint,
- hf_token=app_config.common.hf_token,
- hf_jwt_public_key=hf_jwt_public_key,
- hf_jwt_algorithm=app_config.api.hf_jwt_algorithm,
- external_auth_url=app_config.api.external_auth_url,
- hf_timeout_seconds=app_config.api.hf_timeout_seconds,
- max_age_long=app_config.api.max_age_long,
- max_age_short=app_config.api.max_age_short,
- cache_max_days=app_config.cache.max_days,
- ),
- ),
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index cbc59c89..ea9ef490 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -4 +3,0 @@
-from pathlib import Path
@@ -18,3 +16,0 @@ from api.routes.endpoint import EndpointsDefinition, StepsByInputTypeAndEndpoint
-# Import fixture modules as plugins
-pytest_plugins = ["tests.fixtures.fsspec"]
-
@@ -28 +23,0 @@ def monkeypatch_session() -> Iterator[MonkeyPatch]:
- monkeypatch_session.setenv("CACHED_ASSETS_BASE_URL", "http://localhost/cached-assets")
@@ -142,7 +136,0 @@ def parquet_metadata_directory(app_config: AppConfig) -> StrPath:
-
-
-@fixture
-def image_path() -> str:
- image_path = Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"
- assert image_path.is_file()
- return str(image_path)
diff --git a/services/reverse-proxy/README.md b/services/reverse-proxy/README.md
index d322fda2..2fc6a99b 100644
--- a/services/reverse-proxy/README.md
+++ b/services/reverse-proxy/README.md
@@ -20 +20 @@ It takes various environment variables, all of them are mandatory:
-- `URL_ADMIN`= URL of the admin, eg `http://admin:8080`
+- `URL_ADMIN`= URL of the admin, eg `http://admin:8081`
@@ -21,0 +22 @@ It takes various environment variables, all of them are mandatory:
+- `URL_ROWS`= URL of the rows service, eg `http://rows:8082`
diff --git a/services/rows/.flake8 b/services/rows/.flake8
new file mode 100644
index 00000000..f7d6157c
--- /dev/null
+++ b/services/rows/.flake8
@@ -0,0 +1,5 @@
+[flake8]
+# Recommend matching the black line length (119),
+# rather than using the flake8 default of 79:
+max-line-length = 119
+extend-ignore = "E203"
diff --git a/services/rows/.python-version b/services/rows/.python-version
new file mode 100644
index 00000000..b326afbc
--- /dev/null
+++ b/services/rows/.python-version
@@ -0,0 +1 @@
+3.9.15
diff --git a/services/rows/Dockerfile b/services/rows/Dockerfile
new file mode 100644
index 00000000..6e0db885
--- /dev/null
+++ b/services/rows/Dockerfile
@@ -0,0 +1,35 @@
+# build with
+# docker build -t some_tag_rows -f Dockerfile ../..
+FROM python:3.9.15-slim
+
+ENV PYTHONFAULTHANDLER=1 \
+ PYTHONUNBUFFERED=1 \
+ PYTHONHASHSEED=random \
+ PIP_NO_CACHE_DIR=1 \
+ PIP_DISABLE_PIP_VERSION_CHECK=on \
+ PIP_DEFAULT_TIMEOUT=100 \
+ POETRY_NO_INTERACTION=1 \
+ # Versions:
+ POETRY_VERSION=1.4.0 \
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
+
+# System deps:
+RUN apt-get update \
+ && apt-get install -y build-essential unzip wget \
+ libicu-dev ffmpeg libavcodec-extra libsndfile1 llvm pkg-config \
+ && rm -rf /var/lib/apt/lists/*
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
+
+WORKDIR /src
+COPY services/rows/poetry.lock ./services/rows/poetry.lock
+COPY services/rows/pyproject.toml ./services/rows/pyproject.toml
+COPY libs/libcommon ./libs/libcommon
+COPY libs/libapi ./libs/libapi
+WORKDIR /src/services/rows/
+RUN poetry install --no-cache
+COPY services/rows/src ./src
+RUN poetry install --no-cache
+
+ENTRYPOINT ["poetry", "run", "python", "src/rows/main.py"]
diff --git a/services/rows/Makefile b/services/rows/Makefile
new file mode 100644
index 00000000..02a11c6f
--- /dev/null
+++ b/services/rows/Makefile
@@ -0,0 +1,37 @@
+# environment variables for the commands (docker compose, poetry)
+export COMPOSE_PROJECT_NAME := rows
+export MONGO_PORT := 27032
+export CACHE_MONGO_URL := mongodb://localhost:${MONGO_PORT}
+export QUEUE_MONGO_URL := mongodb://localhost:${MONGO_PORT}
+# makefile variables
+DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml
+TEST_PATH ?= tests
+
+include ../../tools/Python.mk
+include ../../tools/Docker.mk
+
+.PHONY: run
+run:
+ poetry run python src/rows/main.py
+
+.PHONY: watch
+watch:
+ poetry run watchmedo auto-restart -d src/rows -p "*.py" -R python src/rows/main.py
+
+# override the default test target to test prometheus depending on the environment
+# we cannot set the env var with pytest.MonkeyPatch, it's too late
+.PHONY: test
+test:
+ $(MAKE) up
+ poetry run python -m pytest -vv -x ${ADDOPTS} $(TEST_PATH)
+ rm -rf /tmp/rows.prometheus
+ mkdir /tmp/rows.prometheus
+ PROMETHEUS_MULTIPROC_DIR=/tmp/rows.prometheus poetry run python -m pytest -vv -x -k "test_metrics" ${ADDOPTS} $(TEST_PATH)
+ rm -rf /tmp/rows.prometheus
+ $(MAKE) down
+
+.PHONY: coverage
+coverage:
+ $(MAKE) up
+ poetry run python -m pytest -s --cov --cov-report xml:coverage.xml --cov-report=term $(TEST_PATH)
+ $(MAKE) down
diff --git a/services/rows/README.md b/services/rows/README.md
new file mode 100644
index 00000000..ecf8a2ef
--- /dev/null
+++ b/services/rows/README.md
@@ -0,0 +1,23 @@
+# Datasets server API - rows endpoint
+
+> /rows endpoint
+
+## Configuration
+
+The service can be configured using environment variables. They are grouped by scope.
+
+### API service
+
+See [../../libs/libapi/README.md](../../libs/libapi/README.md) for more information about the API configuration.
+
+### Common
+
+See [../../libs/libcommon/README.md](../../libs/libcommon/README.md) for more information about the common configuration.
+
+## Endpoints
+
+See https://huggingface.co/docs/datasets-server
+
+- /healthcheck: ensure the app is running
+- /metrics: return a list of metrics in the Prometheus format
+- /rows: get a slice of rows of a dataset split
diff --git a/services/rows/dev.Dockerfile b/services/rows/dev.Dockerfile
new file mode 100644
index 00000000..6c710bbf
--- /dev/null
+++ b/services/rows/dev.Dockerfile
@@ -0,0 +1,50 @@
+# build with
+# docker build -t some_tag_rows -f Dockerfile ../..
+FROM python:3.9.15-slim
+
+ENV PYTHONFAULTHANDLER=1 \
+ PYTHONUNBUFFERED=1 \
+ PYTHONHASHSEED=random \
+ PIP_NO_CACHE_DIR=1 \
+ PIP_DISABLE_PIP_VERSION_CHECK=on \
+ PIP_DEFAULT_TIMEOUT=100 \
+ POETRY_NO_INTERACTION=1 \
+ # Versions:
+ POETRY_VERSION=1.4.0 \
+ POETRY_VIRTUALENVS_IN_PROJECT=true \
+ PATH="$PATH:/root/.local/bin"
+
+# System deps:
+RUN apt-get update \
+ && apt-get install -y build-essential unzip wget \
+ libicu-dev ffmpeg libavcodec-extra libsndfile1 llvm pkg-config \
+ && rm -rf /var/lib/apt/lists/*
+RUN pip install -U pip
+RUN pip install "poetry==$POETRY_VERSION"
+
+WORKDIR /src
+COPY libs/libcommon/poetry.lock ./libs/libcommon/poetry.lock
+COPY libs/libcommon/pyproject.toml ./libs/libcommon/pyproject.toml
+COPY libs/libapi/poetry.lock ./libs/libapi/poetry.lock
+COPY libs/libapi/pyproject.toml ./libs/libapi/pyproject.toml
+COPY services/rows/poetry.lock ./services/rows/poetry.lock
+COPY services/rows/pyproject.toml ./services/rows/pyproject.toml
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# Initialize an empty libcommon
+# Mapping a volume to ./libs/libcommon/src is required when running this image.
+RUN mkdir ./libs/libcommon/src && mkdir ./libs/libcommon/src/libcommon && touch ./libs/libcommon/src/libcommon/__init__.py
+# Initialize an empty libapi
+# Mapping a volume to ./libs/libapi/src is required when running this image.
+RUN mkdir ./libs/libapi/src && mkdir ./libs/libapi/src/libapi && touch ./libs/libapi/src/libapi/__init__.py
+
+# Install dependencies
+WORKDIR /src/services/rows/
+RUN --mount=type=cache,target=/home/.cache/pypoetry/cache \
+ --mount=type=cache,target=/home/.cache/pypoetry/artifacts \
+ poetry install --no-root
+
+# FOR LOCAL DEVELOPMENT ENVIRONMENT
+# Install the rows package.
+# Mapping a volume to ./services/rows/src is required when running this image.
+ENTRYPOINT ["/bin/sh", "-c" , "poetry install --only-root && poetry run python src/rows/main.py"]
diff --git a/services/rows/poetry.lock b/services/rows/poetry.lock
new file mode 100644
index 00000000..31d00257
--- /dev/null
+++ b/services/rows/poetry.lock
@@ -0,0 +1,3356 @@
+# This file is automatically @generated by Poetry and should not be changed by hand.
+
+[[package]]
+name = "aiohttp"
+version = "3.8.4"
+description = "Async http client/server framework (asyncio)"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"},
+ {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"},
+]
+
+[package.dependencies]
+aiosignal = ">=1.1.2"
+async-timeout = ">=4.0.0a3,<5.0"
+attrs = ">=17.3.0"
+charset-normalizer = ">=2.0,<4.0"
+frozenlist = ">=1.1.1"
+multidict = ">=4.5,<7.0"
+yarl = ">=1.0,<2.0"
+
+[package.extras]
+speedups = ["Brotli", "aiodns", "cchardet"]
+
+[[package]]
+name = "aiosignal"
+version = "1.3.1"
+description = "aiosignal: a list of registered asynchronous callbacks"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"},
+ {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"},
+]
+
+[package.dependencies]
+frozenlist = ">=1.1.0"
+
+[[package]]
+name = "anyio"
+version = "3.7.0"
+description = "High level compatibility layer for multiple asynchronous event loop implementations"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "anyio-3.7.0-py3-none-any.whl", hash = "sha256:eddca883c4175f14df8aedce21054bfca3adb70ffe76a9f607aef9d7fa2ea7f0"},
+ {file = "anyio-3.7.0.tar.gz", hash = "sha256:275d9973793619a5374e1c89a4f4ad3f4b0a5510a2b5b939444bee8f4c4d37ce"},
+]
+
+[package.dependencies]
+exceptiongroup = {version = "*", markers = "python_version < \"3.11\""}
+idna = ">=2.8"
+sniffio = ">=1.1"
+
+[package.extras]
+doc = ["Sphinx (>=6.1.0)", "packaging", "sphinx-autodoc-typehints (>=1.2.0)", "sphinx-rtd-theme", "sphinxcontrib-jquery"]
+test = ["anyio[trio]", "coverage[toml] (>=4.5)", "hypothesis (>=4.0)", "mock (>=4)", "psutil (>=5.9)", "pytest (>=7.0)", "pytest-mock (>=3.6.1)", "trustme", "uvloop (>=0.17)"]
+trio = ["trio (<0.22)"]
+
+[[package]]
+name = "appdirs"
+version = "1.4.4"
+description = "A small Python module for determining appropriate platform-specific dirs, e.g. a \"user data dir\"."
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "appdirs-1.4.4-py2.py3-none-any.whl", hash = "sha256:a841dacd6b99318a741b166adb07e19ee71a274450e68237b4650ca1055ab128"},
+ {file = "appdirs-1.4.4.tar.gz", hash = "sha256:7d5d0167b2b1ba821647616af46a749d1c653740dd0d2415100fe26e27afdf41"},
+]
+
+[[package]]
+name = "async-timeout"
+version = "4.0.2"
+description = "Timeout context manager for asyncio programs"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"},
+ {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"},
+]
+
+[[package]]
+name = "attrs"
+version = "23.1.0"
+description = "Classes Without Boilerplate"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "attrs-23.1.0-py3-none-any.whl", hash = "sha256:1f28b4522cdc2fb4256ac1a020c78acf9cba2c6b461ccd2c126f3aa8e8335d04"},
+ {file = "attrs-23.1.0.tar.gz", hash = "sha256:6279836d581513a26f1bf235f9acd333bc9115683f14f7e8fae46c98fc50e015"},
+]
+
+[package.extras]
+cov = ["attrs[tests]", "coverage[toml] (>=5.3)"]
+dev = ["attrs[docs,tests]", "pre-commit"]
+docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope-interface"]
+tests = ["attrs[tests-no-zope]", "zope-interface"]
+tests-no-zope = ["cloudpickle", "hypothesis", "mypy (>=1.1.1)", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"]
+
+[[package]]
+name = "audioread"
+version = "3.0.0"
+description = "multi-library, cross-platform audio decoding"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"},
+]
+
+[[package]]
+name = "bandit"
+version = "1.7.5"
+description = "Security oriented static analyser for python code."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "bandit-1.7.5-py3-none-any.whl", hash = "sha256:75665181dc1e0096369112541a056c59d1c5f66f9bb74a8d686c3c362b83f549"},
+ {file = "bandit-1.7.5.tar.gz", hash = "sha256:bdfc739baa03b880c2d15d0431b31c658ffc348e907fe197e54e0389dd59e11e"},
+]
+
+[package.dependencies]
+colorama = {version = ">=0.3.9", markers = "platform_system == \"Windows\""}
+GitPython = ">=1.0.1"
+PyYAML = ">=5.3.1"
+rich = "*"
+stevedore = ">=1.20.0"
+
+[package.extras]
+test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "tomli (>=1.1.0)"]
+toml = ["tomli (>=1.1.0)"]
+yaml = ["PyYAML"]
+
+[[package]]
+name = "black"
+version = "22.12.0"
+description = "The uncompromising code formatter."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "black-22.12.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eedd20838bd5d75b80c9f5487dbcb06836a43833a37846cf1d8c1cc01cef59d"},
+ {file = "black-22.12.0-cp310-cp310-win_amd64.whl", hash = "sha256:159a46a4947f73387b4d83e87ea006dbb2337eab6c879620a3ba52699b1f4351"},
+ {file = "black-22.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d30b212bffeb1e252b31dd269dfae69dd17e06d92b87ad26e23890f3efea366f"},
+ {file = "black-22.12.0-cp311-cp311-win_amd64.whl", hash = "sha256:7412e75863aa5c5411886804678b7d083c7c28421210180d67dfd8cf1221e1f4"},
+ {file = "black-22.12.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c116eed0efb9ff870ded8b62fe9f28dd61ef6e9ddd28d83d7d264a38417dcee2"},
+ {file = "black-22.12.0-cp37-cp37m-win_amd64.whl", hash = "sha256:1f58cbe16dfe8c12b7434e50ff889fa479072096d79f0a7f25e4ab8e94cd8350"},
+ {file = "black-22.12.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77d86c9f3db9b1bf6761244bc0b3572a546f5fe37917a044e02f3166d5aafa7d"},
+ {file = "black-22.12.0-cp38-cp38-win_amd64.whl", hash = "sha256:82d9fe8fee3401e02e79767016b4907820a7dc28d70d137eb397b92ef3cc5bfc"},
+ {file = "black-22.12.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:101c69b23df9b44247bd88e1d7e90154336ac4992502d4197bdac35dd7ee3320"},
+ {file = "black-22.12.0-cp39-cp39-win_amd64.whl", hash = "sha256:559c7a1ba9a006226f09e4916060982fd27334ae1998e7a38b3f33a37f7a2148"},
+ {file = "black-22.12.0-py3-none-any.whl", hash = "sha256:436cc9167dd28040ad90d3b404aec22cedf24a6e4d7de221bec2730ec0c97bcf"},
+ {file = "black-22.12.0.tar.gz", hash = "sha256:229351e5a18ca30f447bf724d007f890f97e13af070bb6ad4c0a441cd7596a2f"},
+]
+
+[package.dependencies]
+click = ">=8.0.0"
+mypy-extensions = ">=0.4.3"
+pathspec = ">=0.9.0"
+platformdirs = ">=2"
+tomli = {version = ">=1.1.0", markers = "python_full_version < \"3.11.0a7\""}
+typing-extensions = {version = ">=3.10.0.0", markers = "python_version < \"3.10\""}
+
+[package.extras]
+colorama = ["colorama (>=0.4.3)"]
+d = ["aiohttp (>=3.7.4)"]
+jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"]
+uvloop = ["uvloop (>=0.15.2)"]
+
+[[package]]
+name = "cachecontrol"
+version = "0.13.1"
+description = "httplib2 caching for requests"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
+]
+
+[package.dependencies]
+filelock = {version = ">=3.8.0", optional = true, markers = "extra == \"filecache\""}
+msgpack = ">=0.5.2"
+requests = ">=2.16.0"
+
+[package.extras]
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
+filecache = ["filelock (>=3.8.0)"]
+redis = ["redis (>=2.10.5)"]
+
+[[package]]
+name = "certifi"
+version = "2023.5.7"
+description = "Python package for providing Mozilla's CA Bundle."
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "certifi-2023.5.7-py3-none-any.whl", hash = "sha256:c6c2e98f5c7869efca1f8916fed228dd91539f9f1b444c314c06eef02980c716"},
+ {file = "certifi-2023.5.7.tar.gz", hash = "sha256:0f0d56dc5a6ad56fd4ba36484d6cc34451e1c6548c61daad8c320169f91eddc7"},
+]
+
+[[package]]
+name = "cffi"
+version = "1.15.1"
+description = "Foreign Function Interface for Python calling C code."
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"},
+ {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"},
+ {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"},
+ {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"},
+ {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"},
+ {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"},
+ {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"},
+ {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"},
+ {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"},
+ {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"},
+ {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"},
+ {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"},
+ {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"},
+ {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"},
+ {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"},
+ {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"},
+ {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"},
+ {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"},
+ {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"},
+ {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"},
+ {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"},
+ {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"},
+ {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"},
+ {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"},
+ {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"},
+ {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"},
+ {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"},
+ {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"},
+ {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"},
+ {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"},
+ {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"},
+ {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"},
+ {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"},
+ {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"},
+ {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"},
+]
+
+[package.dependencies]
+pycparser = "*"
+
+[[package]]
+name = "charset-normalizer"
+version = "3.1.0"
+description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet."
+category = "main"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"},
+ {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"},
+]
+
+[[package]]
+name = "click"
+version = "8.1.3"
+description = "Composable command line interface toolkit"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "click-8.1.3-py3-none-any.whl", hash = "sha256:bb4d8133cb15a609f44e8213d9b391b0809795062913b383c62be0ee95b1db48"},
+ {file = "click-8.1.3.tar.gz", hash = "sha256:7682dc8afb30297001674575ea00d1814d808d6a36af415a82bd481d37ba7b8e"},
+]
+
+[package.dependencies]
+colorama = {version = "*", markers = "platform_system == \"Windows\""}
+
+[[package]]
+name = "colorama"
+version = "0.4.6"
+description = "Cross-platform colored terminal text."
+category = "main"
+optional = false
+python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7"
+files = [
+ {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"},
+ {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"},
+]
+
+[[package]]
+name = "coverage"
+version = "7.2.7"
+description = "Code coverage measurement for Python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "coverage-7.2.7-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d39b5b4f2a66ccae8b7263ac3c8170994b65266797fb96cbbfd3fb5b23921db8"},
+ {file = "coverage-7.2.7-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:6d040ef7c9859bb11dfeb056ff5b3872436e3b5e401817d87a31e1750b9ae2fb"},
+ {file = "coverage-7.2.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ba90a9563ba44a72fda2e85302c3abc71c5589cea608ca16c22b9804262aaeb6"},
+ {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e7d9405291c6928619403db1d10bd07888888ec1abcbd9748fdaa971d7d661b2"},
+ {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:31563e97dae5598556600466ad9beea39fb04e0229e61c12eaa206e0aa202063"},
+ {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ebba1cd308ef115925421d3e6a586e655ca5a77b5bf41e02eb0e4562a111f2d1"},
+ {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:cb017fd1b2603ef59e374ba2063f593abe0fc45f2ad9abdde5b4d83bd922a353"},
+ {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d62a5c7dad11015c66fbb9d881bc4caa5b12f16292f857842d9d1871595f4495"},
+ {file = "coverage-7.2.7-cp310-cp310-win32.whl", hash = "sha256:ee57190f24fba796e36bb6d3aa8a8783c643d8fa9760c89f7a98ab5455fbf818"},
+ {file = "coverage-7.2.7-cp310-cp310-win_amd64.whl", hash = "sha256:f75f7168ab25dd93110c8a8117a22450c19976afbc44234cbf71481094c1b850"},
+ {file = "coverage-7.2.7-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:06a9a2be0b5b576c3f18f1a241f0473575c4a26021b52b2a85263a00f034d51f"},
+ {file = "coverage-7.2.7-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5baa06420f837184130752b7c5ea0808762083bf3487b5038d68b012e5937dbe"},
+ {file = "coverage-7.2.7-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fdec9e8cbf13a5bf63290fc6013d216a4c7232efb51548594ca3631a7f13c3a3"},
+ {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:52edc1a60c0d34afa421c9c37078817b2e67a392cab17d97283b64c5833f427f"},
+ {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:63426706118b7f5cf6bb6c895dc215d8a418d5952544042c8a2d9fe87fcf09cb"},
+ {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:afb17f84d56068a7c29f5fa37bfd38d5aba69e3304af08ee94da8ed5b0865833"},
+ {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:48c19d2159d433ccc99e729ceae7d5293fbffa0bdb94952d3579983d1c8c9d97"},
+ {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:0e1f928eaf5469c11e886fe0885ad2bf1ec606434e79842a879277895a50942a"},
+ {file = "coverage-7.2.7-cp311-cp311-win32.whl", hash = "sha256:33d6d3ea29d5b3a1a632b3c4e4f4ecae24ef170b0b9ee493883f2df10039959a"},
+ {file = "coverage-7.2.7-cp311-cp311-win_amd64.whl", hash = "sha256:5b7540161790b2f28143191f5f8ec02fb132660ff175b7747b95dcb77ac26562"},
+ {file = "coverage-7.2.7-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:f2f67fe12b22cd130d34d0ef79206061bfb5eda52feb6ce0dba0644e20a03cf4"},
+ {file = "coverage-7.2.7-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a342242fe22407f3c17f4b499276a02b01e80f861f1682ad1d95b04018e0c0d4"},
+ {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:171717c7cb6b453aebac9a2ef603699da237f341b38eebfee9be75d27dc38e01"},
+ {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:49969a9f7ffa086d973d91cec8d2e31080436ef0fb4a359cae927e742abfaaa6"},
+ {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:b46517c02ccd08092f4fa99f24c3b83d8f92f739b4657b0f146246a0ca6a831d"},
+ {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:a3d33a6b3eae87ceaefa91ffdc130b5e8536182cd6dfdbfc1aa56b46ff8c86de"},
+ {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:976b9c42fb2a43ebf304fa7d4a310e5f16cc99992f33eced91ef6f908bd8f33d"},
+ {file = "coverage-7.2.7-cp312-cp312-win32.whl", hash = "sha256:8de8bb0e5ad103888d65abef8bca41ab93721647590a3f740100cd65c3b00511"},
+ {file = "coverage-7.2.7-cp312-cp312-win_amd64.whl", hash = "sha256:9e31cb64d7de6b6f09702bb27c02d1904b3aebfca610c12772452c4e6c21a0d3"},
+ {file = "coverage-7.2.7-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:58c2ccc2f00ecb51253cbe5d8d7122a34590fac9646a960d1430d5b15321d95f"},
+ {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d22656368f0e6189e24722214ed8d66b8022db19d182927b9a248a2a8a2f67eb"},
+ {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a895fcc7b15c3fc72beb43cdcbdf0ddb7d2ebc959edac9cef390b0d14f39f8a9"},
+ {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e84606b74eb7de6ff581a7915e2dab7a28a0517fbe1c9239eb227e1354064dcd"},
+ {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0a5f9e1dbd7fbe30196578ca36f3fba75376fb99888c395c5880b355e2875f8a"},
+ {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:419bfd2caae268623dd469eff96d510a920c90928b60f2073d79f8fe2bbc5959"},
+ {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:2aee274c46590717f38ae5e4650988d1af340fe06167546cc32fe2f58ed05b02"},
+ {file = "coverage-7.2.7-cp37-cp37m-win32.whl", hash = "sha256:61b9a528fb348373c433e8966535074b802c7a5d7f23c4f421e6c6e2f1697a6f"},
+ {file = "coverage-7.2.7-cp37-cp37m-win_amd64.whl", hash = "sha256:b1c546aca0ca4d028901d825015dc8e4d56aac4b541877690eb76490f1dc8ed0"},
+ {file = "coverage-7.2.7-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:54b896376ab563bd38453cecb813c295cf347cf5906e8b41d340b0321a5433e5"},
+ {file = "coverage-7.2.7-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3d376df58cc111dc8e21e3b6e24606b5bb5dee6024f46a5abca99124b2229ef5"},
+ {file = "coverage-7.2.7-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e330fc79bd7207e46c7d7fd2bb4af2963f5f635703925543a70b99574b0fea9"},
+ {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e9d683426464e4a252bf70c3498756055016f99ddaec3774bf368e76bbe02b6"},
+ {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8d13c64ee2d33eccf7437961b6ea7ad8673e2be040b4f7fd4fd4d4d28d9ccb1e"},
+ {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:b7aa5f8a41217360e600da646004f878250a0d6738bcdc11a0a39928d7dc2050"},
+ {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8fa03bce9bfbeeef9f3b160a8bed39a221d82308b4152b27d82d8daa7041fee5"},
+ {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:245167dd26180ab4c91d5e1496a30be4cd721a5cf2abf52974f965f10f11419f"},
+ {file = "coverage-7.2.7-cp38-cp38-win32.whl", hash = "sha256:d2c2db7fd82e9b72937969bceac4d6ca89660db0a0967614ce2481e81a0b771e"},
+ {file = "coverage-7.2.7-cp38-cp38-win_amd64.whl", hash = "sha256:2e07b54284e381531c87f785f613b833569c14ecacdcb85d56b25c4622c16c3c"},
+ {file = "coverage-7.2.7-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:537891ae8ce59ef63d0123f7ac9e2ae0fc8b72c7ccbe5296fec45fd68967b6c9"},
+ {file = "coverage-7.2.7-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:06fb182e69f33f6cd1d39a6c597294cff3143554b64b9825d1dc69d18cc2fff2"},
+ {file = "coverage-7.2.7-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:201e7389591af40950a6480bd9edfa8ed04346ff80002cec1a66cac4549c1ad7"},
+ {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f6951407391b639504e3b3be51b7ba5f3528adbf1a8ac3302b687ecababf929e"},
+ {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6f48351d66575f535669306aa7d6d6f71bc43372473b54a832222803eb956fd1"},
+ {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b29019c76039dc3c0fd815c41392a044ce555d9bcdd38b0fb60fb4cd8e475ba9"},
+ {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:81c13a1fc7468c40f13420732805a4c38a105d89848b7c10af65a90beff25250"},
+ {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:975d70ab7e3c80a3fe86001d8751f6778905ec723f5b110aed1e450da9d4b7f2"},
+ {file = "coverage-7.2.7-cp39-cp39-win32.whl", hash = "sha256:7ee7d9d4822c8acc74a5e26c50604dff824710bc8de424904c0982e25c39c6cb"},
+ {file = "coverage-7.2.7-cp39-cp39-win_amd64.whl", hash = "sha256:eb393e5ebc85245347950143969b241d08b52b88a3dc39479822e073a1a8eb27"},
+ {file = "coverage-7.2.7-pp37.pp38.pp39-none-any.whl", hash = "sha256:b7b4c971f05e6ae490fef852c218b0e79d4e52f79ef0c8475566584a8fb3e01d"},
+ {file = "coverage-7.2.7.tar.gz", hash = "sha256:924d94291ca674905fe9481f12294eb11f2d3d3fd1adb20314ba89e94f44ed59"},
+]
+
+[package.extras]
+toml = ["tomli"]
+
+[[package]]
+name = "cryptography"
+version = "41.0.1"
+description = "cryptography is a package which provides cryptographic recipes and primitives to Python developers."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_universal2.whl", hash = "sha256:f73bff05db2a3e5974a6fd248af2566134d8981fd7ab012e5dd4ddb1d9a70699"},
+ {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_x86_64.whl", hash = "sha256:1a5472d40c8f8e91ff7a3d8ac6dfa363d8e3138b961529c996f3e2df0c7a411a"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fa01527046ca5facdf973eef2535a27fec4cb651e4daec4d043ef63f6ecd4ca"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b46e37db3cc267b4dea1f56da7346c9727e1209aa98487179ee8ebed09d21e43"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:d198820aba55660b4d74f7b5fd1f17db3aa5eb3e6893b0a41b75e84e4f9e0e4b"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:948224d76c4b6457349d47c0c98657557f429b4e93057cf5a2f71d603e2fc3a3"},
+ {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:059e348f9a3c1950937e1b5d7ba1f8e968508ab181e75fc32b879452f08356db"},
+ {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:b4ceb5324b998ce2003bc17d519080b4ec8d5b7b70794cbd2836101406a9be31"},
+ {file = "cryptography-41.0.1-cp37-abi3-win32.whl", hash = "sha256:8f4ab7021127a9b4323537300a2acfb450124b2def3756f64dc3a3d2160ee4b5"},
+ {file = "cryptography-41.0.1-cp37-abi3-win_amd64.whl", hash = "sha256:1fee5aacc7367487b4e22484d3c7e547992ed726d14864ee33c0176ae43b0d7c"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-macosx_10_12_x86_64.whl", hash = "sha256:9a6c7a3c87d595608a39980ebaa04d5a37f94024c9f24eb7d10262b92f739ddb"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:5d092fdfedaec4cbbffbf98cddc915ba145313a6fdaab83c6e67f4e6c218e6f3"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:1a8e6c2de6fbbcc5e14fd27fb24414507cb3333198ea9ab1258d916f00bc3039"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cb33ccf15e89f7ed89b235cff9d49e2e62c6c981a6061c9c8bb47ed7951190bc"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:5f0ff6e18d13a3de56f609dd1fd11470918f770c6bd5d00d632076c727d35485"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:7bfc55a5eae8b86a287747053140ba221afc65eb06207bedf6e019b8934b477c"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:eb8163f5e549a22888c18b0d53d6bb62a20510060a22fd5a995ec8a05268df8a"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:8dde71c4169ec5ccc1087bb7521d54251c016f126f922ab2dfe6649170a3b8c5"},
+ {file = "cryptography-41.0.1.tar.gz", hash = "sha256:d34579085401d3f49762d2f7d6634d6b6c2ae1242202e860f4d26b046e3a1006"},
+]
+
+[package.dependencies]
+cffi = ">=1.12"
+
+[package.extras]
+docs = ["sphinx (>=5.3.0)", "sphinx-rtd-theme (>=1.1.1)"]
+docstest = ["pyenchant (>=1.6.11)", "sphinxcontrib-spelling (>=4.0.1)", "twine (>=1.12.0)"]
+nox = ["nox"]
+pep8test = ["black", "check-sdist", "mypy", "ruff"]
+sdist = ["build"]
+ssh = ["bcrypt (>=3.1.5)"]
+test = ["pretend", "pytest (>=6.2.0)", "pytest-benchmark", "pytest-cov", "pytest-xdist"]
+test-randomorder = ["pytest-randomly"]
+
+[[package]]
+name = "cyclonedx-python-lib"
+version = "4.0.1"
+description = "A library for producing CycloneDX SBOM (Software Bill of Materials) files."
+category = "dev"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "cyclonedx_python_lib-4.0.1-py3-none-any.whl", hash = "sha256:907b64f00df85d727a425de86604768b248cf19285993729e04f17bec767f692"},
+ {file = "cyclonedx_python_lib-4.0.1.tar.gz", hash = "sha256:878e33b8e0080c786f6cbd4c6f87ad610db65d6a3a686a5698415d9cfcd8925d"},
+]
+
+[package.dependencies]
+packageurl-python = ">=0.11"
+py-serializable = ">=0.11.1,<0.12.0"
+sortedcontainers = ">=2.4.0,<3.0.0"
+
+[[package]]
+name = "datasets"
+version = "2.13.1"
+description = "HuggingFace community-driven open-source library of datasets"
+category = "main"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
+]
+
+[package.dependencies]
+aiohttp = "*"
+dill = ">=0.3.0,<0.3.7"
+fsspec = {version = ">=2021.11.1", extras = ["http"]}
+huggingface-hub = ">=0.11.0,<1.0.0"
+librosa = {version = "*", optional = true, markers = "extra == \"audio\""}
+multiprocess = "*"
+numpy = ">=1.17"
+packaging = "*"
+pandas = "*"
+Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""}
+pyarrow = ">=8.0.0"
+pyyaml = ">=5.1"
+requests = ">=2.19.0"
+soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""}
+tqdm = ">=4.62.1"
+xxhash = "*"
+
+[package.extras]
+apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"]
+audio = ["librosa", "soundfile (>=0.12.1)"]
+benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+docs = ["s3fs"]
+jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"]
+s3 = ["s3fs"]
+tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"]
+tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+torch = ["torch"]
+vision = ["Pillow (>=6.2.1)"]
+
+[[package]]
+name = "decorator"
+version = "5.1.1"
+description = "Decorators for Humans"
+category = "main"
+optional = false
+python-versions = ">=3.5"
+files = [
+ {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"},
+ {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"},
+]
+
+[[package]]
+name = "defusedxml"
+version = "0.7.1"
+description = "XML bomb protection for Python stdlib modules"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "defusedxml-0.7.1-py2.py3-none-any.whl", hash = "sha256:a352e7e428770286cc899e2542b6cdaedb2b4953ff269a210103ec58f6198a61"},
+ {file = "defusedxml-0.7.1.tar.gz", hash = "sha256:1bb3032db185915b62d7c6209c5a8792be6a32ab2fedacc84e01b52c51aa3e69"},
+]
+
+[[package]]
+name = "dill"
+version = "0.3.6"
+description = "serialize all of python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"},
+ {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"},
+]
+
+[package.extras]
+graph = ["objgraph (>=1.7.2)"]
+
+[[package]]
+name = "dnspython"
+version = "1.16.0"
+description = "DNS toolkit"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "dnspython-1.16.0-py2.py3-none-any.whl", hash = "sha256:f69c21288a962f4da86e56c4905b49d11aba7938d3d740e80d9e366ee4f1632d"},
+ {file = "dnspython-1.16.0.zip", hash = "sha256:36c5e8e38d4369a08b6780b7f27d790a292b2b08eea01607865bf0936c558e01"},
+]
+
+[package.extras]
+dnssec = ["ecdsa (>=0.13)", "pycryptodome"]
+idna = ["idna (>=2.1)"]
+
+[[package]]
+name = "environs"
+version = "9.5.0"
+description = "simplified environment variable parsing"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "environs-9.5.0-py2.py3-none-any.whl", hash = "sha256:1e549569a3de49c05f856f40bce86979e7d5ffbbc4398e7f338574c220189124"},
+ {file = "environs-9.5.0.tar.gz", hash = "sha256:a76307b36fbe856bdca7ee9161e6c466fd7fcffc297109a118c59b54e27e30c9"},
+]
+
+[package.dependencies]
+marshmallow = ">=3.0.0"
+python-dotenv = "*"
+
+[package.extras]
+dev = ["dj-database-url", "dj-email-url", "django-cache-url", "flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)", "pytest", "tox"]
+django = ["dj-database-url", "dj-email-url", "django-cache-url"]
+lint = ["flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)"]
+tests = ["dj-database-url", "dj-email-url", "django-cache-url", "pytest"]
+
+[[package]]
+name = "exceptiongroup"
+version = "1.1.2"
+description = "Backport of PEP 654 (exception groups)"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "exceptiongroup-1.1.2-py3-none-any.whl", hash = "sha256:e346e69d186172ca7cf029c8c1d16235aa0e04035e5750b4b95039e65204328f"},
+ {file = "exceptiongroup-1.1.2.tar.gz", hash = "sha256:12c3e887d6485d16943a309616de20ae5582633e0a2eda17f4e10fd61c1e8af5"},
+]
+
+[package.extras]
+test = ["pytest (>=6)"]
+
+[[package]]
+name = "filelock"
+version = "3.12.2"
+description = "A platform independent file lock."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "filelock-3.12.2-py3-none-any.whl", hash = "sha256:cbb791cdea2a72f23da6ac5b5269ab0a0d161e9ef0100e653b69049a7706d1ec"},
+ {file = "filelock-3.12.2.tar.gz", hash = "sha256:002740518d8aa59a26b0c76e10fb8c6e15eae825d34b6fdf670333fd7b938d81"},
+]
+
+[package.extras]
+docs = ["furo (>=2023.5.20)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"]
+testing = ["covdefaults (>=2.3)", "coverage (>=7.2.7)", "diff-cover (>=7.5)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)", "pytest-timeout (>=2.1)"]
+
+[[package]]
+name = "flake8"
+version = "3.9.2"
+description = "the modular source code checker: pep8 pyflakes and co"
+category = "dev"
+optional = false
+python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7"
+files = [
+ {file = "flake8-3.9.2-py2.py3-none-any.whl", hash = "sha256:bf8fd333346d844f616e8d47905ef3a3384edae6b4e9beb0c5101e25e3110907"},
+ {file = "flake8-3.9.2.tar.gz", hash = "sha256:07528381786f2a6237b061f6e96610a4167b226cb926e2aa2b6b1d78057c576b"},
+]
+
+[package.dependencies]
+mccabe = ">=0.6.0,<0.7.0"
+pycodestyle = ">=2.7.0,<2.8.0"
+pyflakes = ">=2.3.0,<2.4.0"
+
+[[package]]
+name = "frozenlist"
+version = "1.3.3"
+description = "A list-like structure which implements collections.abc.MutableSequence"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"},
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"},
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"},
+ {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"},
+ {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"},
+ {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"},
+ {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"},
+ {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"},
+ {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"},
+ {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"},
+ {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"},
+ {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"},
+]
+
+[[package]]
+name = "fsspec"
+version = "2023.6.0"
+description = "File-system specification"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "fsspec-2023.6.0-py3-none-any.whl", hash = "sha256:1cbad1faef3e391fba6dc005ae9b5bdcbf43005c9167ce78c915549c352c869a"},
+ {file = "fsspec-2023.6.0.tar.gz", hash = "sha256:d0b2f935446169753e7a5c5c55681c54ea91996cc67be93c39a154fb3a2742af"},
+]
+
+[package.dependencies]
+aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""}
+requests = {version = "*", optional = true, markers = "extra == \"http\""}
+
+[package.extras]
+abfs = ["adlfs"]
+adl = ["adlfs"]
+arrow = ["pyarrow (>=1)"]
+dask = ["dask", "distributed"]
+devel = ["pytest", "pytest-cov"]
+dropbox = ["dropbox", "dropboxdrivefs", "requests"]
+full = ["adlfs", "aiohttp (!=4.0.0a0,!=4.0.0a1)", "dask", "distributed", "dropbox", "dropboxdrivefs", "fusepy", "gcsfs", "libarchive-c", "ocifs", "panel", "paramiko", "pyarrow (>=1)", "pygit2", "requests", "s3fs", "smbprotocol", "tqdm"]
+fuse = ["fusepy"]
+gcs = ["gcsfs"]
+git = ["pygit2"]
+github = ["requests"]
+gs = ["gcsfs"]
+gui = ["panel"]
+hdfs = ["pyarrow (>=1)"]
+http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"]
+libarchive = ["libarchive-c"]
+oci = ["ocifs"]
+s3 = ["s3fs"]
+sftp = ["paramiko"]
+smb = ["smbprotocol"]
+ssh = ["paramiko"]
+tqdm = ["tqdm"]
+
+[[package]]
+name = "gitdb"
+version = "4.0.10"
+description = "Git Object Database"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "gitdb-4.0.10-py3-none-any.whl", hash = "sha256:c286cf298426064079ed96a9e4a9d39e7f3e9bf15ba60701e95f5492f28415c7"},
+ {file = "gitdb-4.0.10.tar.gz", hash = "sha256:6eb990b69df4e15bad899ea868dc46572c3f75339735663b81de79b06f17eb9a"},
+]
+
+[package.dependencies]
+smmap = ">=3.0.1,<6"
+
+[[package]]
+name = "gitpython"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
+]
+
+[package.dependencies]
+gitdb = ">=4.0.1,<5"
+
+[[package]]
+name = "h11"
+version = "0.14.0"
+description = "A pure-Python, bring-your-own-I/O implementation of HTTP/1.1"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "h11-0.14.0-py3-none-any.whl", hash = "sha256:e3fe4ac4b851c468cc8363d500db52c2ead036020723024a109d37346efaa761"},
+ {file = "h11-0.14.0.tar.gz", hash = "sha256:8f19fbbe99e72420ff35c00b27a34cb9937e902a8b810e2c88300c6f0a3b699d"},
+]
+
+[[package]]
+name = "html5lib"
+version = "1.1"
+description = "HTML parser based on the WHATWG HTML specification"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "html5lib-1.1-py2.py3-none-any.whl", hash = "sha256:0d78f8fde1c230e99fe37986a60526d7049ed4bf8a9fadbad5f00e22e58e041d"},
+ {file = "html5lib-1.1.tar.gz", hash = "sha256:b2e5b40261e20f354d198eae92afc10d750afb487ed5e50f9c4eaf07c184146f"},
+]
+
+[package.dependencies]
+six = ">=1.9"
+webencodings = "*"
+
+[package.extras]
+all = ["chardet (>=2.2)", "genshi", "lxml"]
+chardet = ["chardet (>=2.2)"]
+genshi = ["genshi"]
+lxml = ["lxml"]
+
+[[package]]
+name = "httpcore"
+version = "0.16.3"
+description = "A minimal low-level HTTP client."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "httpcore-0.16.3-py3-none-any.whl", hash = "sha256:da1fb708784a938aa084bde4feb8317056c55037247c787bd7e19eb2c2949dc0"},
+ {file = "httpcore-0.16.3.tar.gz", hash = "sha256:c5d6f04e2fc530f39e0c077e6a30caa53f1451096120f1f38b954afd0b17c0cb"},
+]
+
+[package.dependencies]
+anyio = ">=3.0,<5.0"
+certifi = "*"
+h11 = ">=0.13,<0.15"
+sniffio = ">=1.0.0,<2.0.0"
+
+[package.extras]
+http2 = ["h2 (>=3,<5)"]
+socks = ["socksio (>=1.0.0,<2.0.0)"]
+
+[[package]]
+name = "httpx"
+version = "0.23.3"
+description = "The next generation HTTP client."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "httpx-0.23.3-py3-none-any.whl", hash = "sha256:a211fcce9b1254ea24f0cd6af9869b3d29aba40154e947d2a07bb499b3e310d6"},
+ {file = "httpx-0.23.3.tar.gz", hash = "sha256:9818458eb565bb54898ccb9b8b251a28785dd4a55afbc23d0eb410754fe7d0f9"},
+]
+
+[package.dependencies]
+certifi = "*"
+httpcore = ">=0.15.0,<0.17.0"
+rfc3986 = {version = ">=1.3,<2", extras = ["idna2008"]}
+sniffio = "*"
+
+[package.extras]
+brotli = ["brotli", "brotlicffi"]
+cli = ["click (>=8.0.0,<9.0.0)", "pygments (>=2.0.0,<3.0.0)", "rich (>=10,<13)"]
+http2 = ["h2 (>=3,<5)"]
+socks = ["socksio (>=1.0.0,<2.0.0)"]
+
+[[package]]
+name = "huggingface-hub"
+version = "0.15.1"
+description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub"
+category = "main"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "huggingface_hub-0.15.1-py3-none-any.whl", hash = "sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445"},
+ {file = "huggingface_hub-0.15.1.tar.gz", hash = "sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081"},
+]
+
+[package.dependencies]
+filelock = "*"
+fsspec = "*"
+packaging = ">=20.9"
+pyyaml = ">=5.1"
+requests = "*"
+tqdm = ">=4.42.1"
+typing-extensions = ">=3.7.4.3"
+
+[package.extras]
+all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"]
+cli = ["InquirerPy (==0.3.4)"]
+dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"]
+fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"]
+quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"]
+tensorflow = ["graphviz", "pydot", "tensorflow"]
+testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "gradio", "jedi", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "soundfile", "urllib3 (<2.0)"]
+torch = ["torch"]
+typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"]
+
+[[package]]
+name = "idna"
+version = "3.4"
+description = "Internationalized Domain Names in Applications (IDNA)"
+category = "main"
+optional = false
+python-versions = ">=3.5"
+files = [
+ {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"},
+ {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"},
+]
+
+[[package]]
+name = "iniconfig"
+version = "2.0.0"
+description = "brain-dead simple config-ini parsing"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "iniconfig-2.0.0-py3-none-any.whl", hash = "sha256:b6a85871a79d2e3b22d2d1b94ac2824226a63c6b741c88f7ae975f18b6778374"},
+ {file = "iniconfig-2.0.0.tar.gz", hash = "sha256:2d91e135bf72d31a410b17c16da610a82cb55f6b0477d1a902134b24a455b8b3"},
+]
+
+[[package]]
+name = "isort"
+version = "5.12.0"
+description = "A Python utility / library to sort Python imports."
+category = "dev"
+optional = false
+python-versions = ">=3.8.0"
+files = [
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
+]
+
+[package.extras]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
+plugins = ["setuptools"]
+requirements-deprecated-finder = ["pip-api", "pipreqs"]
+
+[[package]]
+name = "joblib"
+version = "1.3.1"
+description = "Lightweight pipelining with Python functions"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "joblib-1.3.1-py3-none-any.whl", hash = "sha256:89cf0529520e01b3de7ac7b74a8102c90d16d54c64b5dd98cafcd14307fdf915"},
+ {file = "joblib-1.3.1.tar.gz", hash = "sha256:1f937906df65329ba98013dc9692fe22a4c5e4a648112de500508b18a21b41e3"},
+]
+
+[[package]]
+name = "jsonschema"
+version = "4.17.3"
+description = "An implementation of JSON Schema validation for Python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "jsonschema-4.17.3-py3-none-any.whl", hash = "sha256:a870ad254da1a8ca84b6a2905cac29d265f805acc57af304784962a2aa6508f6"},
+ {file = "jsonschema-4.17.3.tar.gz", hash = "sha256:0f864437ab8b6076ba6707453ef8f98a6a0d512a80e93f8abdb676f737ecb60d"},
+]
+
+[package.dependencies]
+attrs = ">=17.4.0"
+pyrsistent = ">=0.14.0,<0.17.0 || >0.17.0,<0.17.1 || >0.17.1,<0.17.2 || >0.17.2"
+
+[package.extras]
+format = ["fqdn", "idna", "isoduration", "jsonpointer (>1.13)", "rfc3339-validator", "rfc3987", "uri-template", "webcolors (>=1.11)"]
+format-nongpl = ["fqdn", "idna", "isoduration", "jsonpointer (>1.13)", "rfc3339-validator", "rfc3986-validator (>0.1.0)", "uri-template", "webcolors (>=1.11)"]
+
+[[package]]
+name = "lazy-loader"
+version = "0.3"
+description = "lazy_loader"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "lazy_loader-0.3-py3-none-any.whl", hash = "sha256:1e9e76ee8631e264c62ce10006718e80b2cfc74340d17d1031e0f84af7478554"},
+ {file = "lazy_loader-0.3.tar.gz", hash = "sha256:3b68898e34f5b2a29daaaac172c6555512d0f32074f147e2254e4a6d9d838f37"},
+]
+
+[package.extras]
+lint = ["pre-commit (>=3.3)"]
+test = ["pytest (>=7.4)", "pytest-cov (>=4.1)"]
+
+[[package]]
+name = "libapi"
+version = "0.1.0"
+description = "Library for the API services"
+category = "main"
+optional = false
+python-versions = "3.9.15"
+files = []
+develop = true
+
+[package.dependencies]
+cryptography = "^41.0.1"
+environs = "^9.5.0"
+libcommon = {path = "../../libs/libcommon", develop = true}
+orjson = "^3.8.6"
+pyjwt = {version = "^2.6.0", extras = ["crypto"]}
+requests = "^2.28.2"
+starlette = "^0.28.0"
+starlette-prometheus = "^0.9.0"
+
+[package.source]
+type = "directory"
+url = "../../libs/libapi"
+
+[[package]]
+name = "libcommon"
+version = "0.6.8"
+description = "Library for utils common to all the services"
+category = "main"
+optional = false
+python-versions = "3.9.15"
+files = []
+develop = true
+
+[package.dependencies]
+appdirs = "^1.4.4"
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
+environs = "^9.5.0"
+huggingface-hub = "^0.15.1"
+mongo-types = "0.15.1"
+mongoengine = "^0.24.2"
+networkx = "^3.0"
+numba = "0.56.4"
+orjson = "^3.8.6"
+pandas = "^2.0.1"
+psutil = "^5.9.4"
+pydub = "^0.25.1"
+pymongo = {version = "^3.13.0", extras = ["srv"]}
+pytz = "^2020.1"
+requests = "^2.31.0"
+soundfile = ">=0.12.1"
+starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
+
+[package.source]
+type = "directory"
+url = "../../libs/libcommon"
+
+[[package]]
+name = "librosa"
+version = "0.10.0.post2"
+description = "Python module for audio and music processing"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"},
+ {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"},
+]
+
+[package.dependencies]
+audioread = ">=2.1.9"
+decorator = ">=4.3.0"
+joblib = ">=0.14"
+lazy-loader = ">=0.1"
+msgpack = ">=1.0"
+numba = ">=0.51.0"
+numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2"
+pooch = ">=1.0,<1.7"
+scikit-learn = ">=0.20.0"
+scipy = ">=1.2.0"
+soundfile = ">=0.12.1"
+soxr = ">=0.3.2"
+typing-extensions = ">=4.1.1"
+
+[package.extras]
+display = ["matplotlib (>=3.3.0)"]
+docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"]
+tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"]
+
+[[package]]
+name = "llvmlite"
+version = "0.39.1"
+description = "lightweight wrapper around basic LLVM functionality"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"},
+ {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"},
+ {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"},
+ {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"},
+ {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"},
+ {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"},
+ {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"},
+ {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"},
+ {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"},
+ {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"},
+ {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"},
+ {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"},
+ {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"},
+ {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"},
+ {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"},
+ {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"},
+ {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"},
+ {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"},
+ {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"},
+ {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"},
+ {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"},
+ {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"},
+]
+
+[[package]]
+name = "markdown-it-py"
+version = "3.0.0"
+description = "Python port of markdown-it. Markdown parsing, done right!"
+category = "dev"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "markdown-it-py-3.0.0.tar.gz", hash = "sha256:e3f60a94fa066dc52ec76661e37c851cb232d92f9886b15cb560aaada2df8feb"},
+ {file = "markdown_it_py-3.0.0-py3-none-any.whl", hash = "sha256:355216845c60bd96232cd8d8c40e8f9765cc86f46880e43a8fd22dc1a1a8cab1"},
+]
+
+[package.dependencies]
+mdurl = ">=0.1,<1.0"
+
+[package.extras]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
+plugins = ["mdit-py-plugins"]
+profiling = ["gprof2dot"]
+rtd = ["jupyter_sphinx", "mdit-py-plugins", "myst-parser", "pyyaml", "sphinx", "sphinx-copybutton", "sphinx-design", "sphinx_book_theme"]
+testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"]
+
+[[package]]
+name = "marshmallow"
+version = "3.19.0"
+description = "A lightweight library for converting complex datatypes to and from native Python datatypes."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "marshmallow-3.19.0-py3-none-any.whl", hash = "sha256:93f0958568da045b0021ec6aeb7ac37c81bfcccbb9a0e7ed8559885070b3a19b"},
+ {file = "marshmallow-3.19.0.tar.gz", hash = "sha256:90032c0fd650ce94b6ec6dc8dfeb0e3ff50c144586462c389b81a07205bedb78"},
+]
+
+[package.dependencies]
+packaging = ">=17.0"
+
+[package.extras]
+dev = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)", "pytest", "pytz", "simplejson", "tox"]
+docs = ["alabaster (==0.7.12)", "autodocsumm (==0.2.9)", "sphinx (==5.3.0)", "sphinx-issues (==3.0.1)", "sphinx-version-warning (==1.1.2)"]
+lint = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)"]
+tests = ["pytest", "pytz", "simplejson"]
+
+[[package]]
+name = "mccabe"
+version = "0.6.1"
+description = "McCabe checker, plugin for flake8"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "mccabe-0.6.1-py2.py3-none-any.whl", hash = "sha256:ab8a6258860da4b6677da4bd2fe5dc2c659cff31b3ee4f7f5d64e79735b80d42"},
+ {file = "mccabe-0.6.1.tar.gz", hash = "sha256:dd8d182285a0fe56bace7f45b5e7d1a6ebcbf524e8f3bd87eb0f125271b8831f"},
+]
+
+[[package]]
+name = "mdurl"
+version = "0.1.2"
+description = "Markdown URL utilities"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8"},
+ {file = "mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba"},
+]
+
+[[package]]
+name = "mongo-types"
+version = "0.15.1"
+description = "Type stubs for mongoengine w/ basic support for bson and pymongo"
+category = "main"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "mongo-types-0.15.1.tar.gz", hash = "sha256:0a9deeb7733ea7da5db3711d92e22d93556b522f860bbff82e5df44c53bd06a9"},
+ {file = "mongo_types-0.15.1-py3-none-any.whl", hash = "sha256:9417ae5b9a759c09630b5ec7d66904cc333c2d2fcfe75e2760a332ed5e267309"},
+]
+
+[[package]]
+name = "mongoengine"
+version = "0.24.2"
+description = "MongoEngine is a Python Object-Document Mapper for working with MongoDB."
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "mongoengine-0.24.2-py3-none-any.whl", hash = "sha256:f5c4e1b206b2ccffe4adc7a6283ed26dd799bd115a5fb1d2e885a075132cdb88"},
+ {file = "mongoengine-0.24.2.tar.gz", hash = "sha256:c76d49658575bb995682e2e77c8ef7cda63faf939415b32ee923745d120f8b02"},
+]
+
+[package.dependencies]
+pymongo = ">=3.4,<5.0"
+
+[[package]]
+name = "msgpack"
+version = "1.0.5"
+description = "MessagePack serializer"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:525228efd79bb831cf6830a732e2e80bc1b05436b086d4264814b4b2955b2fa9"},
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4f8d8b3bf1ff2672567d6b5c725a1b347fe838b912772aa8ae2bf70338d5a198"},
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cdc793c50be3f01106245a61b739328f7dccc2c648b501e237f0699fe1395b81"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb47c21a8a65b165ce29f2bec852790cbc04936f502966768e4aae9fa763cb7"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e42b9594cc3bf4d838d67d6ed62b9e59e201862a25e9a157019e171fbe672dd3"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:55b56a24893105dc52c1253649b60f475f36b3aa0fc66115bffafb624d7cb30b"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:1967f6129fc50a43bfe0951c35acbb729be89a55d849fab7686004da85103f1c"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:20a97bf595a232c3ee6d57ddaadd5453d174a52594bf9c21d10407e2a2d9b3bd"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d25dd59bbbbb996eacf7be6b4ad082ed7eacc4e8f3d2df1ba43822da9bfa122a"},
+ {file = "msgpack-1.0.5-cp310-cp310-win32.whl", hash = "sha256:382b2c77589331f2cb80b67cc058c00f225e19827dbc818d700f61513ab47bea"},
+ {file = "msgpack-1.0.5-cp310-cp310-win_amd64.whl", hash = "sha256:4867aa2df9e2a5fa5f76d7d5565d25ec76e84c106b55509e78c1ede0f152659a"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9f5ae84c5c8a857ec44dc180a8b0cc08238e021f57abdf51a8182e915e6299f0"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:9e6ca5d5699bcd89ae605c150aee83b5321f2115695e741b99618f4856c50898"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5494ea30d517a3576749cad32fa27f7585c65f5f38309c88c6d137877fa28a5a"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ab2f3331cb1b54165976a9d976cb251a83183631c88076613c6c780f0d6e45a"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:28592e20bbb1620848256ebc105fc420436af59515793ed27d5c77a217477705"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe5c63197c55bce6385d9aee16c4d0641684628f63ace85f73571e65ad1c1e8d"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:ed40e926fa2f297e8a653c954b732f125ef97bdd4c889f243182299de27e2aa9"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b2de4c1c0538dcb7010902a2b97f4e00fc4ddf2c8cda9749af0e594d3b7fa3d7"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf22a83f973b50f9d38e55c6aade04c41ddda19b00c4ebc558930d78eecc64ed"},
+ {file = "msgpack-1.0.5-cp311-cp311-win32.whl", hash = "sha256:c396e2cc213d12ce017b686e0f53497f94f8ba2b24799c25d913d46c08ec422c"},
+ {file = "msgpack-1.0.5-cp311-cp311-win_amd64.whl", hash = "sha256:6c4c68d87497f66f96d50142a2b73b97972130d93677ce930718f68828b382e2"},
+ {file = "msgpack-1.0.5-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:a2b031c2e9b9af485d5e3c4520f4220d74f4d222a5b8dc8c1a3ab9448ca79c57"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f837b93669ce4336e24d08286c38761132bc7ab29782727f8557e1eb21b2080"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1d46dfe3832660f53b13b925d4e0fa1432b00f5f7210eb3ad3bb9a13c6204a6"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:366c9a7b9057e1547f4ad51d8facad8b406bab69c7d72c0eb6f529cf76d4b85f"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4c075728a1095efd0634a7dccb06204919a2f67d1893b6aa8e00497258bf926c"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:f933bbda5a3ee63b8834179096923b094b76f0c7a73c1cfe8f07ad608c58844b"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:36961b0568c36027c76e2ae3ca1132e35123dcec0706c4b7992683cc26c1320c"},
+ {file = "msgpack-1.0.5-cp36-cp36m-win32.whl", hash = "sha256:b5ef2f015b95f912c2fcab19c36814963b5463f1fb9049846994b007962743e9"},
+ {file = "msgpack-1.0.5-cp36-cp36m-win_amd64.whl", hash = "sha256:288e32b47e67f7b171f86b030e527e302c91bd3f40fd9033483f2cacc37f327a"},
+ {file = "msgpack-1.0.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:137850656634abddfb88236008339fdaba3178f4751b28f270d2ebe77a563b6c"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c05a4a96585525916b109bb85f8cb6511db1c6f5b9d9cbcbc940dc6b4be944b"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:56a62ec00b636583e5cb6ad313bbed36bb7ead5fa3a3e38938503142c72cba4f"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef8108f8dedf204bb7b42994abf93882da1159728a2d4c5e82012edd92c9da9f"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1835c84d65f46900920b3708f5ba829fb19b1096c1800ad60bae8418652a951d"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:e57916ef1bd0fee4f21c4600e9d1da352d8816b52a599c46460e93a6e9f17086"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:17358523b85973e5f242ad74aa4712b7ee560715562554aa2134d96e7aa4cbbf"},
+ {file = "msgpack-1.0.5-cp37-cp37m-win32.whl", hash = "sha256:cb5aaa8c17760909ec6cb15e744c3ebc2ca8918e727216e79607b7bbce9c8f77"},
+ {file = "msgpack-1.0.5-cp37-cp37m-win_amd64.whl", hash = "sha256:ab31e908d8424d55601ad7075e471b7d0140d4d3dd3272daf39c5c19d936bd82"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:b72d0698f86e8d9ddf9442bdedec15b71df3598199ba33322d9711a19f08145c"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:379026812e49258016dd84ad79ac8446922234d498058ae1d415f04b522d5b2d"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:332360ff25469c346a1c5e47cbe2a725517919892eda5cfaffe6046656f0b7bb"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:476a8fe8fae289fdf273d6d2a6cb6e35b5a58541693e8f9f019bfe990a51e4ba"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a9985b214f33311df47e274eb788a5893a761d025e2b92c723ba4c63936b69b1"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:48296af57cdb1d885843afd73c4656be5c76c0c6328db3440c9601a98f303d87"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:addab7e2e1fcc04bd08e4eb631c2a90960c340e40dfc4a5e24d2ff0d5a3b3edb"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:916723458c25dfb77ff07f4c66aed34e47503b2eb3188b3adbec8d8aa6e00f48"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:821c7e677cc6acf0fd3f7ac664c98803827ae6de594a9f99563e48c5a2f27eb0"},
+ {file = "msgpack-1.0.5-cp38-cp38-win32.whl", hash = "sha256:1c0f7c47f0087ffda62961d425e4407961a7ffd2aa004c81b9c07d9269512f6e"},
+ {file = "msgpack-1.0.5-cp38-cp38-win_amd64.whl", hash = "sha256:bae7de2026cbfe3782c8b78b0db9cbfc5455e079f1937cb0ab8d133496ac55e1"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:20c784e66b613c7f16f632e7b5e8a1651aa5702463d61394671ba07b2fc9e025"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:266fa4202c0eb94d26822d9bfd7af25d1e2c088927fe8de9033d929dd5ba24c5"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:18334484eafc2b1aa47a6d42427da7fa8f2ab3d60b674120bce7a895a0a85bdd"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57e1f3528bd95cc44684beda696f74d3aaa8a5e58c816214b9046512240ef437"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:586d0d636f9a628ddc6a17bfd45aa5b5efaf1606d2b60fa5d87b8986326e933f"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a740fa0e4087a734455f0fc3abf5e746004c9da72fbd541e9b113013c8dc3282"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3055b0455e45810820db1f29d900bf39466df96ddca11dfa6d074fa47054376d"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:a61215eac016f391129a013c9e46f3ab308db5f5ec9f25811e811f96962599a8"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:362d9655cd369b08fda06b6657a303eb7172d5279997abe094512e919cf74b11"},
+ {file = "msgpack-1.0.5-cp39-cp39-win32.whl", hash = "sha256:ac9dd47af78cae935901a9a500104e2dea2e253207c924cc95de149606dc43cc"},
+ {file = "msgpack-1.0.5-cp39-cp39-win_amd64.whl", hash = "sha256:06f5174b5f8ed0ed919da0e62cbd4ffde676a374aba4020034da05fab67b9164"},
+ {file = "msgpack-1.0.5.tar.gz", hash = "sha256:c075544284eadc5cddc70f4757331d99dcbc16b2bbd4849d15f8aae4cf36d31c"},
+]
+
+[[package]]
+name = "multidict"
+version = "6.0.4"
+description = "multidict implementation"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"},
+ {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"},
+ {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"},
+ {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"},
+ {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"},
+ {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"},
+ {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"},
+ {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"},
+ {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"},
+ {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"},
+ {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"},
+ {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"},
+ {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"},
+ {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"},
+ {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"},
+ {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"},
+ {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"},
+ {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"},
+ {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"},
+ {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"},
+ {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"},
+ {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"},
+ {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"},
+ {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"},
+]
+
+[[package]]
+name = "multiprocess"
+version = "0.70.14"
+description = "better multiprocessing and multithreading in python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"},
+ {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"},
+ {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"},
+ {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"},
+ {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"},
+ {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"},
+ {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"},
+ {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"},
+ {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"},
+ {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"},
+ {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"},
+ {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"},
+ {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"},
+ {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"},
+]
+
+[package.dependencies]
+dill = ">=0.3.6"
+
+[[package]]
+name = "mypy"
+version = "1.4.1"
+description = "Optional static typing for Python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "mypy-1.4.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:566e72b0cd6598503e48ea610e0052d1b8168e60a46e0bfd34b3acf2d57f96a8"},
+ {file = "mypy-1.4.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ca637024ca67ab24a7fd6f65d280572c3794665eaf5edcc7e90a866544076878"},
+ {file = "mypy-1.4.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0dde1d180cd84f0624c5dcaaa89c89775550a675aff96b5848de78fb11adabcd"},
+ {file = "mypy-1.4.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:8c4d8e89aa7de683e2056a581ce63c46a0c41e31bd2b6d34144e2c80f5ea53dc"},
+ {file = "mypy-1.4.1-cp310-cp310-win_amd64.whl", hash = "sha256:bfdca17c36ae01a21274a3c387a63aa1aafe72bff976522886869ef131b937f1"},
+ {file = "mypy-1.4.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7549fbf655e5825d787bbc9ecf6028731973f78088fbca3a1f4145c39ef09462"},
+ {file = "mypy-1.4.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:98324ec3ecf12296e6422939e54763faedbfcc502ea4a4c38502082711867258"},
+ {file = "mypy-1.4.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:141dedfdbfe8a04142881ff30ce6e6653c9685b354876b12e4fe6c78598b45e2"},
+ {file = "mypy-1.4.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8207b7105829eca6f3d774f64a904190bb2231de91b8b186d21ffd98005f14a7"},
+ {file = "mypy-1.4.1-cp311-cp311-win_amd64.whl", hash = "sha256:16f0db5b641ba159eff72cff08edc3875f2b62b2fa2bc24f68c1e7a4e8232d01"},
+ {file = "mypy-1.4.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:470c969bb3f9a9efcedbadcd19a74ffb34a25f8e6b0e02dae7c0e71f8372f97b"},
+ {file = "mypy-1.4.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5952d2d18b79f7dc25e62e014fe5a23eb1a3d2bc66318df8988a01b1a037c5b"},
+ {file = "mypy-1.4.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:190b6bab0302cec4e9e6767d3eb66085aef2a1cc98fe04936d8a42ed2ba77bb7"},
+ {file = "mypy-1.4.1-cp37-cp37m-win_amd64.whl", hash = "sha256:9d40652cc4fe33871ad3338581dca3297ff5f2213d0df345bcfbde5162abf0c9"},
+ {file = "mypy-1.4.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:01fd2e9f85622d981fd9063bfaef1aed6e336eaacca00892cd2d82801ab7c042"},
+ {file = "mypy-1.4.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2460a58faeea905aeb1b9b36f5065f2dc9a9c6e4c992a6499a2360c6c74ceca3"},
+ {file = "mypy-1.4.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a2746d69a8196698146a3dbe29104f9eb6a2a4d8a27878d92169a6c0b74435b6"},
+ {file = "mypy-1.4.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:ae704dcfaa180ff7c4cfbad23e74321a2b774f92ca77fd94ce1049175a21c97f"},
+ {file = "mypy-1.4.1-cp38-cp38-win_amd64.whl", hash = "sha256:43d24f6437925ce50139a310a64b2ab048cb2d3694c84c71c3f2a1626d8101dc"},
+ {file = "mypy-1.4.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c482e1246726616088532b5e964e39765b6d1520791348e6c9dc3af25b233828"},
+ {file = "mypy-1.4.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:43b592511672017f5b1a483527fd2684347fdffc041c9ef53428c8dc530f79a3"},
+ {file = "mypy-1.4.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:34a9239d5b3502c17f07fd7c0b2ae6b7dd7d7f6af35fbb5072c6208e76295816"},
+ {file = "mypy-1.4.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5703097c4936bbb9e9bce41478c8d08edd2865e177dc4c52be759f81ee4dd26c"},
+ {file = "mypy-1.4.1-cp39-cp39-win_amd64.whl", hash = "sha256:e02d700ec8d9b1859790c0475df4e4092c7bf3272a4fd2c9f33d87fac4427b8f"},
+ {file = "mypy-1.4.1-py3-none-any.whl", hash = "sha256:45d32cec14e7b97af848bddd97d85ea4f0db4d5a149ed9676caa4eb2f7402bb4"},
+ {file = "mypy-1.4.1.tar.gz", hash = "sha256:9bbcd9ab8ea1f2e1c8031c21445b511442cc45c89951e49bbf852cbb70755b1b"},
+]
+
+[package.dependencies]
+mypy-extensions = ">=1.0.0"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=4.1.0"
+
+[package.extras]
+dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
+
+[[package]]
+name = "mypy-extensions"
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
+category = "dev"
+optional = false
+python-versions = ">=3.5"
+files = [
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
+]
+
+[[package]]
+name = "networkx"
+version = "3.1"
+description = "Python package for creating and manipulating graphs and networks"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"},
+ {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"},
+]
+
+[package.extras]
+default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"]
+developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"]
+doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"]
+extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"]
+test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"]
+
+[[package]]
+name = "numba"
+version = "0.56.4"
+description = "compiling Python code using LLVM"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"},
+ {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"},
+ {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"},
+ {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"},
+ {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"},
+ {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"},
+ {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"},
+ {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"},
+ {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"},
+ {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"},
+ {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"},
+ {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"},
+ {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"},
+ {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"},
+ {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"},
+ {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"},
+ {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"},
+ {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"},
+ {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"},
+ {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"},
+ {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"},
+ {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"},
+ {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"},
+ {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"},
+ {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"},
+ {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"},
+ {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"},
+ {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"},
+]
+
+[package.dependencies]
+llvmlite = ">=0.39.0dev0,<0.40"
+numpy = ">=1.18,<1.24"
+setuptools = "*"
+
+[[package]]
+name = "numpy"
+version = "1.23.5"
+description = "NumPy is the fundamental package for array computing with Python."
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"},
+ {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"},
+ {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"},
+ {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"},
+ {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"},
+ {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"},
+ {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"},
+ {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"},
+ {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"},
+ {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"},
+ {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"},
+ {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"},
+ {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"},
+ {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"},
+ {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"},
+ {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"},
+ {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"},
+ {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"},
+ {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"},
+ {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"},
+ {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"},
+ {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"},
+ {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"},
+ {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"},
+ {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"},
+ {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"},
+ {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"},
+ {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"},
+]
+
+[[package]]
+name = "orjson"
+version = "3.9.1"
+description = "Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "orjson-3.9.1-cp310-cp310-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:c4434b7b786fdc394b95d029fb99949d7c2b05bbd4bf5cb5e3906be96ffeee3b"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09faf14f74ed47e773fa56833be118e04aa534956f661eb491522970b7478e3b"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:503eb86a8d53a187fe66aa80c69295a3ca35475804da89a9547e4fce5f803822"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:20f2804b5a1dbd3609c086041bd243519224d47716efd7429db6c03ed28b7cc3"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0fd828e0656615a711c4cc4da70f3cac142e66a6703ba876c20156a14e28e3fa"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ec53d648176f873203b9c700a0abacab33ca1ab595066e9d616f98cdc56f4434"},
+ {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e186ae76b0d97c505500664193ddf508c13c1e675d9b25f1f4414a7606100da6"},
+ {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d4edee78503016f4df30aeede0d999b3cb11fb56f47e9db0e487bce0aaca9285"},
+ {file = "orjson-3.9.1-cp310-none-win_amd64.whl", hash = "sha256:a4cc5d21e68af982d9a2528ac61e604f092c60eed27aef3324969c68f182ec7e"},
+ {file = "orjson-3.9.1-cp311-cp311-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:761b6efd33c49de20dd73ce64cc59da62c0dab10aa6015f582680e0663cc792c"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:31229f9d0b8dc2ef7ee7e4393f2e4433a28e16582d4b25afbfccc9d68dc768f8"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b7ab18d55ecb1de543d452f0a5f8094b52282b916aa4097ac11a4c79f317b86"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:db774344c39041f4801c7dfe03483df9203cbd6c84e601a65908e5552228dd25"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ae47ef8c0fe89c4677db7e9e1fb2093ca6e66c3acbee5442d84d74e727edad5e"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:103952c21575b9805803c98add2eaecd005580a1e746292ed2ec0d76dd3b9746"},
+ {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:2cb0121e6f2c9da3eddf049b99b95fef0adf8480ea7cb544ce858706cdf916eb"},
+ {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:24d4ddaa2876e657c0fd32902b5c451fd2afc35159d66a58da7837357044b8c2"},
+ {file = "orjson-3.9.1-cp311-none-win_amd64.whl", hash = "sha256:0b53b5f72cf536dd8aa4fc4c95e7e09a7adb119f8ff8ee6cc60f735d7740ad6a"},
+ {file = "orjson-3.9.1-cp37-cp37m-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d4b68d01a506242316a07f1d2f29fb0a8b36cee30a7c35076f1ef59dce0890c1"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d9dd4abe6c6fd352f00f4246d85228f6a9847d0cc14f4d54ee553718c225388f"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9e20bca5e13041e31ceba7a09bf142e6d63c8a7467f5a9c974f8c13377c75af2"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d8ae0467d01eb1e4bcffef4486d964bfd1c2e608103e75f7074ed34be5df48cc"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:06f6ab4697fab090517f295915318763a97a12ee8186054adf21c1e6f6abbd3d"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8515867713301fa065c58ec4c9053ba1a22c35113ab4acad555317b8fd802e50"},
+ {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:393d0697d1dfa18d27d193e980c04fdfb672c87f7765b87952f550521e21b627"},
+ {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d96747662d3666f79119e5d28c124e7d356c7dc195cd4b09faea4031c9079dc9"},
+ {file = "orjson-3.9.1-cp37-none-win_amd64.whl", hash = "sha256:6d173d3921dd58a068c88ec22baea7dbc87a137411501618b1292a9d6252318e"},
+ {file = "orjson-3.9.1-cp38-cp38-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d1c2b0b4246c992ce2529fc610a446b945f1429445ece1c1f826a234c829a918"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:19f70ba1f441e1c4bb1a581f0baa092e8b3e3ce5b2aac2e1e090f0ac097966da"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:375d65f002e686212aac42680aed044872c45ee4bc656cf63d4a215137a6124a"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4751cee4a7b1daeacb90a7f5adf2170ccab893c3ab7c5cea58b45a13f89b30b3"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78d9a2a4b2302d5ebc3695498ebc305c3568e5ad4f3501eb30a6405a32d8af22"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:46b4facc32643b2689dfc292c0c463985dac4b6ab504799cf51fc3c6959ed668"},
+ {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:ec7c8a0f1bf35da0d5fd14f8956f3b82a9a6918a3c6963d718dfd414d6d3b604"},
+ {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:d3a40b0fbe06ccd4d6a99e523d20b47985655bcada8d1eba485b1b32a43e4904"},
+ {file = "orjson-3.9.1-cp38-none-win_amd64.whl", hash = "sha256:402f9d3edfec4560a98880224ec10eba4c5f7b4791e4bc0d4f4d8df5faf2a006"},
+ {file = "orjson-3.9.1-cp39-cp39-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:49c0d78dcd34626e2e934f1192d7c052b94e0ecadc5f386fd2bda6d2e03dadf5"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:125f63e56d38393daa0a1a6dc6fedefca16c538614b66ea5997c3bd3af35ef26"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:08927970365d2e1f3ce4894f9ff928a7b865d53f26768f1bbdd85dd4fee3e966"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f9a744e212d4780ecd67f4b6b128b2e727bee1df03e7059cddb2dfe1083e7dc4"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5d1dbf36db7240c61eec98c8d21545d671bce70be0730deb2c0d772e06b71af3"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:80a1e384626f76b66df615f7bb622a79a25c166d08c5d2151ffd41f24c4cc104"},
+ {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:15d28872fb055bf17ffca913826e618af61b2f689d2b170f72ecae1a86f80d52"},
+ {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:1e4d905338f9ef32c67566929dfbfbb23cc80287af8a2c38930fb0eda3d40b76"},
+ {file = "orjson-3.9.1-cp39-none-win_amd64.whl", hash = "sha256:48a27da6c7306965846565cc385611d03382bbd84120008653aa2f6741e2105d"},
+ {file = "orjson-3.9.1.tar.gz", hash = "sha256:db373a25ec4a4fccf8186f9a72a1b3442837e40807a736a815ab42481e83b7d0"},
+]
+
+[[package]]
+name = "packageurl-python"
+version = "0.11.1"
+description = "A purl aka. Package URL parser and builder"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "packageurl-python-0.11.1.tar.gz", hash = "sha256:bbcc53d2cb5920c815c1626c75992f319bfc450b73893fa7bd8aac5869aa49fe"},
+ {file = "packageurl_python-0.11.1-py3-none-any.whl", hash = "sha256:4bad1d3ea4feb5e7a1db5ca8fb690ac9c82ab18e08d500755947b853df68817d"},
+]
+
+[package.extras]
+build = ["wheel"]
+lint = ["black", "isort", "mypy"]
+test = ["pytest"]
+
+[[package]]
+name = "packaging"
+version = "23.1"
+description = "Core utilities for Python packages"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "packaging-23.1-py3-none-any.whl", hash = "sha256:994793af429502c4ea2ebf6bf664629d07c1a9fe974af92966e4b8d2df7edc61"},
+ {file = "packaging-23.1.tar.gz", hash = "sha256:a392980d2b6cffa644431898be54b0045151319d1e7ec34f0cfed48767dd334f"},
+]
+
+[[package]]
+name = "pandas"
+version = "2.0.3"
+description = "Powerful data structures for data analysis, time series, and statistics"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "pandas-2.0.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e4c7c9f27a4185304c7caf96dc7d91bc60bc162221152de697c98eb0b2648dd8"},
+ {file = "pandas-2.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f167beed68918d62bffb6ec64f2e1d8a7d297a038f86d4aed056b9493fca407f"},
+ {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ce0c6f76a0f1ba361551f3e6dceaff06bde7514a374aa43e33b588ec10420183"},
+ {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ba619e410a21d8c387a1ea6e8a0e49bb42216474436245718d7f2e88a2f8d7c0"},
+ {file = "pandas-2.0.3-cp310-cp310-win32.whl", hash = "sha256:3ef285093b4fe5058eefd756100a367f27029913760773c8bf1d2d8bebe5d210"},
+ {file = "pandas-2.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:9ee1a69328d5c36c98d8e74db06f4ad518a1840e8ccb94a4ba86920986bb617e"},
+ {file = "pandas-2.0.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b084b91d8d66ab19f5bb3256cbd5ea661848338301940e17f4492b2ce0801fe8"},
+ {file = "pandas-2.0.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:37673e3bdf1551b95bf5d4ce372b37770f9529743d2498032439371fc7b7eb26"},
+ {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b9cb1e14fdb546396b7e1b923ffaeeac24e4cedd14266c3497216dd4448e4f2d"},
+ {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9cd88488cceb7635aebb84809d087468eb33551097d600c6dad13602029c2df"},
+ {file = "pandas-2.0.3-cp311-cp311-win32.whl", hash = "sha256:694888a81198786f0e164ee3a581df7d505024fbb1f15202fc7db88a71d84ebd"},
+ {file = "pandas-2.0.3-cp311-cp311-win_amd64.whl", hash = "sha256:6a21ab5c89dcbd57f78d0ae16630b090eec626360085a4148693def5452d8a6b"},
+ {file = "pandas-2.0.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9e4da0d45e7f34c069fe4d522359df7d23badf83abc1d1cef398895822d11061"},
+ {file = "pandas-2.0.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:32fca2ee1b0d93dd71d979726b12b61faa06aeb93cf77468776287f41ff8fdc5"},
+ {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:258d3624b3ae734490e4d63c430256e716f488c4fcb7c8e9bde2d3aa46c29089"},
+ {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eae3dc34fa1aa7772dd3fc60270d13ced7346fcbcfee017d3132ec625e23bb0"},
+ {file = "pandas-2.0.3-cp38-cp38-win32.whl", hash = "sha256:f3421a7afb1a43f7e38e82e844e2bca9a6d793d66c1a7f9f0ff39a795bbc5e02"},
+ {file = "pandas-2.0.3-cp38-cp38-win_amd64.whl", hash = "sha256:69d7f3884c95da3a31ef82b7618af5710dba95bb885ffab339aad925c3e8ce78"},
+ {file = "pandas-2.0.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5247fb1ba347c1261cbbf0fcfba4a3121fbb4029d95d9ef4dc45406620b25c8b"},
+ {file = "pandas-2.0.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:81af086f4543c9d8bb128328b5d32e9986e0c84d3ee673a2ac6fb57fd14f755e"},
+ {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1994c789bf12a7c5098277fb43836ce090f1073858c10f9220998ac74f37c69b"},
+ {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5ec591c48e29226bcbb316e0c1e9423622bc7a4eaf1ef7c3c9fa1a3981f89641"},
+ {file = "pandas-2.0.3-cp39-cp39-win32.whl", hash = "sha256:04dbdbaf2e4d46ca8da896e1805bc04eb85caa9a82e259e8eed00254d5e0c682"},
+ {file = "pandas-2.0.3-cp39-cp39-win_amd64.whl", hash = "sha256:1168574b036cd8b93abc746171c9b4f1b83467438a5e45909fed645cf8692dbc"},
+ {file = "pandas-2.0.3.tar.gz", hash = "sha256:c02f372a88e0d17f36d3093a644c73cfc1788e876a7c4bcb4020a77512e2043c"},
+]
+
+[package.dependencies]
+numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""}
+python-dateutil = ">=2.8.2"
+pytz = ">=2020.1"
+tzdata = ">=2022.1"
+
+[package.extras]
+all = ["PyQt5 (>=5.15.1)", "SQLAlchemy (>=1.4.16)", "beautifulsoup4 (>=4.9.3)", "bottleneck (>=1.3.2)", "brotlipy (>=0.7.0)", "fastparquet (>=0.6.3)", "fsspec (>=2021.07.0)", "gcsfs (>=2021.07.0)", "html5lib (>=1.1)", "hypothesis (>=6.34.2)", "jinja2 (>=3.0.0)", "lxml (>=4.6.3)", "matplotlib (>=3.6.1)", "numba (>=0.53.1)", "numexpr (>=2.7.3)", "odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pandas-gbq (>=0.15.0)", "psycopg2 (>=2.8.6)", "pyarrow (>=7.0.0)", "pymysql (>=1.0.2)", "pyreadstat (>=1.1.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)", "python-snappy (>=0.6.0)", "pyxlsb (>=1.0.8)", "qtpy (>=2.2.0)", "s3fs (>=2021.08.0)", "scipy (>=1.7.1)", "tables (>=3.6.1)", "tabulate (>=0.8.9)", "xarray (>=0.21.0)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)", "zstandard (>=0.15.2)"]
+aws = ["s3fs (>=2021.08.0)"]
+clipboard = ["PyQt5 (>=5.15.1)", "qtpy (>=2.2.0)"]
+compression = ["brotlipy (>=0.7.0)", "python-snappy (>=0.6.0)", "zstandard (>=0.15.2)"]
+computation = ["scipy (>=1.7.1)", "xarray (>=0.21.0)"]
+excel = ["odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pyxlsb (>=1.0.8)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)"]
+feather = ["pyarrow (>=7.0.0)"]
+fss = ["fsspec (>=2021.07.0)"]
+gcp = ["gcsfs (>=2021.07.0)", "pandas-gbq (>=0.15.0)"]
+hdf5 = ["tables (>=3.6.1)"]
+html = ["beautifulsoup4 (>=4.9.3)", "html5lib (>=1.1)", "lxml (>=4.6.3)"]
+mysql = ["SQLAlchemy (>=1.4.16)", "pymysql (>=1.0.2)"]
+output-formatting = ["jinja2 (>=3.0.0)", "tabulate (>=0.8.9)"]
+parquet = ["pyarrow (>=7.0.0)"]
+performance = ["bottleneck (>=1.3.2)", "numba (>=0.53.1)", "numexpr (>=2.7.1)"]
+plot = ["matplotlib (>=3.6.1)"]
+postgresql = ["SQLAlchemy (>=1.4.16)", "psycopg2 (>=2.8.6)"]
+spss = ["pyreadstat (>=1.1.2)"]
+sql-other = ["SQLAlchemy (>=1.4.16)"]
+test = ["hypothesis (>=6.34.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)"]
+xml = ["lxml (>=4.6.3)"]
+
+[[package]]
+name = "pandas-stubs"
+version = "1.5.3.230321"
+description = "Type annotations for pandas"
+category = "dev"
+optional = false
+python-versions = ">=3.8,<3.12"
+files = [
+ {file = "pandas_stubs-1.5.3.230321-py3-none-any.whl", hash = "sha256:4bf36b3071dd55f0e558ac8efe07676a120f2ed89e7a3df0fb78ddf2733bf247"},
+ {file = "pandas_stubs-1.5.3.230321.tar.gz", hash = "sha256:2fa860df9e6058e9f0d2c09bc711c09abb8f0516eee7f0b9f9950d29b835fc6f"},
+]
+
+[package.dependencies]
+types-pytz = ">=2022.1.1"
+
+[[package]]
+name = "pathspec"
+version = "0.11.1"
+description = "Utility library for gitignore style pattern matching of file paths."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pathspec-0.11.1-py3-none-any.whl", hash = "sha256:d8af70af76652554bd134c22b3e8a1cc46ed7d91edcdd721ef1a0c51a84a5293"},
+ {file = "pathspec-0.11.1.tar.gz", hash = "sha256:2798de800fa92780e33acca925945e9a19a133b715067cf165b8866c15a31687"},
+]
+
+[[package]]
+name = "pbr"
+version = "5.11.1"
+description = "Python Build Reasonableness"
+category = "dev"
+optional = false
+python-versions = ">=2.6"
+files = [
+ {file = "pbr-5.11.1-py2.py3-none-any.whl", hash = "sha256:567f09558bae2b3ab53cb3c1e2e33e726ff3338e7bae3db5dc954b3a44eef12b"},
+ {file = "pbr-5.11.1.tar.gz", hash = "sha256:aefc51675b0b533d56bb5fd1c8c6c0522fe31896679882e1c4c63d5e4a0fccb3"},
+]
+
+[[package]]
+name = "pillow"
+version = "10.0.0"
+description = "Python Imaging Library (Fork)"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "Pillow-10.0.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:1f62406a884ae75fb2f818694469519fb685cc7eaff05d3451a9ebe55c646891"},
+ {file = "Pillow-10.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d5db32e2a6ccbb3d34d87c87b432959e0db29755727afb37290e10f6e8e62614"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:edf4392b77bdc81f36e92d3a07a5cd072f90253197f4a52a55a8cec48a12483b"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:520f2a520dc040512699f20fa1c363eed506e94248d71f85412b625026f6142c"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:8c11160913e3dd06c8ffdb5f233a4f254cb449f4dfc0f8f4549eda9e542c93d1"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a74ba0c356aaa3bb8e3eb79606a87669e7ec6444be352870623025d75a14a2bf"},
+ {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d5d0dae4cfd56969d23d94dc8e89fb6a217be461c69090768227beb8ed28c0a3"},
+ {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22c10cc517668d44b211717fd9775799ccec4124b9a7f7b3635fc5386e584992"},
+ {file = "Pillow-10.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:dffe31a7f47b603318c609f378ebcd57f1554a3a6a8effbc59c3c69f804296de"},
+ {file = "Pillow-10.0.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:9fb218c8a12e51d7ead2a7c9e101a04982237d4855716af2e9499306728fb485"},
+ {file = "Pillow-10.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d35e3c8d9b1268cbf5d3670285feb3528f6680420eafe35cccc686b73c1e330f"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3ed64f9ca2f0a95411e88a4efbd7a29e5ce2cea36072c53dd9d26d9c76f753b3"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0b6eb5502f45a60a3f411c63187db83a3d3107887ad0d036c13ce836f8a36f1d"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:c1fbe7621c167ecaa38ad29643d77a9ce7311583761abf7836e1510c580bf3dd"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:cd25d2a9d2b36fcb318882481367956d2cf91329f6892fe5d385c346c0649629"},
+ {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:3b08d4cc24f471b2c8ca24ec060abf4bebc6b144cb89cba638c720546b1cf538"},
+ {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d737a602fbd82afd892ca746392401b634e278cb65d55c4b7a8f48e9ef8d008d"},
+ {file = "Pillow-10.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:3a82c40d706d9aa9734289740ce26460a11aeec2d9c79b7af87bb35f0073c12f"},
+ {file = "Pillow-10.0.0-cp312-cp312-macosx_10_10_x86_64.whl", hash = "sha256:d80cf684b541685fccdd84c485b31ce73fc5c9b5d7523bf1394ce134a60c6883"},
+ {file = "Pillow-10.0.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:76de421f9c326da8f43d690110f0e79fe3ad1e54be811545d7d91898b4c8493e"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81ff539a12457809666fef6624684c008e00ff6bf455b4b89fd00a140eecd640"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce543ed15570eedbb85df19b0a1a7314a9c8141a36ce089c0a894adbfccb4568"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:685ac03cc4ed5ebc15ad5c23bc555d68a87777586d970c2c3e216619a5476223"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:d72e2ecc68a942e8cf9739619b7f408cc7b272b279b56b2c83c6123fcfa5cdff"},
+ {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:d50b6aec14bc737742ca96e85d6d0a5f9bfbded018264b3b70ff9d8c33485551"},
+ {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:00e65f5e822decd501e374b0650146063fbb30a7264b4d2744bdd7b913e0cab5"},
+ {file = "Pillow-10.0.0-cp312-cp312-win_amd64.whl", hash = "sha256:f31f9fdbfecb042d046f9d91270a0ba28368a723302786c0009ee9b9f1f60199"},
+ {file = "Pillow-10.0.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:349930d6e9c685c089284b013478d6f76e3a534e36ddfa912cde493f235372f3"},
+ {file = "Pillow-10.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3a684105f7c32488f7153905a4e3015a3b6c7182e106fe3c37fbb5ef3e6994c3"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b4f69b3700201b80bb82c3a97d5e9254084f6dd5fb5b16fc1a7b974260f89f43"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3f07ea8d2f827d7d2a49ecf1639ec02d75ffd1b88dcc5b3a61bbb37a8759ad8d"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:040586f7d37b34547153fa383f7f9aed68b738992380ac911447bb78f2abe530"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:f88a0b92277de8e3ca715a0d79d68dc82807457dae3ab8699c758f07c20b3c51"},
+ {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:c7cf14a27b0d6adfaebb3ae4153f1e516df54e47e42dcc073d7b3d76111a8d86"},
+ {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:3400aae60685b06bb96f99a21e1ada7bc7a413d5f49bce739828ecd9391bb8f7"},
+ {file = "Pillow-10.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:dbc02381779d412145331789b40cc7b11fdf449e5d94f6bc0b080db0a56ea3f0"},
+ {file = "Pillow-10.0.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9211e7ad69d7c9401cfc0e23d49b69ca65ddd898976d660a2fa5904e3d7a9baa"},
+ {file = "Pillow-10.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:faaf07ea35355b01a35cb442dd950d8f1bb5b040a7787791a535de13db15ed90"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c9f72a021fbb792ce98306ffb0c348b3c9cb967dce0f12a49aa4c3d3fdefa967"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f7c16705f44e0504a3a2a14197c1f0b32a95731d251777dcb060aa83022cb2d"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:76edb0a1fa2b4745fb0c99fb9fb98f8b180a1bbceb8be49b087e0b21867e77d3"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:368ab3dfb5f49e312231b6f27b8820c823652b7cd29cfbd34090565a015e99ba"},
+ {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:608bfdee0d57cf297d32bcbb3c728dc1da0907519d1784962c5f0c68bb93e5a3"},
+ {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5c6e3df6bdd396749bafd45314871b3d0af81ff935b2d188385e970052091017"},
+ {file = "Pillow-10.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:7be600823e4c8631b74e4a0d38384c73f680e6105a7d3c6824fcf226c178c7e6"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-macosx_10_10_x86_64.whl", hash = "sha256:92be919bbc9f7d09f7ae343c38f5bb21c973d2576c1d45600fce4b74bafa7ac0"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f8182b523b2289f7c415f589118228d30ac8c355baa2f3194ced084dac2dbba"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:38250a349b6b390ee6047a62c086d3817ac69022c127f8a5dc058c31ccef17f3"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:88af2003543cc40c80f6fca01411892ec52b11021b3dc22ec3bc9d5afd1c5334"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:c189af0545965fa8d3b9613cfdb0cd37f9d71349e0f7750e1fd704648d475ed2"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce7b031a6fc11365970e6a5686d7ba8c63e4c1cf1ea143811acbb524295eabed"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:db24668940f82321e746773a4bc617bfac06ec831e5c88b643f91f122a785684"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:efe8c0681042536e0d06c11f48cebe759707c9e9abf880ee213541c5b46c5bf3"},
+ {file = "Pillow-10.0.0.tar.gz", hash = "sha256:9c82b5b3e043c7af0d95792d0d20ccf68f61a1fec6b3530e718b688422727396"},
+]
+
+[package.extras]
+docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-removed-in", "sphinxext-opengraph"]
+tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"]
+
+[[package]]
+name = "pip"
+version = "23.1.2"
+description = "The PyPA recommended tool for installing Python packages."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pip-23.1.2-py3-none-any.whl", hash = "sha256:3ef6ac33239e4027d9a5598a381b9d30880a1477e50039db2eac6e8a8f6d1b18"},
+ {file = "pip-23.1.2.tar.gz", hash = "sha256:0e7c86f486935893c708287b30bd050a36ac827ec7fe5e43fe7cb198dd835fba"},
+]
+
+[[package]]
+name = "pip-api"
+version = "0.0.30"
+description = "An unofficial, importable pip API"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pip-api-0.0.30.tar.gz", hash = "sha256:a05df2c7aa9b7157374bcf4273544201a0c7bae60a9c65bcf84f3959ef3896f3"},
+ {file = "pip_api-0.0.30-py3-none-any.whl", hash = "sha256:2a0314bd31522eb9ffe8a99668b0d07fee34ebc537931e7b6483001dbedcbdc9"},
+]
+
+[package.dependencies]
+pip = "*"
+
+[[package]]
+name = "pip-audit"
+version = "2.6.0"
+description = "A tool for scanning Python environments for known vulnerabilities"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pip_audit-2.6.0-py3-none-any.whl", hash = "sha256:49e97e3d6663d2ed0c00b7a7c468afcb816beb3988f32f8496d3fe3927cfd627"},
+ {file = "pip_audit-2.6.0.tar.gz", hash = "sha256:6431c363efa80ef52c2599197c5b8a39ff8708ce316624b97fa35b5cdf493118"},
+]
+
+[package.dependencies]
+CacheControl = {version = ">=0.13.0", extras = ["filecache"]}
+cyclonedx-python-lib = ">=4.0,<5.0"
+html5lib = ">=1.1"
+packaging = ">=23.0.0"
+pip-api = ">=0.0.28"
+pip-requirements-parser = ">=32.0.0"
+requests = ">=2.31.0"
+rich = ">=12.4"
+toml = ">=0.10"
+
+[package.extras]
+dev = ["build", "bump (>=1.3.2)", "pip-audit[doc,lint,test]"]
+doc = ["pdoc"]
+lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.276)", "types-html5lib", "types-requests", "types-toml"]
+test = ["coverage[toml]", "pretend", "pytest", "pytest-cov"]
+
+[[package]]
+name = "pip-requirements-parser"
+version = "32.0.1"
+description = "pip requirements parser - a mostly correct pip requirements parsing library because it uses pip's own code."
+category = "dev"
+optional = false
+python-versions = ">=3.6.0"
+files = [
+ {file = "pip-requirements-parser-32.0.1.tar.gz", hash = "sha256:b4fa3a7a0be38243123cf9d1f3518da10c51bdb165a2b2985566247f9155a7d3"},
+ {file = "pip_requirements_parser-32.0.1-py3-none-any.whl", hash = "sha256:4659bc2a667783e7a15d190f6fccf8b2486685b6dba4c19c3876314769c57526"},
+]
+
+[package.dependencies]
+packaging = "*"
+pyparsing = "*"
+
+[package.extras]
+docs = ["Sphinx (>=3.3.1)", "doc8 (>=0.8.1)", "sphinx-rtd-theme (>=0.5.0)"]
+testing = ["aboutcode-toolkit (>=6.0.0)", "black", "pytest (>=6,!=7.0.0)", "pytest-xdist (>=2)"]
+
+[[package]]
+name = "platformdirs"
+version = "3.8.0"
+description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "platformdirs-3.8.0-py3-none-any.whl", hash = "sha256:ca9ed98ce73076ba72e092b23d3c93ea6c4e186b3f1c3dad6edd98ff6ffcca2e"},
+ {file = "platformdirs-3.8.0.tar.gz", hash = "sha256:b0cabcb11063d21a0b261d557acb0a9d2126350e63b70cdf7db6347baea456dc"},
+]
+
+[package.extras]
+docs = ["furo (>=2023.5.20)", "proselint (>=0.13)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.3)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)"]
+
+[[package]]
+name = "pluggy"
+version = "1.2.0"
+description = "plugin and hook calling mechanisms for python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pluggy-1.2.0-py3-none-any.whl", hash = "sha256:c2fd55a7d7a3863cba1a013e4e2414658b1d07b6bc57b3919e0c63c9abb99849"},
+ {file = "pluggy-1.2.0.tar.gz", hash = "sha256:d12f0c4b579b15f5e054301bb226ee85eeeba08ffec228092f8defbaa3a4c4b3"},
+]
+
+[package.extras]
+dev = ["pre-commit", "tox"]
+testing = ["pytest", "pytest-benchmark"]
+
+[[package]]
+name = "pooch"
+version = "1.6.0"
+description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\""
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"},
+ {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"},
+]
+
+[package.dependencies]
+appdirs = ">=1.3.0"
+packaging = ">=20.0"
+requests = ">=2.19.0"
+
+[package.extras]
+progress = ["tqdm (>=4.41.0,<5.0.0)"]
+sftp = ["paramiko (>=2.7.0)"]
+xxhash = ["xxhash (>=1.4.3)"]
+
+[[package]]
+name = "prometheus-client"
+version = "0.12.0"
+description = "Python client for the Prometheus monitoring system."
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "prometheus_client-0.12.0-py2.py3-none-any.whl", hash = "sha256:317453ebabff0a1b02df7f708efbab21e3489e7072b61cb6957230dd004a0af0"},
+ {file = "prometheus_client-0.12.0.tar.gz", hash = "sha256:1b12ba48cee33b9b0b9de64a1047cbd3c5f2d0ab6ebcead7ddda613a750ec3c5"},
+]
+
+[package.extras]
+twisted = ["twisted"]
+
+[[package]]
+name = "psutil"
+version = "5.9.5"
+description = "Cross-platform lib for process and system monitoring in Python."
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "psutil-5.9.5-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:be8929ce4313f9f8146caad4272f6abb8bf99fc6cf59344a3167ecd74f4f203f"},
+ {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:ab8ed1a1d77c95453db1ae00a3f9c50227ebd955437bcf2a574ba8adbf6a74d5"},
+ {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:4aef137f3345082a3d3232187aeb4ac4ef959ba3d7c10c33dd73763fbc063da4"},
+ {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:ea8518d152174e1249c4f2a1c89e3e6065941df2fa13a1ab45327716a23c2b48"},
+ {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:acf2aef9391710afded549ff602b5887d7a2349831ae4c26be7c807c0a39fac4"},
+ {file = "psutil-5.9.5-cp27-none-win32.whl", hash = "sha256:5b9b8cb93f507e8dbaf22af6a2fd0ccbe8244bf30b1baad6b3954e935157ae3f"},
+ {file = "psutil-5.9.5-cp27-none-win_amd64.whl", hash = "sha256:8c5f7c5a052d1d567db4ddd231a9d27a74e8e4a9c3f44b1032762bd7b9fdcd42"},
+ {file = "psutil-5.9.5-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:3c6f686f4225553615612f6d9bc21f1c0e305f75d7d8454f9b46e901778e7217"},
+ {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7a7dd9997128a0d928ed4fb2c2d57e5102bb6089027939f3b722f3a210f9a8da"},
+ {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:89518112647f1276b03ca97b65cc7f64ca587b1eb0278383017c2a0dcc26cbe4"},
+ {file = "psutil-5.9.5-cp36-abi3-win32.whl", hash = "sha256:104a5cc0e31baa2bcf67900be36acde157756b9c44017b86b2c049f11957887d"},
+ {file = "psutil-5.9.5-cp36-abi3-win_amd64.whl", hash = "sha256:b258c0c1c9d145a1d5ceffab1134441c4c5113b2417fafff7315a917a026c3c9"},
+ {file = "psutil-5.9.5-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:c607bb3b57dc779d55e1554846352b4e358c10fff3abf3514a7a6601beebdb30"},
+ {file = "psutil-5.9.5.tar.gz", hash = "sha256:5410638e4df39c54d957fc51ce03048acd8e6d60abc0f5107af51e5fb566eb3c"},
+]
+
+[package.extras]
+test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"]
+
+[[package]]
+name = "py-serializable"
+version = "0.11.1"
+description = "Library for serializing and deserializing Python Objects to and from JSON and XML."
+category = "dev"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "py-serializable-0.11.1.tar.gz", hash = "sha256:ba0e1287b9e4f645a5334f1913abd8e647e7250209f84f55dce3909498a6f586"},
+ {file = "py_serializable-0.11.1-py3-none-any.whl", hash = "sha256:79e21f0672822e6200b15f45ce9f636e8126466f62dbd7d488c67313c72b5c3e"},
+]
+
+[package.dependencies]
+defusedxml = ">=0.7.1,<0.8.0"
+
+[[package]]
+name = "pyarrow"
+version = "11.0.0"
+description = "Python library for Apache Arrow"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pyarrow-11.0.0-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:40bb42afa1053c35c749befbe72f6429b7b5f45710e85059cdd534553ebcf4f2"},
+ {file = "pyarrow-11.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:7c28b5f248e08dea3b3e0c828b91945f431f4202f1a9fe84d1012a761324e1ba"},
+ {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a37bc81f6c9435da3c9c1e767324ac3064ffbe110c4e460660c43e144be4ed85"},
+ {file = "pyarrow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad7c53def8dbbc810282ad308cc46a523ec81e653e60a91c609c2233ae407689"},
+ {file = "pyarrow-11.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:25aa11c443b934078bfd60ed63e4e2d42461682b5ac10f67275ea21e60e6042c"},
+ {file = "pyarrow-11.0.0-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:e217d001e6389b20a6759392a5ec49d670757af80101ee6b5f2c8ff0172e02ca"},
+ {file = "pyarrow-11.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ad42bb24fc44c48f74f0d8c72a9af16ba9a01a2ccda5739a517aa860fa7e3d56"},
+ {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2d942c690ff24a08b07cb3df818f542a90e4d359381fbff71b8f2aea5bf58841"},
+ {file = "pyarrow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f010ce497ca1b0f17a8243df3048055c0d18dcadbcc70895d5baf8921f753de5"},
+ {file = "pyarrow-11.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:2f51dc7ca940fdf17893227edb46b6784d37522ce08d21afc56466898cb213b2"},
+ {file = "pyarrow-11.0.0-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:1cbcfcbb0e74b4d94f0b7dde447b835a01bc1d16510edb8bb7d6224b9bf5bafc"},
+ {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaee8f79d2a120bf3e032d6d64ad20b3af6f56241b0ffc38d201aebfee879d00"},
+ {file = "pyarrow-11.0.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:410624da0708c37e6a27eba321a72f29d277091c8f8d23f72c92bada4092eb5e"},
+ {file = "pyarrow-11.0.0-cp37-cp37m-win_amd64.whl", hash = "sha256:2d53ba72917fdb71e3584ffc23ee4fcc487218f8ff29dd6df3a34c5c48fe8c06"},
+ {file = "pyarrow-11.0.0-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:f12932e5a6feb5c58192209af1d2607d488cb1d404fbc038ac12ada60327fa34"},
+ {file = "pyarrow-11.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:41a1451dd895c0b2964b83d91019e46f15b5564c7ecd5dcb812dadd3f05acc97"},
+ {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:becc2344be80e5dce4e1b80b7c650d2fc2061b9eb339045035a1baa34d5b8f1c"},
+ {file = "pyarrow-11.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f40be0d7381112a398b93c45a7e69f60261e7b0269cc324e9f739ce272f4f70"},
+ {file = "pyarrow-11.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:362a7c881b32dc6b0eccf83411a97acba2774c10edcec715ccaab5ebf3bb0835"},
+ {file = "pyarrow-11.0.0-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:ccbf29a0dadfcdd97632b4f7cca20a966bb552853ba254e874c66934931b9841"},
+ {file = "pyarrow-11.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3e99be85973592051e46412accea31828da324531a060bd4585046a74ba45854"},
+ {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:69309be84dcc36422574d19c7d3a30a7ea43804f12552356d1ab2a82a713c418"},
+ {file = "pyarrow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:da93340fbf6f4e2a62815064383605b7ffa3e9eeb320ec839995b1660d69f89b"},
+ {file = "pyarrow-11.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:caad867121f182d0d3e1a0d36f197df604655d0b466f1bc9bafa903aa95083e4"},
+ {file = "pyarrow-11.0.0.tar.gz", hash = "sha256:5461c57dbdb211a632a48facb9b39bbeb8a7905ec95d768078525283caef5f6d"},
+]
+
+[package.dependencies]
+numpy = ">=1.16.6"
+
+[[package]]
+name = "pycodestyle"
+version = "2.7.0"
+description = "Python style guide checker"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "pycodestyle-2.7.0-py2.py3-none-any.whl", hash = "sha256:514f76d918fcc0b55c6680472f0a37970994e07bbb80725808c17089be302068"},
+ {file = "pycodestyle-2.7.0.tar.gz", hash = "sha256:c389c1d06bf7904078ca03399a4816f974a1d590090fecea0c63ec26ebaf1cef"},
+]
+
+[[package]]
+name = "pycparser"
+version = "2.21"
+description = "C parser in Python"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"},
+ {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"},
+]
+
+[[package]]
+name = "pydub"
+version = "0.25.1"
+description = "Manipulate audio with an simple and easy high level interface"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"},
+ {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"},
+]
+
+[[package]]
+name = "pyflakes"
+version = "2.3.1"
+description = "passive checker of Python programs"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "pyflakes-2.3.1-py2.py3-none-any.whl", hash = "sha256:7893783d01b8a89811dd72d7dfd4d84ff098e5eed95cfa8905b22bbffe52efc3"},
+ {file = "pyflakes-2.3.1.tar.gz", hash = "sha256:f5bc8ecabc05bb9d291eb5203d6810b49040f6ff446a756326104746cc00c1db"},
+]
+
+[[package]]
+name = "pygments"
+version = "2.15.1"
+description = "Pygments is a syntax highlighting package written in Python."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "Pygments-2.15.1-py3-none-any.whl", hash = "sha256:db2db3deb4b4179f399a09054b023b6a586b76499d36965813c71aa8ed7b5fd1"},
+ {file = "Pygments-2.15.1.tar.gz", hash = "sha256:8ace4d3c1dd481894b2005f560ead0f9f19ee64fe983366be1a21e171d12775c"},
+]
+
+[package.extras]
+plugins = ["importlib-metadata"]
+
+[[package]]
+name = "pyjwt"
+version = "2.7.0"
+description = "JSON Web Token implementation in Python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "PyJWT-2.7.0-py3-none-any.whl", hash = "sha256:ba2b425b15ad5ef12f200dc67dd56af4e26de2331f965c5439994dad075876e1"},
+ {file = "PyJWT-2.7.0.tar.gz", hash = "sha256:bd6ca4a3c4285c1a2d4349e5a035fdf8fb94e04ccd0fcbe6ba289dae9cc3e074"},
+]
+
+[package.dependencies]
+cryptography = {version = ">=3.4.0", optional = true, markers = "extra == \"crypto\""}
+
+[package.extras]
+crypto = ["cryptography (>=3.4.0)"]
+dev = ["coverage[toml] (==5.0.4)", "cryptography (>=3.4.0)", "pre-commit", "pytest (>=6.0.0,<7.0.0)", "sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"]
+docs = ["sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"]
+tests = ["coverage[toml] (==5.0.4)", "pytest (>=6.0.0,<7.0.0)"]
+
+[[package]]
+name = "pymongo"
+version = "3.13.0"
+description = "Python driver for MongoDB <http://www.mongodb.org>"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl", hash = "sha256:3ad3a3df830f7df7e0856c2bdb54d19f5bf188bd7420985e18643b8e4d2a075f"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:b96e0e9d2d48948240b510bac81614458fc10adcd3a93240c2fd96448b4efd35"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9f592b202d77923498b32ddc5b376e5fa9ba280d3e16ed56cb8c932fe6d6a478"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:851f2bb52b5cb2f4711171ca925e0e05344a8452972a748a8a8ffdda1e1d72a7"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:1c9d23f62a3fa7523d849c4942acc0d9ff7081ebc00c808ee7cfdc070df0687f"},
+ {file = "pymongo-3.13.0-cp27-cp27m-win32.whl", hash = "sha256:a17b81f22398e3e0f72bdf938e98c810286994b2bcc0a125cd5ad8fd4ea54ad7"},
+ {file = "pymongo-3.13.0-cp27-cp27m-win_amd64.whl", hash = "sha256:4f6dd55dab77adf60b445c11f426ee5cdfa1b86f6d54cb937bfcbf09572333ab"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:776f90bf2252f90a4ae838e7917638894c6356bef7265f424592e2fd1f577d05"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:50b99f4d3eee6f03778fe841d6f470e6c18e744dc665156da6da3bc6e65b398d"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50a81b2d9f188c7909e0a1084fa969bb92a788076809c437ac1ae80393f46df9"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c7c45a8a1a752002b0a7c81ab3a4c5e3b6f67f9826b16fbe3943f5329f565f24"},
+ {file = "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1037097708498bdc85f23c8798a5c46c7bce432d77d23608ff14e0d831f1a971"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl", hash = "sha256:b5b733694e7df22d5c049581acfc487695a6ff813322318bed8dd66f79978636"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:d7c91747ec8dde51440dd594603158cc98abb3f7df84b2ed8a836f138285e4fb"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl", hash = "sha256:f4175fcdddf764d371ee52ec4505a40facee2533e84abf2953cda86d050cfa1f"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl", hash = "sha256:93d4e9a02c17813b34e4bd9f6fbf07310c140c8f74341537c24d07c1cdeb24d1"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl", hash = "sha256:3b261d593f2563299062733ae003a925420a86ff4ddda68a69097d67204e43f3"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl", hash = "sha256:172db03182a22e9002157b262c1ea3b0045c73d4ff465adc152ce5b4b0e7b8d4"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09de3bfc995ae8cb955abb0c9ae963c134dba1b5622be3bcc527b89b0fd4091c"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c0379447587ee4b8f983ba183202496e86c0358f47c45612619d634d1fcd82bd"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30245a8747dc90019a3c9ad9df987e0280a3ea632ad36227cde7d1d8dcba0830"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65b6fddf6a7b91da044f202771a38e71bbb9bf42720a406b26b25fe2256e7102"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5831a377d15a626fbec10890ffebc4c6abcd37e4126737932cd780a171eabdc1"},
+ {file = "pymongo-3.13.0-cp310-cp310-win32.whl", hash = "sha256:944249aa83dee314420c37d0f40c30a8f6dc4a3877566017b87062e53af449f4"},
+ {file = "pymongo-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea8824ebc9a1a5c8269e8f1e3989b5a6bec876726e2f3c33ebd036cb488277f0"},
+ {file = "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:bdd34c57b4da51a7961beb33645646d197e41f8517801dc76b37c1441e7a4e10"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26f9cc42a162faa241c82e117ac85734ae9f14343dc2df1c90c6b2181f791b22"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a82a1c10f5608e6494913faa169e213d703194bfca0aa710901f303be212414"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8927f22ef6a16229da7f18944deac8605bdc2c0858be5184259f2f7ce7fd4459"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e6f8191a282ef77e526f8f8f63753a437e4aa4bc78f5edd8b6b6ed0eaebd5363"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4d9ed67c987bf9ac2ac684590ba3d2599cdfb0f331ee3db607f9684469b3b59d"},
+ {file = "pymongo-3.13.0-cp311-cp311-win32.whl", hash = "sha256:e8f6979664ff477cd61b06bf8aba206df7b2334209815ab3b1019931dab643d6"},
+ {file = "pymongo-3.13.0-cp311-cp311-win_amd64.whl", hash = "sha256:174fd1000e896d0dfbc7f6d7e6a1992a4868796c7dec31679e38218c78d6a942"},
+ {file = "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl", hash = "sha256:d1ee773fb72ba024e7e3bb6ea8907fe52bccafcb5184aaced6bad995bd30ea20"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:28565e3dbd69fe5fe35a210067064dbb6ed5abe997079f653c19c873c3896fe6"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:5c1db7d366004d6c699eb08c716a63ae0a3e946d061cbebea65d7ce361950265"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e1956f3338c10308e2f99c2c9ff46ae412035cbcd7aaa76c39ccdb806854a247"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:10f0fddc1d63ba3d4a4bffcc7720184c1b7efd570726ad5e2f55818da320239f"},
+ {file = "pymongo-3.13.0-cp35-cp35m-win32.whl", hash = "sha256:570ae3365b23d4fd8c669cb57613b1a90b2757e993588d3370ef90945dbeec4b"},
+ {file = "pymongo-3.13.0-cp35-cp35m-win_amd64.whl", hash = "sha256:79f777eaf3f5b2c6d81f9ef00d87837001d7063302503bbcbfdbf3e9bc27c96f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl", hash = "sha256:d42eb29ba314adfd9c11234b4b646f61b0448bf9b00f14db4b317e6e4b947e77"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:e5e87c0eb774561c546f979342a8ff36ebee153c60a0b6c6b03ba989ceb9538c"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:0f2c5a5984599a88d087a15859860579b825098b473d8c843f1979a83d159f2e"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:59c98e86c5e861032b71e6e5b65f23e6afaacea6e82483b66f1191a5021a7b4f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:70b67390e27e58876853efbb87e43c85252de2515e2887f7dd901b4fa3d21973"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:42ba8606492d76e6f9e4c7a458ed4bc712603be393259a52450345f0945da2cf"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:0e5536994cf2d8488c6fd9dea71df3c4dbb3e0d2ba5e695da06d9142a29a0969"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:fe8194f107f0fa3cabd14e9e809f174eca335993c1db72d1e74e0f496e7afe1f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d593d50815771f517d3ac4367ff716e3f3c78edae51d98e1e25791459f8848ff"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5136ebe8da6a1604998a8eb96be55935aa5f7129c41cc7bddc400d48e8df43be"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a424bdedfd84454d2905a861e0d4bb947cc5bd024fdeb3600c1a97d2be0f4255"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5161167b3840e9c84c80f2534ea6a099f51749d5673b662a3dd248be17c3208"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:644470442beaf969df99c4e00367a817eee05f0bba5d888f1ba6fe97b5e1c102"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2406df90b2335371706c59b7d79e9633b81ed2a7ecd48c1faf8584552bdf2d90"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:222591b828de10ac90064047b5d4916953f38c38b155009c4b8b5e0d33117c2b"},
+ {file = "pymongo-3.13.0-cp36-cp36m-win32.whl", hash = "sha256:7cb987b199fa223ad78eebaa9fbc183d5a5944bfe568a9d6f617316ca1c1f32f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-win_amd64.whl", hash = "sha256:a6cbb73d9fc2282677e2b7a137d13da987bd0b13abd88ed27bba5534c226db06"},
+ {file = "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl", hash = "sha256:b1223b826acbef07a7f5eb9bf37247b0b580119916dca9eae19d92b1290f5855"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:398fb86d374dc351a4abc2e24cd15e5e14b2127f6d90ce0df3fdf2adcc55ac1b"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:9c3d07ea19cd2856d9943dce37e75d69ecbb5baf93c3e4c82f73b6075c481292"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:2943d739715f265a2983ac43747595b6af3312d0a370614040959fd293763adf"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:c3b70ed82f20d18d22eafc9bda0ea656605071762f7d31f3c5afc35c59d3393b"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:7ec2bb598847569ae34292f580842d37619eea3e546005042f485e15710180d5"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:8cc37b437cba909bef06499dadd91a39c15c14225e8d8c7870020049f8a549fe"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:65a063970e15a4f338f14b820561cf6cdaf2839691ac0adb2474ddff9d0b8b0b"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02f0e1a75d3bc0e16c7e15daf9c56185642be055e425f3b34888fc6eb1b22401"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16e74b9c2aca2734c7f49f00fe68d6830a30d26df60e2ace7fe40ccb92087b94"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:24e954be35ad4537840f20bbc8d75320ae647d3cb4fab12cb8fcd2d55f408e76"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a149377d1ff766fd618500798d0d94637f66d0ae222bb6d28f41f3e15c626297"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61660710b054ae52c8fc10368e91d74719eb05554b631d7f8ca93d21d2bff2e6"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4bbc0d27dfef7689285e54f2e0a224f0c7cd9d5c46d2638fabad5500b951c92f"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:9b2ed9c3b30f11cd4a3fbfc22167af7987b01b444215c2463265153fe7cf66d6"},
+ {file = "pymongo-3.13.0-cp37-cp37m-win32.whl", hash = "sha256:1c2c5e2b00e2fadcd590c0b2e293d71215e98ed1cb635cfca2be4998d197e534"},
+ {file = "pymongo-3.13.0-cp37-cp37m-win_amd64.whl", hash = "sha256:32eac95bbb030b2376ffd897376c6f870222a3457f01a9ce466b9057876132f8"},
+ {file = "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a796ef39dadf9d73af05d24937644d386495e43a7d13617aa3651d836da542c8"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:b6793baf4639c72a500698a49e9250b293e17ae1faf11ac1699d8141194786fe"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:80d8576b04d0824f63bf803190359c0d3bcb6e7fa63fefbd4bc0ceaa7faae38c"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:db2e11507fe9cc2a722be21ccc62c1b1295398fe9724c1f14900cdc7166fc0d7"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:b01ce58eec5edeededf1992d2dce63fb8565e437be12d6f139d75b15614c4d08"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:d1a19d6c5098f1f4e11430cd74621699453cbc534dd7ade9167e582f50814b19"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:7219b1a726ced3bacecabef9bd114529bbb69477901373e800d7d0140baadc95"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:2dae3b353a10c3767e0aa1c1492f2af388f1012b08117695ab3fd1f219e5814e"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:12721d926d43d33dd3318e58dce9b0250e8a9c6e1093fa8e09f4805193ff4b43"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6af0a4b17faf26779d5caee8542a4f2cba040cea27d3bffc476cbc6ccbd4c8ee"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:09b9d0f5a445c7e0ddcc021b09835aa6556f0166afc498f57dfdd72cdf6f02ad"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db5b4f8ad8607a3d612da1d4c89a84e4cf5c88f98b46365820d9babe5884ba45"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:34dbf5fecf653c152edb75a35a8b15dfdc4549473484ee768aeb12c97983cead"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:34cd48df7e1fc69222f296d8f69e3957eb7c6b5aa0709d3467184880ed7538c0"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c8f755ff1f4ab4ca790d1d6d3229006100b301475948021b6b2757822e0d6c97"},
+ {file = "pymongo-3.13.0-cp38-cp38-win32.whl", hash = "sha256:b0746d0d4535f56bbaa63a8f6da362f330804d578e66e126b226eebe76c2bf00"},
+ {file = "pymongo-3.13.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ad0515abb132f52ce9d8abd1a29681a1e65dba7b7fe13ea01e1a8db5715bf80"},
+ {file = "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:3c5cb6c93c94df76a879bad4b89db0104b01806d17c2b803c1316ba50962b6d6"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2e0854170813238f0c3131050c67cb1fb1ade75c93bf6cd156c1bd9a16095528"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:1410faa51ce835cc1234c99ec42e98ab4f3c6f50d92d86a2d4f6e11c97ee7a4e"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:d7910135f5de1c5c3578e61d6f4b087715b15e365f11d4fa51a9cee92988b2bd"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:028175dd8d2979a889153a2308e8e500b3df7d9e3fd1c33ca7fdeadf61cc87a2"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:2bfc39276c0e6d07c95bd1088b5003f049e986e089509f7dbd68bb7a4b1e65ac"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:4092b660ec720d44d3ca81074280dc25c7a3718df1b6c0fe9fe36ac6ed2833e4"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:5bdeb71a610a7b801416268e500e716d0fe693fb10d809e17f0fb3dac5be5a34"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aa3bca8e76f5c00ed2bb4325e0e383a547d71595926d5275d7c88175aaf7435e"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7c7cab8155f430ca460a6fc7ae8a705b34f3e279a57adb5f900eb81943ec777c"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4a32f3dfcca4a4816373bdb6256c18c78974ebb3430e7da988516cd95b2bd6e4"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30ed2788a6ec68743e2040ab1d16573d7d9f6e7333e45070ce9268cbc93d148c"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21e61a536ffed84d10376c21c13a6ed1ebefb61989a844952547c229d6aeedf3"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0665412dce26b2318092a33bd2d2327d487c4490cfcde158d6946d39b1e28d78"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:64ed1a5ce5e5926727eb0f87c698c4d9a7a9f7b0953683a65e9ce2b7cc5f8e91"},
+ {file = "pymongo-3.13.0-cp39-cp39-win32.whl", hash = "sha256:7593cb1214185a0c5b43b96effc51ce82ddc933298ee36db7dc2bd45d61b4adc"},
+ {file = "pymongo-3.13.0-cp39-cp39-win_amd64.whl", hash = "sha256:3cfc9bc1e8b5667bc1f3dbe46d2f85b3f24ff7533893bdc1203058012db2c046"},
+ {file = "pymongo-3.13.0.tar.gz", hash = "sha256:e22d6cf5802cd09b674c307cc9e03870b8c37c503ebec3d25b86f2ce8c535dc7"},
+]
+
+[package.dependencies]
+dnspython = {version = ">=1.16.0,<1.17.0", optional = true, markers = "extra == \"srv\""}
+
+[package.extras]
+aws = ["pymongo-auth-aws (<2.0.0)"]
+encryption = ["pymongocrypt (>=1.1.0,<2.0.0)"]
+gssapi = ["pykerberos"]
+ocsp = ["certifi", "pyopenssl (>=17.2.0)", "requests (<3.0.0)", "service-identity (>=18.1.0)"]
+snappy = ["python-snappy"]
+srv = ["dnspython (>=1.16.0,<1.17.0)"]
+tls = ["ipaddress"]
+zstd = ["zstandard"]
+
+[[package]]
+name = "pyparsing"
+version = "3.1.0"
+description = "pyparsing module - Classes and methods to define and execute parsing grammars"
+category = "dev"
+optional = false
+python-versions = ">=3.6.8"
+files = [
+ {file = "pyparsing-3.1.0-py3-none-any.whl", hash = "sha256:d554a96d1a7d3ddaf7183104485bc19fd80543ad6ac5bdb6426719d766fb06c1"},
+ {file = "pyparsing-3.1.0.tar.gz", hash = "sha256:edb662d6fe322d6e990b1594b5feaeadf806803359e3d4d42f11e295e588f0ea"},
+]
+
+[package.extras]
+diagrams = ["jinja2", "railroad-diagrams"]
+
+[[package]]
+name = "pyrsistent"
+version = "0.19.3"
+description = "Persistent/Functional/Immutable data structures"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pyrsistent-0.19.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:20460ac0ea439a3e79caa1dbd560344b64ed75e85d8703943e0b66c2a6150e4a"},
+ {file = "pyrsistent-0.19.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4c18264cb84b5e68e7085a43723f9e4c1fd1d935ab240ce02c0324a8e01ccb64"},
+ {file = "pyrsistent-0.19.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4b774f9288dda8d425adb6544e5903f1fb6c273ab3128a355c6b972b7df39dcf"},
+ {file = "pyrsistent-0.19.3-cp310-cp310-win32.whl", hash = "sha256:5a474fb80f5e0d6c9394d8db0fc19e90fa540b82ee52dba7d246a7791712f74a"},
+ {file = "pyrsistent-0.19.3-cp310-cp310-win_amd64.whl", hash = "sha256:49c32f216c17148695ca0e02a5c521e28a4ee6c5089f97e34fe24163113722da"},
+ {file = "pyrsistent-0.19.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:f0774bf48631f3a20471dd7c5989657b639fd2d285b861237ea9e82c36a415a9"},
+ {file = "pyrsistent-0.19.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3ab2204234c0ecd8b9368dbd6a53e83c3d4f3cab10ecaf6d0e772f456c442393"},
+ {file = "pyrsistent-0.19.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e42296a09e83028b3476f7073fcb69ffebac0e66dbbfd1bd847d61f74db30f19"},
+ {file = "pyrsistent-0.19.3-cp311-cp311-win32.whl", hash = "sha256:64220c429e42a7150f4bfd280f6f4bb2850f95956bde93c6fda1b70507af6ef3"},
+ {file = "pyrsistent-0.19.3-cp311-cp311-win_amd64.whl", hash = "sha256:016ad1afadf318eb7911baa24b049909f7f3bb2c5b1ed7b6a8f21db21ea3faa8"},
+ {file = "pyrsistent-0.19.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:c4db1bd596fefd66b296a3d5d943c94f4fac5bcd13e99bffe2ba6a759d959a28"},
+ {file = "pyrsistent-0.19.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aeda827381f5e5d65cced3024126529ddc4289d944f75e090572c77ceb19adbf"},
+ {file = "pyrsistent-0.19.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:42ac0b2f44607eb92ae88609eda931a4f0dfa03038c44c772e07f43e738bcac9"},
+ {file = "pyrsistent-0.19.3-cp37-cp37m-win32.whl", hash = "sha256:e8f2b814a3dc6225964fa03d8582c6e0b6650d68a232df41e3cc1b66a5d2f8d1"},
+ {file = "pyrsistent-0.19.3-cp37-cp37m-win_amd64.whl", hash = "sha256:c9bb60a40a0ab9aba40a59f68214eed5a29c6274c83b2cc206a359c4a89fa41b"},
+ {file = "pyrsistent-0.19.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:a2471f3f8693101975b1ff85ffd19bb7ca7dd7c38f8a81701f67d6b4f97b87d8"},
+ {file = "pyrsistent-0.19.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cc5d149f31706762c1f8bda2e8c4f8fead6e80312e3692619a75301d3dbb819a"},
+ {file = "pyrsistent-0.19.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3311cb4237a341aa52ab8448c27e3a9931e2ee09561ad150ba94e4cfd3fc888c"},
+ {file = "pyrsistent-0.19.3-cp38-cp38-win32.whl", hash = "sha256:f0e7c4b2f77593871e918be000b96c8107da48444d57005b6a6bc61fb4331b2c"},
+ {file = "pyrsistent-0.19.3-cp38-cp38-win_amd64.whl", hash = "sha256:c147257a92374fde8498491f53ffa8f4822cd70c0d85037e09028e478cababb7"},
+ {file = "pyrsistent-0.19.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:b735e538f74ec31378f5a1e3886a26d2ca6351106b4dfde376a26fc32a044edc"},
+ {file = "pyrsistent-0.19.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:99abb85579e2165bd8522f0c0138864da97847875ecbd45f3e7e2af569bfc6f2"},
+ {file = "pyrsistent-0.19.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a8cb235fa6d3fd7aae6a4f1429bbb1fec1577d978098da1252f0489937786f3"},
+ {file = "pyrsistent-0.19.3-cp39-cp39-win32.whl", hash = "sha256:c74bed51f9b41c48366a286395c67f4e894374306b197e62810e0fdaf2364da2"},
+ {file = "pyrsistent-0.19.3-cp39-cp39-win_amd64.whl", hash = "sha256:878433581fc23e906d947a6814336eee031a00e6defba224234169ae3d3d6a98"},
+ {file = "pyrsistent-0.19.3-py3-none-any.whl", hash = "sha256:ccf0d6bd208f8111179f0c26fdf84ed7c3891982f2edaeae7422575f47e66b64"},
+ {file = "pyrsistent-0.19.3.tar.gz", hash = "sha256:1a2994773706bbb4995c31a97bc94f1418314923bd1048c6d964837040376440"},
+]
+
+[[package]]
+name = "pytest"
+version = "7.4.0"
+description = "pytest: simple powerful testing with Python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pytest-7.4.0-py3-none-any.whl", hash = "sha256:78bf16451a2eb8c7a2ea98e32dc119fd2aa758f1d5d66dbf0a59d69a3969df32"},
+ {file = "pytest-7.4.0.tar.gz", hash = "sha256:b4bf8c45bd59934ed84001ad51e11b4ee40d40a1229d2c79f9c592b0a3f6bd8a"},
+]
+
+[package.dependencies]
+colorama = {version = "*", markers = "sys_platform == \"win32\""}
+exceptiongroup = {version = ">=1.0.0rc8", markers = "python_version < \"3.11\""}
+iniconfig = "*"
+packaging = "*"
+pluggy = ">=0.12,<2.0"
+tomli = {version = ">=1.0.0", markers = "python_version < \"3.11\""}
+
+[package.extras]
+testing = ["argcomplete", "attrs (>=19.2.0)", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.7.2)", "requests", "setuptools", "xmlschema"]
+
+[[package]]
+name = "pytest-cov"
+version = "2.12.1"
+description = "Pytest plugin for measuring coverage."
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"},
+ {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"},
+]
+
+[package.dependencies]
+coverage = ">=5.2.1"
+pytest = ">=4.6"
+toml = "*"
+
+[package.extras]
+testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"]
+
+[[package]]
+name = "python-dateutil"
+version = "2.8.2"
+description = "Extensions to the standard Python datetime module"
+category = "main"
+optional = false
+python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7"
+files = [
+ {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"},
+ {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"},
+]
+
+[package.dependencies]
+six = ">=1.5"
+
+[[package]]
+name = "python-dotenv"
+version = "1.0.0"
+description = "Read key-value pairs from a .env file and set them as environment variables"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"},
+ {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"},
+]
+
+[package.extras]
+cli = ["click (>=5.0)"]
+
+[[package]]
+name = "pytz"
+version = "2020.5"
+description = "World timezone definitions, modern and historical"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"},
+ {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"},
+]
+
+[[package]]
+name = "pyyaml"
+version = "6.0"
+description = "YAML parser and emitter for Python"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d4db7c7aef085872ef65a8fd7d6d09a14ae91f691dec3e87ee5ee0539d516f53"},
+ {file = "PyYAML-6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9df7ed3b3d2e0ecfe09e14741b857df43adb5a3ddadc919a2d94fbdf78fea53c"},
+ {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77f396e6ef4c73fdc33a9157446466f1cff553d979bd00ecb64385760c6babdc"},
+ {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a80a78046a72361de73f8f395f1f1e49f956c6be882eed58505a15f3e430962b"},
+ {file = "PyYAML-6.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f84fbc98b019fef2ee9a1cb3ce93e3187a6df0b2538a651bfb890254ba9f90b5"},
+ {file = "PyYAML-6.0-cp310-cp310-win32.whl", hash = "sha256:2cd5df3de48857ed0544b34e2d40e9fac445930039f3cfe4bcc592a1f836d513"},
+ {file = "PyYAML-6.0-cp310-cp310-win_amd64.whl", hash = "sha256:daf496c58a8c52083df09b80c860005194014c3698698d1a57cbcfa182142a3a"},
+ {file = "PyYAML-6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d4b0ba9512519522b118090257be113b9468d804b19d63c71dbcf4a48fa32358"},
+ {file = "PyYAML-6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:81957921f441d50af23654aa6c5e5eaf9b06aba7f0a19c18a538dc7ef291c5a1"},
+ {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:afa17f5bc4d1b10afd4466fd3a44dc0e245382deca5b3c353d8b757f9e3ecb8d"},
+ {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dbad0e9d368bb989f4515da330b88a057617d16b6a8245084f1b05400f24609f"},
+ {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:432557aa2c09802be39460360ddffd48156e30721f5e8d917f01d31694216782"},
+ {file = "PyYAML-6.0-cp311-cp311-win32.whl", hash = "sha256:bfaef573a63ba8923503d27530362590ff4f576c626d86a9fed95822a8255fd7"},
+ {file = "PyYAML-6.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b45c0191e6d66c470b6cf1b9531a771a83c1c4208272ead47a3ae4f2f603bf"},
+ {file = "PyYAML-6.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:897b80890765f037df3403d22bab41627ca8811ae55e9a722fd0392850ec4d86"},
+ {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50602afada6d6cbfad699b0c7bb50d5ccffa7e46a3d738092afddc1f9758427f"},
+ {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:48c346915c114f5fdb3ead70312bd042a953a8ce5c7106d5bfb1a5254e47da92"},
+ {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:98c4d36e99714e55cfbaaee6dd5badbc9a1ec339ebfc3b1f52e293aee6bb71a4"},
+ {file = "PyYAML-6.0-cp36-cp36m-win32.whl", hash = "sha256:0283c35a6a9fbf047493e3a0ce8d79ef5030852c51e9d911a27badfde0605293"},
+ {file = "PyYAML-6.0-cp36-cp36m-win_amd64.whl", hash = "sha256:07751360502caac1c067a8132d150cf3d61339af5691fe9e87803040dbc5db57"},
+ {file = "PyYAML-6.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:819b3830a1543db06c4d4b865e70ded25be52a2e0631ccd2f6a47a2822f2fd7c"},
+ {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:473f9edb243cb1935ab5a084eb238d842fb8f404ed2193a915d1784b5a6b5fc0"},
+ {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0ce82d761c532fe4ec3f87fc45688bdd3a4c1dc5e0b4a19814b9009a29baefd4"},
+ {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:231710d57adfd809ef5d34183b8ed1eeae3f76459c18fb4a0b373ad56bedcdd9"},
+ {file = "PyYAML-6.0-cp37-cp37m-win32.whl", hash = "sha256:c5687b8d43cf58545ade1fe3e055f70eac7a5a1a0bf42824308d868289a95737"},
+ {file = "PyYAML-6.0-cp37-cp37m-win_amd64.whl", hash = "sha256:d15a181d1ecd0d4270dc32edb46f7cb7733c7c508857278d3d378d14d606db2d"},
+ {file = "PyYAML-6.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b4624f379dab24d3725ffde76559cff63d9ec94e1736b556dacdfebe5ab6d4b"},
+ {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:213c60cd50106436cc818accf5baa1aba61c0189ff610f64f4a3e8c6726218ba"},
+ {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9fa600030013c4de8165339db93d182b9431076eb98eb40ee068700c9c813e34"},
+ {file = "PyYAML-6.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:277a0ef2981ca40581a47093e9e2d13b3f1fbbeffae064c1d21bfceba2030287"},
+ {file = "PyYAML-6.0-cp38-cp38-win32.whl", hash = "sha256:d4eccecf9adf6fbcc6861a38015c2a64f38b9d94838ac1810a9023a0609e1b78"},
+ {file = "PyYAML-6.0-cp38-cp38-win_amd64.whl", hash = "sha256:1e4747bc279b4f613a09eb64bba2ba602d8a6664c6ce6396a4d0cd413a50ce07"},
+ {file = "PyYAML-6.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:055d937d65826939cb044fc8c9b08889e8c743fdc6a32b33e2390f66013e449b"},
+ {file = "PyYAML-6.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e61ceaab6f49fb8bdfaa0f92c4b57bcfbea54c09277b1b4f7ac376bfb7a7c174"},
+ {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d67d839ede4ed1b28a4e8909735fc992a923cdb84e618544973d7dfc71540803"},
+ {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba8c411ef271aa037d7357a2bc8f9ee8b58b9965831d9e51baf703280dc73d3"},
+ {file = "PyYAML-6.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:40527857252b61eacd1d9af500c3337ba8deb8fc298940291486c465c8b46ec0"},
+ {file = "PyYAML-6.0-cp39-cp39-win32.whl", hash = "sha256:b5b9eccad747aabaaffbc6064800670f0c297e52c12754eb1d976c57e4f74dcb"},
+ {file = "PyYAML-6.0-cp39-cp39-win_amd64.whl", hash = "sha256:b3d267842bf12586ba6c734f89d1f5b871df0273157918b0ccefa29deb05c21c"},
+ {file = "PyYAML-6.0.tar.gz", hash = "sha256:68fb519c14306fec9720a2a5b45bc9f0c8d1b9c72adf45c37baedfcd949c35a2"},
+]
+
+[[package]]
+name = "requests"
+version = "2.31.0"
+description = "Python HTTP for Humans."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "requests-2.31.0-py3-none-any.whl", hash = "sha256:58cd2187c01e70e6e26505bca751777aa9f2ee0b7f4300988b709f44e013003f"},
+ {file = "requests-2.31.0.tar.gz", hash = "sha256:942c5a758f98d790eaed1a29cb6eefc7ffb0d1cf7af05c3d2791656dbd6ad1e1"},
+]
+
+[package.dependencies]
+certifi = ">=2017.4.17"
+charset-normalizer = ">=2,<4"
+idna = ">=2.5,<4"
+urllib3 = ">=1.21.1,<3"
+
+[package.extras]
+socks = ["PySocks (>=1.5.6,!=1.5.7)"]
+use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"]
+
+[[package]]
+name = "rfc3986"
+version = "1.5.0"
+description = "Validating URI References per RFC 3986"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "rfc3986-1.5.0-py2.py3-none-any.whl", hash = "sha256:a86d6e1f5b1dc238b218b012df0aa79409667bb209e58da56d0b94704e712a97"},
+ {file = "rfc3986-1.5.0.tar.gz", hash = "sha256:270aaf10d87d0d4e095063c65bf3ddbc6ee3d0b226328ce21e036f946e421835"},
+]
+
+[package.dependencies]
+idna = {version = "*", optional = true, markers = "extra == \"idna2008\""}
+
+[package.extras]
+idna2008 = ["idna"]
+
+[[package]]
+name = "rich"
+version = "13.4.2"
+description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal"
+category = "dev"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "rich-13.4.2-py3-none-any.whl", hash = "sha256:8f87bc7ee54675732fa66a05ebfe489e27264caeeff3728c945d25971b6485ec"},
+ {file = "rich-13.4.2.tar.gz", hash = "sha256:d653d6bccede5844304c605d5aac802c7cf9621efd700b46c7ec2b51ea914898"},
+]
+
+[package.dependencies]
+markdown-it-py = ">=2.2.0"
+pygments = ">=2.13.0,<3.0.0"
+
+[package.extras]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
+
+[[package]]
+name = "scikit-learn"
+version = "1.3.0"
+description = "A set of python modules for machine learning and data mining"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "scikit-learn-1.3.0.tar.gz", hash = "sha256:8be549886f5eda46436b6e555b0e4873b4f10aa21c07df45c4bc1735afbccd7a"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:981287869e576d42c682cf7ca96af0c6ac544ed9316328fd0d9292795c742cf5"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:436aaaae2c916ad16631142488e4c82f4296af2404f480e031d866863425d2a2"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c7e28d8fa47a0b30ae1bd7a079519dd852764e31708a7804da6cb6f8b36e3630"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ae80c08834a473d08a204d966982a62e11c976228d306a2648c575e3ead12111"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:552fd1b6ee22900cf1780d7386a554bb96949e9a359999177cf30211e6b20df6"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:79970a6d759eb00a62266a31e2637d07d2d28446fca8079cf9afa7c07b0427f8"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:850a00b559e636b23901aabbe79b73dc604b4e4248ba9e2d6e72f95063765603"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee04835fb016e8062ee9fe9074aef9b82e430504e420bff51e3e5fffe72750ca"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9d953531f5d9f00c90c34fa3b7d7cfb43ecff4c605dac9e4255a20b114a27369"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:151ac2bf65ccf363664a689b8beafc9e6aae36263db114b4ca06fbbbf827444a"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:6a885a9edc9c0a341cab27ec4f8a6c58b35f3d449c9d2503a6fd23e06bbd4f6a"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:9877af9c6d1b15486e18a94101b742e9d0d2f343d35a634e337411ddb57783f3"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c470f53cea065ff3d588050955c492793bb50c19a92923490d18fcb637f6383a"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fd6e2d7389542eae01077a1ee0318c4fec20c66c957f45c7aac0c6eb0fe3c612"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-win_amd64.whl", hash = "sha256:3a11936adbc379a6061ea32fa03338d4ca7248d86dd507c81e13af428a5bc1db"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:998d38fcec96584deee1e79cd127469b3ad6fefd1ea6c2dfc54e8db367eb396b"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:ded35e810438a527e17623ac6deae3b360134345b7c598175ab7741720d7ffa7"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0e8102d5036e28d08ab47166b48c8d5e5810704daecf3a476a4282d562be9a28"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7617164951c422747e7c32be4afa15d75ad8044f42e7d70d3e2e0429a50e6718"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-win_amd64.whl", hash = "sha256:1d54fb9e6038284548072df22fd34777e434153f7ffac72c8596f2d6987110dd"},
+]
+
+[package.dependencies]
+joblib = ">=1.1.1"
+numpy = ">=1.17.3"
+scipy = ">=1.5.0"
+threadpoolctl = ">=2.0.0"
+
+[package.extras]
+benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"]
+docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=6.0.0)", "sphinx-copybutton (>=0.5.2)", "sphinx-gallery (>=0.10.1)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"]
+examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"]
+tests = ["black (>=23.3.0)", "matplotlib (>=3.1.3)", "mypy (>=1.3)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=7.1.2)", "pytest-cov (>=2.9.0)", "ruff (>=0.0.272)", "scikit-image (>=0.16.2)"]
+
+[[package]]
+name = "scipy"
+version = "1.11.1"
+description = "Fundamental algorithms for scientific computing in Python"
+category = "main"
+optional = false
+python-versions = "<3.13,>=3.9"
+files = [
+ {file = "scipy-1.11.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:aec8c62fbe52914f9cf28d846cf0401dd80ab80788bbab909434eb336ed07c04"},
+ {file = "scipy-1.11.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:3b9963798df1d8a52db41a6fc0e6fa65b1c60e85d73da27ae8bb754de4792481"},
+ {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3e8eb42db36526b130dfbc417609498a6192381abc1975b91e3eb238e0b41c1a"},
+ {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:366a6a937110d80dca4f63b3f5b00cc89d36f678b2d124a01067b154e692bab1"},
+ {file = "scipy-1.11.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:08d957ca82d3535b3b9ba6c8ff355d78fe975271874e2af267cb5add5bd78625"},
+ {file = "scipy-1.11.1-cp310-cp310-win_amd64.whl", hash = "sha256:e866514bc2d660608447b6ba95c8900d591f2865c07cca0aa4f7ff3c4ca70f30"},
+ {file = "scipy-1.11.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ba94eeef3c9caa4cea7b402a35bb02a5714ee1ee77eb98aca1eed4543beb0f4c"},
+ {file = "scipy-1.11.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:512fdc18c65f76dadaca139348e525646d440220d8d05f6d21965b8d4466bccd"},
+ {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cce154372f0ebe88556ed06d7b196e9c2e0c13080ecb58d0f35062dc7cc28b47"},
+ {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b4bb943010203465ac81efa392e4645265077b4d9e99b66cf3ed33ae12254173"},
+ {file = "scipy-1.11.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:249cfa465c379c9bb2c20123001e151ff5e29b351cbb7f9c91587260602c58d0"},
+ {file = "scipy-1.11.1-cp311-cp311-win_amd64.whl", hash = "sha256:ffb28e3fa31b9c376d0fb1f74c1f13911c8c154a760312fbee87a21eb21efe31"},
+ {file = "scipy-1.11.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:39154437654260a52871dfde852adf1b93b1d1bc5dc0ffa70068f16ec0be2624"},
+ {file = "scipy-1.11.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:b588311875c58d1acd4ef17c983b9f1ab5391755a47c3d70b6bd503a45bfaf71"},
+ {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d51565560565a0307ed06fa0ec4c6f21ff094947d4844d6068ed04400c72d0c3"},
+ {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b41a0f322b4eb51b078cb3441e950ad661ede490c3aca66edef66f4b37ab1877"},
+ {file = "scipy-1.11.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:396fae3f8c12ad14c5f3eb40499fd06a6fef8393a6baa352a652ecd51e74e029"},
+ {file = "scipy-1.11.1-cp39-cp39-win_amd64.whl", hash = "sha256:be8c962a821957fdde8c4044efdab7a140c13294997a407eaee777acf63cbf0c"},
+ {file = "scipy-1.11.1.tar.gz", hash = "sha256:fb5b492fa035334fd249f0973cc79ecad8b09c604b42a127a677b45a9a3d4289"},
+]
+
+[package.dependencies]
+numpy = ">=1.21.6,<1.28.0"
+
+[package.extras]
+dev = ["click", "cython-lint (>=0.12.2)", "doit (>=0.36.0)", "mypy", "pycodestyle", "pydevtool", "rich-click", "ruff", "types-psutil", "typing_extensions"]
+doc = ["jupytext", "matplotlib (>2)", "myst-nb", "numpydoc", "pooch", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"]
+test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"]
+
+[[package]]
+name = "setuptools"
+version = "68.0.0"
+description = "Easily download, build, install, upgrade, and uninstall Python packages"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "setuptools-68.0.0-py3-none-any.whl", hash = "sha256:11e52c67415a381d10d6b462ced9cfb97066179f0e871399e006c4ab101fc85f"},
+ {file = "setuptools-68.0.0.tar.gz", hash = "sha256:baf1fdb41c6da4cd2eae722e135500da913332ab3f2f5c7d33af9b492acb5235"},
+]
+
+[package.extras]
+docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-hoverxref (<2)", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (==0.8.3)", "sphinx-reredirects", "sphinxcontrib-towncrier"]
+testing = ["build[virtualenv]", "filelock (>=3.4.0)", "flake8-2020", "ini2toml[lite] (>=0.9)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pip (>=19.1)", "pip-run (>=8.8)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-mypy (>=0.9.1)", "pytest-perf", "pytest-ruff", "pytest-timeout", "pytest-xdist", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel"]
+testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pytest", "pytest-enabler", "pytest-xdist", "tomli", "virtualenv (>=13.0.0)", "wheel"]
+
+[[package]]
+name = "six"
+version = "1.16.0"
+description = "Python 2 and 3 compatibility utilities"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*"
+files = [
+ {file = "six-1.16.0-py2.py3-none-any.whl", hash = "sha256:8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254"},
+ {file = "six-1.16.0.tar.gz", hash = "sha256:1e61c37477a1626458e36f7b1d82aa5c9b094fa4802892072e49de9c60c4c926"},
+]
+
+[[package]]
+name = "smmap"
+version = "5.0.0"
+description = "A pure Python implementation of a sliding window memory map manager"
+category = "dev"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "smmap-5.0.0-py3-none-any.whl", hash = "sha256:2aba19d6a040e78d8b09de5c57e96207b09ed71d8e55ce0959eeee6c8e190d94"},
+ {file = "smmap-5.0.0.tar.gz", hash = "sha256:c840e62059cd3be204b0c9c9f74be2c09d5648eddd4580d9314c3ecde0b30936"},
+]
+
+[[package]]
+name = "sniffio"
+version = "1.3.0"
+description = "Sniff out which async library your code is running under"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "sniffio-1.3.0-py3-none-any.whl", hash = "sha256:eecefdce1e5bbfb7ad2eeaabf7c1eeb404d7757c379bd1f7e5cce9d8bf425384"},
+ {file = "sniffio-1.3.0.tar.gz", hash = "sha256:e60305c5e5d314f5389259b7f22aaa33d8f7dee49763119234af3755c55b9101"},
+]
+
+[[package]]
+name = "sortedcontainers"
+version = "2.4.0"
+description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"},
+ {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"},
+]
+
+[[package]]
+name = "soundfile"
+version = "0.12.1"
+description = "An audio library based on libsndfile, CFFI and NumPy"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"},
+ {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"},
+ {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"},
+ {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
+ {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"},
+ {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"},
+ {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"},
+ {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"},
+]
+
+[package.dependencies]
+cffi = ">=1.0"
+
+[package.extras]
+numpy = ["numpy"]
+
+[[package]]
+name = "soxr"
+version = "0.3.5"
+description = "High quality, one-dimensional sample-rate conversion library"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "soxr-0.3.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:21c3aa3b2e12351b4310eea9d56cf52ec0769e6832f911ee6ba32f85b7c92baa"},
+ {file = "soxr-0.3.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ac3d7abc96082ff18a31fb1d678ddc0562f0c5e6d91f1cf0024b044989f63e93"},
+ {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:145e1e9d1b873a59ce0b5aa463ccacc40cf4bb74d9d8e6cef23433c752bfecea"},
+ {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4a376b3678801ffc1d0b9ae918b958be29d5884ca1b4bbeab32e29c567723bb3"},
+ {file = "soxr-0.3.5-cp310-cp310-win32.whl", hash = "sha256:907e2eb176bdefec40cc8f6015b7cef7f3d525a34219b3580b603ee696cb25c6"},
+ {file = "soxr-0.3.5-cp310-cp310-win_amd64.whl", hash = "sha256:0a6dbf9c7b7a3642916aba264c1d0b872b2e173be56204ed1895dbe381a32077"},
+ {file = "soxr-0.3.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:22c08a41e8eee99241fc0e9afb510f9bc7ada4a149d469b8891b596281a27db3"},
+ {file = "soxr-0.3.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bdacbe4ce4a1001043f1f8f0744480e294f5c5106e7861fd7033a83a869ba371"},
+ {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4b9acd5c42159eac4a90807524d9aa450d6ea0c750df94455c151165896d922e"},
+ {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:44b5d30f4e0d98b6d0034c00b04d5571ad070ce5cf3772f93193095b01b373de"},
+ {file = "soxr-0.3.5-cp311-cp311-win32.whl", hash = "sha256:677d5f44e85fdf0fdef33cd0e6087470732dd2e08fa73286c3659814110d1183"},
+ {file = "soxr-0.3.5-cp311-cp311-win_amd64.whl", hash = "sha256:a479984dd17bf0b50fb9fd659eba54a2dc59bf6eba9c29bb3a4a79ecec7dc9a4"},
+ {file = "soxr-0.3.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a2eb4f273ca14d7cfa882b234a03497d0e5dfd6f769a488a0962fe500450838c"},
+ {file = "soxr-0.3.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:a254c5e1adddb1204d8f327158b6c11a854908a10b5782103f38a67156108334"},
+ {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5766727dfee4d3616edd2a866a9a0d2f272c01545bed165c5a2676fbfd278723"},
+ {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2578664c6f94329685d864cdcae59794121bcbd808441572b2ffd01e7adc45dd"},
+ {file = "soxr-0.3.5-cp38-cp38-win32.whl", hash = "sha256:8a6f03804f48d986610eab8ca2b52e50b495f40ec13507741cd95f00ef7c2cb6"},
+ {file = "soxr-0.3.5-cp38-cp38-win_amd64.whl", hash = "sha256:592e9393e433501769a7e36b10460f4578c8e4ec3cddeec1aaaea4688e3558ef"},
+ {file = "soxr-0.3.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:93adbf04f51c7a5113059395633c2647f73bf195fa820256e1dd4da78af59275"},
+ {file = "soxr-0.3.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:37c4ec7ce275f284b0bf9741e5e6844a211ba1a850b2bf1c6a47769cdd3d109e"},
+ {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:18d5f3151fe4a88dfc37447bc6c397072aedcf36aeffb325cc817350ac5ad78e"},
+ {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:549a8358ba3b99a75588453c96aaa802e0c84d40957bdbe1f820f14f83a052ca"},
+ {file = "soxr-0.3.5-cp39-cp39-win32.whl", hash = "sha256:799df1875803dc9c4a4d3a7c285b8c1cb34b40dc39dba7ac7bac85d072f936a5"},
+ {file = "soxr-0.3.5-cp39-cp39-win_amd64.whl", hash = "sha256:4dd3f61929eb304c109f1f3b6cc8243e3a1a46d636d5bd86b5a7f50609ecd7d6"},
+ {file = "soxr-0.3.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:028af32bd4ce4b4c8183bb36da99e23ae954a114034d74538b4cae1bf40a0555"},
+ {file = "soxr-0.3.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1299e2aae4d659e222bcbbaca69a51ee99571486070ed49a393725ea6010a8e9"},
+ {file = "soxr-0.3.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:162f4e8b9a014c6819b4db6def2d43f7f4d97432ae33f2edfc8e5d0c97cf1cb3"},
+ {file = "soxr-0.3.5.tar.gz", hash = "sha256:b6b60f6381c98249a2f2a594e9234b647b78856c76c060597d53ed27b6efd249"},
+]
+
+[package.dependencies]
+numpy = "*"
+
+[package.extras]
+docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"]
+test = ["pytest"]
+
+[[package]]
+name = "starlette"
+version = "0.28.0"
+description = "The little ASGI library that shines."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "starlette-0.28.0-py3-none-any.whl", hash = "sha256:e58b9fc402c579950260fbb6d57173395c4e62804c40d3ede7e9ef1074f0c579"},
+ {file = "starlette-0.28.0.tar.gz", hash = "sha256:7bf3da5e997e796cc202cef2bd3f96a7d9b1e1943203c2fe2b42e020bc658482"},
+]
+
+[package.dependencies]
+anyio = ">=3.4.0,<5"
+typing-extensions = {version = ">=3.10.0", markers = "python_version < \"3.10\""}
+
+[package.extras]
+full = ["httpx (>=0.22.0)", "itsdangerous", "jinja2", "python-multipart", "pyyaml"]
+
+[[package]]
+name = "starlette-prometheus"
+version = "0.9.0"
+description = "Prometheus integration for Starlette"
+category = "main"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "starlette-prometheus-0.9.0.tar.gz", hash = "sha256:a52fb0f1df52b44a7a677a792759337ef0ce0d59ddf3e684a7d6459a93a90e99"},
+ {file = "starlette_prometheus-0.9.0-py3-none-any.whl", hash = "sha256:b4702e4ec67dce508d28551db0e45f12f58411afdb5d1078c92ff74331915381"},
+]
+
+[package.dependencies]
+prometheus_client = ">=0.12,<0.13"
+starlette = ">=0.12.2"
+
+[[package]]
+name = "stevedore"
+version = "5.1.0"
+description = "Manage dynamic plugins for Python applications"
+category = "dev"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "stevedore-5.1.0-py3-none-any.whl", hash = "sha256:8cc040628f3cea5d7128f2e76cf486b2251a4e543c7b938f58d9a377f6694a2d"},
+ {file = "stevedore-5.1.0.tar.gz", hash = "sha256:a54534acf9b89bc7ed264807013b505bf07f74dbe4bcfa37d32bd063870b087c"},
+]
+
+[package.dependencies]
+pbr = ">=2.0.0,<2.1.0 || >2.1.0"
+
+[[package]]
+name = "threadpoolctl"
+version = "3.1.0"
+description = "threadpoolctl"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"},
+ {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"},
+]
+
+[[package]]
+name = "toml"
+version = "0.10.2"
+description = "Python Library for Tom's Obvious, Minimal Language"
+category = "dev"
+optional = false
+python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*"
+files = [
+ {file = "toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b"},
+ {file = "toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f"},
+]
+
+[[package]]
+name = "tomli"
+version = "2.0.1"
+description = "A lil' TOML parser"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "tomli-2.0.1-py3-none-any.whl", hash = "sha256:939de3e7a6161af0c887ef91b7d41a53e7c5a1ca976325f429cb46ea9bc30ecc"},
+ {file = "tomli-2.0.1.tar.gz", hash = "sha256:de526c12914f0c550d15924c62d72abc48d6fe7364aa87328337a31007fe8a4f"},
+]
+
+[[package]]
+name = "tqdm"
+version = "4.65.0"
+description = "Fast, Extensible Progress Meter"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"},
+ {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"},
+]
+
+[package.dependencies]
+colorama = {version = "*", markers = "platform_system == \"Windows\""}
+
+[package.extras]
+dev = ["py-make (>=0.1.0)", "twine", "wheel"]
+notebook = ["ipywidgets (>=6)"]
+slack = ["slack-sdk"]
+telegram = ["requests"]
+
+[[package]]
+name = "types-jsonschema"
+version = "4.17.0.8"
+description = "Typing stubs for jsonschema"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-jsonschema-4.17.0.8.tar.gz", hash = "sha256:96a56990910f405e62de58862c0bbb3ac29ee6dba6d3d99aa0ba7f874cc547de"},
+ {file = "types_jsonschema-4.17.0.8-py3-none-any.whl", hash = "sha256:f5958eb7b53217dfb5125f0412aeaef226a8a9013eac95816c95b5b523f6796b"},
+]
+
+[[package]]
+name = "types-psutil"
+version = "5.9.5.15"
+description = "Typing stubs for psutil"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-psutil-5.9.5.15.tar.gz", hash = "sha256:943a5f9556c1995097c89f50274330dd8d71d17bd73cbc6b09c61dbf4e114cc6"},
+ {file = "types_psutil-5.9.5.15-py3-none-any.whl", hash = "sha256:7f183072d1fdb4e092fd6dd88ea017b957056bf10e58eafcdca26bf4372c0742"},
+]
+
+[[package]]
+name = "types-pytz"
+version = "2023.3.0.0"
+description = "Typing stubs for pytz"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-pytz-2023.3.0.0.tar.gz", hash = "sha256:ecdc70d543aaf3616a7e48631543a884f74205f284cefd6649ddf44c6a820aac"},
+ {file = "types_pytz-2023.3.0.0-py3-none-any.whl", hash = "sha256:4fc2a7fbbc315f0b6630e0b899fd6c743705abe1094d007b0e612d10da15e0f3"},
+]
+
+[[package]]
+name = "types-requests"
+version = "2.31.0.1"
+description = "Typing stubs for requests"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-requests-2.31.0.1.tar.gz", hash = "sha256:3de667cffa123ce698591de0ad7db034a5317457a596eb0b4944e5a9d9e8d1ac"},
+ {file = "types_requests-2.31.0.1-py3-none-any.whl", hash = "sha256:afb06ef8f25ba83d59a1d424bd7a5a939082f94b94e90ab5e6116bd2559deaa3"},
+]
+
+[package.dependencies]
+types-urllib3 = "*"
+
+[[package]]
+name = "types-urllib3"
+version = "1.26.25.13"
+description = "Typing stubs for urllib3"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-urllib3-1.26.25.13.tar.gz", hash = "sha256:3300538c9dc11dad32eae4827ac313f5d986b8b21494801f1bf97a1ac6c03ae5"},
+ {file = "types_urllib3-1.26.25.13-py3-none-any.whl", hash = "sha256:5dbd1d2bef14efee43f5318b5d36d805a489f6600252bb53626d4bfafd95e27c"},
+]
+
+[[package]]
+name = "typing-extensions"
+version = "4.7.1"
+description = "Backported and Experimental Type Hints for Python 3.7+"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "typing_extensions-4.7.1-py3-none-any.whl", hash = "sha256:440d5dd3af93b060174bf433bccd69b0babc3b15b1a8dca43789fd7f61514b36"},
+ {file = "typing_extensions-4.7.1.tar.gz", hash = "sha256:b75ddc264f0ba5615db7ba217daeb99701ad295353c45f9e95963337ceeeffb2"},
+]
+
+[[package]]
+name = "tzdata"
+version = "2023.3"
+description = "Provider of IANA time zone data"
+category = "main"
+optional = false
+python-versions = ">=2"
+files = [
+ {file = "tzdata-2023.3-py2.py3-none-any.whl", hash = "sha256:7e65763eef3120314099b6939b5546db7adce1e7d6f2e179e3df563c70511eda"},
+ {file = "tzdata-2023.3.tar.gz", hash = "sha256:11ef1e08e54acb0d4f95bdb1be05da659673de4acbd21bf9c69e94cc5e907a3a"},
+]
+
+[[package]]
+name = "urllib3"
+version = "2.0.3"
+description = "HTTP library with thread-safe connection pooling, file post, and more."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "urllib3-2.0.3-py3-none-any.whl", hash = "sha256:48e7fafa40319d358848e1bc6809b208340fafe2096f1725d05d67443d0483d1"},
+ {file = "urllib3-2.0.3.tar.gz", hash = "sha256:bee28b5e56addb8226c96f7f13ac28cb4c301dd5ea8a6ca179c0b9835e032825"},
+]
+
+[package.extras]
+brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)"]
+secure = ["certifi", "cryptography (>=1.9)", "idna (>=2.0.0)", "pyopenssl (>=17.1.0)", "urllib3-secure-extra"]
+socks = ["pysocks (>=1.5.6,!=1.5.7,<2.0)"]
+zstd = ["zstandard (>=0.18.0)"]
+
+[[package]]
+name = "uvicorn"
+version = "0.20.0"
+description = "The lightning-fast ASGI server."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "uvicorn-0.20.0-py3-none-any.whl", hash = "sha256:c3ed1598a5668208723f2bb49336f4509424ad198d6ab2615b7783db58d919fd"},
+ {file = "uvicorn-0.20.0.tar.gz", hash = "sha256:a4e12017b940247f836bc90b72e725d7dfd0c8ed1c51eb365f5ba30d9f5127d8"},
+]
+
+[package.dependencies]
+click = ">=7.0"
+h11 = ">=0.8"
+
+[package.extras]
+standard = ["colorama (>=0.4)", "httptools (>=0.5.0)", "python-dotenv (>=0.13)", "pyyaml (>=5.1)", "uvloop (>=0.14.0,!=0.15.0,!=0.15.1)", "watchfiles (>=0.13)", "websockets (>=10.4)"]
+
+[[package]]
+name = "watchdog"
+version = "2.3.1"
+description = "Filesystem events monitoring"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:d1f1200d4ec53b88bf04ab636f9133cb703eb19768a39351cee649de21a33697"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:564e7739abd4bd348aeafbf71cc006b6c0ccda3160c7053c4a53b67d14091d42"},
+ {file = "watchdog-2.3.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:95ad708a9454050a46f741ba5e2f3468655ea22da1114e4c40b8cbdaca572565"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:a073c91a6ef0dda488087669586768195c3080c66866144880f03445ca23ef16"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aa8b028750b43e80eea9946d01925168eeadb488dfdef1d82be4b1e28067f375"},
+ {file = "watchdog-2.3.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:964fd236cd443933268ae49b59706569c8b741073dbfd7ca705492bae9d39aab"},
+ {file = "watchdog-2.3.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:91fd146d723392b3e6eb1ac21f122fcce149a194a2ba0a82c5e4d0ee29cd954c"},
+ {file = "watchdog-2.3.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:efe3252137392a471a2174d721e1037a0e6a5da7beb72a021e662b7000a9903f"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:85bf2263290591b7c5fa01140601b64c831be88084de41efbcba6ea289874f44"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:8f2df370cd8e4e18499dd0bfdef476431bcc396108b97195d9448d90924e3131"},
+ {file = "watchdog-2.3.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ea5d86d1bcf4a9d24610aa2f6f25492f441960cf04aed2bd9a97db439b643a7b"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:6f5d0f7eac86807275eba40b577c671b306f6f335ba63a5c5a348da151aba0fc"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5b848c71ef2b15d0ef02f69da8cc120d335cec0ed82a3fa7779e27a5a8527225"},
+ {file = "watchdog-2.3.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0d9878be36d2b9271e3abaa6f4f051b363ff54dbbe7e7df1af3c920e4311ee43"},
+ {file = "watchdog-2.3.1-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:4cd61f98cb37143206818cb1786d2438626aa78d682a8f2ecee239055a9771d5"},
+ {file = "watchdog-2.3.1-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:3d2dbcf1acd96e7a9c9aefed201c47c8e311075105d94ce5e899f118155709fd"},
+ {file = "watchdog-2.3.1-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:03f342a9432fe08107defbe8e405a2cb922c5d00c4c6c168c68b633c64ce6190"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_aarch64.whl", hash = "sha256:7a596f9415a378d0339681efc08d2249e48975daae391d58f2e22a3673b977cf"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_armv7l.whl", hash = "sha256:0e1dd6d449267cc7d6935d7fe27ee0426af6ee16578eed93bacb1be9ff824d2d"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_i686.whl", hash = "sha256:7a1876f660e32027a1a46f8a0fa5747ad4fcf86cb451860eae61a26e102c8c79"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64.whl", hash = "sha256:2caf77ae137935c1466f8cefd4a3aec7017b6969f425d086e6a528241cba7256"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_ppc64le.whl", hash = "sha256:53f3e95081280898d9e4fc51c5c69017715929e4eea1ab45801d5e903dd518ad"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_s390x.whl", hash = "sha256:9da7acb9af7e4a272089bd2af0171d23e0d6271385c51d4d9bde91fe918c53ed"},
+ {file = "watchdog-2.3.1-py3-none-manylinux2014_x86_64.whl", hash = "sha256:8a4d484e846dcd75e96b96d80d80445302621be40e293bfdf34a631cab3b33dc"},
+ {file = "watchdog-2.3.1-py3-none-win32.whl", hash = "sha256:a74155398434937ac2780fd257c045954de5b11b5c52fc844e2199ce3eecf4cf"},
+ {file = "watchdog-2.3.1-py3-none-win_amd64.whl", hash = "sha256:5defe4f0918a2a1a4afbe4dbb967f743ac3a93d546ea4674567806375b024adb"},
+ {file = "watchdog-2.3.1-py3-none-win_ia64.whl", hash = "sha256:4109cccf214b7e3462e8403ab1e5b17b302ecce6c103eb2fc3afa534a7f27b96"},
+ {file = "watchdog-2.3.1.tar.gz", hash = "sha256:d9f9ed26ed22a9d331820a8432c3680707ea8b54121ddcc9dc7d9f2ceeb36906"},
+]
+
+[package.dependencies]
+PyYAML = {version = ">=3.10", optional = true, markers = "extra == \"watchmedo\""}
+
+[package.extras]
+watchmedo = ["PyYAML (>=3.10)"]
+
+[[package]]
+name = "webencodings"
+version = "0.5.1"
+description = "Character encoding aliases for legacy web content"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78"},
+ {file = "webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923"},
+]
+
+[[package]]
+name = "xxhash"
+version = "3.2.0"
+description = "Python binding for xxHash"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"},
+ {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"},
+ {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"},
+ {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"},
+ {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"},
+ {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"},
+ {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"},
+ {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"},
+ {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"},
+ {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"},
+ {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"},
+ {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"},
+ {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"},
+ {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"},
+ {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"},
+ {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"},
+ {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"},
+ {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"},
+ {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"},
+ {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"},
+ {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"},
+ {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"},
+ {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"},
+]
+
+[[package]]
+name = "yarl"
+version = "1.9.2"
+description = "Yet another URL library"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:8c2ad583743d16ddbdf6bb14b5cd76bf43b0d0006e918809d5d4ddf7bde8dd82"},
+ {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:82aa6264b36c50acfb2424ad5ca537a2060ab6de158a5bd2a72a032cc75b9eb8"},
+ {file = "yarl-1.9.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c0c77533b5ed4bcc38e943178ccae29b9bcf48ffd1063f5821192f23a1bd27b9"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee4afac41415d52d53a9833ebae7e32b344be72835bbb589018c9e938045a560"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9bf345c3a4f5ba7f766430f97f9cc1320786f19584acc7086491f45524a551ac"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2a96c19c52ff442a808c105901d0bdfd2e28575b3d5f82e2f5fd67e20dc5f4ea"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:891c0e3ec5ec881541f6c5113d8df0315ce5440e244a716b95f2525b7b9f3608"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c3a53ba34a636a256d767c086ceb111358876e1fb6b50dfc4d3f4951d40133d5"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:566185e8ebc0898b11f8026447eacd02e46226716229cea8db37496c8cdd26e0"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:2b0738fb871812722a0ac2154be1f049c6223b9f6f22eec352996b69775b36d4"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:32f1d071b3f362c80f1a7d322bfd7b2d11e33d2adf395cc1dd4df36c9c243095"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e9fdc7ac0d42bc3ea78818557fab03af6181e076a2944f43c38684b4b6bed8e3"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:56ff08ab5df8429901ebdc5d15941b59f6253393cb5da07b4170beefcf1b2528"},
+ {file = "yarl-1.9.2-cp310-cp310-win32.whl", hash = "sha256:8ea48e0a2f931064469bdabca50c2f578b565fc446f302a79ba6cc0ee7f384d3"},
+ {file = "yarl-1.9.2-cp310-cp310-win_amd64.whl", hash = "sha256:50f33040f3836e912ed16d212f6cc1efb3231a8a60526a407aeb66c1c1956dde"},
+ {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:646d663eb2232d7909e6601f1a9107e66f9791f290a1b3dc7057818fe44fc2b6"},
+ {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aff634b15beff8902d1f918012fc2a42e0dbae6f469fce134c8a0dc51ca423bb"},
+ {file = "yarl-1.9.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:a83503934c6273806aed765035716216cc9ab4e0364f7f066227e1aaea90b8d0"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b25322201585c69abc7b0e89e72790469f7dad90d26754717f3310bfe30331c2"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:22a94666751778629f1ec4280b08eb11815783c63f52092a5953faf73be24191"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8ec53a0ea2a80c5cd1ab397925f94bff59222aa3cf9c6da938ce05c9ec20428d"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:159d81f22d7a43e6eabc36d7194cb53f2f15f498dbbfa8edc8a3239350f59fe7"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:832b7e711027c114d79dffb92576acd1bd2decc467dec60e1cac96912602d0e6"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:95d2ecefbcf4e744ea952d073c6922e72ee650ffc79028eb1e320e732898d7e8"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:d4e2c6d555e77b37288eaf45b8f60f0737c9efa3452c6c44626a5455aeb250b9"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:783185c75c12a017cc345015ea359cc801c3b29a2966c2655cd12b233bf5a2be"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:b8cc1863402472f16c600e3e93d542b7e7542a540f95c30afd472e8e549fc3f7"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:822b30a0f22e588b32d3120f6d41e4ed021806418b4c9f0bc3048b8c8cb3f92a"},
+ {file = "yarl-1.9.2-cp311-cp311-win32.whl", hash = "sha256:a60347f234c2212a9f0361955007fcf4033a75bf600a33c88a0a8e91af77c0e8"},
+ {file = "yarl-1.9.2-cp311-cp311-win_amd64.whl", hash = "sha256:be6b3fdec5c62f2a67cb3f8c6dbf56bbf3f61c0f046f84645cd1ca73532ea051"},
+ {file = "yarl-1.9.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:38a3928ae37558bc1b559f67410df446d1fbfa87318b124bf5032c31e3447b74"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ac9bb4c5ce3975aeac288cfcb5061ce60e0d14d92209e780c93954076c7c4367"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3da8a678ca8b96c8606bbb8bfacd99a12ad5dd288bc6f7979baddd62f71c63ef"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:13414591ff516e04fcdee8dc051c13fd3db13b673c7a4cb1350e6b2ad9639ad3"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf74d08542c3a9ea97bb8f343d4fcbd4d8f91bba5ec9d5d7f792dbe727f88938"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6e7221580dc1db478464cfeef9b03b95c5852cc22894e418562997df0d074ccc"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:494053246b119b041960ddcd20fd76224149cfea8ed8777b687358727911dd33"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:52a25809fcbecfc63ac9ba0c0fb586f90837f5425edfd1ec9f3372b119585e45"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:e65610c5792870d45d7b68c677681376fcf9cc1c289f23e8e8b39c1485384185"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1b1bba902cba32cdec51fca038fd53f8beee88b77efc373968d1ed021024cc04"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:662e6016409828ee910f5d9602a2729a8a57d74b163c89a837de3fea050c7582"},
+ {file = "yarl-1.9.2-cp37-cp37m-win32.whl", hash = "sha256:f364d3480bffd3aa566e886587eaca7c8c04d74f6e8933f3f2c996b7f09bee1b"},
+ {file = "yarl-1.9.2-cp37-cp37m-win_amd64.whl", hash = "sha256:6a5883464143ab3ae9ba68daae8e7c5c95b969462bbe42e2464d60e7e2698368"},
+ {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5610f80cf43b6202e2c33ba3ec2ee0a2884f8f423c8f4f62906731d876ef4fac"},
+ {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b9a4e67ad7b646cd6f0938c7ebfd60e481b7410f574c560e455e938d2da8e0f4"},
+ {file = "yarl-1.9.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:83fcc480d7549ccebe9415d96d9263e2d4226798c37ebd18c930fce43dfb9574"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5fcd436ea16fee7d4207c045b1e340020e58a2597301cfbcfdbe5abd2356c2fb"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:84e0b1599334b1e1478db01b756e55937d4614f8654311eb26012091be109d59"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:3458a24e4ea3fd8930e934c129b676c27452e4ebda80fbe47b56d8c6c7a63a9e"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:838162460b3a08987546e881a2bfa573960bb559dfa739e7800ceeec92e64417"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f4e2d08f07a3d7d3e12549052eb5ad3eab1c349c53ac51c209a0e5991bbada78"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:de119f56f3c5f0e2fb4dee508531a32b069a5f2c6e827b272d1e0ff5ac040333"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:149ddea5abf329752ea5051b61bd6c1d979e13fbf122d3a1f9f0c8be6cb6f63c"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:674ca19cbee4a82c9f54e0d1eee28116e63bc6fd1e96c43031d11cbab8b2afd5"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:9b3152f2f5677b997ae6c804b73da05a39daa6a9e85a512e0e6823d81cdad7cc"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:5415d5a4b080dc9612b1b63cba008db84e908b95848369aa1da3686ae27b6d2b"},
+ {file = "yarl-1.9.2-cp38-cp38-win32.whl", hash = "sha256:f7a3d8146575e08c29ed1cd287068e6d02f1c7bdff8970db96683b9591b86ee7"},
+ {file = "yarl-1.9.2-cp38-cp38-win_amd64.whl", hash = "sha256:63c48f6cef34e6319a74c727376e95626f84ea091f92c0250a98e53e62c77c72"},
+ {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:75df5ef94c3fdc393c6b19d80e6ef1ecc9ae2f4263c09cacb178d871c02a5ba9"},
+ {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c027a6e96ef77d401d8d5a5c8d6bc478e8042f1e448272e8d9752cb0aff8b5c8"},
+ {file = "yarl-1.9.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f3b078dbe227f79be488ffcfc7a9edb3409d018e0952cf13f15fd6512847f3f7"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:59723a029760079b7d991a401386390c4be5bfec1e7dd83e25a6a0881859e716"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b03917871bf859a81ccb180c9a2e6c1e04d2f6a51d953e6a5cdd70c93d4e5a2a"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c1012fa63eb6c032f3ce5d2171c267992ae0c00b9e164efe4d73db818465fac3"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a74dcbfe780e62f4b5a062714576f16c2f3493a0394e555ab141bf0d746bb955"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8c56986609b057b4839968ba901944af91b8e92f1725d1a2d77cbac6972b9ed1"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:2c315df3293cd521033533d242d15eab26583360b58f7ee5d9565f15fee1bef4"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:b7232f8dfbd225d57340e441d8caf8652a6acd06b389ea2d3222b8bc89cbfca6"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:53338749febd28935d55b41bf0bcc79d634881195a39f6b2f767870b72514caf"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:066c163aec9d3d073dc9ffe5dd3ad05069bcb03fcaab8d221290ba99f9f69ee3"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:8288d7cd28f8119b07dd49b7230d6b4562f9b61ee9a4ab02221060d21136be80"},
+ {file = "yarl-1.9.2-cp39-cp39-win32.whl", hash = "sha256:b124e2a6d223b65ba8768d5706d103280914d61f5cae3afbc50fc3dfcc016623"},
+ {file = "yarl-1.9.2-cp39-cp39-win_amd64.whl", hash = "sha256:61016e7d582bc46a5378ffdd02cd0314fb8ba52f40f9cf4d9a5e7dbef88dee18"},
+ {file = "yarl-1.9.2.tar.gz", hash = "sha256:04ab9d4b9f587c06d801c2abfe9317b77cdf996c65a90d5e84ecc45010823571"},
+]
+
+[package.dependencies]
+idna = ">=2.0"
+multidict = ">=4.0"
+
+[metadata]
+lock-version = "2.0"
+python-versions = "3.9.15"
+content-hash = "c46efd608bb9fde12d1d6aec4385e7b26da6d08011d84e287d5a1663f3c93217"
diff --git a/services/rows/poetry.toml b/services/rows/poetry.toml
new file mode 100644
index 00000000..5fcef8cd
--- /dev/null
+++ b/services/rows/poetry.toml
@@ -0,0 +1,3 @@
+[virtualenvs]
+in-project = true
+prefer-active-python = true
diff --git a/services/rows/pyproject.toml b/services/rows/pyproject.toml
new file mode 100644
index 00000000..7072a9cf
--- /dev/null
+++ b/services/rows/pyproject.toml
@@ -0,0 +1,69 @@
+[tool.poetry]
+authors = ["The HuggingFace Authors."]
+description = "REST API app - /rows endpoint"
+name = "rows"
+version = "0.1.0"
+license = "Apache-2.0"
+
+[tool.poetry.dependencies]
+environs = "^9.5.0"
+jsonschema = "^4.17.0"
+libapi = {path = "../../libs/libapi", develop = true}
+pyarrow = "^11.0.0"
+python = "3.9.15"
+soundfile = ">=0.12.1"
+uvicorn = "^0.20.0"
+watchdog = { extras = ["watchmedo"], version = "^2.2.1" }
+
+[tool.poetry.group.dev.dependencies]
+bandit = "^1.7.4"
+black = "^22.12.0"
+flake8 = "^3.9.2"
+httpx = "^0.23.3"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pandas-stubs = "^1.5.3"
+pip-audit = "^2.5.4"
+pytest = "^7.2.1"
+pytest-cov = "^2.12.1"
+types-psutil = "^5.9.5"
+types-jsonschema = "^4.17.0.4"
+types-requests = "^2.28.11"
+
+[build-system]
+build-backend = "poetry.core.masonry.api"
+requires = ["poetry-core>=1.0.0"]
+
+[tool.pytest.ini_options]
+filterwarnings = ["ignore::DeprecationWarning"]
+markers = [
+ "real_dataset: tests on the Hub",
+ "wip: tests being developed"
+]
+
+[tool.coverage.run]
+source = ["rows"]
+
+[tool.isort]
+profile = "black"
+
+[tool.black]
+line-length = 119
+preview = true
+
+[tool.mypy]
+strict = true
+disallow_untyped_calls = false
+# ^ call to expected_algorithm.from_jwk forces to set this to false
+
+[[tool.mypy.overrides]]
+module = [
+ "datasets.*",
+ "huggingface_hub.*",
+ "prometheus_client.*",
+ "pyarrow.*",
+ "tqdm.*",
+ "fsspec.*"
+]
+# ^ prometheus_client is now typed, but starlette-prometheus requires an old version
+ignore_missing_imports = true
diff --git a/services/rows/src/rows/__init__.py b/services/rows/src/rows/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/services/rows/src/rows/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/services/rows/src/rows/app.py b/services/rows/src/rows/app.py
new file mode 100644
index 00000000..d792fc80
--- /dev/null
+++ b/services/rows/src/rows/app.py
@@ -0,0 +1,101 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+import uvicorn
+from libapi.config import UvicornConfig
+from libapi.jwt_token import fetch_jwt_public_key
+from libapi.routes.healthcheck import healthcheck_endpoint
+from libapi.routes.metrics import create_metrics_endpoint
+from libcommon.log import init_logging
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.resources import CacheMongoResource, QueueMongoResource, Resource
+from libcommon.storage import exists, init_cached_assets_dir, init_parquet_metadata_dir
+from starlette.applications import Starlette
+from starlette.middleware import Middleware
+from starlette.middleware.cors import CORSMiddleware
+from starlette.middleware.gzip import GZipMiddleware
+from starlette.routing import Route
+from starlette_prometheus import PrometheusMiddleware
+
+from rows.config import AppConfig
+from rows.routes.rows import create_rows_endpoint
+
+
+def create_app() -> Starlette:
+ app_config = AppConfig.from_env()
+ return create_app_with_config(app_config=app_config)
+
+
+def create_app_with_config(app_config: AppConfig) -> Starlette:
+ init_logging(level=app_config.log.level)
+ # ^ set first to have logs as soon as possible
+ cached_assets_directory = init_cached_assets_dir(directory=app_config.cached_assets.storage_directory)
+ if not exists(cached_assets_directory):
+ raise RuntimeError("The assets storage directory could not be accessed. Exiting.")
+ parquet_metadata_directory = init_parquet_metadata_dir(directory=app_config.parquet_metadata.storage_directory)
+ if not exists(parquet_metadata_directory):
+ raise RuntimeError("The parquet metadata storage directory could not be accessed. Exiting.")
+
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
+ hf_jwt_public_key = (
+ fetch_jwt_public_key(
+ url=app_config.api.hf_jwt_public_key_url,
+ hf_jwt_algorithm=app_config.api.hf_jwt_algorithm,
+ hf_timeout_seconds=app_config.api.hf_timeout_seconds,
+ )
+ if app_config.api.hf_jwt_public_key_url and app_config.api.hf_jwt_algorithm
+ else None
+ )
+
+ middleware = [
+ Middleware(
+ CORSMiddleware, allow_origins=["*"], allow_methods=["*"], allow_headers=["*"], allow_credentials=True
+ ),
+ Middleware(GZipMiddleware),
+ Middleware(PrometheusMiddleware, filter_unhandled_paths=True),
+ ]
+
+ cache_resource = CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url)
+ queue_resource = QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url)
+ resources: list[Resource] = [cache_resource, queue_resource]
+ if not cache_resource.is_available():
+ raise RuntimeError("The connection to the cache database could not be established. Exiting.")
+ if not queue_resource.is_available():
+ raise RuntimeError("The connection to the queue database could not be established. Exiting.")
+
+ routes = [
+ Route("/healthcheck", endpoint=healthcheck_endpoint),
+ Route("/metrics", endpoint=create_metrics_endpoint()),
+ # ^ called by Prometheus
+ Route(
+ "/rows",
+ endpoint=create_rows_endpoint(
+ processing_graph=processing_graph,
+ cached_assets_base_url=app_config.cached_assets.base_url,
+ cached_assets_directory=cached_assets_directory,
+ parquet_metadata_directory=parquet_metadata_directory,
+ hf_endpoint=app_config.common.hf_endpoint,
+ hf_token=app_config.common.hf_token,
+ hf_jwt_public_key=hf_jwt_public_key,
+ hf_jwt_algorithm=app_config.api.hf_jwt_algorithm,
+ external_auth_url=app_config.api.external_auth_url,
+ hf_timeout_seconds=app_config.api.hf_timeout_seconds,
+ max_age_long=app_config.api.max_age_long,
+ max_age_short=app_config.api.max_age_short,
+ cache_max_days=app_config.cache.max_days,
+ ),
+ ),
+ ]
+
+ return Starlette(routes=routes, middleware=middleware, on_shutdown=[resource.release for resource in resources])
+
+
+def start() -> None:
+ uvicorn_config = UvicornConfig.from_env()
+ uvicorn.run(
+ "app:create_app",
+ host=uvicorn_config.hostname,
+ port=uvicorn_config.port,
+ factory=True,
+ workers=uvicorn_config.num_workers,
+ )
diff --git a/services/rows/src/rows/config.py b/services/rows/src/rows/config.py
new file mode 100644
index 00000000..722f4b93
--- /dev/null
+++ b/services/rows/src/rows/config.py
@@ -0,0 +1,41 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+from dataclasses import dataclass, field
+
+from libapi.config import ApiConfig
+from libcommon.config import (
+ CacheConfig,
+ CachedAssetsConfig,
+ CommonConfig,
+ LogConfig,
+ ParquetMetadataConfig,
+ ProcessingGraphConfig,
+ QueueConfig,
+)
+
+
+@dataclass(frozen=True)
+class AppConfig:
+ api: ApiConfig = field(default_factory=ApiConfig)
+ cached_assets: CachedAssetsConfig = field(default_factory=CachedAssetsConfig)
+ cache: CacheConfig = field(default_factory=CacheConfig)
+ common: CommonConfig = field(default_factory=CommonConfig)
+ log: LogConfig = field(default_factory=LogConfig)
+ queue: QueueConfig = field(default_factory=QueueConfig)
+ processing_graph: ProcessingGraphConfig = field(default_factory=ProcessingGraphConfig)
+ parquet_metadata: ParquetMetadataConfig = field(default_factory=ParquetMetadataConfig)
+
+ @classmethod
+ def from_env(cls) -> "AppConfig":
+ common_config = CommonConfig.from_env()
+ return cls(
+ common=common_config,
+ cached_assets=CachedAssetsConfig.from_env(),
+ cache=CacheConfig.from_env(),
+ log=LogConfig.from_env(),
+ processing_graph=ProcessingGraphConfig.from_env(),
+ queue=QueueConfig.from_env(),
+ api=ApiConfig.from_env(hf_endpoint=common_config.hf_endpoint),
+ parquet_metadata=ParquetMetadataConfig.from_env(),
+ )
diff --git a/services/rows/src/rows/main.py b/services/rows/src/rows/main.py
new file mode 100644
index 00000000..b4790c09
--- /dev/null
+++ b/services/rows/src/rows/main.py
@@ -0,0 +1,7 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from rows.app import start
+
+if __name__ == "__main__":
+ start()
diff --git a/services/rows/src/rows/py.typed b/services/rows/src/rows/py.typed
new file mode 100644
index 00000000..e69de29b
diff --git a/services/rows/src/rows/routes/__init__.py b/services/rows/src/rows/routes/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/services/rows/src/rows/routes/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/services/api/src/api/routes/rows.py b/services/rows/src/rows/routes/rows.py
similarity index 100%
rename from services/api/src/api/routes/rows.py
rename to services/rows/src/rows/routes/rows.py
diff --git a/services/rows/tests/__init__.py b/services/rows/tests/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/services/rows/tests/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/services/rows/tests/conftest.py b/services/rows/tests/conftest.py
new file mode 100644
index 00000000..e6f5323b
--- /dev/null
+++ b/services/rows/tests/conftest.py
@@ -0,0 +1,104 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from pathlib import Path
+from typing import Iterator
+
+from libapi.config import UvicornConfig
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.queue import _clean_queue_database
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import _clean_cache_database
+from libcommon.storage import StrPath, init_cached_assets_dir, init_parquet_metadata_dir
+from pytest import MonkeyPatch, fixture
+
+from rows.config import AppConfig
+
+# Import fixture modules as plugins
+pytest_plugins = ["tests.fixtures.fsspec"]
+
+
+# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200
+@fixture(scope="session")
+def monkeypatch_session() -> Iterator[MonkeyPatch]:
+ monkeypatch_session = MonkeyPatch()
+ monkeypatch_session.setenv("CACHE_MONGO_DATABASE", "datasets_server_cache_test")
+ monkeypatch_session.setenv("QUEUE_MONGO_DATABASE", "datasets_server_queue_test")
+ monkeypatch_session.setenv("CACHED_ASSETS_BASE_URL", "http://localhost/cached-assets")
+ hostname = "localhost"
+ port = "8888"
+ monkeypatch_session.setenv("API_HF_TIMEOUT_SECONDS", "10")
+ monkeypatch_session.setenv("API_UVICORN_HOSTNAME", hostname)
+ monkeypatch_session.setenv("API_UVICORN_PORT", port)
+ monkeypatch_session.setenv("COMMON_HF_ENDPOINT", f"http://{hostname}:{port}")
+ yield monkeypatch_session
+ monkeypatch_session.undo()
+
+
+@fixture(scope="session")
+def app_config(monkeypatch_session: MonkeyPatch) -> AppConfig:
+ app_config = AppConfig.from_env()
+ if "test" not in app_config.cache.mongo_database or "test" not in app_config.queue.mongo_database:
+ raise ValueError("Test must be launched on a test mongo database")
+ return app_config
+
+
+@fixture(scope="session")
+def processing_graph(app_config: AppConfig) -> ProcessingGraph:
+ return ProcessingGraph(app_config.processing_graph.specification)
+
+
+@fixture(scope="session")
+def rows_endpoint() -> str:
+ return "/rows"
+
+
+@fixture(autouse=True)
+def cache_mongo_resource(app_config: AppConfig) -> Iterator[CacheMongoResource]:
+ with CacheMongoResource(database=app_config.cache.mongo_database, host=app_config.cache.mongo_url) as resource:
+ yield resource
+ _clean_cache_database()
+
+
+@fixture(autouse=True)
+def queue_mongo_resource(app_config: AppConfig) -> Iterator[QueueMongoResource]:
+ with QueueMongoResource(database=app_config.queue.mongo_database, host=app_config.queue.mongo_url) as resource:
+ yield resource
+ _clean_queue_database()
+
+
+@fixture(scope="session")
+def uvicorn_config(monkeypatch_session: MonkeyPatch) -> UvicornConfig:
+ return UvicornConfig.from_env()
+
+
+@fixture(scope="session")
+def httpserver_listen_address(uvicorn_config: UvicornConfig) -> tuple[str, int]:
+ return (uvicorn_config.hostname, uvicorn_config.port)
+
+
+@fixture(scope="session")
+def hf_endpoint(app_config: AppConfig) -> str:
+ return app_config.common.hf_endpoint
+
+
+@fixture(scope="session")
+def hf_auth_path(app_config: AppConfig) -> str:
+ return app_config.api.hf_auth_path
+
+
+@fixture
+def cached_assets_directory(app_config: AppConfig) -> StrPath:
+ return init_cached_assets_dir(app_config.cached_assets.storage_directory)
+
+
+@fixture
+def parquet_metadata_directory(app_config: AppConfig) -> StrPath:
+ return init_parquet_metadata_dir(app_config.parquet_metadata.storage_directory)
+
+
+@fixture
+def image_path() -> str:
+ image_path = Path(__file__).resolve().parent / "data" / "test_image_rgb.jpg"
+ assert image_path.is_file()
+ return str(image_path)
diff --git a/services/api/tests/data/test_image_rgb.jpg b/services/rows/tests/data/test_image_rgb.jpg
similarity index 100%
rename from services/api/tests/data/test_image_rgb.jpg
rename to services/rows/tests/data/test_image_rgb.jpg
diff --git a/services/api/tests/fixtures/fsspec.py b/services/rows/tests/fixtures/fsspec.py
similarity index 100%
rename from services/api/tests/fixtures/fsspec.py
rename to services/rows/tests/fixtures/fsspec.py
diff --git a/services/rows/tests/routes/__init__.py b/services/rows/tests/routes/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/services/rows/tests/routes/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/services/api/tests/routes/test_rows.py b/services/rows/tests/routes/test_rows.py
similarity index 98%
rename from services/api/tests/routes/test_rows.py
rename to services/rows/tests/routes/test_rows.py
index e8c944db..fdb55d06 100644
--- a/services/api/tests/routes/test_rows.py
+++ b/services/rows/tests/routes/test_rows.py
@@ -30,2 +30,2 @@ from libcommon.viewer_utils.asset import update_last_modified_date_of_rows_in_as
-from api.config import AppConfig
-from api.routes.rows import clean_cached_assets, create_response
+from rows.config import AppConfig
+from rows.routes.rows import clean_cached_assets, create_response
@@ -41 +41 @@ def enable_parquet_metadata_on_all_datasets() -> Generator[None, None, None]:
- with patch("api.routes.rows.ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST", "all"):
+ with patch("rows.routes.rows.ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST", "all"):
@@ -470 +470 @@ def test_clean_cached_assets(
- with patch("api.routes.rows.glob_rows_in_assets_dir", deterministic_glob_rows_in_assets_dir):
+ with patch("rows.routes.rows.glob_rows_in_assets_dir", deterministic_glob_rows_in_assets_dir):
diff --git a/services/rows/tests/test_app.py b/services/rows/tests/test_app.py
new file mode 100644
index 00000000..cfde7342
--- /dev/null
+++ b/services/rows/tests/test_app.py
@@ -0,0 +1,96 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from typing import Optional
+
+import pytest
+from starlette.testclient import TestClient
+
+from rows.app import create_app_with_config
+from rows.config import AppConfig
+
+
[email protected](scope="module")
+def client(monkeypatch_session: pytest.MonkeyPatch, app_config: AppConfig) -> TestClient:
+ return TestClient(create_app_with_config(app_config=app_config))
+
+
+def test_cors(client: TestClient) -> None:
+ origin = "http://localhost:3000"
+ method = "GET"
+ header = "X-Requested-With"
+ response = client.options(
+ "/rows?dataset=dataset1&config=config1&split=train",
+ headers={
+ "Origin": origin,
+ "Access-Control-Request-Method": method,
+ "Access-Control-Request-Headers": header,
+ },
+ )
+ assert response.status_code == 200
+ assert (
+ origin in [o.strip() for o in response.headers["Access-Control-Allow-Origin"].split(",")]
+ or response.headers["Access-Control-Allow-Origin"] == "*"
+ )
+ assert (
+ header in [o.strip() for o in response.headers["Access-Control-Allow-Headers"].split(",")]
+ or response.headers["Access-Control-Expose-Headers"] == "*"
+ )
+ assert (
+ method in [o.strip() for o in response.headers["Access-Control-Allow-Methods"].split(",")]
+ or response.headers["Access-Control-Expose-Headers"] == "*"
+ )
+ assert response.headers["Access-Control-Allow-Credentials"] == "true"
+
+
+def test_get_healthcheck(client: TestClient) -> None:
+ response = client.get("/healthcheck")
+ assert response.status_code == 200
+ assert response.text == "ok"
+
+
+def test_get_rows(client: TestClient) -> None:
+ # missing parameter
+ response = client.get("/rows")
+ assert response.status_code == 422
+
+
[email protected](
+ "dataset,config,split",
+ [
+ (None, None, None),
+ ("a", None, None),
+ ("a", "b", None),
+ ("a", "b", ""),
+ ],
+)
+def test_get_split_missing_parameter(
+ client: TestClient,
+ dataset: Optional[str],
+ config: Optional[str],
+ split: Optional[str],
+) -> None:
+ response = client.get("/rows", params={"dataset": dataset, "config": config, "split": split})
+ assert response.status_code == 422
+
+
+def test_metrics(client: TestClient) -> None:
+ response = client.get("/healthcheck")
+ response = client.get("/metrics")
+ assert response.status_code == 200
+ text = response.text
+ lines = text.split("\n")
+ # examples:
+ # starlette_requests_total{method="GET",path_template="/metrics"} 1.0
+ # method_steps_processing_time_seconds_sum{method="healthcheck_endpoint",step="all"} 1.6772013623267412e-05
+ metrics = {
+ parts[0]: float(parts[1]) for line in lines if line and line[0] != "#" and (parts := line.rsplit(" ", 1))
+ }
+
+ # the metrics should contain at least the following
+ for name in [
+ 'starlette_requests_total{method="GET",path_template="/metrics"}',
+ 'method_steps_processing_time_seconds_sum{context="None",method="healthcheck_endpoint",step="all"}',
+ ]:
+ assert name in metrics, metrics
+ assert metrics[name] > 0, metrics
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index 37b1c87d..542101d5 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -19,0 +20 @@ services:
+ URL_ROWS: http://rows:${ROWS_UVICORN_PORT-8082}
@@ -21 +21,0 @@ services:
- - api
@@ -22,0 +23,2 @@ services:
+ - api
+ - rows
@@ -56,0 +59,27 @@ services:
+ environment:
+ # service
+ API_HF_AUTH_PATH: ${API_HF_AUTH_PATH-/api/datasets/%s/auth-check}
+ API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://huggingface.co/api/keys/jwt}
+ API_HF_JWT_ALGORITHM: ${API_HF_JWT_ALGORITHM-EdDSA}
+ API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-0.2}
+ API_MAX_AGE_LONG: ${API_MAX_AGE_LONG-120}
+ API_MAX_AGE_SHORT: ${API_MAX_AGE_SHORT-10}
+ # prometheus
+ PROMETHEUS_MULTIPROC_DIR: ${PROMETHEUS_MULTIPROC_DIR-}
+ # uvicorn
+ API_UVICORN_HOSTNAME: 0.0.0.0 # required for docker compose
+ API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2}
+ API_UVICORN_PORT: ${API_UVICORN_PORT-8080}
+ ports:
+ # for debug
+ - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080}
+ depends_on:
+ - mongodb
+ restart: unless-stopped
+ rows:
+ build:
+ context: ..
+ dockerfile: services/rows/Dockerfile
+ extends:
+ file: docker-compose-base.yml
+ service: api
@@ -79,2 +108,2 @@ services:
- API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2}
- API_UVICORN_PORT: ${API_UVICORN_PORT-8080}
+ API_UVICORN_NUM_WORKERS: ${ROWS_UVICORN_NUM_WORKERS-2}
+ API_UVICORN_PORT: ${ROWS_UVICORN_PORT-8082}
@@ -83 +112 @@ services:
- - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080}
+ - ${ROWS_UVICORN_PORT-8082}:${ROWS_UVICORN_PORT-8082}
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 233e90f2..3a5cb0f4 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -19,0 +20 @@ services:
+ URL_ROWS: http://host.docker.internal:${ROWS_UVICORN_PORT-8082}
@@ -23 +23,0 @@ services:
- - api
@@ -24,0 +25,2 @@ services:
+ - api
+ - rows
@@ -58,0 +61,27 @@ services:
+ environment:
+ # service
+ API_HF_AUTH_PATH: ${API_HF_AUTH_PATH-/api/datasets/%s/auth-check}
+ API_HF_JWT_PUBLIC_KEY_URL: ${API_HF_JWT_PUBLIC_KEY_URL-https://hub-ci.huggingface.co/api/keys/jwt}
+ API_HF_JWT_ALGORITHM: ${API_HF_JWT_ALGORITHM-EdDSA}
+ API_HF_TIMEOUT_SECONDS: ${API_HF_TIMEOUT_SECONDS-1.0}
+ API_MAX_AGE_LONG: ${API_MAX_AGE_LONG-120}
+ API_MAX_AGE_SHORT: ${API_MAX_AGE_SHORT-10}
+ # prometheus
+ PROMETHEUS_MULTIPROC_DIR: ${PROMETHEUS_MULTIPROC_DIR-}
+ # uvicorn
+ API_UVICORN_HOSTNAME: 0.0.0.0 # required for docker compose
+ API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2}
+ API_UVICORN_PORT: ${API_UVICORN_PORT-8080}
+ network_mode: ${DEV_NETWORK_MODE}
+ ports:
+ - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080}
+ depends_on:
+ - mongodb
+ restart: unless-stopped
+ rows:
+ build:
+ context: ..
+ dockerfile: services/rows/dev.Dockerfile
+ extends:
+ file: docker-compose-dev-base.yml
+ service: api
@@ -81,2 +110,2 @@ services:
- API_UVICORN_NUM_WORKERS: ${API_UVICORN_NUM_WORKERS-2}
- API_UVICORN_PORT: ${API_UVICORN_PORT-8080}
+ API_UVICORN_NUM_WORKERS: ${ROWS_UVICORN_NUM_WORKERS-2}
+ API_UVICORN_PORT: ${ROWS_UVICORN_PORT-8082}
@@ -85 +114,2 @@ services:
- - ${API_UVICORN_PORT-8080}:${API_UVICORN_PORT-8080}
+ # for debug
+ - ${ROWS_UVICORN_PORT-8082}:${ROWS_UVICORN_PORT-8082}
|
|
6343f2cdd06772efbc44a1e184f9cab2761745b0
|
Andrea Francis Soria Jimenez
| 2023-07-04T13:32:37 |
Moving configs and adding doc in readme (#1474)
|
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 3a713a1f..59006344 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -106,36 +105,0 @@ class ParquetMetadataConfig:
-
-
-DUCKDB_INDEX_STORAGE_DIRECTORY = None
-DUCKDB_INDEX_COMMIT_MESSAGE = "Update duckdb index file"
-DUCKDB_INDEX_COMMITTER_HF_TOKEN = None
-DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES = 100_000_000
-DUCKDB_INDEX_TARGET_REVISION = "refs/convert/parquet"
-DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s"
-DUCKDB_INDEX_EXTENSIONS_DIRECTORY: Optional[str] = None
-
-
-@dataclass(frozen=True)
-class DuckDbIndexConfig:
- storage_directory: Optional[str] = DUCKDB_INDEX_STORAGE_DIRECTORY
- commit_message: str = DUCKDB_INDEX_COMMIT_MESSAGE
- committer_hf_token: Optional[str] = DUCKDB_INDEX_COMMITTER_HF_TOKEN
- target_revision: str = DUCKDB_INDEX_TARGET_REVISION
- url_template: str = DUCKDB_INDEX_URL_TEMPLATE
- max_parquet_size_bytes: int = DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES
- extensions_directory: Optional[str] = DUCKDB_INDEX_EXTENSIONS_DIRECTORY
-
- @classmethod
- def from_env(cls) -> "DuckDbIndexConfig":
- env = Env(expand_vars=True)
- with env.prefixed("DUCKDB_INDEX_"):
- return cls(
- storage_directory=env.str(name="STORAGE_DIRECTORY", default=DUCKDB_INDEX_STORAGE_DIRECTORY),
- commit_message=env.str(name="COMMIT_MESSAGE", default=DUCKDB_INDEX_COMMIT_MESSAGE),
- committer_hf_token=env.str(name="COMMITTER_HF_TOKEN", default=DUCKDB_INDEX_COMMITTER_HF_TOKEN),
- target_revision=env.str(name="TARGET_REVISION", default=DUCKDB_INDEX_TARGET_REVISION),
- url_template=env.str(name="URL_TEMPLATE", default=DUCKDB_INDEX_URL_TEMPLATE),
- max_parquet_size_bytes=env.int(
- name="MAX_PARQUET_SIZE_BYTES", default=DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES
- ),
- extensions_directory=env.str(name="EXTENSIONS_DIRECTORY", default=DUCKDB_INDEX_EXTENSIONS_DIRECTORY),
- )
diff --git a/services/worker/README.md b/services/worker/README.md
index 00bd203e..b6299326 100644
--- a/services/worker/README.md
+++ b/services/worker/README.md
@@ -77,0 +78,12 @@ Set environment variables to configure the parquet worker (`PARQUET_AND_INFO_` p
+### Duckdb Index worker
+
+Set environment variables to configure the duckdb index worker (`DUCKDB_INDEX_` prefix):
+
+- `DUCKDB_INDEX_STORAGE_DIRECTORY`: directory where the temporal duckdb index files are stored. Defaults to empty.
+- `DUCKDB_INDEX_COMMIT_MESSAGE`: the git commit message when the worker uploads the duckdb index file to the Hub. Defaults to `Update duckdb index file`.
+- `DUCKDB_INDEX_COMMITTER_HF_TOKEN`: the HuggingFace token to commit the duckdb index file to the Hub. The token must be an app token associated with a user that has the right to 1. create the `refs/convert/parquet` branch (see `DUCKDB_INDEX_TARGET_REVISION`) and 2. push commits to it on any dataset. [Datasets maintainers](https://huggingface.co/datasets-maintainers) members have these rights. The token must have permission to write. If not set, the worker will fail. Defaults to None.
+- `DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES`: the maximum size in bytes of the dataset's parquet files to index. Datasets with bigger size are ignored. Defaults to `100_000_000`.
+- `DUCKDB_INDEX_TARGET_REVISION`: the git revision of the dataset where to store the duckdb index file. Make sure the committer token (`DUCKDB_INDEX_COMMITTER_HF_TOKEN`) has the permission to write there. Defaults to `refs/convert/parquet`.
+- `DUCKDB_INDEX_URL_TEMPLATE`: the URL template to build the duckdb index file URL. Defaults to `/datasets/%s/resolve/%s/%s`.
+- `DUCKDB_INDEX_EXTENSIONS_DIRECTORY`: directory where the duckdb extensions will be downloaded. Defaults to empty.
+
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index 0a57557e..d39a3c03 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -12 +11,0 @@ from libcommon.config import (
- DuckDbIndexConfig,
@@ -236,0 +236,36 @@ class ConfigNamesConfig:
+DUCKDB_INDEX_STORAGE_DIRECTORY = None
+DUCKDB_INDEX_COMMIT_MESSAGE = "Update duckdb index file"
+DUCKDB_INDEX_COMMITTER_HF_TOKEN = None
+DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES = 100_000_000
+DUCKDB_INDEX_TARGET_REVISION = "refs/convert/parquet"
+DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s"
+DUCKDB_INDEX_EXTENSIONS_DIRECTORY: Optional[str] = None
+
+
+@dataclass(frozen=True)
+class DuckDbIndexConfig:
+ storage_directory: Optional[str] = DUCKDB_INDEX_STORAGE_DIRECTORY
+ commit_message: str = DUCKDB_INDEX_COMMIT_MESSAGE
+ committer_hf_token: Optional[str] = DUCKDB_INDEX_COMMITTER_HF_TOKEN
+ target_revision: str = DUCKDB_INDEX_TARGET_REVISION
+ url_template: str = DUCKDB_INDEX_URL_TEMPLATE
+ max_parquet_size_bytes: int = DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES
+ extensions_directory: Optional[str] = DUCKDB_INDEX_EXTENSIONS_DIRECTORY
+
+ @classmethod
+ def from_env(cls) -> "DuckDbIndexConfig":
+ env = Env(expand_vars=True)
+ with env.prefixed("DUCKDB_INDEX_"):
+ return cls(
+ storage_directory=env.str(name="STORAGE_DIRECTORY", default=DUCKDB_INDEX_STORAGE_DIRECTORY),
+ commit_message=env.str(name="COMMIT_MESSAGE", default=DUCKDB_INDEX_COMMIT_MESSAGE),
+ committer_hf_token=env.str(name="COMMITTER_HF_TOKEN", default=DUCKDB_INDEX_COMMITTER_HF_TOKEN),
+ target_revision=env.str(name="TARGET_REVISION", default=DUCKDB_INDEX_TARGET_REVISION),
+ url_template=env.str(name="URL_TEMPLATE", default=DUCKDB_INDEX_URL_TEMPLATE),
+ max_parquet_size_bytes=env.int(
+ name="MAX_PARQUET_SIZE_BYTES", default=DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES
+ ),
+ extensions_directory=env.str(name="EXTENSIONS_DIRECTORY", default=DUCKDB_INDEX_EXTENSIONS_DIRECTORY),
+ )
+
+
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py
index 97fbb673..26ff821b 100644
--- a/services/worker/src/worker/job_runners/split/duckdb_index.py
+++ b/services/worker/src/worker/job_runners/split/duckdb_index.py
@@ -16 +15,0 @@ from huggingface_hub.utils._errors import HfHubHTTPError, RepositoryNotFoundErro
-from libcommon.config import DuckDbIndexConfig
@@ -36 +35 @@ from libcommon.utils import JobInfo, SplitHubFile
-from worker.config import AppConfig
+from worker.config import AppConfig, DuckDbIndexConfig
|
|
4c81b9616c0f570985a7589eb7a15b0e35b445f9
|
Sylvain Lesage
| 2023-07-04T13:32:18 |
Create libapi (#1475)
|
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml
index c285df42..b282f8e3 100644
--- a/.github/workflows/_quality-python.yml
+++ b/.github/workflows/_quality-python.yml
@@ -52 +52 @@ jobs:
- run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed 's/^huggingface-hub @ git.*/huggingface-hub==0.15.1 ; python_full_version == \"3.9.15\" --hash=sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445 --hash=sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081/')"
+ run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libapi @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed 's/^huggingface-hub @ git.*/huggingface-hub==0.15.1 ; python_full_version == \"3.9.15\" --hash=sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445 --hash=sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081/')"
diff --git a/.github/workflows/l-libapi.yml b/.github/workflows/l-libapi.yml
new file mode 100644
index 00000000..e7f75a41
--- /dev/null
+++ b/.github/workflows/l-libapi.yml
@@ -0,0 +1,31 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+name: libs/libapi
+on:
+ workflow_dispatch:
+ push:
+ branches:
+ - main
+ paths:
+ - "libs/libapi/**"
+ - ".github/workflows/l-libapi.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
+ pull_request:
+ paths:
+ - "libs/libapi/**"
+ - ".github/workflows/l-libapi.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
+jobs:
+ quality:
+ uses: ./.github/workflows/_quality-python.yml
+ with:
+ working-directory: libs/libapi
+ unit-tests:
+ uses: ./.github/workflows/_unit-tests-python.yml
+ with:
+ working-directory: libs/libapi
diff --git a/.vscode/monorepo.code-workspace b/.vscode/monorepo.code-workspace
index d73dd79a..597c0f2f 100644
--- a/.vscode/monorepo.code-workspace
+++ b/.vscode/monorepo.code-workspace
@@ -22,0 +23,4 @@
+ {
+ "name": "libs/libapi",
+ "path": "../libs/libapi"
+ },
diff --git a/libs/libapi/.flake8 b/libs/libapi/.flake8
new file mode 100644
index 00000000..f7d6157c
--- /dev/null
+++ b/libs/libapi/.flake8
@@ -0,0 +1,5 @@
+[flake8]
+# Recommend matching the black line length (119),
+# rather than using the flake8 default of 79:
+max-line-length = 119
+extend-ignore = "E203"
diff --git a/libs/libapi/.python-version b/libs/libapi/.python-version
new file mode 100644
index 00000000..b326afbc
--- /dev/null
+++ b/libs/libapi/.python-version
@@ -0,0 +1 @@
+3.9.15
diff --git a/libs/libapi/Makefile b/libs/libapi/Makefile
new file mode 100644
index 00000000..ae7d6714
--- /dev/null
+++ b/libs/libapi/Makefile
@@ -0,0 +1,10 @@
+# environment variables for the commands (docker compose, poetry)
+export MONGO_PORT := 27021
+export METRICS_MONGO_URL := mongodb://localhost:${MONGO_PORT}
+export COMPOSE_PROJECT_NAME := libapi
+# makefile variables
+DOCKER_COMPOSE := ../../tools/docker-compose-mongo.yml
+
+include ../../tools/Python.mk
+include ../../tools/PythonTest.mk
+include ../../tools/Docker.mk
diff --git a/libs/libapi/README.md b/libs/libapi/README.md
new file mode 100644
index 00000000..8ab47fe8
--- /dev/null
+++ b/libs/libapi/README.md
@@ -0,0 +1,31 @@
+# libapi
+
+A Python library for the API services
+
+## Configuration
+
+The APIs can be configured using environment variables. They are grouped by scope.
+
+### API service
+
+Set environment variables to configure the application (`API_` prefix):
+
+- `API_HF_AUTH_PATH`: the path of the external authentication service, on the hub (see `HF_ENDPOINT`). The string must contain `%s` which will be replaced with the dataset name. The external authentication service must return 200, 401, 403 or 404. Defaults to "/api/datasets/%s/auth-check".
+- `API_HF_JWT_PUBLIC_KEY_URL`: the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format. It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored. Defaults to empty.
+- `API_HF_JWT_ALGORITHM`: the algorithm used to encode the JWT. Defaults to `"EdDSA"`.
+- `API_HF_TIMEOUT_SECONDS`: the timeout in seconds for the requests to the Hugging Face Hub. Defaults to `0.2` (200 ms).
+- `API_HF_WEBHOOK_SECRET`: a shared secret sent by the Hub in the "X-Webhook-Secret" header of POST requests sent to /webhook, to authenticate the originator and bypass some validation of the content (avoiding roundtrip to the Hub). If not set, all the validations are done. Defaults to empty.
+- `API_MAX_AGE_LONG`: number of seconds to set in the `max-age` header on data endpoints. Defaults to `120` (2 minutes).
+- `API_MAX_AGE_SHORT`: number of seconds to set in the `max-age` header on technical endpoints. Defaults to `10` (10 seconds).
+
+### Uvicorn
+
+The following environment variables are used to configure the Uvicorn server (`API_UVICORN_` prefix):
+
+- `API_UVICORN_HOSTNAME`: the hostname. Defaults to `"localhost"`.
+- `API_UVICORN_NUM_WORKERS`: the number of uvicorn workers. Defaults to `2`.
+- `API_UVICORN_PORT`: the port. Defaults to `8000`.
+
+### Prometheus
+
+- `PROMETHEUS_MULTIPROC_DIR`: the directory where the uvicorn workers share their prometheus metrics. See https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn. Defaults to empty, in which case every worker manages its own metrics, and the /metrics endpoint returns the metrics of a random worker.
diff --git a/libs/libapi/poetry.lock b/libs/libapi/poetry.lock
new file mode 100644
index 00000000..c970667c
--- /dev/null
+++ b/libs/libapi/poetry.lock
@@ -0,0 +1,3178 @@
+# This file is automatically @generated by Poetry and should not be changed by hand.
+
+[[package]]
+name = "aiohttp"
+version = "3.8.4"
+description = "Async http client/server framework (asyncio)"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:5ce45967538fb747370308d3145aa68a074bdecb4f3a300869590f725ced69c1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b744c33b6f14ca26b7544e8d8aadff6b765a80ad6164fb1a430bbadd593dfb1a"},
+ {file = "aiohttp-3.8.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1a45865451439eb320784918617ba54b7a377e3501fb70402ab84d38c2cd891b"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a86d42d7cba1cec432d47ab13b6637bee393a10f664c425ea7b305d1301ca1a3"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ee3c36df21b5714d49fc4580247947aa64bcbe2939d1b77b4c8dcb8f6c9faecc"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:176a64b24c0935869d5bbc4c96e82f89f643bcdf08ec947701b9dbb3c956b7dd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c844fd628851c0bc309f3c801b3a3d58ce430b2ce5b359cd918a5a76d0b20cb5"},
+ {file = "aiohttp-3.8.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5393fb786a9e23e4799fec788e7e735de18052f83682ce2dfcabaf1c00c2c08e"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e4b09863aae0dc965c3ef36500d891a3ff495a2ea9ae9171e4519963c12ceefd"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:adfbc22e87365a6e564c804c58fc44ff7727deea782d175c33602737b7feadb6"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:147ae376f14b55f4f3c2b118b95be50a369b89b38a971e80a17c3fd623f280c9"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:eafb3e874816ebe2a92f5e155f17260034c8c341dad1df25672fb710627c6949"},
+ {file = "aiohttp-3.8.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c6cc15d58053c76eacac5fa9152d7d84b8d67b3fde92709195cb984cfb3475ea"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win32.whl", hash = "sha256:59f029a5f6e2d679296db7bee982bb3d20c088e52a2977e3175faf31d6fb75d1"},
+ {file = "aiohttp-3.8.4-cp310-cp310-win_amd64.whl", hash = "sha256:fe7ba4a51f33ab275515f66b0a236bcde4fb5561498fe8f898d4e549b2e4509f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d8ef1a630519a26d6760bc695842579cb09e373c5f227a21b67dc3eb16cfea4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5b3f2e06a512e94722886c0827bee9807c86a9f698fac6b3aee841fab49bbfb4"},
+ {file = "aiohttp-3.8.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3a80464982d41b1fbfe3154e440ba4904b71c1a53e9cd584098cd41efdb188ef"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b631e26df63e52f7cce0cce6507b7a7f1bc9b0c501fcde69742130b32e8782f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3f43255086fe25e36fd5ed8f2ee47477408a73ef00e804cb2b5cba4bf2ac7f5e"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4d347a172f866cd1d93126d9b239fcbe682acb39b48ee0873c73c933dd23bd0f"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3fec6a4cb5551721cdd70473eb009d90935b4063acc5f40905d40ecfea23e05"},
+ {file = "aiohttp-3.8.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80a37fe8f7c1e6ce8f2d9c411676e4bc633a8462844e38f46156d07a7d401654"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:d1e6a862b76f34395a985b3cd39a0d949ca80a70b6ebdea37d3ab39ceea6698a"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:cd468460eefef601ece4428d3cf4562459157c0f6523db89365202c31b6daebb"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:618c901dd3aad4ace71dfa0f5e82e88b46ef57e3239fc7027773cb6d4ed53531"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:652b1bff4f15f6287550b4670546a2947f2a4575b6c6dff7760eafb22eacbf0b"},
+ {file = "aiohttp-3.8.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:80575ba9377c5171407a06d0196b2310b679dc752d02a1fcaa2bc20b235dbf24"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win32.whl", hash = "sha256:bbcf1a76cf6f6dacf2c7f4d2ebd411438c275faa1dc0c68e46eb84eebd05dd7d"},
+ {file = "aiohttp-3.8.4-cp311-cp311-win_amd64.whl", hash = "sha256:6e74dd54f7239fcffe07913ff8b964e28b712f09846e20de78676ce2a3dc0bfc"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:880e15bb6dad90549b43f796b391cfffd7af373f4646784795e20d92606b7a51"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb96fa6b56bb536c42d6a4a87dfca570ff8e52de2d63cabebfd6fb67049c34b6"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a6cadebe132e90cefa77e45f2d2f1a4b2ce5c6b1bfc1656c1ddafcfe4ba8131"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f352b62b45dff37b55ddd7b9c0c8672c4dd2eb9c0f9c11d395075a84e2c40f75"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ab43061a0c81198d88f39aaf90dae9a7744620978f7ef3e3708339b8ed2ef01"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c9cb1565a7ad52e096a6988e2ee0397f72fe056dadf75d17fa6b5aebaea05622"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:1b3ea7edd2d24538959c1c1abf97c744d879d4e541d38305f9bd7d9b10c9ec41"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:7c7837fe8037e96b6dd5cfcf47263c1620a9d332a87ec06a6ca4564e56bd0f36"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:3b90467ebc3d9fa5b0f9b6489dfb2c304a1db7b9946fa92aa76a831b9d587e99"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:cab9401de3ea52b4b4c6971db5fb5c999bd4260898af972bf23de1c6b5dd9d71"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:d1f9282c5f2b5e241034a009779e7b2a1aa045f667ff521e7948ea9b56e0c5ff"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win32.whl", hash = "sha256:5e14f25765a578a0a634d5f0cd1e2c3f53964553a00347998dfdf96b8137f777"},
+ {file = "aiohttp-3.8.4-cp36-cp36m-win_amd64.whl", hash = "sha256:4c745b109057e7e5f1848c689ee4fb3a016c8d4d92da52b312f8a509f83aa05e"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:aede4df4eeb926c8fa70de46c340a1bc2c6079e1c40ccf7b0eae1313ffd33519"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ddaae3f3d32fc2cb4c53fab020b69a05c8ab1f02e0e59665c6f7a0d3a5be54f"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c4eb3b82ca349cf6fadcdc7abcc8b3a50ab74a62e9113ab7a8ebc268aad35bb9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9bcb89336efa095ea21b30f9e686763f2be4478f1b0a616969551982c4ee4c3b"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c08e8ed6fa3d477e501ec9db169bfac8140e830aa372d77e4a43084d8dd91ab"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c6cd05ea06daca6ad6a4ca3ba7fe7dc5b5de063ff4daec6170ec0f9979f6c332"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7a00a9ed8d6e725b55ef98b1b35c88013245f35f68b1b12c5cd4100dddac333"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:de04b491d0e5007ee1b63a309956eaed959a49f5bb4e84b26c8f5d49de140fa9"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:40653609b3bf50611356e6b6554e3a331f6879fa7116f3959b20e3528783e699"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dbf3a08a06b3f433013c143ebd72c15cac33d2914b8ea4bea7ac2c23578815d6"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:854f422ac44af92bfe172d8e73229c270dc09b96535e8a548f99c84f82dde241"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win32.whl", hash = "sha256:aeb29c84bb53a84b1a81c6c09d24cf33bb8432cc5c39979021cc0f98c1292a1a"},
+ {file = "aiohttp-3.8.4-cp37-cp37m-win_amd64.whl", hash = "sha256:db3fc6120bce9f446d13b1b834ea5b15341ca9ff3f335e4a951a6ead31105480"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:fabb87dd8850ef0f7fe2b366d44b77d7e6fa2ea87861ab3844da99291e81e60f"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:91f6d540163f90bbaef9387e65f18f73ffd7c79f5225ac3d3f61df7b0d01ad15"},
+ {file = "aiohttp-3.8.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:d265f09a75a79a788237d7f9054f929ced2e69eb0bb79de3798c468d8a90f945"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d89efa095ca7d442a6d0cbc755f9e08190ba40069b235c9886a8763b03785da"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4dac314662f4e2aa5009977b652d9b8db7121b46c38f2073bfeed9f4049732cd"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:fe11310ae1e4cd560035598c3f29d86cef39a83d244c7466f95c27ae04850f10"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6ddb2a2026c3f6a68c3998a6c47ab6795e4127315d2e35a09997da21865757f8"},
+ {file = "aiohttp-3.8.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e75b89ac3bd27d2d043b234aa7b734c38ba1b0e43f07787130a0ecac1e12228a"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6e601588f2b502c93c30cd5a45bfc665faaf37bbe835b7cfd461753068232074"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:a5d794d1ae64e7753e405ba58e08fcfa73e3fad93ef9b7e31112ef3c9a0efb52"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:a1f4689c9a1462f3df0a1f7e797791cd6b124ddbee2b570d34e7f38ade0e2c71"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:3032dcb1c35bc330134a5b8a5d4f68c1a87252dfc6e1262c65a7e30e62298275"},
+ {file = "aiohttp-3.8.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:8189c56eb0ddbb95bfadb8f60ea1b22fcfa659396ea36f6adcc521213cd7b44d"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win32.whl", hash = "sha256:33587f26dcee66efb2fff3c177547bd0449ab7edf1b73a7f5dea1e38609a0c54"},
+ {file = "aiohttp-3.8.4-cp38-cp38-win_amd64.whl", hash = "sha256:e595432ac259af2d4630008bf638873d69346372d38255774c0e286951e8b79f"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:5a7bdf9e57126dc345b683c3632e8ba317c31d2a41acd5800c10640387d193ed"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:22f6eab15b6db242499a16de87939a342f5a950ad0abaf1532038e2ce7d31567"},
+ {file = "aiohttp-3.8.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7235604476a76ef249bd64cb8274ed24ccf6995c4a8b51a237005ee7a57e8643"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea9eb976ffdd79d0e893869cfe179a8f60f152d42cb64622fca418cd9b18dc2a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:92c0cea74a2a81c4c76b62ea1cac163ecb20fb3ba3a75c909b9fa71b4ad493cf"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:493f5bc2f8307286b7799c6d899d388bbaa7dfa6c4caf4f97ef7521b9cb13719"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0a63f03189a6fa7c900226e3ef5ba4d3bd047e18f445e69adbd65af433add5a2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:10c8cefcff98fd9168cdd86c4da8b84baaa90bf2da2269c6161984e6737bf23e"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:bca5f24726e2919de94f047739d0a4fc01372801a3672708260546aa2601bf57"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:03baa76b730e4e15a45f81dfe29a8d910314143414e528737f8589ec60cf7391"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8c29c77cc57e40f84acef9bfb904373a4e89a4e8b74e71aa8075c021ec9078c2"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:03543dcf98a6619254b409be2d22b51f21ec66272be4ebda7b04e6412e4b2e14"},
+ {file = "aiohttp-3.8.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:17b79c2963db82086229012cff93ea55196ed31f6493bb1ccd2c62f1724324e4"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win32.whl", hash = "sha256:34ce9f93a4a68d1272d26030655dd1b58ff727b3ed2a33d80ec433561b03d67a"},
+ {file = "aiohttp-3.8.4-cp39-cp39-win_amd64.whl", hash = "sha256:41a86a69bb63bb2fc3dc9ad5ea9f10f1c9c8e282b471931be0268ddd09430b04"},
+ {file = "aiohttp-3.8.4.tar.gz", hash = "sha256:bf2e1a9162c1e441bf805a1fd166e249d574ca04e03b34f97e2928769e91ab5c"},
+]
+
+[package.dependencies]
+aiosignal = ">=1.1.2"
+async-timeout = ">=4.0.0a3,<5.0"
+attrs = ">=17.3.0"
+charset-normalizer = ">=2.0,<4.0"
+frozenlist = ">=1.1.1"
+multidict = ">=4.5,<7.0"
+yarl = ">=1.0,<2.0"
+
+[package.extras]
+speedups = ["Brotli", "aiodns", "cchardet"]
+
+[[package]]
+name = "aiosignal"
+version = "1.3.1"
+description = "aiosignal: a list of registered asynchronous callbacks"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"},
+ {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"},
+]
+
+[package.dependencies]
+frozenlist = ">=1.1.0"
+
+[[package]]
+name = "anyio"
+version = "3.7.0"
+description = "High level compatibility layer for multiple asynchronous event loop implementations"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "anyio-3.7.0-py3-none-any.whl", hash = "sha256:eddca883c4175f14df8aedce21054bfca3adb70ffe76a9f607aef9d7fa2ea7f0"},
+ {file = "anyio-3.7.0.tar.gz", hash = "sha256:275d9973793619a5374e1c89a4f4ad3f4b0a5510a2b5b939444bee8f4c4d37ce"},
+]
+
+[package.dependencies]
+exceptiongroup = {version = "*", markers = "python_version < \"3.11\""}
+idna = ">=2.8"
+sniffio = ">=1.1"
+
+[package.extras]
+doc = ["Sphinx (>=6.1.0)", "packaging", "sphinx-autodoc-typehints (>=1.2.0)", "sphinx-rtd-theme", "sphinxcontrib-jquery"]
+test = ["anyio[trio]", "coverage[toml] (>=4.5)", "hypothesis (>=4.0)", "mock (>=4)", "psutil (>=5.9)", "pytest (>=7.0)", "pytest-mock (>=3.6.1)", "trustme", "uvloop (>=0.17)"]
+trio = ["trio (<0.22)"]
+
+[[package]]
+name = "appdirs"
+version = "1.4.4"
+description = "A small Python module for determining appropriate platform-specific dirs, e.g. a \"user data dir\"."
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "appdirs-1.4.4-py2.py3-none-any.whl", hash = "sha256:a841dacd6b99318a741b166adb07e19ee71a274450e68237b4650ca1055ab128"},
+ {file = "appdirs-1.4.4.tar.gz", hash = "sha256:7d5d0167b2b1ba821647616af46a749d1c653740dd0d2415100fe26e27afdf41"},
+]
+
+[[package]]
+name = "async-timeout"
+version = "4.0.2"
+description = "Timeout context manager for asyncio programs"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "async-timeout-4.0.2.tar.gz", hash = "sha256:2163e1640ddb52b7a8c80d0a67a08587e5d245cc9c553a74a847056bc2976b15"},
+ {file = "async_timeout-4.0.2-py3-none-any.whl", hash = "sha256:8ca1e4fcf50d07413d66d1a5e416e42cfdf5851c981d679a09851a6853383b3c"},
+]
+
+[[package]]
+name = "attrs"
+version = "23.1.0"
+description = "Classes Without Boilerplate"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "attrs-23.1.0-py3-none-any.whl", hash = "sha256:1f28b4522cdc2fb4256ac1a020c78acf9cba2c6b461ccd2c126f3aa8e8335d04"},
+ {file = "attrs-23.1.0.tar.gz", hash = "sha256:6279836d581513a26f1bf235f9acd333bc9115683f14f7e8fae46c98fc50e015"},
+]
+
+[package.extras]
+cov = ["attrs[tests]", "coverage[toml] (>=5.3)"]
+dev = ["attrs[docs,tests]", "pre-commit"]
+docs = ["furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier", "zope-interface"]
+tests = ["attrs[tests-no-zope]", "zope-interface"]
+tests-no-zope = ["cloudpickle", "hypothesis", "mypy (>=1.1.1)", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"]
+
+[[package]]
+name = "audioread"
+version = "3.0.0"
+description = "multi-library, cross-platform audio decoding"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "audioread-3.0.0.tar.gz", hash = "sha256:121995bd207eb1fda3d566beb851d3534275925bc35a4fb6da0cb11de0f7251a"},
+]
+
+[[package]]
+name = "bandit"
+version = "1.7.5"
+description = "Security oriented static analyser for python code."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "bandit-1.7.5-py3-none-any.whl", hash = "sha256:75665181dc1e0096369112541a056c59d1c5f66f9bb74a8d686c3c362b83f549"},
+ {file = "bandit-1.7.5.tar.gz", hash = "sha256:bdfc739baa03b880c2d15d0431b31c658ffc348e907fe197e54e0389dd59e11e"},
+]
+
+[package.dependencies]
+colorama = {version = ">=0.3.9", markers = "platform_system == \"Windows\""}
+GitPython = ">=1.0.1"
+PyYAML = ">=5.3.1"
+rich = "*"
+stevedore = ">=1.20.0"
+
+[package.extras]
+test = ["beautifulsoup4 (>=4.8.0)", "coverage (>=4.5.4)", "fixtures (>=3.0.0)", "flake8 (>=4.0.0)", "pylint (==1.9.4)", "stestr (>=2.5.0)", "testscenarios (>=0.5.0)", "testtools (>=2.3.0)", "tomli (>=1.1.0)"]
+toml = ["tomli (>=1.1.0)"]
+yaml = ["PyYAML"]
+
+[[package]]
+name = "black"
+version = "22.12.0"
+description = "The uncompromising code formatter."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "black-22.12.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eedd20838bd5d75b80c9f5487dbcb06836a43833a37846cf1d8c1cc01cef59d"},
+ {file = "black-22.12.0-cp310-cp310-win_amd64.whl", hash = "sha256:159a46a4947f73387b4d83e87ea006dbb2337eab6c879620a3ba52699b1f4351"},
+ {file = "black-22.12.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d30b212bffeb1e252b31dd269dfae69dd17e06d92b87ad26e23890f3efea366f"},
+ {file = "black-22.12.0-cp311-cp311-win_amd64.whl", hash = "sha256:7412e75863aa5c5411886804678b7d083c7c28421210180d67dfd8cf1221e1f4"},
+ {file = "black-22.12.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c116eed0efb9ff870ded8b62fe9f28dd61ef6e9ddd28d83d7d264a38417dcee2"},
+ {file = "black-22.12.0-cp37-cp37m-win_amd64.whl", hash = "sha256:1f58cbe16dfe8c12b7434e50ff889fa479072096d79f0a7f25e4ab8e94cd8350"},
+ {file = "black-22.12.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77d86c9f3db9b1bf6761244bc0b3572a546f5fe37917a044e02f3166d5aafa7d"},
+ {file = "black-22.12.0-cp38-cp38-win_amd64.whl", hash = "sha256:82d9fe8fee3401e02e79767016b4907820a7dc28d70d137eb397b92ef3cc5bfc"},
+ {file = "black-22.12.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:101c69b23df9b44247bd88e1d7e90154336ac4992502d4197bdac35dd7ee3320"},
+ {file = "black-22.12.0-cp39-cp39-win_amd64.whl", hash = "sha256:559c7a1ba9a006226f09e4916060982fd27334ae1998e7a38b3f33a37f7a2148"},
+ {file = "black-22.12.0-py3-none-any.whl", hash = "sha256:436cc9167dd28040ad90d3b404aec22cedf24a6e4d7de221bec2730ec0c97bcf"},
+ {file = "black-22.12.0.tar.gz", hash = "sha256:229351e5a18ca30f447bf724d007f890f97e13af070bb6ad4c0a441cd7596a2f"},
+]
+
+[package.dependencies]
+click = ">=8.0.0"
+mypy-extensions = ">=0.4.3"
+pathspec = ">=0.9.0"
+platformdirs = ">=2"
+tomli = {version = ">=1.1.0", markers = "python_full_version < \"3.11.0a7\""}
+typing-extensions = {version = ">=3.10.0.0", markers = "python_version < \"3.10\""}
+
+[package.extras]
+colorama = ["colorama (>=0.4.3)"]
+d = ["aiohttp (>=3.7.4)"]
+jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"]
+uvloop = ["uvloop (>=0.15.2)"]
+
+[[package]]
+name = "cachecontrol"
+version = "0.13.1"
+description = "httplib2 caching for requests"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
+]
+
+[package.dependencies]
+filelock = {version = ">=3.8.0", optional = true, markers = "extra == \"filecache\""}
+msgpack = ">=0.5.2"
+requests = ">=2.16.0"
+
+[package.extras]
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
+filecache = ["filelock (>=3.8.0)"]
+redis = ["redis (>=2.10.5)"]
+
+[[package]]
+name = "certifi"
+version = "2023.5.7"
+description = "Python package for providing Mozilla's CA Bundle."
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "certifi-2023.5.7-py3-none-any.whl", hash = "sha256:c6c2e98f5c7869efca1f8916fed228dd91539f9f1b444c314c06eef02980c716"},
+ {file = "certifi-2023.5.7.tar.gz", hash = "sha256:0f0d56dc5a6ad56fd4ba36484d6cc34451e1c6548c61daad8c320169f91eddc7"},
+]
+
+[[package]]
+name = "cffi"
+version = "1.15.1"
+description = "Foreign Function Interface for Python calling C code."
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "cffi-1.15.1-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:a66d3508133af6e8548451b25058d5812812ec3798c886bf38ed24a98216fab2"},
+ {file = "cffi-1.15.1-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:470c103ae716238bbe698d67ad020e1db9d9dba34fa5a899b5e21577e6d52ed2"},
+ {file = "cffi-1.15.1-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9ad5db27f9cabae298d151c85cf2bad1d359a1b9c686a275df03385758e2f914"},
+ {file = "cffi-1.15.1-cp27-cp27m-win32.whl", hash = "sha256:b3bbeb01c2b273cca1e1e0c5df57f12dce9a4dd331b4fa1635b8bec26350bde3"},
+ {file = "cffi-1.15.1-cp27-cp27m-win_amd64.whl", hash = "sha256:e00b098126fd45523dd056d2efba6c5a63b71ffe9f2bbe1a4fe1716e1d0c331e"},
+ {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:d61f4695e6c866a23a21acab0509af1cdfd2c013cf256bbf5b6b5e2695827162"},
+ {file = "cffi-1.15.1-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:ed9cb427ba5504c1dc15ede7d516b84757c3e3d7868ccc85121d9310d27eed0b"},
+ {file = "cffi-1.15.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:39d39875251ca8f612b6f33e6b1195af86d1b3e60086068be9cc053aa4376e21"},
+ {file = "cffi-1.15.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:285d29981935eb726a4399badae8f0ffdff4f5050eaa6d0cfc3f64b857b77185"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3eb6971dcff08619f8d91607cfc726518b6fa2a9eba42856be181c6d0d9515fd"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:21157295583fe8943475029ed5abdcf71eb3911894724e360acff1d61c1d54bc"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5635bd9cb9731e6d4a1132a498dd34f764034a8ce60cef4f5319c0541159392f"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2012c72d854c2d03e45d06ae57f40d78e5770d252f195b93f581acf3ba44496e"},
+ {file = "cffi-1.15.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd86c085fae2efd48ac91dd7ccffcfc0571387fe1193d33b6394db7ef31fe2a4"},
+ {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:fa6693661a4c91757f4412306191b6dc88c1703f780c8234035eac011922bc01"},
+ {file = "cffi-1.15.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:59c0b02d0a6c384d453fece7566d1c7e6b7bae4fc5874ef2ef46d56776d61c9e"},
+ {file = "cffi-1.15.1-cp310-cp310-win32.whl", hash = "sha256:cba9d6b9a7d64d4bd46167096fc9d2f835e25d7e4c121fb2ddfc6528fb0413b2"},
+ {file = "cffi-1.15.1-cp310-cp310-win_amd64.whl", hash = "sha256:ce4bcc037df4fc5e3d184794f27bdaab018943698f4ca31630bc7f84a7b69c6d"},
+ {file = "cffi-1.15.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3d08afd128ddaa624a48cf2b859afef385b720bb4b43df214f85616922e6a5ac"},
+ {file = "cffi-1.15.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:3799aecf2e17cf585d977b780ce79ff0dc9b78d799fc694221ce814c2c19db83"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a591fe9e525846e4d154205572a029f653ada1a78b93697f3b5a8f1f2bc055b9"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3548db281cd7d2561c9ad9984681c95f7b0e38881201e157833a2342c30d5e8c"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:91fc98adde3d7881af9b59ed0294046f3806221863722ba7d8d120c575314325"},
+ {file = "cffi-1.15.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:94411f22c3985acaec6f83c6df553f2dbe17b698cc7f8ae751ff2237d96b9e3c"},
+ {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:03425bdae262c76aad70202debd780501fabeaca237cdfddc008987c0e0f59ef"},
+ {file = "cffi-1.15.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc4d65aeeaa04136a12677d3dd0b1c0c94dc43abac5860ab33cceb42b801c1e8"},
+ {file = "cffi-1.15.1-cp311-cp311-win32.whl", hash = "sha256:a0f100c8912c114ff53e1202d0078b425bee3649ae34d7b070e9697f93c5d52d"},
+ {file = "cffi-1.15.1-cp311-cp311-win_amd64.whl", hash = "sha256:04ed324bda3cda42b9b695d51bb7d54b680b9719cfab04227cdd1e04e5de3104"},
+ {file = "cffi-1.15.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:50a74364d85fd319352182ef59c5c790484a336f6db772c1a9231f1c3ed0cbd7"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e263d77ee3dd201c3a142934a086a4450861778baaeeb45db4591ef65550b0a6"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cec7d9412a9102bdc577382c3929b337320c4c4c4849f2c5cdd14d7368c5562d"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4289fc34b2f5316fbb762d75362931e351941fa95fa18789191b33fc4cf9504a"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:173379135477dc8cac4bc58f45db08ab45d228b3363adb7af79436135d028405"},
+ {file = "cffi-1.15.1-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:6975a3fac6bc83c4a65c9f9fcab9e47019a11d3d2cf7f3c0d03431bf145a941e"},
+ {file = "cffi-1.15.1-cp36-cp36m-win32.whl", hash = "sha256:2470043b93ff09bf8fb1d46d1cb756ce6132c54826661a32d4e4d132e1977adf"},
+ {file = "cffi-1.15.1-cp36-cp36m-win_amd64.whl", hash = "sha256:30d78fbc8ebf9c92c9b7823ee18eb92f2e6ef79b45ac84db507f52fbe3ec4497"},
+ {file = "cffi-1.15.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:198caafb44239b60e252492445da556afafc7d1e3ab7a1fb3f0584ef6d742375"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5ef34d190326c3b1f822a5b7a45f6c4535e2f47ed06fec77d3d799c450b2651e"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8102eaf27e1e448db915d08afa8b41d6c7ca7a04b7d73af6514df10a3e74bd82"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5df2768244d19ab7f60546d0c7c63ce1581f7af8b5de3eb3004b9b6fc8a9f84b"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a8c4917bd7ad33e8eb21e9a5bbba979b49d9a97acb3a803092cbc1133e20343c"},
+ {file = "cffi-1.15.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e2642fe3142e4cc4af0799748233ad6da94c62a8bec3a6648bf8ee68b1c7426"},
+ {file = "cffi-1.15.1-cp37-cp37m-win32.whl", hash = "sha256:e229a521186c75c8ad9490854fd8bbdd9a0c9aa3a524326b55be83b54d4e0ad9"},
+ {file = "cffi-1.15.1-cp37-cp37m-win_amd64.whl", hash = "sha256:a0b71b1b8fbf2b96e41c4d990244165e2c9be83d54962a9a1d118fd8657d2045"},
+ {file = "cffi-1.15.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:320dab6e7cb2eacdf0e658569d2575c4dad258c0fcc794f46215e1e39f90f2c3"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e74c6b51a9ed6589199c787bf5f9875612ca4a8a0785fb2d4a84429badaf22a"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a5c84c68147988265e60416b57fc83425a78058853509c1b0629c180094904a5"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b926aa83d1edb5aa5b427b4053dc420ec295a08e40911296b9eb1b6170f6cca"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:87c450779d0914f2861b8526e035c5e6da0a3199d8f1add1a665e1cbc6fc6d02"},
+ {file = "cffi-1.15.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f2c9f67e9821cad2e5f480bc8d83b8742896f1242dba247911072d4fa94c192"},
+ {file = "cffi-1.15.1-cp38-cp38-win32.whl", hash = "sha256:8b7ee99e510d7b66cdb6c593f21c043c248537a32e0bedf02e01e9553a172314"},
+ {file = "cffi-1.15.1-cp38-cp38-win_amd64.whl", hash = "sha256:00a9ed42e88df81ffae7a8ab6d9356b371399b91dbdf0c3cb1e84c03a13aceb5"},
+ {file = "cffi-1.15.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:54a2db7b78338edd780e7ef7f9f6c442500fb0d41a5a4ea24fff1c929d5af585"},
+ {file = "cffi-1.15.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:fcd131dd944808b5bdb38e6f5b53013c5aa4f334c5cad0c72742f6eba4b73db0"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7473e861101c9e72452f9bf8acb984947aa1661a7704553a9f6e4baa5ba64415"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6c9a799e985904922a4d207a94eae35c78ebae90e128f0c4e521ce339396be9d"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3bcde07039e586f91b45c88f8583ea7cf7a0770df3a1649627bf598332cb6984"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:33ab79603146aace82c2427da5ca6e58f2b3f2fb5da893ceac0c42218a40be35"},
+ {file = "cffi-1.15.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5d598b938678ebf3c67377cdd45e09d431369c3b1a5b331058c338e201f12b27"},
+ {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:db0fbb9c62743ce59a9ff687eb5f4afbe77e5e8403d6697f7446e5f609976f76"},
+ {file = "cffi-1.15.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:98d85c6a2bef81588d9227dde12db8a7f47f639f4a17c9ae08e773aa9c697bf3"},
+ {file = "cffi-1.15.1-cp39-cp39-win32.whl", hash = "sha256:40f4774f5a9d4f5e344f31a32b5096977b5d48560c5592e2f3d2c4374bd543ee"},
+ {file = "cffi-1.15.1-cp39-cp39-win_amd64.whl", hash = "sha256:70df4e3b545a17496c9b3f41f5115e69a4f2e77e94e1d2a8e1070bc0c38c8a3c"},
+ {file = "cffi-1.15.1.tar.gz", hash = "sha256:d400bfb9a37b1351253cb402671cea7e89bdecc294e8016a707f6d1d8ac934f9"},
+]
+
+[package.dependencies]
+pycparser = "*"
+
+[[package]]
+name = "charset-normalizer"
+version = "3.1.0"
+description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet."
+category = "main"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "charset-normalizer-3.1.0.tar.gz", hash = "sha256:34e0a2f9c370eb95597aae63bf85eb5e96826d81e3dcf88b8886012906f509b5"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e0ac8959c929593fee38da1c2b64ee9778733cdf03c482c9ff1d508b6b593b2b"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d7fc3fca01da18fbabe4625d64bb612b533533ed10045a2ac3dd194bfa656b60"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:04eefcee095f58eaabe6dc3cc2262f3bcd776d2c67005880894f447b3f2cb9c1"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20064ead0717cf9a73a6d1e779b23d149b53daf971169289ed2ed43a71e8d3b0"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1435ae15108b1cb6fffbcea2af3d468683b7afed0169ad718451f8db5d1aff6f"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c84132a54c750fda57729d1e2599bb598f5fa0344085dbde5003ba429a4798c0"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75f2568b4189dda1c567339b48cba4ac7384accb9c2a7ed655cd86b04055c795"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:11d3bcb7be35e7b1bba2c23beedac81ee893ac9871d0ba79effc7fc01167db6c"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:891cf9b48776b5c61c700b55a598621fdb7b1e301a550365571e9624f270c203"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5f008525e02908b20e04707a4f704cd286d94718f48bb33edddc7d7b584dddc1"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:b06f0d3bf045158d2fb8837c5785fe9ff9b8c93358be64461a1089f5da983137"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:49919f8400b5e49e961f320c735388ee686a62327e773fa5b3ce6721f7e785ce"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22908891a380d50738e1f978667536f6c6b526a2064156203d418f4856d6e86a"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-win32.whl", hash = "sha256:12d1a39aa6b8c6f6248bb54550efcc1c38ce0d8096a146638fd4738e42284448"},
+ {file = "charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:65ed923f84a6844de5fd29726b888e58c62820e0769b76565480e1fdc3d062f8"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9a3267620866c9d17b959a84dd0bd2d45719b817245e49371ead79ed4f710d19"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6734e606355834f13445b6adc38b53c0fd45f1a56a9ba06c2058f86893ae8017"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:f8303414c7b03f794347ad062c0516cee0e15f7a612abd0ce1e25caf6ceb47df"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aaf53a6cebad0eae578f062c7d462155eada9c172bd8c4d250b8c1d8eb7f916a"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3dc5b6a8ecfdc5748a7e429782598e4f17ef378e3e272eeb1340ea57c9109f41"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1b25e3ad6c909f398df8921780d6a3d120d8c09466720226fc621605b6f92b1"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ca564606d2caafb0abe6d1b5311c2649e8071eb241b2d64e75a0d0065107e62"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b82fab78e0b1329e183a65260581de4375f619167478dddab510c6c6fb04d9b6"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:bd7163182133c0c7701b25e604cf1611c0d87712e56e88e7ee5d72deab3e76b5"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:11d117e6c63e8f495412d37e7dc2e2fff09c34b2d09dbe2bee3c6229577818be"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cf6511efa4801b9b38dc5546d7547d5b5c6ef4b081c60b23e4d941d0eba9cbeb"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:abc1185d79f47c0a7aaf7e2412a0eb2c03b724581139193d2d82b3ad8cbb00ac"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cb7b2ab0188829593b9de646545175547a70d9a6e2b63bf2cd87a0a391599324"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-win32.whl", hash = "sha256:c36bcbc0d5174a80d6cccf43a0ecaca44e81d25be4b7f90f0ed7bcfbb5a00909"},
+ {file = "charset_normalizer-3.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:cca4def576f47a09a943666b8f829606bcb17e2bc2d5911a46c8f8da45f56755"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:0c95f12b74681e9ae127728f7e5409cbbef9cd914d5896ef238cc779b8152373"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fca62a8301b605b954ad2e9c3666f9d97f63872aa4efcae5492baca2056b74ab"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ac0aa6cd53ab9a31d397f8303f92c42f534693528fafbdb997c82bae6e477ad9"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c3af8e0f07399d3176b179f2e2634c3ce9c1301379a6b8c9c9aeecd481da494f"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3a5fc78f9e3f501a1614a98f7c54d3969f3ad9bba8ba3d9b438c3bc5d047dd28"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:628c985afb2c7d27a4800bfb609e03985aaecb42f955049957814e0491d4006d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:74db0052d985cf37fa111828d0dd230776ac99c740e1a758ad99094be4f1803d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:1e8fcdd8f672a1c4fc8d0bd3a2b576b152d2a349782d1eb0f6b8e52e9954731d"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:04afa6387e2b282cf78ff3dbce20f0cc071c12dc8f685bd40960cc68644cfea6"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:dd5653e67b149503c68c4018bf07e42eeed6b4e956b24c00ccdf93ac79cdff84"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d2686f91611f9e17f4548dbf050e75b079bbc2a82be565832bc8ea9047b61c8c"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-win32.whl", hash = "sha256:4155b51ae05ed47199dc5b2a4e62abccb274cee6b01da5b895099b61b1982974"},
+ {file = "charset_normalizer-3.1.0-cp37-cp37m-win_amd64.whl", hash = "sha256:322102cdf1ab682ecc7d9b1c5eed4ec59657a65e1c146a0da342b78f4112db23"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:e633940f28c1e913615fd624fcdd72fdba807bf53ea6925d6a588e84e1151531"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a06f32c9634a8705f4ca9946d667609f52cf130d5548881401f1eb2c39b1e2c"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:7381c66e0561c5757ffe616af869b916c8b4e42b367ab29fedc98481d1e74e14"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3573d376454d956553c356df45bb824262c397c6e26ce43e8203c4c540ee0acb"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e89df2958e5159b811af9ff0f92614dabf4ff617c03a4c1c6ff53bf1c399e0e1"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78cacd03e79d009d95635e7d6ff12c21eb89b894c354bd2b2ed0b4763373693b"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:de5695a6f1d8340b12a5d6d4484290ee74d61e467c39ff03b39e30df62cf83a0"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c60b9c202d00052183c9be85e5eaf18a4ada0a47d188a83c8f5c5b23252f649"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:f645caaf0008bacf349875a974220f1f1da349c5dbe7c4ec93048cdc785a3326"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ea9f9c6034ea2d93d9147818f17c2a0860d41b71c38b9ce4d55f21b6f9165a11"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:80d1543d58bd3d6c271b66abf454d437a438dff01c3e62fdbcd68f2a11310d4b"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:73dc03a6a7e30b7edc5b01b601e53e7fc924b04e1835e8e407c12c037e81adbd"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:6f5c2e7bc8a4bf7c426599765b1bd33217ec84023033672c1e9a8b35eaeaaaf8"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-win32.whl", hash = "sha256:12a2b561af122e3d94cdb97fe6fb2bb2b82cef0cdca131646fdb940a1eda04f0"},
+ {file = "charset_normalizer-3.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:3160a0fd9754aab7d47f95a6b63ab355388d890163eb03b2d2b87ab0a30cfa59"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:38e812a197bf8e71a59fe55b757a84c1f946d0ac114acafaafaf21667a7e169e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6baf0baf0d5d265fa7944feb9f7451cc316bfe30e8df1a61b1bb08577c554f31"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:8f25e17ab3039b05f762b0a55ae0b3632b2e073d9c8fc88e89aca31a6198e88f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3747443b6a904001473370d7810aa19c3a180ccd52a7157aacc264a5ac79265e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b116502087ce8a6b7a5f1814568ccbd0e9f6cfd99948aa59b0e241dc57cf739f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d16fd5252f883eb074ca55cb622bc0bee49b979ae4e8639fff6ca3ff44f9f854"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:21fa558996782fc226b529fdd2ed7866c2c6ec91cee82735c98a197fae39f706"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f6c7a8a57e9405cad7485f4c9d3172ae486cfef1344b5ddd8e5239582d7355e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ac3775e3311661d4adace3697a52ac0bab17edd166087d493b52d4f4f553f9f0"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:10c93628d7497c81686e8e5e557aafa78f230cd9e77dd0c40032ef90c18f2230"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:6f4f4668e1831850ebcc2fd0b1cd11721947b6dc7c00bf1c6bd3c929ae14f2c7"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:0be65ccf618c1e7ac9b849c315cc2e8a8751d9cfdaa43027d4f6624bd587ab7e"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:53d0a3fa5f8af98a1e261de6a3943ca631c526635eb5817a87a59d9a57ebf48f"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-win32.whl", hash = "sha256:a04f86f41a8916fe45ac5024ec477f41f886b3c435da2d4e3d2709b22ab02af1"},
+ {file = "charset_normalizer-3.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:830d2948a5ec37c386d3170c483063798d7879037492540f10a475e3fd6f244b"},
+ {file = "charset_normalizer-3.1.0-py3-none-any.whl", hash = "sha256:3d9098b479e78c85080c98e1e35ff40b4a31d8953102bb0fd7d1b6f8a2111a3d"},
+]
+
+[[package]]
+name = "click"
+version = "8.1.3"
+description = "Composable command line interface toolkit"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "click-8.1.3-py3-none-any.whl", hash = "sha256:bb4d8133cb15a609f44e8213d9b391b0809795062913b383c62be0ee95b1db48"},
+ {file = "click-8.1.3.tar.gz", hash = "sha256:7682dc8afb30297001674575ea00d1814d808d6a36af415a82bd481d37ba7b8e"},
+]
+
+[package.dependencies]
+colorama = {version = "*", markers = "platform_system == \"Windows\""}
+
+[[package]]
+name = "colorama"
+version = "0.4.6"
+description = "Cross-platform colored terminal text."
+category = "main"
+optional = false
+python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7"
+files = [
+ {file = "colorama-0.4.6-py2.py3-none-any.whl", hash = "sha256:4f1d9991f5acc0ca119f9d443620b77f9d6b33703e51011c16baf57afb285fc6"},
+ {file = "colorama-0.4.6.tar.gz", hash = "sha256:08695f5cb7ed6e0531a20572697297273c47b8cae5a63ffc6d6ed5c201be6e44"},
+]
+
+[[package]]
+name = "coverage"
+version = "7.2.7"
+description = "Code coverage measurement for Python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "coverage-7.2.7-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d39b5b4f2a66ccae8b7263ac3c8170994b65266797fb96cbbfd3fb5b23921db8"},
+ {file = "coverage-7.2.7-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:6d040ef7c9859bb11dfeb056ff5b3872436e3b5e401817d87a31e1750b9ae2fb"},
+ {file = "coverage-7.2.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ba90a9563ba44a72fda2e85302c3abc71c5589cea608ca16c22b9804262aaeb6"},
+ {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e7d9405291c6928619403db1d10bd07888888ec1abcbd9748fdaa971d7d661b2"},
+ {file = "coverage-7.2.7-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:31563e97dae5598556600466ad9beea39fb04e0229e61c12eaa206e0aa202063"},
+ {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ebba1cd308ef115925421d3e6a586e655ca5a77b5bf41e02eb0e4562a111f2d1"},
+ {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:cb017fd1b2603ef59e374ba2063f593abe0fc45f2ad9abdde5b4d83bd922a353"},
+ {file = "coverage-7.2.7-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d62a5c7dad11015c66fbb9d881bc4caa5b12f16292f857842d9d1871595f4495"},
+ {file = "coverage-7.2.7-cp310-cp310-win32.whl", hash = "sha256:ee57190f24fba796e36bb6d3aa8a8783c643d8fa9760c89f7a98ab5455fbf818"},
+ {file = "coverage-7.2.7-cp310-cp310-win_amd64.whl", hash = "sha256:f75f7168ab25dd93110c8a8117a22450c19976afbc44234cbf71481094c1b850"},
+ {file = "coverage-7.2.7-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:06a9a2be0b5b576c3f18f1a241f0473575c4a26021b52b2a85263a00f034d51f"},
+ {file = "coverage-7.2.7-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5baa06420f837184130752b7c5ea0808762083bf3487b5038d68b012e5937dbe"},
+ {file = "coverage-7.2.7-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fdec9e8cbf13a5bf63290fc6013d216a4c7232efb51548594ca3631a7f13c3a3"},
+ {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:52edc1a60c0d34afa421c9c37078817b2e67a392cab17d97283b64c5833f427f"},
+ {file = "coverage-7.2.7-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:63426706118b7f5cf6bb6c895dc215d8a418d5952544042c8a2d9fe87fcf09cb"},
+ {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:afb17f84d56068a7c29f5fa37bfd38d5aba69e3304af08ee94da8ed5b0865833"},
+ {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:48c19d2159d433ccc99e729ceae7d5293fbffa0bdb94952d3579983d1c8c9d97"},
+ {file = "coverage-7.2.7-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:0e1f928eaf5469c11e886fe0885ad2bf1ec606434e79842a879277895a50942a"},
+ {file = "coverage-7.2.7-cp311-cp311-win32.whl", hash = "sha256:33d6d3ea29d5b3a1a632b3c4e4f4ecae24ef170b0b9ee493883f2df10039959a"},
+ {file = "coverage-7.2.7-cp311-cp311-win_amd64.whl", hash = "sha256:5b7540161790b2f28143191f5f8ec02fb132660ff175b7747b95dcb77ac26562"},
+ {file = "coverage-7.2.7-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:f2f67fe12b22cd130d34d0ef79206061bfb5eda52feb6ce0dba0644e20a03cf4"},
+ {file = "coverage-7.2.7-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a342242fe22407f3c17f4b499276a02b01e80f861f1682ad1d95b04018e0c0d4"},
+ {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:171717c7cb6b453aebac9a2ef603699da237f341b38eebfee9be75d27dc38e01"},
+ {file = "coverage-7.2.7-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:49969a9f7ffa086d973d91cec8d2e31080436ef0fb4a359cae927e742abfaaa6"},
+ {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:b46517c02ccd08092f4fa99f24c3b83d8f92f739b4657b0f146246a0ca6a831d"},
+ {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:a3d33a6b3eae87ceaefa91ffdc130b5e8536182cd6dfdbfc1aa56b46ff8c86de"},
+ {file = "coverage-7.2.7-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:976b9c42fb2a43ebf304fa7d4a310e5f16cc99992f33eced91ef6f908bd8f33d"},
+ {file = "coverage-7.2.7-cp312-cp312-win32.whl", hash = "sha256:8de8bb0e5ad103888d65abef8bca41ab93721647590a3f740100cd65c3b00511"},
+ {file = "coverage-7.2.7-cp312-cp312-win_amd64.whl", hash = "sha256:9e31cb64d7de6b6f09702bb27c02d1904b3aebfca610c12772452c4e6c21a0d3"},
+ {file = "coverage-7.2.7-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:58c2ccc2f00ecb51253cbe5d8d7122a34590fac9646a960d1430d5b15321d95f"},
+ {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d22656368f0e6189e24722214ed8d66b8022db19d182927b9a248a2a8a2f67eb"},
+ {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a895fcc7b15c3fc72beb43cdcbdf0ddb7d2ebc959edac9cef390b0d14f39f8a9"},
+ {file = "coverage-7.2.7-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e84606b74eb7de6ff581a7915e2dab7a28a0517fbe1c9239eb227e1354064dcd"},
+ {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0a5f9e1dbd7fbe30196578ca36f3fba75376fb99888c395c5880b355e2875f8a"},
+ {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:419bfd2caae268623dd469eff96d510a920c90928b60f2073d79f8fe2bbc5959"},
+ {file = "coverage-7.2.7-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:2aee274c46590717f38ae5e4650988d1af340fe06167546cc32fe2f58ed05b02"},
+ {file = "coverage-7.2.7-cp37-cp37m-win32.whl", hash = "sha256:61b9a528fb348373c433e8966535074b802c7a5d7f23c4f421e6c6e2f1697a6f"},
+ {file = "coverage-7.2.7-cp37-cp37m-win_amd64.whl", hash = "sha256:b1c546aca0ca4d028901d825015dc8e4d56aac4b541877690eb76490f1dc8ed0"},
+ {file = "coverage-7.2.7-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:54b896376ab563bd38453cecb813c295cf347cf5906e8b41d340b0321a5433e5"},
+ {file = "coverage-7.2.7-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3d376df58cc111dc8e21e3b6e24606b5bb5dee6024f46a5abca99124b2229ef5"},
+ {file = "coverage-7.2.7-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e330fc79bd7207e46c7d7fd2bb4af2963f5f635703925543a70b99574b0fea9"},
+ {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1e9d683426464e4a252bf70c3498756055016f99ddaec3774bf368e76bbe02b6"},
+ {file = "coverage-7.2.7-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8d13c64ee2d33eccf7437961b6ea7ad8673e2be040b4f7fd4fd4d4d28d9ccb1e"},
+ {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:b7aa5f8a41217360e600da646004f878250a0d6738bcdc11a0a39928d7dc2050"},
+ {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:8fa03bce9bfbeeef9f3b160a8bed39a221d82308b4152b27d82d8daa7041fee5"},
+ {file = "coverage-7.2.7-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:245167dd26180ab4c91d5e1496a30be4cd721a5cf2abf52974f965f10f11419f"},
+ {file = "coverage-7.2.7-cp38-cp38-win32.whl", hash = "sha256:d2c2db7fd82e9b72937969bceac4d6ca89660db0a0967614ce2481e81a0b771e"},
+ {file = "coverage-7.2.7-cp38-cp38-win_amd64.whl", hash = "sha256:2e07b54284e381531c87f785f613b833569c14ecacdcb85d56b25c4622c16c3c"},
+ {file = "coverage-7.2.7-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:537891ae8ce59ef63d0123f7ac9e2ae0fc8b72c7ccbe5296fec45fd68967b6c9"},
+ {file = "coverage-7.2.7-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:06fb182e69f33f6cd1d39a6c597294cff3143554b64b9825d1dc69d18cc2fff2"},
+ {file = "coverage-7.2.7-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:201e7389591af40950a6480bd9edfa8ed04346ff80002cec1a66cac4549c1ad7"},
+ {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f6951407391b639504e3b3be51b7ba5f3528adbf1a8ac3302b687ecababf929e"},
+ {file = "coverage-7.2.7-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6f48351d66575f535669306aa7d6d6f71bc43372473b54a832222803eb956fd1"},
+ {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b29019c76039dc3c0fd815c41392a044ce555d9bcdd38b0fb60fb4cd8e475ba9"},
+ {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:81c13a1fc7468c40f13420732805a4c38a105d89848b7c10af65a90beff25250"},
+ {file = "coverage-7.2.7-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:975d70ab7e3c80a3fe86001d8751f6778905ec723f5b110aed1e450da9d4b7f2"},
+ {file = "coverage-7.2.7-cp39-cp39-win32.whl", hash = "sha256:7ee7d9d4822c8acc74a5e26c50604dff824710bc8de424904c0982e25c39c6cb"},
+ {file = "coverage-7.2.7-cp39-cp39-win_amd64.whl", hash = "sha256:eb393e5ebc85245347950143969b241d08b52b88a3dc39479822e073a1a8eb27"},
+ {file = "coverage-7.2.7-pp37.pp38.pp39-none-any.whl", hash = "sha256:b7b4c971f05e6ae490fef852c218b0e79d4e52f79ef0c8475566584a8fb3e01d"},
+ {file = "coverage-7.2.7.tar.gz", hash = "sha256:924d94291ca674905fe9481f12294eb11f2d3d3fd1adb20314ba89e94f44ed59"},
+]
+
+[package.extras]
+toml = ["tomli"]
+
+[[package]]
+name = "cryptography"
+version = "41.0.1"
+description = "cryptography is a package which provides cryptographic recipes and primitives to Python developers."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_universal2.whl", hash = "sha256:f73bff05db2a3e5974a6fd248af2566134d8981fd7ab012e5dd4ddb1d9a70699"},
+ {file = "cryptography-41.0.1-cp37-abi3-macosx_10_12_x86_64.whl", hash = "sha256:1a5472d40c8f8e91ff7a3d8ac6dfa363d8e3138b961529c996f3e2df0c7a411a"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fa01527046ca5facdf973eef2535a27fec4cb651e4daec4d043ef63f6ecd4ca"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b46e37db3cc267b4dea1f56da7346c9727e1209aa98487179ee8ebed09d21e43"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:d198820aba55660b4d74f7b5fd1f17db3aa5eb3e6893b0a41b75e84e4f9e0e4b"},
+ {file = "cryptography-41.0.1-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:948224d76c4b6457349d47c0c98657557f429b4e93057cf5a2f71d603e2fc3a3"},
+ {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:059e348f9a3c1950937e1b5d7ba1f8e968508ab181e75fc32b879452f08356db"},
+ {file = "cryptography-41.0.1-cp37-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:b4ceb5324b998ce2003bc17d519080b4ec8d5b7b70794cbd2836101406a9be31"},
+ {file = "cryptography-41.0.1-cp37-abi3-win32.whl", hash = "sha256:8f4ab7021127a9b4323537300a2acfb450124b2def3756f64dc3a3d2160ee4b5"},
+ {file = "cryptography-41.0.1-cp37-abi3-win_amd64.whl", hash = "sha256:1fee5aacc7367487b4e22484d3c7e547992ed726d14864ee33c0176ae43b0d7c"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-macosx_10_12_x86_64.whl", hash = "sha256:9a6c7a3c87d595608a39980ebaa04d5a37f94024c9f24eb7d10262b92f739ddb"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:5d092fdfedaec4cbbffbf98cddc915ba145313a6fdaab83c6e67f4e6c218e6f3"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:1a8e6c2de6fbbcc5e14fd27fb24414507cb3333198ea9ab1258d916f00bc3039"},
+ {file = "cryptography-41.0.1-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:cb33ccf15e89f7ed89b235cff9d49e2e62c6c981a6061c9c8bb47ed7951190bc"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:5f0ff6e18d13a3de56f609dd1fd11470918f770c6bd5d00d632076c727d35485"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:7bfc55a5eae8b86a287747053140ba221afc65eb06207bedf6e019b8934b477c"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:eb8163f5e549a22888c18b0d53d6bb62a20510060a22fd5a995ec8a05268df8a"},
+ {file = "cryptography-41.0.1-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:8dde71c4169ec5ccc1087bb7521d54251c016f126f922ab2dfe6649170a3b8c5"},
+ {file = "cryptography-41.0.1.tar.gz", hash = "sha256:d34579085401d3f49762d2f7d6634d6b6c2ae1242202e860f4d26b046e3a1006"},
+]
+
+[package.dependencies]
+cffi = ">=1.12"
+
+[package.extras]
+docs = ["sphinx (>=5.3.0)", "sphinx-rtd-theme (>=1.1.1)"]
+docstest = ["pyenchant (>=1.6.11)", "sphinxcontrib-spelling (>=4.0.1)", "twine (>=1.12.0)"]
+nox = ["nox"]
+pep8test = ["black", "check-sdist", "mypy", "ruff"]
+sdist = ["build"]
+ssh = ["bcrypt (>=3.1.5)"]
+test = ["pretend", "pytest (>=6.2.0)", "pytest-benchmark", "pytest-cov", "pytest-xdist"]
+test-randomorder = ["pytest-randomly"]
+
+[[package]]
+name = "cyclonedx-python-lib"
+version = "4.0.1"
+description = "A library for producing CycloneDX SBOM (Software Bill of Materials) files."
+category = "dev"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "cyclonedx_python_lib-4.0.1-py3-none-any.whl", hash = "sha256:907b64f00df85d727a425de86604768b248cf19285993729e04f17bec767f692"},
+ {file = "cyclonedx_python_lib-4.0.1.tar.gz", hash = "sha256:878e33b8e0080c786f6cbd4c6f87ad610db65d6a3a686a5698415d9cfcd8925d"},
+]
+
+[package.dependencies]
+packageurl-python = ">=0.11"
+py-serializable = ">=0.11.1,<0.12.0"
+sortedcontainers = ">=2.4.0,<3.0.0"
+
+[[package]]
+name = "datasets"
+version = "2.13.1"
+description = "HuggingFace community-driven open-source library of datasets"
+category = "main"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
+]
+
+[package.dependencies]
+aiohttp = "*"
+dill = ">=0.3.0,<0.3.7"
+fsspec = {version = ">=2021.11.1", extras = ["http"]}
+huggingface-hub = ">=0.11.0,<1.0.0"
+librosa = {version = "*", optional = true, markers = "extra == \"audio\""}
+multiprocess = "*"
+numpy = ">=1.17"
+packaging = "*"
+pandas = "*"
+Pillow = {version = ">=6.2.1", optional = true, markers = "extra == \"vision\""}
+pyarrow = ">=8.0.0"
+pyyaml = ">=5.1"
+requests = ">=2.19.0"
+soundfile = {version = ">=0.12.1", optional = true, markers = "extra == \"audio\""}
+tqdm = ">=4.62.1"
+xxhash = "*"
+
+[package.extras]
+apache-beam = ["apache-beam (>=2.26.0,<2.44.0)"]
+audio = ["librosa", "soundfile (>=0.12.1)"]
+benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)", "torch (==1.7.1)", "transformers (==3.0.2)"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+docs = ["s3fs"]
+jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+quality = ["black (>=23.1,<24.0)", "pyyaml (>=5.3.1)", "ruff (>=0.0.241)"]
+s3 = ["s3fs"]
+tensorflow = ["tensorflow (>=2.2.0,!=2.6.0,!=2.6.1)", "tensorflow-macos"]
+tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+torch = ["torch"]
+vision = ["Pillow (>=6.2.1)"]
+
+[[package]]
+name = "decorator"
+version = "5.1.1"
+description = "Decorators for Humans"
+category = "main"
+optional = false
+python-versions = ">=3.5"
+files = [
+ {file = "decorator-5.1.1-py3-none-any.whl", hash = "sha256:b8c3f85900b9dc423225913c5aace94729fe1fa9763b38939a95226f02d37186"},
+ {file = "decorator-5.1.1.tar.gz", hash = "sha256:637996211036b6385ef91435e4fae22989472f9d571faba8927ba8253acbc330"},
+]
+
+[[package]]
+name = "defusedxml"
+version = "0.7.1"
+description = "XML bomb protection for Python stdlib modules"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "defusedxml-0.7.1-py2.py3-none-any.whl", hash = "sha256:a352e7e428770286cc899e2542b6cdaedb2b4953ff269a210103ec58f6198a61"},
+ {file = "defusedxml-0.7.1.tar.gz", hash = "sha256:1bb3032db185915b62d7c6209c5a8792be6a32ab2fedacc84e01b52c51aa3e69"},
+]
+
+[[package]]
+name = "dill"
+version = "0.3.6"
+description = "serialize all of python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "dill-0.3.6-py3-none-any.whl", hash = "sha256:a07ffd2351b8c678dfc4a856a3005f8067aea51d6ba6c700796a4d9e280f39f0"},
+ {file = "dill-0.3.6.tar.gz", hash = "sha256:e5db55f3687856d8fbdab002ed78544e1c4559a130302693d839dfe8f93f2373"},
+]
+
+[package.extras]
+graph = ["objgraph (>=1.7.2)"]
+
+[[package]]
+name = "dnspython"
+version = "1.16.0"
+description = "DNS toolkit"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "dnspython-1.16.0-py2.py3-none-any.whl", hash = "sha256:f69c21288a962f4da86e56c4905b49d11aba7938d3d740e80d9e366ee4f1632d"},
+ {file = "dnspython-1.16.0.zip", hash = "sha256:36c5e8e38d4369a08b6780b7f27d790a292b2b08eea01607865bf0936c558e01"},
+]
+
+[package.extras]
+dnssec = ["ecdsa (>=0.13)", "pycryptodome"]
+idna = ["idna (>=2.1)"]
+
+[[package]]
+name = "environs"
+version = "9.5.0"
+description = "simplified environment variable parsing"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "environs-9.5.0-py2.py3-none-any.whl", hash = "sha256:1e549569a3de49c05f856f40bce86979e7d5ffbbc4398e7f338574c220189124"},
+ {file = "environs-9.5.0.tar.gz", hash = "sha256:a76307b36fbe856bdca7ee9161e6c466fd7fcffc297109a118c59b54e27e30c9"},
+]
+
+[package.dependencies]
+marshmallow = ">=3.0.0"
+python-dotenv = "*"
+
+[package.extras]
+dev = ["dj-database-url", "dj-email-url", "django-cache-url", "flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)", "pytest", "tox"]
+django = ["dj-database-url", "dj-email-url", "django-cache-url"]
+lint = ["flake8 (==4.0.1)", "flake8-bugbear (==21.9.2)", "mypy (==0.910)", "pre-commit (>=2.4,<3.0)"]
+tests = ["dj-database-url", "dj-email-url", "django-cache-url", "pytest"]
+
+[[package]]
+name = "exceptiongroup"
+version = "1.1.2"
+description = "Backport of PEP 654 (exception groups)"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "exceptiongroup-1.1.2-py3-none-any.whl", hash = "sha256:e346e69d186172ca7cf029c8c1d16235aa0e04035e5750b4b95039e65204328f"},
+ {file = "exceptiongroup-1.1.2.tar.gz", hash = "sha256:12c3e887d6485d16943a309616de20ae5582633e0a2eda17f4e10fd61c1e8af5"},
+]
+
+[package.extras]
+test = ["pytest (>=6)"]
+
+[[package]]
+name = "filelock"
+version = "3.12.2"
+description = "A platform independent file lock."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "filelock-3.12.2-py3-none-any.whl", hash = "sha256:cbb791cdea2a72f23da6ac5b5269ab0a0d161e9ef0100e653b69049a7706d1ec"},
+ {file = "filelock-3.12.2.tar.gz", hash = "sha256:002740518d8aa59a26b0c76e10fb8c6e15eae825d34b6fdf670333fd7b938d81"},
+]
+
+[package.extras]
+docs = ["furo (>=2023.5.20)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"]
+testing = ["covdefaults (>=2.3)", "coverage (>=7.2.7)", "diff-cover (>=7.5)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)", "pytest-timeout (>=2.1)"]
+
+[[package]]
+name = "flake8"
+version = "3.9.2"
+description = "the modular source code checker: pep8 pyflakes and co"
+category = "dev"
+optional = false
+python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,>=2.7"
+files = [
+ {file = "flake8-3.9.2-py2.py3-none-any.whl", hash = "sha256:bf8fd333346d844f616e8d47905ef3a3384edae6b4e9beb0c5101e25e3110907"},
+ {file = "flake8-3.9.2.tar.gz", hash = "sha256:07528381786f2a6237b061f6e96610a4167b226cb926e2aa2b6b1d78057c576b"},
+]
+
+[package.dependencies]
+mccabe = ">=0.6.0,<0.7.0"
+pycodestyle = ">=2.7.0,<2.8.0"
+pyflakes = ">=2.3.0,<2.4.0"
+
+[[package]]
+name = "frozenlist"
+version = "1.3.3"
+description = "A list-like structure which implements collections.abc.MutableSequence"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:ff8bf625fe85e119553b5383ba0fb6aa3d0ec2ae980295aaefa552374926b3f4"},
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:dfbac4c2dfcc082fcf8d942d1e49b6aa0766c19d3358bd86e2000bf0fa4a9cf0"},
+ {file = "frozenlist-1.3.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:b1c63e8d377d039ac769cd0926558bb7068a1f7abb0f003e3717ee003ad85530"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7fdfc24dcfce5b48109867c13b4cb15e4660e7bd7661741a391f821f23dfdca7"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2c926450857408e42f0bbc295e84395722ce74bae69a3b2aa2a65fe22cb14b99"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1841e200fdafc3d51f974d9d377c079a0694a8f06de2e67b48150328d66d5483"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f470c92737afa7d4c3aacc001e335062d582053d4dbe73cda126f2d7031068dd"},
+ {file = "frozenlist-1.3.3-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:783263a4eaad7c49983fe4b2e7b53fa9770c136c270d2d4bbb6d2192bf4d9caf"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:924620eef691990dfb56dc4709f280f40baee568c794b5c1885800c3ecc69816"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:ae4dc05c465a08a866b7a1baf360747078b362e6a6dbeb0c57f234db0ef88ae0"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:bed331fe18f58d844d39ceb398b77d6ac0b010d571cba8267c2e7165806b00ce"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:02c9ac843e3390826a265e331105efeab489ffaf4dd86384595ee8ce6d35ae7f"},
+ {file = "frozenlist-1.3.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:9545a33965d0d377b0bc823dcabf26980e77f1b6a7caa368a365a9497fb09420"},
+ {file = "frozenlist-1.3.3-cp310-cp310-win32.whl", hash = "sha256:d5cd3ab21acbdb414bb6c31958d7b06b85eeb40f66463c264a9b343a4e238642"},
+ {file = "frozenlist-1.3.3-cp310-cp310-win_amd64.whl", hash = "sha256:b756072364347cb6aa5b60f9bc18e94b2f79632de3b0190253ad770c5df17db1"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4395e2f8d83fbe0c627b2b696acce67868793d7d9750e90e39592b3626691b7"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:14143ae966a6229350021384870458e4777d1eae4c28d1a7aa47f24d030e6678"},
+ {file = "frozenlist-1.3.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5d8860749e813a6f65bad8285a0520607c9500caa23fea6ee407e63debcdbef6"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:23d16d9f477bb55b6154654e0e74557040575d9d19fe78a161bd33d7d76808e8"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:eb82dbba47a8318e75f679690190c10a5e1f447fbf9df41cbc4c3afd726d88cb"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9309869032abb23d196cb4e4db574232abe8b8be1339026f489eeb34a4acfd91"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a97b4fe50b5890d36300820abd305694cb865ddb7885049587a5678215782a6b"},
+ {file = "frozenlist-1.3.3-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c188512b43542b1e91cadc3c6c915a82a5eb95929134faf7fd109f14f9892ce4"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:303e04d422e9b911a09ad499b0368dc551e8c3cd15293c99160c7f1f07b59a48"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:0771aed7f596c7d73444c847a1c16288937ef988dc04fb9f7be4b2aa91db609d"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:66080ec69883597e4d026f2f71a231a1ee9887835902dbe6b6467d5a89216cf6"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:41fe21dc74ad3a779c3d73a2786bdf622ea81234bdd4faf90b8b03cad0c2c0b4"},
+ {file = "frozenlist-1.3.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:f20380df709d91525e4bee04746ba612a4df0972c1b8f8e1e8af997e678c7b81"},
+ {file = "frozenlist-1.3.3-cp311-cp311-win32.whl", hash = "sha256:f30f1928162e189091cf4d9da2eac617bfe78ef907a761614ff577ef4edfb3c8"},
+ {file = "frozenlist-1.3.3-cp311-cp311-win_amd64.whl", hash = "sha256:a6394d7dadd3cfe3f4b3b186e54d5d8504d44f2d58dcc89d693698e8b7132b32"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8df3de3a9ab8325f94f646609a66cbeeede263910c5c0de0101079ad541af332"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0693c609e9742c66ba4870bcee1ad5ff35462d5ffec18710b4ac89337ff16e27"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:cd4210baef299717db0a600d7a3cac81d46ef0e007f88c9335db79f8979c0d3d"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:394c9c242113bfb4b9aa36e2b80a05ffa163a30691c7b5a29eba82e937895d5e"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6327eb8e419f7d9c38f333cde41b9ae348bec26d840927332f17e887a8dcb70d"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2e24900aa13212e75e5b366cb9065e78bbf3893d4baab6052d1aca10d46d944c"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:3843f84a6c465a36559161e6c59dce2f2ac10943040c2fd021cfb70d58c4ad56"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:84610c1502b2461255b4c9b7d5e9c48052601a8957cd0aea6ec7a7a1e1fb9420"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:c21b9aa40e08e4f63a2f92ff3748e6b6c84d717d033c7b3438dd3123ee18f70e"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:efce6ae830831ab6a22b9b4091d411698145cb9b8fc869e1397ccf4b4b6455cb"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:40de71985e9042ca00b7953c4f41eabc3dc514a2d1ff534027f091bc74416401"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-win32.whl", hash = "sha256:180c00c66bde6146a860cbb81b54ee0df350d2daf13ca85b275123bbf85de18a"},
+ {file = "frozenlist-1.3.3-cp37-cp37m-win_amd64.whl", hash = "sha256:9bbbcedd75acdfecf2159663b87f1bb5cfc80e7cd99f7ddd9d66eb98b14a8411"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:034a5c08d36649591be1cbb10e09da9f531034acfe29275fc5454a3b101ce41a"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:ba64dc2b3b7b158c6660d49cdb1d872d1d0bf4e42043ad8d5006099479a194e5"},
+ {file = "frozenlist-1.3.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47df36a9fe24054b950bbc2db630d508cca3aa27ed0566c0baf661225e52c18e"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:008a054b75d77c995ea26629ab3a0c0d7281341f2fa7e1e85fa6153ae29ae99c"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:841ea19b43d438a80b4de62ac6ab21cfe6827bb8a9dc62b896acc88eaf9cecba"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e235688f42b36be2b6b06fc37ac2126a73b75fb8d6bc66dd632aa35286238703"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca713d4af15bae6e5d79b15c10c8522859a9a89d3b361a50b817c98c2fb402a2"},
+ {file = "frozenlist-1.3.3-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9ac5995f2b408017b0be26d4a1d7c61bce106ff3d9e3324374d66b5964325448"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:a4ae8135b11652b08a8baf07631d3ebfe65a4c87909dbef5fa0cdde440444ee4"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4ea42116ceb6bb16dbb7d526e242cb6747b08b7710d9782aa3d6732bd8d27649"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:810860bb4bdce7557bc0febb84bbd88198b9dbc2022d8eebe5b3590b2ad6c842"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:ee78feb9d293c323b59a6f2dd441b63339a30edf35abcb51187d2fc26e696d13"},
+ {file = "frozenlist-1.3.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:0af2e7c87d35b38732e810befb9d797a99279cbb85374d42ea61c1e9d23094b3"},
+ {file = "frozenlist-1.3.3-cp38-cp38-win32.whl", hash = "sha256:899c5e1928eec13fd6f6d8dc51be23f0d09c5281e40d9cf4273d188d9feeaf9b"},
+ {file = "frozenlist-1.3.3-cp38-cp38-win_amd64.whl", hash = "sha256:7f44e24fa70f6fbc74aeec3e971f60a14dde85da364aa87f15d1be94ae75aeef"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:2b07ae0c1edaa0a36339ec6cce700f51b14a3fc6545fdd32930d2c83917332cf"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:ebb86518203e12e96af765ee89034a1dbb0c3c65052d1b0c19bbbd6af8a145e1"},
+ {file = "frozenlist-1.3.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:5cf820485f1b4c91e0417ea0afd41ce5cf5965011b3c22c400f6d144296ccbc0"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5c11e43016b9024240212d2a65043b70ed8dfd3b52678a1271972702d990ac6d"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8fa3c6e3305aa1146b59a09b32b2e04074945ffcfb2f0931836d103a2c38f936"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:352bd4c8c72d508778cf05ab491f6ef36149f4d0cb3c56b1b4302852255d05d5"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65a5e4d3aa679610ac6e3569e865425b23b372277f89b5ef06cf2cdaf1ebf22b"},
+ {file = "frozenlist-1.3.3-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e2c1185858d7e10ff045c496bbf90ae752c28b365fef2c09cf0fa309291669"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f163d2fd041c630fed01bc48d28c3ed4a3b003c00acd396900e11ee5316b56bb"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:05cdb16d09a0832eedf770cb7bd1fe57d8cf4eaf5aced29c4e41e3f20b30a784"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:8bae29d60768bfa8fb92244b74502b18fae55a80eac13c88eb0b496d4268fd2d"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:eedab4c310c0299961ac285591acd53dc6723a1ebd90a57207c71f6e0c2153ab"},
+ {file = "frozenlist-1.3.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3bbdf44855ed8f0fbcd102ef05ec3012d6a4fd7c7562403f76ce6a52aeffb2b1"},
+ {file = "frozenlist-1.3.3-cp39-cp39-win32.whl", hash = "sha256:efa568b885bca461f7c7b9e032655c0c143d305bf01c30caf6db2854a4532b38"},
+ {file = "frozenlist-1.3.3-cp39-cp39-win_amd64.whl", hash = "sha256:cfe33efc9cb900a4c46f91a5ceba26d6df370ffddd9ca386eb1d4f0ad97b9ea9"},
+ {file = "frozenlist-1.3.3.tar.gz", hash = "sha256:58bcc55721e8a90b88332d6cd441261ebb22342e238296bb330968952fbb3a6a"},
+]
+
+[[package]]
+name = "fsspec"
+version = "2023.6.0"
+description = "File-system specification"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "fsspec-2023.6.0-py3-none-any.whl", hash = "sha256:1cbad1faef3e391fba6dc005ae9b5bdcbf43005c9167ce78c915549c352c869a"},
+ {file = "fsspec-2023.6.0.tar.gz", hash = "sha256:d0b2f935446169753e7a5c5c55681c54ea91996cc67be93c39a154fb3a2742af"},
+]
+
+[package.dependencies]
+aiohttp = {version = "<4.0.0a0 || >4.0.0a0,<4.0.0a1 || >4.0.0a1", optional = true, markers = "extra == \"http\""}
+requests = {version = "*", optional = true, markers = "extra == \"http\""}
+
+[package.extras]
+abfs = ["adlfs"]
+adl = ["adlfs"]
+arrow = ["pyarrow (>=1)"]
+dask = ["dask", "distributed"]
+devel = ["pytest", "pytest-cov"]
+dropbox = ["dropbox", "dropboxdrivefs", "requests"]
+full = ["adlfs", "aiohttp (!=4.0.0a0,!=4.0.0a1)", "dask", "distributed", "dropbox", "dropboxdrivefs", "fusepy", "gcsfs", "libarchive-c", "ocifs", "panel", "paramiko", "pyarrow (>=1)", "pygit2", "requests", "s3fs", "smbprotocol", "tqdm"]
+fuse = ["fusepy"]
+gcs = ["gcsfs"]
+git = ["pygit2"]
+github = ["requests"]
+gs = ["gcsfs"]
+gui = ["panel"]
+hdfs = ["pyarrow (>=1)"]
+http = ["aiohttp (!=4.0.0a0,!=4.0.0a1)", "requests"]
+libarchive = ["libarchive-c"]
+oci = ["ocifs"]
+s3 = ["s3fs"]
+sftp = ["paramiko"]
+smb = ["smbprotocol"]
+ssh = ["paramiko"]
+tqdm = ["tqdm"]
+
+[[package]]
+name = "gitdb"
+version = "4.0.10"
+description = "Git Object Database"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "gitdb-4.0.10-py3-none-any.whl", hash = "sha256:c286cf298426064079ed96a9e4a9d39e7f3e9bf15ba60701e95f5492f28415c7"},
+ {file = "gitdb-4.0.10.tar.gz", hash = "sha256:6eb990b69df4e15bad899ea868dc46572c3f75339735663b81de79b06f17eb9a"},
+]
+
+[package.dependencies]
+smmap = ">=3.0.1,<6"
+
+[[package]]
+name = "gitpython"
+version = "3.1.31"
+description = "GitPython is a Python library used to interact with Git repositories"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "GitPython-3.1.31-py3-none-any.whl", hash = "sha256:f04893614f6aa713a60cbbe1e6a97403ef633103cdd0ef5eb6efe0deb98dbe8d"},
+ {file = "GitPython-3.1.31.tar.gz", hash = "sha256:8ce3bcf69adfdf7c7d503e78fd3b1c492af782d58893b650adb2ac8912ddd573"},
+]
+
+[package.dependencies]
+gitdb = ">=4.0.1,<5"
+
+[[package]]
+name = "html5lib"
+version = "1.1"
+description = "HTML parser based on the WHATWG HTML specification"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "html5lib-1.1-py2.py3-none-any.whl", hash = "sha256:0d78f8fde1c230e99fe37986a60526d7049ed4bf8a9fadbad5f00e22e58e041d"},
+ {file = "html5lib-1.1.tar.gz", hash = "sha256:b2e5b40261e20f354d198eae92afc10d750afb487ed5e50f9c4eaf07c184146f"},
+]
+
+[package.dependencies]
+six = ">=1.9"
+webencodings = "*"
+
+[package.extras]
+all = ["chardet (>=2.2)", "genshi", "lxml"]
+chardet = ["chardet (>=2.2)"]
+genshi = ["genshi"]
+lxml = ["lxml"]
+
+[[package]]
+name = "huggingface-hub"
+version = "0.15.1"
+description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub"
+category = "main"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "huggingface_hub-0.15.1-py3-none-any.whl", hash = "sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445"},
+ {file = "huggingface_hub-0.15.1.tar.gz", hash = "sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081"},
+]
+
+[package.dependencies]
+filelock = "*"
+fsspec = "*"
+packaging = ">=20.9"
+pyyaml = ">=5.1"
+requests = "*"
+tqdm = ">=4.42.1"
+typing-extensions = ">=3.7.4.3"
+
+[package.extras]
+all = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"]
+cli = ["InquirerPy (==0.3.4)"]
+dev = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "black (>=23.1,<24.0)", "gradio", "jedi", "mypy (==0.982)", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "ruff (>=0.0.241)", "soundfile", "types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3", "urllib3 (<2.0)"]
+fastai = ["fastai (>=2.4)", "fastcore (>=1.3.27)", "toml"]
+quality = ["black (>=23.1,<24.0)", "mypy (==0.982)", "ruff (>=0.0.241)"]
+tensorflow = ["graphviz", "pydot", "tensorflow"]
+testing = ["InquirerPy (==0.3.4)", "Jinja2", "Pillow", "gradio", "jedi", "numpy", "pytest", "pytest-cov", "pytest-env", "pytest-vcr", "pytest-xdist", "soundfile", "urllib3 (<2.0)"]
+torch = ["torch"]
+typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "types-tqdm", "types-urllib3"]
+
+[[package]]
+name = "idna"
+version = "3.4"
+description = "Internationalized Domain Names in Applications (IDNA)"
+category = "main"
+optional = false
+python-versions = ">=3.5"
+files = [
+ {file = "idna-3.4-py3-none-any.whl", hash = "sha256:90b77e79eaa3eba6de819a0c442c0b4ceefc341a7a2ab77d7562bf49f425c5c2"},
+ {file = "idna-3.4.tar.gz", hash = "sha256:814f528e8dead7d329833b91c5faa87d60bf71824cd12a7530b5526063d02cb4"},
+]
+
+[[package]]
+name = "iniconfig"
+version = "2.0.0"
+description = "brain-dead simple config-ini parsing"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "iniconfig-2.0.0-py3-none-any.whl", hash = "sha256:b6a85871a79d2e3b22d2d1b94ac2824226a63c6b741c88f7ae975f18b6778374"},
+ {file = "iniconfig-2.0.0.tar.gz", hash = "sha256:2d91e135bf72d31a410b17c16da610a82cb55f6b0477d1a902134b24a455b8b3"},
+]
+
+[[package]]
+name = "isort"
+version = "5.12.0"
+description = "A Python utility / library to sort Python imports."
+category = "dev"
+optional = false
+python-versions = ">=3.8.0"
+files = [
+ {file = "isort-5.12.0-py3-none-any.whl", hash = "sha256:f84c2818376e66cf843d497486ea8fed8700b340f308f076c6fb1229dff318b6"},
+ {file = "isort-5.12.0.tar.gz", hash = "sha256:8bef7dde241278824a6d83f44a544709b065191b95b6e50894bdc722fcba0504"},
+]
+
+[package.extras]
+colors = ["colorama (>=0.4.3)"]
+pipfile-deprecated-finder = ["pip-shims (>=0.5.2)", "pipreqs", "requirementslib"]
+plugins = ["setuptools"]
+requirements-deprecated-finder = ["pip-api", "pipreqs"]
+
+[[package]]
+name = "joblib"
+version = "1.3.1"
+description = "Lightweight pipelining with Python functions"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "joblib-1.3.1-py3-none-any.whl", hash = "sha256:89cf0529520e01b3de7ac7b74a8102c90d16d54c64b5dd98cafcd14307fdf915"},
+ {file = "joblib-1.3.1.tar.gz", hash = "sha256:1f937906df65329ba98013dc9692fe22a4c5e4a648112de500508b18a21b41e3"},
+]
+
+[[package]]
+name = "lazy-loader"
+version = "0.3"
+description = "lazy_loader"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "lazy_loader-0.3-py3-none-any.whl", hash = "sha256:1e9e76ee8631e264c62ce10006718e80b2cfc74340d17d1031e0f84af7478554"},
+ {file = "lazy_loader-0.3.tar.gz", hash = "sha256:3b68898e34f5b2a29daaaac172c6555512d0f32074f147e2254e4a6d9d838f37"},
+]
+
+[package.extras]
+lint = ["pre-commit (>=3.3)"]
+test = ["pytest (>=7.4)", "pytest-cov (>=4.1)"]
+
+[[package]]
+name = "libcommon"
+version = "0.6.8"
+description = "Library for utils common to all the services"
+category = "main"
+optional = false
+python-versions = "3.9.15"
+files = []
+develop = true
+
+[package.dependencies]
+appdirs = "^1.4.4"
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
+environs = "^9.5.0"
+huggingface-hub = "^0.15.1"
+mongo-types = "0.15.1"
+mongoengine = "^0.24.2"
+networkx = "^3.0"
+numba = "0.56.4"
+orjson = "^3.8.6"
+pandas = "^2.0.1"
+psutil = "^5.9.4"
+pydub = "^0.25.1"
+pymongo = {version = "^3.13.0", extras = ["srv"]}
+pytz = "^2020.1"
+requests = "^2.31.0"
+soundfile = ">=0.12.1"
+starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
+
+[package.source]
+type = "directory"
+url = "../libcommon"
+
+[[package]]
+name = "librosa"
+version = "0.10.0.post2"
+description = "Python module for audio and music processing"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "librosa-0.10.0.post2-py3-none-any.whl", hash = "sha256:0f3b56118cb01ea89df4b04e924c7f48c5c13d42cc55a12540eb04ae87ab5848"},
+ {file = "librosa-0.10.0.post2.tar.gz", hash = "sha256:6623673da30773beaae962cb4685f188155582f25bc60fc52da968f59eea8567"},
+]
+
+[package.dependencies]
+audioread = ">=2.1.9"
+decorator = ">=4.3.0"
+joblib = ">=0.14"
+lazy-loader = ">=0.1"
+msgpack = ">=1.0"
+numba = ">=0.51.0"
+numpy = ">=1.20.3,<1.22.0 || >1.22.0,<1.22.1 || >1.22.1,<1.22.2 || >1.22.2"
+pooch = ">=1.0,<1.7"
+scikit-learn = ">=0.20.0"
+scipy = ">=1.2.0"
+soundfile = ">=0.12.1"
+soxr = ">=0.3.2"
+typing-extensions = ">=4.1.1"
+
+[package.extras]
+display = ["matplotlib (>=3.3.0)"]
+docs = ["ipython (>=7.0)", "matplotlib (>=3.3.0)", "mir-eval (>=0.5)", "numba (>=0.51)", "numpydoc", "presets", "sphinx (!=1.3.1,<6)", "sphinx-gallery (>=0.7)", "sphinx-multiversion (>=0.2.3)", "sphinx-rtd-theme (>=1.0.0,<2.0.0)", "sphinxcontrib-svg2pdfconverter"]
+tests = ["matplotlib (>=3.3.0)", "packaging (>=20.0)", "pytest", "pytest-cov", "pytest-mpl", "resampy (>=0.2.2)", "samplerate", "types-decorator"]
+
+[[package]]
+name = "llvmlite"
+version = "0.39.1"
+description = "lightweight wrapper around basic LLVM functionality"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "llvmlite-0.39.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:6717c7a6e93c9d2c3d07c07113ec80ae24af45cde536b34363d4bcd9188091d9"},
+ {file = "llvmlite-0.39.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ddab526c5a2c4ccb8c9ec4821fcea7606933dc53f510e2a6eebb45a418d3488a"},
+ {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a3f331a323d0f0ada6b10d60182ef06c20a2f01be21699999d204c5750ffd0b4"},
+ {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c00ff204afa721b0bb9835b5bf1ba7fba210eefcec5552a9e05a63219ba0dc"},
+ {file = "llvmlite-0.39.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:16f56eb1eec3cda3a5c526bc3f63594fc24e0c8d219375afeb336f289764c6c7"},
+ {file = "llvmlite-0.39.1-cp310-cp310-win32.whl", hash = "sha256:d0bfd18c324549c0fec2c5dc610fd024689de6f27c6cc67e4e24a07541d6e49b"},
+ {file = "llvmlite-0.39.1-cp310-cp310-win_amd64.whl", hash = "sha256:7ebf1eb9badc2a397d4f6a6c8717447c81ac011db00064a00408bc83c923c0e4"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:6546bed4e02a1c3d53a22a0bced254b3b6894693318b16c16c8e43e29d6befb6"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1578f5000fdce513712e99543c50e93758a954297575610f48cb1fd71b27c08a"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3803f11ad5f6f6c3d2b545a303d68d9fabb1d50e06a8d6418e6fcd2d0df00959"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50aea09a2b933dab7c9df92361b1844ad3145bfb8dd2deb9cd8b8917d59306fb"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-win32.whl", hash = "sha256:b1a0bbdb274fb683f993198775b957d29a6f07b45d184c571ef2a721ce4388cf"},
+ {file = "llvmlite-0.39.1-cp37-cp37m-win_amd64.whl", hash = "sha256:e172c73fccf7d6db4bd6f7de963dedded900d1a5c6778733241d878ba613980e"},
+ {file = "llvmlite-0.39.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:e31f4b799d530255aaf0566e3da2df5bfc35d3cd9d6d5a3dcc251663656c27b1"},
+ {file = "llvmlite-0.39.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62c0ea22e0b9dffb020601bb65cb11dd967a095a488be73f07d8867f4e327ca5"},
+ {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ffc84ade195abd4abcf0bd3b827b9140ae9ef90999429b9ea84d5df69c9058c"},
+ {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c0f158e4708dda6367d21cf15afc58de4ebce979c7a1aa2f6b977aae737e2a54"},
+ {file = "llvmlite-0.39.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:22d36591cd5d02038912321d9ab8e4668e53ae2211da5523f454e992b5e13c36"},
+ {file = "llvmlite-0.39.1-cp38-cp38-win32.whl", hash = "sha256:4c6ebace910410daf0bebda09c1859504fc2f33d122e9a971c4c349c89cca630"},
+ {file = "llvmlite-0.39.1-cp38-cp38-win_amd64.whl", hash = "sha256:fb62fc7016b592435d3e3a8f680e3ea8897c3c9e62e6e6cc58011e7a4801439e"},
+ {file = "llvmlite-0.39.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fa9b26939ae553bf30a9f5c4c754db0fb2d2677327f2511e674aa2f5df941789"},
+ {file = "llvmlite-0.39.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e4f212c018db951da3e1dc25c2651abc688221934739721f2dad5ff1dd5f90e7"},
+ {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39dc2160aed36e989610fc403487f11b8764b6650017ff367e45384dff88ffbf"},
+ {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1ec3d70b3e507515936e475d9811305f52d049281eaa6c8273448a61c9b5b7e2"},
+ {file = "llvmlite-0.39.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60f8dd1e76f47b3dbdee4b38d9189f3e020d22a173c00f930b52131001d801f9"},
+ {file = "llvmlite-0.39.1-cp39-cp39-win32.whl", hash = "sha256:03aee0ccd81735696474dc4f8b6be60774892a2929d6c05d093d17392c237f32"},
+ {file = "llvmlite-0.39.1-cp39-cp39-win_amd64.whl", hash = "sha256:3fc14e757bc07a919221f0cbaacb512704ce5774d7fcada793f1996d6bc75f2a"},
+ {file = "llvmlite-0.39.1.tar.gz", hash = "sha256:b43abd7c82e805261c425d50335be9a6c4f84264e34d6d6e475207300005d572"},
+]
+
+[[package]]
+name = "markdown-it-py"
+version = "3.0.0"
+description = "Python port of markdown-it. Markdown parsing, done right!"
+category = "dev"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "markdown-it-py-3.0.0.tar.gz", hash = "sha256:e3f60a94fa066dc52ec76661e37c851cb232d92f9886b15cb560aaada2df8feb"},
+ {file = "markdown_it_py-3.0.0-py3-none-any.whl", hash = "sha256:355216845c60bd96232cd8d8c40e8f9765cc86f46880e43a8fd22dc1a1a8cab1"},
+]
+
+[package.dependencies]
+mdurl = ">=0.1,<1.0"
+
+[package.extras]
+benchmarking = ["psutil", "pytest", "pytest-benchmark"]
+code-style = ["pre-commit (>=3.0,<4.0)"]
+compare = ["commonmark (>=0.9,<1.0)", "markdown (>=3.4,<4.0)", "mistletoe (>=1.0,<2.0)", "mistune (>=2.0,<3.0)", "panflute (>=2.3,<3.0)"]
+linkify = ["linkify-it-py (>=1,<3)"]
+plugins = ["mdit-py-plugins"]
+profiling = ["gprof2dot"]
+rtd = ["jupyter_sphinx", "mdit-py-plugins", "myst-parser", "pyyaml", "sphinx", "sphinx-copybutton", "sphinx-design", "sphinx_book_theme"]
+testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"]
+
+[[package]]
+name = "markupsafe"
+version = "2.1.3"
+description = "Safely add untrusted strings to HTML/XML markup."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cd0f502fe016460680cd20aaa5a76d241d6f35a1c3350c474bac1273803893fa"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e09031c87a1e51556fdcb46e5bd4f59dfb743061cf93c4d6831bf894f125eb57"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:68e78619a61ecf91e76aa3e6e8e33fc4894a2bebe93410754bd28fce0a8a4f9f"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65c1a9bcdadc6c28eecee2c119465aebff8f7a584dd719facdd9e825ec61ab52"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:525808b8019e36eb524b8c68acdd63a37e75714eac50e988180b169d64480a00"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:962f82a3086483f5e5f64dbad880d31038b698494799b097bc59c2edf392fce6"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:aa7bd130efab1c280bed0f45501b7c8795f9fdbeb02e965371bbef3523627779"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c9c804664ebe8f83a211cace637506669e7890fec1b4195b505c214e50dd4eb7"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-win32.whl", hash = "sha256:10bbfe99883db80bdbaff2dcf681dfc6533a614f700da1287707e8a5d78a8431"},
+ {file = "MarkupSafe-2.1.3-cp310-cp310-win_amd64.whl", hash = "sha256:1577735524cdad32f9f694208aa75e422adba74f1baee7551620e43a3141f559"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:ad9e82fb8f09ade1c3e1b996a6337afac2b8b9e365f926f5a61aacc71adc5b3c"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3c0fae6c3be832a0a0473ac912810b2877c8cb9d76ca48de1ed31e1c68386575"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b076b6226fb84157e3f7c971a47ff3a679d837cf338547532ab866c57930dbee"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bfce63a9e7834b12b87c64d6b155fdd9b3b96191b6bd334bf37db7ff1fe457f2"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:338ae27d6b8745585f87218a3f23f1512dbf52c26c28e322dbe54bcede54ccb9"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:e4dd52d80b8c83fdce44e12478ad2e85c64ea965e75d66dbeafb0a3e77308fcc"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:df0be2b576a7abbf737b1575f048c23fb1d769f267ec4358296f31c2479db8f9"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:5bbe06f8eeafd38e5d0a4894ffec89378b6c6a625ff57e3028921f8ff59318ac"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-win32.whl", hash = "sha256:dd15ff04ffd7e05ffcb7fe79f1b98041b8ea30ae9234aed2a9168b5797c3effb"},
+ {file = "MarkupSafe-2.1.3-cp311-cp311-win_amd64.whl", hash = "sha256:134da1eca9ec0ae528110ccc9e48041e0828d79f24121a1a146161103c76e686"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8e254ae696c88d98da6555f5ace2279cf7cd5b3f52be2b5cf97feafe883b58d2"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cb0932dc158471523c9637e807d9bfb93e06a95cbf010f1a38b98623b929ef2b"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9402b03f1a1b4dc4c19845e5c749e3ab82d5078d16a2a4c2cd2df62d57bb0707"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca379055a47383d02a5400cb0d110cef0a776fc644cda797db0c5696cfd7e18e"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7ff0f54cb4ff66dd38bebd335a38e2c22c41a8ee45aa608efc890ac3e3931bc"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c011a4149cfbcf9f03994ec2edffcb8b1dc2d2aede7ca243746df97a5d41ce48"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:56d9f2ecac662ca1611d183feb03a3fa4406469dafe241673d521dd5ae92a155"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-win32.whl", hash = "sha256:8758846a7e80910096950b67071243da3e5a20ed2546e6392603c096778d48e0"},
+ {file = "MarkupSafe-2.1.3-cp37-cp37m-win_amd64.whl", hash = "sha256:787003c0ddb00500e49a10f2844fac87aa6ce977b90b0feaaf9de23c22508b24"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:2ef12179d3a291be237280175b542c07a36e7f60718296278d8593d21ca937d4"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:2c1b19b3aaacc6e57b7e25710ff571c24d6c3613a45e905b1fde04d691b98ee0"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8afafd99945ead6e075b973fefa56379c5b5c53fd8937dad92c662da5d8fd5ee"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8c41976a29d078bb235fea9b2ecd3da465df42a562910f9022f1a03107bd02be"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d080e0a5eb2529460b30190fcfcc4199bd7f827663f858a226a81bc27beaa97e"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:69c0f17e9f5a7afdf2cc9fb2d1ce6aabdb3bafb7f38017c0b77862bcec2bbad8"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:504b320cd4b7eff6f968eddf81127112db685e81f7e36e75f9f84f0df46041c3"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:42de32b22b6b804f42c5d98be4f7e5e977ecdd9ee9b660fda1a3edf03b11792d"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-win32.whl", hash = "sha256:ceb01949af7121f9fc39f7d27f91be8546f3fb112c608bc4029aef0bab86a2a5"},
+ {file = "MarkupSafe-2.1.3-cp38-cp38-win_amd64.whl", hash = "sha256:1b40069d487e7edb2676d3fbdb2b0829ffa2cd63a2ec26c4938b2d34391b4ecc"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:8023faf4e01efadfa183e863fefde0046de576c6f14659e8782065bcece22198"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6b2b56950d93e41f33b4223ead100ea0fe11f8e6ee5f641eb753ce4b77a7042b"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9dcdfd0eaf283af041973bff14a2e143b8bd64e069f4c383416ecd79a81aab58"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:05fb21170423db021895e1ea1e1f3ab3adb85d1c2333cbc2310f2a26bc77272e"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:282c2cb35b5b673bbcadb33a585408104df04f14b2d9b01d4c345a3b92861c2c"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ab4a0df41e7c16a1392727727e7998a467472d0ad65f3ad5e6e765015df08636"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:7ef3cb2ebbf91e330e3bb937efada0edd9003683db6b57bb108c4001f37a02ea"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:0a4e4a1aff6c7ac4cd55792abf96c915634c2b97e3cc1c7129578aa68ebd754e"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-win32.whl", hash = "sha256:fec21693218efe39aa7f8599346e90c705afa52c5b31ae019b2e57e8f6542bb2"},
+ {file = "MarkupSafe-2.1.3-cp39-cp39-win_amd64.whl", hash = "sha256:3fd4abcb888d15a94f32b75d8fd18ee162ca0c064f35b11134be77050296d6ba"},
+ {file = "MarkupSafe-2.1.3.tar.gz", hash = "sha256:af598ed32d6ae86f1b747b82783958b1a4ab8f617b06fe68795c7f026abbdcad"},
+]
+
+[[package]]
+name = "marshmallow"
+version = "3.19.0"
+description = "A lightweight library for converting complex datatypes to and from native Python datatypes."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "marshmallow-3.19.0-py3-none-any.whl", hash = "sha256:93f0958568da045b0021ec6aeb7ac37c81bfcccbb9a0e7ed8559885070b3a19b"},
+ {file = "marshmallow-3.19.0.tar.gz", hash = "sha256:90032c0fd650ce94b6ec6dc8dfeb0e3ff50c144586462c389b81a07205bedb78"},
+]
+
+[package.dependencies]
+packaging = ">=17.0"
+
+[package.extras]
+dev = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)", "pytest", "pytz", "simplejson", "tox"]
+docs = ["alabaster (==0.7.12)", "autodocsumm (==0.2.9)", "sphinx (==5.3.0)", "sphinx-issues (==3.0.1)", "sphinx-version-warning (==1.1.2)"]
+lint = ["flake8 (==5.0.4)", "flake8-bugbear (==22.10.25)", "mypy (==0.990)", "pre-commit (>=2.4,<3.0)"]
+tests = ["pytest", "pytz", "simplejson"]
+
+[[package]]
+name = "mccabe"
+version = "0.6.1"
+description = "McCabe checker, plugin for flake8"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "mccabe-0.6.1-py2.py3-none-any.whl", hash = "sha256:ab8a6258860da4b6677da4bd2fe5dc2c659cff31b3ee4f7f5d64e79735b80d42"},
+ {file = "mccabe-0.6.1.tar.gz", hash = "sha256:dd8d182285a0fe56bace7f45b5e7d1a6ebcbf524e8f3bd87eb0f125271b8831f"},
+]
+
+[[package]]
+name = "mdurl"
+version = "0.1.2"
+description = "Markdown URL utilities"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8"},
+ {file = "mdurl-0.1.2.tar.gz", hash = "sha256:bb413d29f5eea38f31dd4754dd7377d4465116fb207585f97bf925588687c1ba"},
+]
+
+[[package]]
+name = "mongo-types"
+version = "0.15.1"
+description = "Type stubs for mongoengine w/ basic support for bson and pymongo"
+category = "main"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "mongo-types-0.15.1.tar.gz", hash = "sha256:0a9deeb7733ea7da5db3711d92e22d93556b522f860bbff82e5df44c53bd06a9"},
+ {file = "mongo_types-0.15.1-py3-none-any.whl", hash = "sha256:9417ae5b9a759c09630b5ec7d66904cc333c2d2fcfe75e2760a332ed5e267309"},
+]
+
+[[package]]
+name = "mongoengine"
+version = "0.24.2"
+description = "MongoEngine is a Python Object-Document Mapper for working with MongoDB."
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "mongoengine-0.24.2-py3-none-any.whl", hash = "sha256:f5c4e1b206b2ccffe4adc7a6283ed26dd799bd115a5fb1d2e885a075132cdb88"},
+ {file = "mongoengine-0.24.2.tar.gz", hash = "sha256:c76d49658575bb995682e2e77c8ef7cda63faf939415b32ee923745d120f8b02"},
+]
+
+[package.dependencies]
+pymongo = ">=3.4,<5.0"
+
+[[package]]
+name = "msgpack"
+version = "1.0.5"
+description = "MessagePack serializer"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:525228efd79bb831cf6830a732e2e80bc1b05436b086d4264814b4b2955b2fa9"},
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:4f8d8b3bf1ff2672567d6b5c725a1b347fe838b912772aa8ae2bf70338d5a198"},
+ {file = "msgpack-1.0.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:cdc793c50be3f01106245a61b739328f7dccc2c648b501e237f0699fe1395b81"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5cb47c21a8a65b165ce29f2bec852790cbc04936f502966768e4aae9fa763cb7"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e42b9594cc3bf4d838d67d6ed62b9e59e201862a25e9a157019e171fbe672dd3"},
+ {file = "msgpack-1.0.5-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:55b56a24893105dc52c1253649b60f475f36b3aa0fc66115bffafb624d7cb30b"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:1967f6129fc50a43bfe0951c35acbb729be89a55d849fab7686004da85103f1c"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:20a97bf595a232c3ee6d57ddaadd5453d174a52594bf9c21d10407e2a2d9b3bd"},
+ {file = "msgpack-1.0.5-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d25dd59bbbbb996eacf7be6b4ad082ed7eacc4e8f3d2df1ba43822da9bfa122a"},
+ {file = "msgpack-1.0.5-cp310-cp310-win32.whl", hash = "sha256:382b2c77589331f2cb80b67cc058c00f225e19827dbc818d700f61513ab47bea"},
+ {file = "msgpack-1.0.5-cp310-cp310-win_amd64.whl", hash = "sha256:4867aa2df9e2a5fa5f76d7d5565d25ec76e84c106b55509e78c1ede0f152659a"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:9f5ae84c5c8a857ec44dc180a8b0cc08238e021f57abdf51a8182e915e6299f0"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:9e6ca5d5699bcd89ae605c150aee83b5321f2115695e741b99618f4856c50898"},
+ {file = "msgpack-1.0.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5494ea30d517a3576749cad32fa27f7585c65f5f38309c88c6d137877fa28a5a"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1ab2f3331cb1b54165976a9d976cb251a83183631c88076613c6c780f0d6e45a"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:28592e20bbb1620848256ebc105fc420436af59515793ed27d5c77a217477705"},
+ {file = "msgpack-1.0.5-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe5c63197c55bce6385d9aee16c4d0641684628f63ace85f73571e65ad1c1e8d"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:ed40e926fa2f297e8a653c954b732f125ef97bdd4c889f243182299de27e2aa9"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b2de4c1c0538dcb7010902a2b97f4e00fc4ddf2c8cda9749af0e594d3b7fa3d7"},
+ {file = "msgpack-1.0.5-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:bf22a83f973b50f9d38e55c6aade04c41ddda19b00c4ebc558930d78eecc64ed"},
+ {file = "msgpack-1.0.5-cp311-cp311-win32.whl", hash = "sha256:c396e2cc213d12ce017b686e0f53497f94f8ba2b24799c25d913d46c08ec422c"},
+ {file = "msgpack-1.0.5-cp311-cp311-win_amd64.whl", hash = "sha256:6c4c68d87497f66f96d50142a2b73b97972130d93677ce930718f68828b382e2"},
+ {file = "msgpack-1.0.5-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:a2b031c2e9b9af485d5e3c4520f4220d74f4d222a5b8dc8c1a3ab9448ca79c57"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f837b93669ce4336e24d08286c38761132bc7ab29782727f8557e1eb21b2080"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1d46dfe3832660f53b13b925d4e0fa1432b00f5f7210eb3ad3bb9a13c6204a6"},
+ {file = "msgpack-1.0.5-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:366c9a7b9057e1547f4ad51d8facad8b406bab69c7d72c0eb6f529cf76d4b85f"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:4c075728a1095efd0634a7dccb06204919a2f67d1893b6aa8e00497258bf926c"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:f933bbda5a3ee63b8834179096923b094b76f0c7a73c1cfe8f07ad608c58844b"},
+ {file = "msgpack-1.0.5-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:36961b0568c36027c76e2ae3ca1132e35123dcec0706c4b7992683cc26c1320c"},
+ {file = "msgpack-1.0.5-cp36-cp36m-win32.whl", hash = "sha256:b5ef2f015b95f912c2fcab19c36814963b5463f1fb9049846994b007962743e9"},
+ {file = "msgpack-1.0.5-cp36-cp36m-win_amd64.whl", hash = "sha256:288e32b47e67f7b171f86b030e527e302c91bd3f40fd9033483f2cacc37f327a"},
+ {file = "msgpack-1.0.5-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:137850656634abddfb88236008339fdaba3178f4751b28f270d2ebe77a563b6c"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0c05a4a96585525916b109bb85f8cb6511db1c6f5b9d9cbcbc940dc6b4be944b"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:56a62ec00b636583e5cb6ad313bbed36bb7ead5fa3a3e38938503142c72cba4f"},
+ {file = "msgpack-1.0.5-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef8108f8dedf204bb7b42994abf93882da1159728a2d4c5e82012edd92c9da9f"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:1835c84d65f46900920b3708f5ba829fb19b1096c1800ad60bae8418652a951d"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:e57916ef1bd0fee4f21c4600e9d1da352d8816b52a599c46460e93a6e9f17086"},
+ {file = "msgpack-1.0.5-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:17358523b85973e5f242ad74aa4712b7ee560715562554aa2134d96e7aa4cbbf"},
+ {file = "msgpack-1.0.5-cp37-cp37m-win32.whl", hash = "sha256:cb5aaa8c17760909ec6cb15e744c3ebc2ca8918e727216e79607b7bbce9c8f77"},
+ {file = "msgpack-1.0.5-cp37-cp37m-win_amd64.whl", hash = "sha256:ab31e908d8424d55601ad7075e471b7d0140d4d3dd3272daf39c5c19d936bd82"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:b72d0698f86e8d9ddf9442bdedec15b71df3598199ba33322d9711a19f08145c"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:379026812e49258016dd84ad79ac8446922234d498058ae1d415f04b522d5b2d"},
+ {file = "msgpack-1.0.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:332360ff25469c346a1c5e47cbe2a725517919892eda5cfaffe6046656f0b7bb"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:476a8fe8fae289fdf273d6d2a6cb6e35b5a58541693e8f9f019bfe990a51e4ba"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a9985b214f33311df47e274eb788a5893a761d025e2b92c723ba4c63936b69b1"},
+ {file = "msgpack-1.0.5-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:48296af57cdb1d885843afd73c4656be5c76c0c6328db3440c9601a98f303d87"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:addab7e2e1fcc04bd08e4eb631c2a90960c340e40dfc4a5e24d2ff0d5a3b3edb"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:916723458c25dfb77ff07f4c66aed34e47503b2eb3188b3adbec8d8aa6e00f48"},
+ {file = "msgpack-1.0.5-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:821c7e677cc6acf0fd3f7ac664c98803827ae6de594a9f99563e48c5a2f27eb0"},
+ {file = "msgpack-1.0.5-cp38-cp38-win32.whl", hash = "sha256:1c0f7c47f0087ffda62961d425e4407961a7ffd2aa004c81b9c07d9269512f6e"},
+ {file = "msgpack-1.0.5-cp38-cp38-win_amd64.whl", hash = "sha256:bae7de2026cbfe3782c8b78b0db9cbfc5455e079f1937cb0ab8d133496ac55e1"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:20c784e66b613c7f16f632e7b5e8a1651aa5702463d61394671ba07b2fc9e025"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:266fa4202c0eb94d26822d9bfd7af25d1e2c088927fe8de9033d929dd5ba24c5"},
+ {file = "msgpack-1.0.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:18334484eafc2b1aa47a6d42427da7fa8f2ab3d60b674120bce7a895a0a85bdd"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:57e1f3528bd95cc44684beda696f74d3aaa8a5e58c816214b9046512240ef437"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:586d0d636f9a628ddc6a17bfd45aa5b5efaf1606d2b60fa5d87b8986326e933f"},
+ {file = "msgpack-1.0.5-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a740fa0e4087a734455f0fc3abf5e746004c9da72fbd541e9b113013c8dc3282"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3055b0455e45810820db1f29d900bf39466df96ddca11dfa6d074fa47054376d"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:a61215eac016f391129a013c9e46f3ab308db5f5ec9f25811e811f96962599a8"},
+ {file = "msgpack-1.0.5-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:362d9655cd369b08fda06b6657a303eb7172d5279997abe094512e919cf74b11"},
+ {file = "msgpack-1.0.5-cp39-cp39-win32.whl", hash = "sha256:ac9dd47af78cae935901a9a500104e2dea2e253207c924cc95de149606dc43cc"},
+ {file = "msgpack-1.0.5-cp39-cp39-win_amd64.whl", hash = "sha256:06f5174b5f8ed0ed919da0e62cbd4ffde676a374aba4020034da05fab67b9164"},
+ {file = "msgpack-1.0.5.tar.gz", hash = "sha256:c075544284eadc5cddc70f4757331d99dcbc16b2bbd4849d15f8aae4cf36d31c"},
+]
+
+[[package]]
+name = "multidict"
+version = "6.0.4"
+description = "multidict implementation"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0b1a97283e0c85772d613878028fec909f003993e1007eafa715b24b377cb9b8"},
+ {file = "multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:eeb6dcc05e911516ae3d1f207d4b0520d07f54484c49dfc294d6e7d63b734171"},
+ {file = "multidict-6.0.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d6d635d5209b82a3492508cf5b365f3446afb65ae7ebd755e70e18f287b0adf7"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c048099e4c9e9d615545e2001d3d8a4380bd403e1a0578734e0d31703d1b0c0b"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ea20853c6dbbb53ed34cb4d080382169b6f4554d394015f1bef35e881bf83547"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:16d232d4e5396c2efbbf4f6d4df89bfa905eb0d4dc5b3549d872ab898451f569"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:36c63aaa167f6c6b04ef2c85704e93af16c11d20de1d133e39de6a0e84582a93"},
+ {file = "multidict-6.0.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:64bdf1086b6043bf519869678f5f2757f473dee970d7abf6da91ec00acb9cb98"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:43644e38f42e3af682690876cff722d301ac585c5b9e1eacc013b7a3f7b696a0"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7582a1d1030e15422262de9f58711774e02fa80df0d1578995c76214f6954988"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:ddff9c4e225a63a5afab9dd15590432c22e8057e1a9a13d28ed128ecf047bbdc"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:ee2a1ece51b9b9e7752e742cfb661d2a29e7bcdba2d27e66e28a99f1890e4fa0"},
+ {file = "multidict-6.0.4-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:a2e4369eb3d47d2034032a26c7a80fcb21a2cb22e1173d761a162f11e562caa5"},
+ {file = "multidict-6.0.4-cp310-cp310-win32.whl", hash = "sha256:574b7eae1ab267e5f8285f0fe881f17efe4b98c39a40858247720935b893bba8"},
+ {file = "multidict-6.0.4-cp310-cp310-win_amd64.whl", hash = "sha256:4dcbb0906e38440fa3e325df2359ac6cb043df8e58c965bb45f4e406ecb162cc"},
+ {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:0dfad7a5a1e39c53ed00d2dd0c2e36aed4650936dc18fd9a1826a5ae1cad6f03"},
+ {file = "multidict-6.0.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:64da238a09d6039e3bd39bb3aee9c21a5e34f28bfa5aa22518581f910ff94af3"},
+ {file = "multidict-6.0.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:ff959bee35038c4624250473988b24f846cbeb2c6639de3602c073f10410ceba"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01a3a55bd90018c9c080fbb0b9f4891db37d148a0a18722b42f94694f8b6d4c9"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c5cb09abb18c1ea940fb99360ea0396f34d46566f157122c92dfa069d3e0e982"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:666daae833559deb2d609afa4490b85830ab0dfca811a98b70a205621a6109fe"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:11bdf3f5e1518b24530b8241529d2050014c884cf18b6fc69c0c2b30ca248710"},
+ {file = "multidict-6.0.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7d18748f2d30f94f498e852c67d61261c643b349b9d2a581131725595c45ec6c"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:458f37be2d9e4c95e2d8866a851663cbc76e865b78395090786f6cd9b3bbf4f4"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:b1a2eeedcead3a41694130495593a559a668f382eee0727352b9a41e1c45759a"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:7d6ae9d593ef8641544d6263c7fa6408cc90370c8cb2bbb65f8d43e5b0351d9c"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:5979b5632c3e3534e42ca6ff856bb24b2e3071b37861c2c727ce220d80eee9ed"},
+ {file = "multidict-6.0.4-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:dcfe792765fab89c365123c81046ad4103fcabbc4f56d1c1997e6715e8015461"},
+ {file = "multidict-6.0.4-cp311-cp311-win32.whl", hash = "sha256:3601a3cece3819534b11d4efc1eb76047488fddd0c85a3948099d5da4d504636"},
+ {file = "multidict-6.0.4-cp311-cp311-win_amd64.whl", hash = "sha256:81a4f0b34bd92df3da93315c6a59034df95866014ac08535fc819f043bfd51f0"},
+ {file = "multidict-6.0.4-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:67040058f37a2a51ed8ea8f6b0e6ee5bd78ca67f169ce6122f3e2ec80dfe9b78"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:853888594621e6604c978ce2a0444a1e6e70c8d253ab65ba11657659dcc9100f"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:39ff62e7d0f26c248b15e364517a72932a611a9b75f35b45be078d81bdb86603"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:af048912e045a2dc732847d33821a9d84ba553f5c5f028adbd364dd4765092ac"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b1e8b901e607795ec06c9e42530788c45ac21ef3aaa11dbd0c69de543bfb79a9"},
+ {file = "multidict-6.0.4-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:62501642008a8b9871ddfccbf83e4222cf8ac0d5aeedf73da36153ef2ec222d2"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:99b76c052e9f1bc0721f7541e5e8c05db3941eb9ebe7b8553c625ef88d6eefde"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:509eac6cf09c794aa27bcacfd4d62c885cce62bef7b2c3e8b2e49d365b5003fe"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:21a12c4eb6ddc9952c415f24eef97e3e55ba3af61f67c7bc388dcdec1404a067"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:5cad9430ab3e2e4fa4a2ef4450f548768400a2ac635841bc2a56a2052cdbeb87"},
+ {file = "multidict-6.0.4-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:ab55edc2e84460694295f401215f4a58597f8f7c9466faec545093045476327d"},
+ {file = "multidict-6.0.4-cp37-cp37m-win32.whl", hash = "sha256:5a4dcf02b908c3b8b17a45fb0f15b695bf117a67b76b7ad18b73cf8e92608775"},
+ {file = "multidict-6.0.4-cp37-cp37m-win_amd64.whl", hash = "sha256:6ed5f161328b7df384d71b07317f4d8656434e34591f20552c7bcef27b0ab88e"},
+ {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5fc1b16f586f049820c5c5b17bb4ee7583092fa0d1c4e28b5239181ff9532e0c"},
+ {file = "multidict-6.0.4-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:1502e24330eb681bdaa3eb70d6358e818e8e8f908a22a1851dfd4e15bc2f8161"},
+ {file = "multidict-6.0.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:b692f419760c0e65d060959df05f2a531945af31fda0c8a3b3195d4efd06de11"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45e1ecb0379bfaab5eef059f50115b54571acfbe422a14f668fc8c27ba410e7e"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ddd3915998d93fbcd2566ddf9cf62cdb35c9e093075f862935573d265cf8f65d"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:59d43b61c59d82f2effb39a93c48b845efe23a3852d201ed2d24ba830d0b4cf2"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cc8e1d0c705233c5dd0c5e6460fbad7827d5d36f310a0fadfd45cc3029762258"},
+ {file = "multidict-6.0.4-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6aa0418fcc838522256761b3415822626f866758ee0bc6632c9486b179d0b52"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:6748717bb10339c4760c1e63da040f5f29f5ed6e59d76daee30305894069a660"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:4d1a3d7ef5e96b1c9e92f973e43aa5e5b96c659c9bc3124acbbd81b0b9c8a951"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:4372381634485bec7e46718edc71528024fcdc6f835baefe517b34a33c731d60"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:fc35cb4676846ef752816d5be2193a1e8367b4c1397b74a565a9d0389c433a1d"},
+ {file = "multidict-6.0.4-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:4b9d9e4e2b37daddb5c23ea33a3417901fa7c7b3dee2d855f63ee67a0b21e5b1"},
+ {file = "multidict-6.0.4-cp38-cp38-win32.whl", hash = "sha256:e41b7e2b59679edfa309e8db64fdf22399eec4b0b24694e1b2104fb789207779"},
+ {file = "multidict-6.0.4-cp38-cp38-win_amd64.whl", hash = "sha256:d6c254ba6e45d8e72739281ebc46ea5eb5f101234f3ce171f0e9f5cc86991480"},
+ {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:16ab77bbeb596e14212e7bab8429f24c1579234a3a462105cda4a66904998664"},
+ {file = "multidict-6.0.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:bc779e9e6f7fda81b3f9aa58e3a6091d49ad528b11ed19f6621408806204ad35"},
+ {file = "multidict-6.0.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:4ceef517eca3e03c1cceb22030a3e39cb399ac86bff4e426d4fc6ae49052cc60"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:281af09f488903fde97923c7744bb001a9b23b039a909460d0f14edc7bf59706"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:52f2dffc8acaba9a2f27174c41c9e57f60b907bb9f096b36b1a1f3be71c6284d"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b41156839806aecb3641f3208c0dafd3ac7775b9c4c422d82ee2a45c34ba81ca"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d5e3fc56f88cc98ef8139255cf8cd63eb2c586531e43310ff859d6bb3a6b51f1"},
+ {file = "multidict-6.0.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8316a77808c501004802f9beebde51c9f857054a0c871bd6da8280e718444449"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:f70b98cd94886b49d91170ef23ec5c0e8ebb6f242d734ed7ed677b24d50c82cf"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bf6774e60d67a9efe02b3616fee22441d86fab4c6d335f9d2051d19d90a40063"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:e69924bfcdda39b722ef4d9aa762b2dd38e4632b3641b1d9a57ca9cd18f2f83a"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:6b181d8c23da913d4ff585afd1155a0e1194c0b50c54fcfe286f70cdaf2b7176"},
+ {file = "multidict-6.0.4-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:52509b5be062d9eafc8170e53026fbc54cf3b32759a23d07fd935fb04fc22d95"},
+ {file = "multidict-6.0.4-cp39-cp39-win32.whl", hash = "sha256:27c523fbfbdfd19c6867af7346332b62b586eed663887392cff78d614f9ec313"},
+ {file = "multidict-6.0.4-cp39-cp39-win_amd64.whl", hash = "sha256:33029f5734336aa0d4c0384525da0387ef89148dc7191aae00ca5fb23d7aafc2"},
+ {file = "multidict-6.0.4.tar.gz", hash = "sha256:3666906492efb76453c0e7b97f2cf459b0682e7402c0489a95484965dbc1da49"},
+]
+
+[[package]]
+name = "multiprocess"
+version = "0.70.14"
+description = "better multiprocessing and multithreading in python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "multiprocess-0.70.14-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:560a27540daef4ce8b24ed3cc2496a3c670df66c96d02461a4da67473685adf3"},
+ {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_i686.whl", hash = "sha256:bfbbfa36f400b81d1978c940616bc77776424e5e34cb0c94974b178d727cfcd5"},
+ {file = "multiprocess-0.70.14-pp37-pypy37_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:89fed99553a04ec4f9067031f83a886d7fdec5952005551a896a4b6a59575bb9"},
+ {file = "multiprocess-0.70.14-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:40a5e3685462079e5fdee7c6789e3ef270595e1755199f0d50685e72523e1d2a"},
+ {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_i686.whl", hash = "sha256:44936b2978d3f2648727b3eaeab6d7fa0bedf072dc5207bf35a96d5ee7c004cf"},
+ {file = "multiprocess-0.70.14-pp38-pypy38_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:e628503187b5d494bf29ffc52d3e1e57bb770ce7ce05d67c4bbdb3a0c7d3b05f"},
+ {file = "multiprocess-0.70.14-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:0d5da0fc84aacb0e4bd69c41b31edbf71b39fe2fb32a54eaedcaea241050855c"},
+ {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:6a7b03a5b98e911a7785b9116805bd782815c5e2bd6c91c6a320f26fd3e7b7ad"},
+ {file = "multiprocess-0.70.14-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:cea5bdedd10aace3c660fedeac8b087136b4366d4ee49a30f1ebf7409bce00ae"},
+ {file = "multiprocess-0.70.14-py310-none-any.whl", hash = "sha256:7dc1f2f6a1d34894c8a9a013fbc807971e336e7cc3f3ff233e61b9dc679b3b5c"},
+ {file = "multiprocess-0.70.14-py37-none-any.whl", hash = "sha256:93a8208ca0926d05cdbb5b9250a604c401bed677579e96c14da3090beb798193"},
+ {file = "multiprocess-0.70.14-py38-none-any.whl", hash = "sha256:6725bc79666bbd29a73ca148a0fb5f4ea22eed4a8f22fce58296492a02d18a7b"},
+ {file = "multiprocess-0.70.14-py39-none-any.whl", hash = "sha256:63cee628b74a2c0631ef15da5534c8aedbc10c38910b9c8b18dcd327528d1ec7"},
+ {file = "multiprocess-0.70.14.tar.gz", hash = "sha256:3eddafc12f2260d27ae03fe6069b12570ab4764ab59a75e81624fac453fbf46a"},
+]
+
+[package.dependencies]
+dill = ">=0.3.6"
+
+[[package]]
+name = "mypy"
+version = "1.4.1"
+description = "Optional static typing for Python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "mypy-1.4.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:566e72b0cd6598503e48ea610e0052d1b8168e60a46e0bfd34b3acf2d57f96a8"},
+ {file = "mypy-1.4.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ca637024ca67ab24a7fd6f65d280572c3794665eaf5edcc7e90a866544076878"},
+ {file = "mypy-1.4.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0dde1d180cd84f0624c5dcaaa89c89775550a675aff96b5848de78fb11adabcd"},
+ {file = "mypy-1.4.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:8c4d8e89aa7de683e2056a581ce63c46a0c41e31bd2b6d34144e2c80f5ea53dc"},
+ {file = "mypy-1.4.1-cp310-cp310-win_amd64.whl", hash = "sha256:bfdca17c36ae01a21274a3c387a63aa1aafe72bff976522886869ef131b937f1"},
+ {file = "mypy-1.4.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:7549fbf655e5825d787bbc9ecf6028731973f78088fbca3a1f4145c39ef09462"},
+ {file = "mypy-1.4.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:98324ec3ecf12296e6422939e54763faedbfcc502ea4a4c38502082711867258"},
+ {file = "mypy-1.4.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:141dedfdbfe8a04142881ff30ce6e6653c9685b354876b12e4fe6c78598b45e2"},
+ {file = "mypy-1.4.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:8207b7105829eca6f3d774f64a904190bb2231de91b8b186d21ffd98005f14a7"},
+ {file = "mypy-1.4.1-cp311-cp311-win_amd64.whl", hash = "sha256:16f0db5b641ba159eff72cff08edc3875f2b62b2fa2bc24f68c1e7a4e8232d01"},
+ {file = "mypy-1.4.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:470c969bb3f9a9efcedbadcd19a74ffb34a25f8e6b0e02dae7c0e71f8372f97b"},
+ {file = "mypy-1.4.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5952d2d18b79f7dc25e62e014fe5a23eb1a3d2bc66318df8988a01b1a037c5b"},
+ {file = "mypy-1.4.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:190b6bab0302cec4e9e6767d3eb66085aef2a1cc98fe04936d8a42ed2ba77bb7"},
+ {file = "mypy-1.4.1-cp37-cp37m-win_amd64.whl", hash = "sha256:9d40652cc4fe33871ad3338581dca3297ff5f2213d0df345bcfbde5162abf0c9"},
+ {file = "mypy-1.4.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:01fd2e9f85622d981fd9063bfaef1aed6e336eaacca00892cd2d82801ab7c042"},
+ {file = "mypy-1.4.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:2460a58faeea905aeb1b9b36f5065f2dc9a9c6e4c992a6499a2360c6c74ceca3"},
+ {file = "mypy-1.4.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a2746d69a8196698146a3dbe29104f9eb6a2a4d8a27878d92169a6c0b74435b6"},
+ {file = "mypy-1.4.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:ae704dcfaa180ff7c4cfbad23e74321a2b774f92ca77fd94ce1049175a21c97f"},
+ {file = "mypy-1.4.1-cp38-cp38-win_amd64.whl", hash = "sha256:43d24f6437925ce50139a310a64b2ab048cb2d3694c84c71c3f2a1626d8101dc"},
+ {file = "mypy-1.4.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c482e1246726616088532b5e964e39765b6d1520791348e6c9dc3af25b233828"},
+ {file = "mypy-1.4.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:43b592511672017f5b1a483527fd2684347fdffc041c9ef53428c8dc530f79a3"},
+ {file = "mypy-1.4.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:34a9239d5b3502c17f07fd7c0b2ae6b7dd7d7f6af35fbb5072c6208e76295816"},
+ {file = "mypy-1.4.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5703097c4936bbb9e9bce41478c8d08edd2865e177dc4c52be759f81ee4dd26c"},
+ {file = "mypy-1.4.1-cp39-cp39-win_amd64.whl", hash = "sha256:e02d700ec8d9b1859790c0475df4e4092c7bf3272a4fd2c9f33d87fac4427b8f"},
+ {file = "mypy-1.4.1-py3-none-any.whl", hash = "sha256:45d32cec14e7b97af848bddd97d85ea4f0db4d5a149ed9676caa4eb2f7402bb4"},
+ {file = "mypy-1.4.1.tar.gz", hash = "sha256:9bbcd9ab8ea1f2e1c8031c21445b511442cc45c89951e49bbf852cbb70755b1b"},
+]
+
+[package.dependencies]
+mypy-extensions = ">=1.0.0"
+tomli = {version = ">=1.1.0", markers = "python_version < \"3.11\""}
+typing-extensions = ">=4.1.0"
+
+[package.extras]
+dmypy = ["psutil (>=4.0)"]
+install-types = ["pip"]
+python2 = ["typed-ast (>=1.4.0,<2)"]
+reports = ["lxml"]
+
+[[package]]
+name = "mypy-extensions"
+version = "1.0.0"
+description = "Type system extensions for programs checked with the mypy type checker."
+category = "dev"
+optional = false
+python-versions = ">=3.5"
+files = [
+ {file = "mypy_extensions-1.0.0-py3-none-any.whl", hash = "sha256:4392f6c0eb8a5668a69e23d168ffa70f0be9ccfd32b5cc2d26a34ae5b844552d"},
+ {file = "mypy_extensions-1.0.0.tar.gz", hash = "sha256:75dbf8955dc00442a438fc4d0666508a9a97b6bd41aa2f0ffe9d2f2725af0782"},
+]
+
+[[package]]
+name = "networkx"
+version = "3.1"
+description = "Python package for creating and manipulating graphs and networks"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "networkx-3.1-py3-none-any.whl", hash = "sha256:4f33f68cb2afcf86f28a45f43efc27a9386b535d567d2127f8f61d51dec58d36"},
+ {file = "networkx-3.1.tar.gz", hash = "sha256:de346335408f84de0eada6ff9fafafff9bcda11f0a0dfaa931133debb146ab61"},
+]
+
+[package.extras]
+default = ["matplotlib (>=3.4)", "numpy (>=1.20)", "pandas (>=1.3)", "scipy (>=1.8)"]
+developer = ["mypy (>=1.1)", "pre-commit (>=3.2)"]
+doc = ["nb2plots (>=0.6)", "numpydoc (>=1.5)", "pillow (>=9.4)", "pydata-sphinx-theme (>=0.13)", "sphinx (>=6.1)", "sphinx-gallery (>=0.12)", "texext (>=0.6.7)"]
+extra = ["lxml (>=4.6)", "pydot (>=1.4.2)", "pygraphviz (>=1.10)", "sympy (>=1.10)"]
+test = ["codecov (>=2.1)", "pytest (>=7.2)", "pytest-cov (>=4.0)"]
+
+[[package]]
+name = "numba"
+version = "0.56.4"
+description = "compiling Python code using LLVM"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "numba-0.56.4-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:9f62672145f8669ec08762895fe85f4cf0ead08ce3164667f2b94b2f62ab23c3"},
+ {file = "numba-0.56.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c602d015478b7958408d788ba00a50272649c5186ea8baa6cf71d4a1c761bba1"},
+ {file = "numba-0.56.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:85dbaed7a05ff96492b69a8900c5ba605551afb9b27774f7f10511095451137c"},
+ {file = "numba-0.56.4-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:f4cfc3a19d1e26448032049c79fc60331b104f694cf570a9e94f4e2c9d0932bb"},
+ {file = "numba-0.56.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:4e08e203b163ace08bad500b0c16f6092b1eb34fd1fce4feaf31a67a3a5ecf3b"},
+ {file = "numba-0.56.4-cp310-cp310-win32.whl", hash = "sha256:0611e6d3eebe4cb903f1a836ffdb2bda8d18482bcd0a0dcc56e79e2aa3fefef5"},
+ {file = "numba-0.56.4-cp310-cp310-win_amd64.whl", hash = "sha256:fbfb45e7b297749029cb28694abf437a78695a100e7c2033983d69f0ba2698d4"},
+ {file = "numba-0.56.4-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:3cb1a07a082a61df80a468f232e452d818f5ae254b40c26390054e4e868556e0"},
+ {file = "numba-0.56.4-cp37-cp37m-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:d69ad934e13c15684e7887100a8f5f0f61d7a8e57e0fd29d9993210089a5b531"},
+ {file = "numba-0.56.4-cp37-cp37m-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:dbcc847bac2d225265d054993a7f910fda66e73d6662fe7156452cac0325b073"},
+ {file = "numba-0.56.4-cp37-cp37m-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:8a95ca9cc77ea4571081f6594e08bd272b66060634b8324e99cd1843020364f9"},
+ {file = "numba-0.56.4-cp37-cp37m-win32.whl", hash = "sha256:fcdf84ba3ed8124eb7234adfbb8792f311991cbf8aed1cad4b1b1a7ee08380c1"},
+ {file = "numba-0.56.4-cp37-cp37m-win_amd64.whl", hash = "sha256:42f9e1be942b215df7e6cc9948cf9c15bb8170acc8286c063a9e57994ef82fd1"},
+ {file = "numba-0.56.4-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:553da2ce74e8862e18a72a209ed3b6d2924403bdd0fb341fa891c6455545ba7c"},
+ {file = "numba-0.56.4-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:4373da9757049db7c90591e9ec55a2e97b2b36ba7ae3bf9c956a513374077470"},
+ {file = "numba-0.56.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:3a993349b90569518739009d8f4b523dfedd7e0049e6838c0e17435c3e70dcc4"},
+ {file = "numba-0.56.4-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:720886b852a2d62619ae3900fe71f1852c62db4f287d0c275a60219e1643fc04"},
+ {file = "numba-0.56.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:e64d338b504c9394a4a34942df4627e1e6cb07396ee3b49fe7b8d6420aa5104f"},
+ {file = "numba-0.56.4-cp38-cp38-win32.whl", hash = "sha256:03fe94cd31e96185cce2fae005334a8cc712fc2ba7756e52dff8c9400718173f"},
+ {file = "numba-0.56.4-cp38-cp38-win_amd64.whl", hash = "sha256:91f021145a8081f881996818474ef737800bcc613ffb1e618a655725a0f9e246"},
+ {file = "numba-0.56.4-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:d0ae9270a7a5cc0ede63cd234b4ff1ce166c7a749b91dbbf45e0000c56d3eade"},
+ {file = "numba-0.56.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:c75e8a5f810ce80a0cfad6e74ee94f9fde9b40c81312949bf356b7304ef20740"},
+ {file = "numba-0.56.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl", hash = "sha256:a12ef323c0f2101529d455cfde7f4135eaa147bad17afe10b48634f796d96abd"},
+ {file = "numba-0.56.4-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:03634579d10a6129181129de293dd6b5eaabee86881369d24d63f8fe352dd6cb"},
+ {file = "numba-0.56.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:0240f9026b015e336069329839208ebd70ec34ae5bfbf402e4fcc8e06197528e"},
+ {file = "numba-0.56.4-cp39-cp39-win32.whl", hash = "sha256:14dbbabf6ffcd96ee2ac827389afa59a70ffa9f089576500434c34abf9b054a4"},
+ {file = "numba-0.56.4-cp39-cp39-win_amd64.whl", hash = "sha256:0da583c532cd72feefd8e551435747e0e0fbb3c0530357e6845fcc11e38d6aea"},
+ {file = "numba-0.56.4.tar.gz", hash = "sha256:32d9fef412c81483d7efe0ceb6cf4d3310fde8b624a9cecca00f790573ac96ee"},
+]
+
+[package.dependencies]
+llvmlite = ">=0.39.0dev0,<0.40"
+numpy = ">=1.18,<1.24"
+setuptools = "*"
+
+[[package]]
+name = "numpy"
+version = "1.23.5"
+description = "NumPy is the fundamental package for array computing with Python."
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "numpy-1.23.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:9c88793f78fca17da0145455f0d7826bcb9f37da4764af27ac945488116efe63"},
+ {file = "numpy-1.23.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e9f4c4e51567b616be64e05d517c79a8a22f3606499941d97bb76f2ca59f982d"},
+ {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7903ba8ab592b82014713c491f6c5d3a1cde5b4a3bf116404e08f5b52f6daf43"},
+ {file = "numpy-1.23.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e05b1c973a9f858c74367553e236f287e749465f773328c8ef31abe18f691e1"},
+ {file = "numpy-1.23.5-cp310-cp310-win32.whl", hash = "sha256:522e26bbf6377e4d76403826ed689c295b0b238f46c28a7251ab94716da0b280"},
+ {file = "numpy-1.23.5-cp310-cp310-win_amd64.whl", hash = "sha256:dbee87b469018961d1ad79b1a5d50c0ae850000b639bcb1b694e9981083243b6"},
+ {file = "numpy-1.23.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ce571367b6dfe60af04e04a1834ca2dc5f46004ac1cc756fb95319f64c095a96"},
+ {file = "numpy-1.23.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:56e454c7833e94ec9769fa0f86e6ff8e42ee38ce0ce1fa4cbb747ea7e06d56aa"},
+ {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5039f55555e1eab31124a5768898c9e22c25a65c1e0037f4d7c495a45778c9f2"},
+ {file = "numpy-1.23.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58f545efd1108e647604a1b5aa809591ccd2540f468a880bedb97247e72db387"},
+ {file = "numpy-1.23.5-cp311-cp311-win32.whl", hash = "sha256:b2a9ab7c279c91974f756c84c365a669a887efa287365a8e2c418f8b3ba73fb0"},
+ {file = "numpy-1.23.5-cp311-cp311-win_amd64.whl", hash = "sha256:0cbe9848fad08baf71de1a39e12d1b6310f1d5b2d0ea4de051058e6e1076852d"},
+ {file = "numpy-1.23.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:f063b69b090c9d918f9df0a12116029e274daf0181df392839661c4c7ec9018a"},
+ {file = "numpy-1.23.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0aaee12d8883552fadfc41e96b4c82ee7d794949e2a7c3b3a7201e968c7ecab9"},
+ {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:92c8c1e89a1f5028a4c6d9e3ccbe311b6ba53694811269b992c0b224269e2398"},
+ {file = "numpy-1.23.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d208a0f8729f3fb790ed18a003f3a57895b989b40ea4dce4717e9cf4af62c6bb"},
+ {file = "numpy-1.23.5-cp38-cp38-win32.whl", hash = "sha256:06005a2ef6014e9956c09ba07654f9837d9e26696a0470e42beedadb78c11b07"},
+ {file = "numpy-1.23.5-cp38-cp38-win_amd64.whl", hash = "sha256:ca51fcfcc5f9354c45f400059e88bc09215fb71a48d3768fb80e357f3b457e1e"},
+ {file = "numpy-1.23.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:8969bfd28e85c81f3f94eb4a66bc2cf1dbdc5c18efc320af34bffc54d6b1e38f"},
+ {file = "numpy-1.23.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:a7ac231a08bb37f852849bbb387a20a57574a97cfc7b6cabb488a4fc8be176de"},
+ {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf837dc63ba5c06dc8797c398db1e223a466c7ece27a1f7b5232ba3466aafe3d"},
+ {file = "numpy-1.23.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:33161613d2269025873025b33e879825ec7b1d831317e68f4f2f0f84ed14c719"},
+ {file = "numpy-1.23.5-cp39-cp39-win32.whl", hash = "sha256:af1da88f6bc3d2338ebbf0e22fe487821ea4d8e89053e25fa59d1d79786e7481"},
+ {file = "numpy-1.23.5-cp39-cp39-win_amd64.whl", hash = "sha256:09b7847f7e83ca37c6e627682f145856de331049013853f344f37b0c9690e3df"},
+ {file = "numpy-1.23.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:abdde9f795cf292fb9651ed48185503a2ff29be87770c3b8e2a14b0cd7aa16f8"},
+ {file = "numpy-1.23.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f9a909a8bae284d46bbfdefbdd4a262ba19d3bc9921b1e76126b1d21c3c34135"},
+ {file = "numpy-1.23.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:01dd17cbb340bf0fc23981e52e1d18a9d4050792e8fb8363cecbf066a84b827d"},
+ {file = "numpy-1.23.5.tar.gz", hash = "sha256:1b1766d6f397c18153d40015ddfc79ddb715cabadc04d2d228d4e5a8bc4ded1a"},
+]
+
+[[package]]
+name = "orjson"
+version = "3.9.1"
+description = "Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "orjson-3.9.1-cp310-cp310-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:c4434b7b786fdc394b95d029fb99949d7c2b05bbd4bf5cb5e3906be96ffeee3b"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09faf14f74ed47e773fa56833be118e04aa534956f661eb491522970b7478e3b"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:503eb86a8d53a187fe66aa80c69295a3ca35475804da89a9547e4fce5f803822"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:20f2804b5a1dbd3609c086041bd243519224d47716efd7429db6c03ed28b7cc3"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0fd828e0656615a711c4cc4da70f3cac142e66a6703ba876c20156a14e28e3fa"},
+ {file = "orjson-3.9.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ec53d648176f873203b9c700a0abacab33ca1ab595066e9d616f98cdc56f4434"},
+ {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:e186ae76b0d97c505500664193ddf508c13c1e675d9b25f1f4414a7606100da6"},
+ {file = "orjson-3.9.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d4edee78503016f4df30aeede0d999b3cb11fb56f47e9db0e487bce0aaca9285"},
+ {file = "orjson-3.9.1-cp310-none-win_amd64.whl", hash = "sha256:a4cc5d21e68af982d9a2528ac61e604f092c60eed27aef3324969c68f182ec7e"},
+ {file = "orjson-3.9.1-cp311-cp311-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:761b6efd33c49de20dd73ce64cc59da62c0dab10aa6015f582680e0663cc792c"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:31229f9d0b8dc2ef7ee7e4393f2e4433a28e16582d4b25afbfccc9d68dc768f8"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:0b7ab18d55ecb1de543d452f0a5f8094b52282b916aa4097ac11a4c79f317b86"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:db774344c39041f4801c7dfe03483df9203cbd6c84e601a65908e5552228dd25"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ae47ef8c0fe89c4677db7e9e1fb2093ca6e66c3acbee5442d84d74e727edad5e"},
+ {file = "orjson-3.9.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:103952c21575b9805803c98add2eaecd005580a1e746292ed2ec0d76dd3b9746"},
+ {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:2cb0121e6f2c9da3eddf049b99b95fef0adf8480ea7cb544ce858706cdf916eb"},
+ {file = "orjson-3.9.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:24d4ddaa2876e657c0fd32902b5c451fd2afc35159d66a58da7837357044b8c2"},
+ {file = "orjson-3.9.1-cp311-none-win_amd64.whl", hash = "sha256:0b53b5f72cf536dd8aa4fc4c95e7e09a7adb119f8ff8ee6cc60f735d7740ad6a"},
+ {file = "orjson-3.9.1-cp37-cp37m-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d4b68d01a506242316a07f1d2f29fb0a8b36cee30a7c35076f1ef59dce0890c1"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d9dd4abe6c6fd352f00f4246d85228f6a9847d0cc14f4d54ee553718c225388f"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:9e20bca5e13041e31ceba7a09bf142e6d63c8a7467f5a9c974f8c13377c75af2"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d8ae0467d01eb1e4bcffef4486d964bfd1c2e608103e75f7074ed34be5df48cc"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:06f6ab4697fab090517f295915318763a97a12ee8186054adf21c1e6f6abbd3d"},
+ {file = "orjson-3.9.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8515867713301fa065c58ec4c9053ba1a22c35113ab4acad555317b8fd802e50"},
+ {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:393d0697d1dfa18d27d193e980c04fdfb672c87f7765b87952f550521e21b627"},
+ {file = "orjson-3.9.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:d96747662d3666f79119e5d28c124e7d356c7dc195cd4b09faea4031c9079dc9"},
+ {file = "orjson-3.9.1-cp37-none-win_amd64.whl", hash = "sha256:6d173d3921dd58a068c88ec22baea7dbc87a137411501618b1292a9d6252318e"},
+ {file = "orjson-3.9.1-cp38-cp38-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:d1c2b0b4246c992ce2529fc610a446b945f1429445ece1c1f826a234c829a918"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:19f70ba1f441e1c4bb1a581f0baa092e8b3e3ce5b2aac2e1e090f0ac097966da"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:375d65f002e686212aac42680aed044872c45ee4bc656cf63d4a215137a6124a"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4751cee4a7b1daeacb90a7f5adf2170ccab893c3ab7c5cea58b45a13f89b30b3"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:78d9a2a4b2302d5ebc3695498ebc305c3568e5ad4f3501eb30a6405a32d8af22"},
+ {file = "orjson-3.9.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:46b4facc32643b2689dfc292c0c463985dac4b6ab504799cf51fc3c6959ed668"},
+ {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:ec7c8a0f1bf35da0d5fd14f8956f3b82a9a6918a3c6963d718dfd414d6d3b604"},
+ {file = "orjson-3.9.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:d3a40b0fbe06ccd4d6a99e523d20b47985655bcada8d1eba485b1b32a43e4904"},
+ {file = "orjson-3.9.1-cp38-none-win_amd64.whl", hash = "sha256:402f9d3edfec4560a98880224ec10eba4c5f7b4791e4bc0d4f4d8df5faf2a006"},
+ {file = "orjson-3.9.1-cp39-cp39-macosx_10_15_x86_64.macosx_11_0_arm64.macosx_10_15_universal2.whl", hash = "sha256:49c0d78dcd34626e2e934f1192d7c052b94e0ecadc5f386fd2bda6d2e03dadf5"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:125f63e56d38393daa0a1a6dc6fedefca16c538614b66ea5997c3bd3af35ef26"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:08927970365d2e1f3ce4894f9ff928a7b865d53f26768f1bbdd85dd4fee3e966"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f9a744e212d4780ecd67f4b6b128b2e727bee1df03e7059cddb2dfe1083e7dc4"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5d1dbf36db7240c61eec98c8d21545d671bce70be0730deb2c0d772e06b71af3"},
+ {file = "orjson-3.9.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:80a1e384626f76b66df615f7bb622a79a25c166d08c5d2151ffd41f24c4cc104"},
+ {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:15d28872fb055bf17ffca913826e618af61b2f689d2b170f72ecae1a86f80d52"},
+ {file = "orjson-3.9.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:1e4d905338f9ef32c67566929dfbfbb23cc80287af8a2c38930fb0eda3d40b76"},
+ {file = "orjson-3.9.1-cp39-none-win_amd64.whl", hash = "sha256:48a27da6c7306965846565cc385611d03382bbd84120008653aa2f6741e2105d"},
+ {file = "orjson-3.9.1.tar.gz", hash = "sha256:db373a25ec4a4fccf8186f9a72a1b3442837e40807a736a815ab42481e83b7d0"},
+]
+
+[[package]]
+name = "packageurl-python"
+version = "0.11.1"
+description = "A purl aka. Package URL parser and builder"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "packageurl-python-0.11.1.tar.gz", hash = "sha256:bbcc53d2cb5920c815c1626c75992f319bfc450b73893fa7bd8aac5869aa49fe"},
+ {file = "packageurl_python-0.11.1-py3-none-any.whl", hash = "sha256:4bad1d3ea4feb5e7a1db5ca8fb690ac9c82ab18e08d500755947b853df68817d"},
+]
+
+[package.extras]
+build = ["wheel"]
+lint = ["black", "isort", "mypy"]
+test = ["pytest"]
+
+[[package]]
+name = "packaging"
+version = "23.1"
+description = "Core utilities for Python packages"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "packaging-23.1-py3-none-any.whl", hash = "sha256:994793af429502c4ea2ebf6bf664629d07c1a9fe974af92966e4b8d2df7edc61"},
+ {file = "packaging-23.1.tar.gz", hash = "sha256:a392980d2b6cffa644431898be54b0045151319d1e7ec34f0cfed48767dd334f"},
+]
+
+[[package]]
+name = "pandas"
+version = "2.0.3"
+description = "Powerful data structures for data analysis, time series, and statistics"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "pandas-2.0.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e4c7c9f27a4185304c7caf96dc7d91bc60bc162221152de697c98eb0b2648dd8"},
+ {file = "pandas-2.0.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:f167beed68918d62bffb6ec64f2e1d8a7d297a038f86d4aed056b9493fca407f"},
+ {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ce0c6f76a0f1ba361551f3e6dceaff06bde7514a374aa43e33b588ec10420183"},
+ {file = "pandas-2.0.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ba619e410a21d8c387a1ea6e8a0e49bb42216474436245718d7f2e88a2f8d7c0"},
+ {file = "pandas-2.0.3-cp310-cp310-win32.whl", hash = "sha256:3ef285093b4fe5058eefd756100a367f27029913760773c8bf1d2d8bebe5d210"},
+ {file = "pandas-2.0.3-cp310-cp310-win_amd64.whl", hash = "sha256:9ee1a69328d5c36c98d8e74db06f4ad518a1840e8ccb94a4ba86920986bb617e"},
+ {file = "pandas-2.0.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b084b91d8d66ab19f5bb3256cbd5ea661848338301940e17f4492b2ce0801fe8"},
+ {file = "pandas-2.0.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:37673e3bdf1551b95bf5d4ce372b37770f9529743d2498032439371fc7b7eb26"},
+ {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b9cb1e14fdb546396b7e1b923ffaeeac24e4cedd14266c3497216dd4448e4f2d"},
+ {file = "pandas-2.0.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d9cd88488cceb7635aebb84809d087468eb33551097d600c6dad13602029c2df"},
+ {file = "pandas-2.0.3-cp311-cp311-win32.whl", hash = "sha256:694888a81198786f0e164ee3a581df7d505024fbb1f15202fc7db88a71d84ebd"},
+ {file = "pandas-2.0.3-cp311-cp311-win_amd64.whl", hash = "sha256:6a21ab5c89dcbd57f78d0ae16630b090eec626360085a4148693def5452d8a6b"},
+ {file = "pandas-2.0.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:9e4da0d45e7f34c069fe4d522359df7d23badf83abc1d1cef398895822d11061"},
+ {file = "pandas-2.0.3-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:32fca2ee1b0d93dd71d979726b12b61faa06aeb93cf77468776287f41ff8fdc5"},
+ {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:258d3624b3ae734490e4d63c430256e716f488c4fcb7c8e9bde2d3aa46c29089"},
+ {file = "pandas-2.0.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9eae3dc34fa1aa7772dd3fc60270d13ced7346fcbcfee017d3132ec625e23bb0"},
+ {file = "pandas-2.0.3-cp38-cp38-win32.whl", hash = "sha256:f3421a7afb1a43f7e38e82e844e2bca9a6d793d66c1a7f9f0ff39a795bbc5e02"},
+ {file = "pandas-2.0.3-cp38-cp38-win_amd64.whl", hash = "sha256:69d7f3884c95da3a31ef82b7618af5710dba95bb885ffab339aad925c3e8ce78"},
+ {file = "pandas-2.0.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5247fb1ba347c1261cbbf0fcfba4a3121fbb4029d95d9ef4dc45406620b25c8b"},
+ {file = "pandas-2.0.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:81af086f4543c9d8bb128328b5d32e9986e0c84d3ee673a2ac6fb57fd14f755e"},
+ {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1994c789bf12a7c5098277fb43836ce090f1073858c10f9220998ac74f37c69b"},
+ {file = "pandas-2.0.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5ec591c48e29226bcbb316e0c1e9423622bc7a4eaf1ef7c3c9fa1a3981f89641"},
+ {file = "pandas-2.0.3-cp39-cp39-win32.whl", hash = "sha256:04dbdbaf2e4d46ca8da896e1805bc04eb85caa9a82e259e8eed00254d5e0c682"},
+ {file = "pandas-2.0.3-cp39-cp39-win_amd64.whl", hash = "sha256:1168574b036cd8b93abc746171c9b4f1b83467438a5e45909fed645cf8692dbc"},
+ {file = "pandas-2.0.3.tar.gz", hash = "sha256:c02f372a88e0d17f36d3093a644c73cfc1788e876a7c4bcb4020a77512e2043c"},
+]
+
+[package.dependencies]
+numpy = {version = ">=1.20.3", markers = "python_version < \"3.10\""}
+python-dateutil = ">=2.8.2"
+pytz = ">=2020.1"
+tzdata = ">=2022.1"
+
+[package.extras]
+all = ["PyQt5 (>=5.15.1)", "SQLAlchemy (>=1.4.16)", "beautifulsoup4 (>=4.9.3)", "bottleneck (>=1.3.2)", "brotlipy (>=0.7.0)", "fastparquet (>=0.6.3)", "fsspec (>=2021.07.0)", "gcsfs (>=2021.07.0)", "html5lib (>=1.1)", "hypothesis (>=6.34.2)", "jinja2 (>=3.0.0)", "lxml (>=4.6.3)", "matplotlib (>=3.6.1)", "numba (>=0.53.1)", "numexpr (>=2.7.3)", "odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pandas-gbq (>=0.15.0)", "psycopg2 (>=2.8.6)", "pyarrow (>=7.0.0)", "pymysql (>=1.0.2)", "pyreadstat (>=1.1.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)", "python-snappy (>=0.6.0)", "pyxlsb (>=1.0.8)", "qtpy (>=2.2.0)", "s3fs (>=2021.08.0)", "scipy (>=1.7.1)", "tables (>=3.6.1)", "tabulate (>=0.8.9)", "xarray (>=0.21.0)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)", "zstandard (>=0.15.2)"]
+aws = ["s3fs (>=2021.08.0)"]
+clipboard = ["PyQt5 (>=5.15.1)", "qtpy (>=2.2.0)"]
+compression = ["brotlipy (>=0.7.0)", "python-snappy (>=0.6.0)", "zstandard (>=0.15.2)"]
+computation = ["scipy (>=1.7.1)", "xarray (>=0.21.0)"]
+excel = ["odfpy (>=1.4.1)", "openpyxl (>=3.0.7)", "pyxlsb (>=1.0.8)", "xlrd (>=2.0.1)", "xlsxwriter (>=1.4.3)"]
+feather = ["pyarrow (>=7.0.0)"]
+fss = ["fsspec (>=2021.07.0)"]
+gcp = ["gcsfs (>=2021.07.0)", "pandas-gbq (>=0.15.0)"]
+hdf5 = ["tables (>=3.6.1)"]
+html = ["beautifulsoup4 (>=4.9.3)", "html5lib (>=1.1)", "lxml (>=4.6.3)"]
+mysql = ["SQLAlchemy (>=1.4.16)", "pymysql (>=1.0.2)"]
+output-formatting = ["jinja2 (>=3.0.0)", "tabulate (>=0.8.9)"]
+parquet = ["pyarrow (>=7.0.0)"]
+performance = ["bottleneck (>=1.3.2)", "numba (>=0.53.1)", "numexpr (>=2.7.1)"]
+plot = ["matplotlib (>=3.6.1)"]
+postgresql = ["SQLAlchemy (>=1.4.16)", "psycopg2 (>=2.8.6)"]
+spss = ["pyreadstat (>=1.1.2)"]
+sql-other = ["SQLAlchemy (>=1.4.16)"]
+test = ["hypothesis (>=6.34.2)", "pytest (>=7.3.2)", "pytest-asyncio (>=0.17.0)", "pytest-xdist (>=2.2.0)"]
+xml = ["lxml (>=4.6.3)"]
+
+[[package]]
+name = "pathspec"
+version = "0.11.1"
+description = "Utility library for gitignore style pattern matching of file paths."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pathspec-0.11.1-py3-none-any.whl", hash = "sha256:d8af70af76652554bd134c22b3e8a1cc46ed7d91edcdd721ef1a0c51a84a5293"},
+ {file = "pathspec-0.11.1.tar.gz", hash = "sha256:2798de800fa92780e33acca925945e9a19a133b715067cf165b8866c15a31687"},
+]
+
+[[package]]
+name = "pbr"
+version = "5.11.1"
+description = "Python Build Reasonableness"
+category = "dev"
+optional = false
+python-versions = ">=2.6"
+files = [
+ {file = "pbr-5.11.1-py2.py3-none-any.whl", hash = "sha256:567f09558bae2b3ab53cb3c1e2e33e726ff3338e7bae3db5dc954b3a44eef12b"},
+ {file = "pbr-5.11.1.tar.gz", hash = "sha256:aefc51675b0b533d56bb5fd1c8c6c0522fe31896679882e1c4c63d5e4a0fccb3"},
+]
+
+[[package]]
+name = "pillow"
+version = "10.0.0"
+description = "Python Imaging Library (Fork)"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "Pillow-10.0.0-cp310-cp310-macosx_10_10_x86_64.whl", hash = "sha256:1f62406a884ae75fb2f818694469519fb685cc7eaff05d3451a9ebe55c646891"},
+ {file = "Pillow-10.0.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:d5db32e2a6ccbb3d34d87c87b432959e0db29755727afb37290e10f6e8e62614"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:edf4392b77bdc81f36e92d3a07a5cd072f90253197f4a52a55a8cec48a12483b"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:520f2a520dc040512699f20fa1c363eed506e94248d71f85412b625026f6142c"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:8c11160913e3dd06c8ffdb5f233a4f254cb449f4dfc0f8f4549eda9e542c93d1"},
+ {file = "Pillow-10.0.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a74ba0c356aaa3bb8e3eb79606a87669e7ec6444be352870623025d75a14a2bf"},
+ {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d5d0dae4cfd56969d23d94dc8e89fb6a217be461c69090768227beb8ed28c0a3"},
+ {file = "Pillow-10.0.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:22c10cc517668d44b211717fd9775799ccec4124b9a7f7b3635fc5386e584992"},
+ {file = "Pillow-10.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:dffe31a7f47b603318c609f378ebcd57f1554a3a6a8effbc59c3c69f804296de"},
+ {file = "Pillow-10.0.0-cp311-cp311-macosx_10_10_x86_64.whl", hash = "sha256:9fb218c8a12e51d7ead2a7c9e101a04982237d4855716af2e9499306728fb485"},
+ {file = "Pillow-10.0.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:d35e3c8d9b1268cbf5d3670285feb3528f6680420eafe35cccc686b73c1e330f"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3ed64f9ca2f0a95411e88a4efbd7a29e5ce2cea36072c53dd9d26d9c76f753b3"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0b6eb5502f45a60a3f411c63187db83a3d3107887ad0d036c13ce836f8a36f1d"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:c1fbe7621c167ecaa38ad29643d77a9ce7311583761abf7836e1510c580bf3dd"},
+ {file = "Pillow-10.0.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:cd25d2a9d2b36fcb318882481367956d2cf91329f6892fe5d385c346c0649629"},
+ {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:3b08d4cc24f471b2c8ca24ec060abf4bebc6b144cb89cba638c720546b1cf538"},
+ {file = "Pillow-10.0.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d737a602fbd82afd892ca746392401b634e278cb65d55c4b7a8f48e9ef8d008d"},
+ {file = "Pillow-10.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:3a82c40d706d9aa9734289740ce26460a11aeec2d9c79b7af87bb35f0073c12f"},
+ {file = "Pillow-10.0.0-cp312-cp312-macosx_10_10_x86_64.whl", hash = "sha256:d80cf684b541685fccdd84c485b31ce73fc5c9b5d7523bf1394ce134a60c6883"},
+ {file = "Pillow-10.0.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:76de421f9c326da8f43d690110f0e79fe3ad1e54be811545d7d91898b4c8493e"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81ff539a12457809666fef6624684c008e00ff6bf455b4b89fd00a140eecd640"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce543ed15570eedbb85df19b0a1a7314a9c8141a36ce089c0a894adbfccb4568"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:685ac03cc4ed5ebc15ad5c23bc555d68a87777586d970c2c3e216619a5476223"},
+ {file = "Pillow-10.0.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:d72e2ecc68a942e8cf9739619b7f408cc7b272b279b56b2c83c6123fcfa5cdff"},
+ {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:d50b6aec14bc737742ca96e85d6d0a5f9bfbded018264b3b70ff9d8c33485551"},
+ {file = "Pillow-10.0.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:00e65f5e822decd501e374b0650146063fbb30a7264b4d2744bdd7b913e0cab5"},
+ {file = "Pillow-10.0.0-cp312-cp312-win_amd64.whl", hash = "sha256:f31f9fdbfecb042d046f9d91270a0ba28368a723302786c0009ee9b9f1f60199"},
+ {file = "Pillow-10.0.0-cp38-cp38-macosx_10_10_x86_64.whl", hash = "sha256:349930d6e9c685c089284b013478d6f76e3a534e36ddfa912cde493f235372f3"},
+ {file = "Pillow-10.0.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:3a684105f7c32488f7153905a4e3015a3b6c7182e106fe3c37fbb5ef3e6994c3"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b4f69b3700201b80bb82c3a97d5e9254084f6dd5fb5b16fc1a7b974260f89f43"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3f07ea8d2f827d7d2a49ecf1639ec02d75ffd1b88dcc5b3a61bbb37a8759ad8d"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_aarch64.whl", hash = "sha256:040586f7d37b34547153fa383f7f9aed68b738992380ac911447bb78f2abe530"},
+ {file = "Pillow-10.0.0-cp38-cp38-manylinux_2_28_x86_64.whl", hash = "sha256:f88a0b92277de8e3ca715a0d79d68dc82807457dae3ab8699c758f07c20b3c51"},
+ {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:c7cf14a27b0d6adfaebb3ae4153f1e516df54e47e42dcc073d7b3d76111a8d86"},
+ {file = "Pillow-10.0.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:3400aae60685b06bb96f99a21e1ada7bc7a413d5f49bce739828ecd9391bb8f7"},
+ {file = "Pillow-10.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:dbc02381779d412145331789b40cc7b11fdf449e5d94f6bc0b080db0a56ea3f0"},
+ {file = "Pillow-10.0.0-cp39-cp39-macosx_10_10_x86_64.whl", hash = "sha256:9211e7ad69d7c9401cfc0e23d49b69ca65ddd898976d660a2fa5904e3d7a9baa"},
+ {file = "Pillow-10.0.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:faaf07ea35355b01a35cb442dd950d8f1bb5b040a7787791a535de13db15ed90"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c9f72a021fbb792ce98306ffb0c348b3c9cb967dce0f12a49aa4c3d3fdefa967"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f7c16705f44e0504a3a2a14197c1f0b32a95731d251777dcb060aa83022cb2d"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:76edb0a1fa2b4745fb0c99fb9fb98f8b180a1bbceb8be49b087e0b21867e77d3"},
+ {file = "Pillow-10.0.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:368ab3dfb5f49e312231b6f27b8820c823652b7cd29cfbd34090565a015e99ba"},
+ {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:608bfdee0d57cf297d32bcbb3c728dc1da0907519d1784962c5f0c68bb93e5a3"},
+ {file = "Pillow-10.0.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5c6e3df6bdd396749bafd45314871b3d0af81ff935b2d188385e970052091017"},
+ {file = "Pillow-10.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:7be600823e4c8631b74e4a0d38384c73f680e6105a7d3c6824fcf226c178c7e6"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-macosx_10_10_x86_64.whl", hash = "sha256:92be919bbc9f7d09f7ae343c38f5bb21c973d2576c1d45600fce4b74bafa7ac0"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8f8182b523b2289f7c415f589118228d30ac8c355baa2f3194ced084dac2dbba"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:38250a349b6b390ee6047a62c086d3817ac69022c127f8a5dc058c31ccef17f3"},
+ {file = "Pillow-10.0.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:88af2003543cc40c80f6fca01411892ec52b11021b3dc22ec3bc9d5afd1c5334"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-macosx_10_10_x86_64.whl", hash = "sha256:c189af0545965fa8d3b9613cfdb0cd37f9d71349e0f7750e1fd704648d475ed2"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce7b031a6fc11365970e6a5686d7ba8c63e4c1cf1ea143811acbb524295eabed"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:db24668940f82321e746773a4bc617bfac06ec831e5c88b643f91f122a785684"},
+ {file = "Pillow-10.0.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:efe8c0681042536e0d06c11f48cebe759707c9e9abf880ee213541c5b46c5bf3"},
+ {file = "Pillow-10.0.0.tar.gz", hash = "sha256:9c82b5b3e043c7af0d95792d0d20ccf68f61a1fec6b3530e718b688422727396"},
+]
+
+[package.extras]
+docs = ["furo", "olefile", "sphinx (>=2.4)", "sphinx-copybutton", "sphinx-inline-tabs", "sphinx-removed-in", "sphinxext-opengraph"]
+tests = ["check-manifest", "coverage", "defusedxml", "markdown2", "olefile", "packaging", "pyroma", "pytest", "pytest-cov", "pytest-timeout"]
+
+[[package]]
+name = "pip"
+version = "23.1.2"
+description = "The PyPA recommended tool for installing Python packages."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pip-23.1.2-py3-none-any.whl", hash = "sha256:3ef6ac33239e4027d9a5598a381b9d30880a1477e50039db2eac6e8a8f6d1b18"},
+ {file = "pip-23.1.2.tar.gz", hash = "sha256:0e7c86f486935893c708287b30bd050a36ac827ec7fe5e43fe7cb198dd835fba"},
+]
+
+[[package]]
+name = "pip-api"
+version = "0.0.30"
+description = "An unofficial, importable pip API"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pip-api-0.0.30.tar.gz", hash = "sha256:a05df2c7aa9b7157374bcf4273544201a0c7bae60a9c65bcf84f3959ef3896f3"},
+ {file = "pip_api-0.0.30-py3-none-any.whl", hash = "sha256:2a0314bd31522eb9ffe8a99668b0d07fee34ebc537931e7b6483001dbedcbdc9"},
+]
+
+[package.dependencies]
+pip = "*"
+
+[[package]]
+name = "pip-audit"
+version = "2.6.0"
+description = "A tool for scanning Python environments for known vulnerabilities"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pip_audit-2.6.0-py3-none-any.whl", hash = "sha256:49e97e3d6663d2ed0c00b7a7c468afcb816beb3988f32f8496d3fe3927cfd627"},
+ {file = "pip_audit-2.6.0.tar.gz", hash = "sha256:6431c363efa80ef52c2599197c5b8a39ff8708ce316624b97fa35b5cdf493118"},
+]
+
+[package.dependencies]
+CacheControl = {version = ">=0.13.0", extras = ["filecache"]}
+cyclonedx-python-lib = ">=4.0,<5.0"
+html5lib = ">=1.1"
+packaging = ">=23.0.0"
+pip-api = ">=0.0.28"
+pip-requirements-parser = ">=32.0.0"
+requests = ">=2.31.0"
+rich = ">=12.4"
+toml = ">=0.10"
+
+[package.extras]
+dev = ["build", "bump (>=1.3.2)", "pip-audit[doc,lint,test]"]
+doc = ["pdoc"]
+lint = ["black (>=22.3.0)", "interrogate", "isort", "mypy", "ruff (<0.0.276)", "types-html5lib", "types-requests", "types-toml"]
+test = ["coverage[toml]", "pretend", "pytest", "pytest-cov"]
+
+[[package]]
+name = "pip-requirements-parser"
+version = "32.0.1"
+description = "pip requirements parser - a mostly correct pip requirements parsing library because it uses pip's own code."
+category = "dev"
+optional = false
+python-versions = ">=3.6.0"
+files = [
+ {file = "pip-requirements-parser-32.0.1.tar.gz", hash = "sha256:b4fa3a7a0be38243123cf9d1f3518da10c51bdb165a2b2985566247f9155a7d3"},
+ {file = "pip_requirements_parser-32.0.1-py3-none-any.whl", hash = "sha256:4659bc2a667783e7a15d190f6fccf8b2486685b6dba4c19c3876314769c57526"},
+]
+
+[package.dependencies]
+packaging = "*"
+pyparsing = "*"
+
+[package.extras]
+docs = ["Sphinx (>=3.3.1)", "doc8 (>=0.8.1)", "sphinx-rtd-theme (>=0.5.0)"]
+testing = ["aboutcode-toolkit (>=6.0.0)", "black", "pytest (>=6,!=7.0.0)", "pytest-xdist (>=2)"]
+
+[[package]]
+name = "platformdirs"
+version = "3.8.0"
+description = "A small Python package for determining appropriate platform-specific dirs, e.g. a \"user data dir\"."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "platformdirs-3.8.0-py3-none-any.whl", hash = "sha256:ca9ed98ce73076ba72e092b23d3c93ea6c4e186b3f1c3dad6edd98ff6ffcca2e"},
+ {file = "platformdirs-3.8.0.tar.gz", hash = "sha256:b0cabcb11063d21a0b261d557acb0a9d2126350e63b70cdf7db6347baea456dc"},
+]
+
+[package.extras]
+docs = ["furo (>=2023.5.20)", "proselint (>=0.13)", "sphinx (>=7.0.1)", "sphinx-autodoc-typehints (>=1.23,!=1.23.4)"]
+test = ["appdirs (==1.4.4)", "covdefaults (>=2.3)", "pytest (>=7.3.1)", "pytest-cov (>=4.1)", "pytest-mock (>=3.10)"]
+
+[[package]]
+name = "pluggy"
+version = "1.2.0"
+description = "plugin and hook calling mechanisms for python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pluggy-1.2.0-py3-none-any.whl", hash = "sha256:c2fd55a7d7a3863cba1a013e4e2414658b1d07b6bc57b3919e0c63c9abb99849"},
+ {file = "pluggy-1.2.0.tar.gz", hash = "sha256:d12f0c4b579b15f5e054301bb226ee85eeeba08ffec228092f8defbaa3a4c4b3"},
+]
+
+[package.extras]
+dev = ["pre-commit", "tox"]
+testing = ["pytest", "pytest-benchmark"]
+
+[[package]]
+name = "pooch"
+version = "1.6.0"
+description = "\"Pooch manages your Python library's sample data files: it automatically downloads and stores them in a local directory, with support for versioning and corruption checks.\""
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "pooch-1.6.0-py3-none-any.whl", hash = "sha256:3bf0e20027096836b8dbce0152dbb785a269abeb621618eb4bdd275ff1e23c9c"},
+ {file = "pooch-1.6.0.tar.gz", hash = "sha256:57d20ec4b10dd694d2b05bb64bc6b109c6e85a6c1405794ce87ed8b341ab3f44"},
+]
+
+[package.dependencies]
+appdirs = ">=1.3.0"
+packaging = ">=20.0"
+requests = ">=2.19.0"
+
+[package.extras]
+progress = ["tqdm (>=4.41.0,<5.0.0)"]
+sftp = ["paramiko (>=2.7.0)"]
+xxhash = ["xxhash (>=1.4.3)"]
+
+[[package]]
+name = "prometheus-client"
+version = "0.12.0"
+description = "Python client for the Prometheus monitoring system."
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "prometheus_client-0.12.0-py2.py3-none-any.whl", hash = "sha256:317453ebabff0a1b02df7f708efbab21e3489e7072b61cb6957230dd004a0af0"},
+ {file = "prometheus_client-0.12.0.tar.gz", hash = "sha256:1b12ba48cee33b9b0b9de64a1047cbd3c5f2d0ab6ebcead7ddda613a750ec3c5"},
+]
+
+[package.extras]
+twisted = ["twisted"]
+
+[[package]]
+name = "psutil"
+version = "5.9.5"
+description = "Cross-platform lib for process and system monitoring in Python."
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "psutil-5.9.5-cp27-cp27m-macosx_10_9_x86_64.whl", hash = "sha256:be8929ce4313f9f8146caad4272f6abb8bf99fc6cf59344a3167ecd74f4f203f"},
+ {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_i686.whl", hash = "sha256:ab8ed1a1d77c95453db1ae00a3f9c50227ebd955437bcf2a574ba8adbf6a74d5"},
+ {file = "psutil-5.9.5-cp27-cp27m-manylinux2010_x86_64.whl", hash = "sha256:4aef137f3345082a3d3232187aeb4ac4ef959ba3d7c10c33dd73763fbc063da4"},
+ {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_i686.whl", hash = "sha256:ea8518d152174e1249c4f2a1c89e3e6065941df2fa13a1ab45327716a23c2b48"},
+ {file = "psutil-5.9.5-cp27-cp27mu-manylinux2010_x86_64.whl", hash = "sha256:acf2aef9391710afded549ff602b5887d7a2349831ae4c26be7c807c0a39fac4"},
+ {file = "psutil-5.9.5-cp27-none-win32.whl", hash = "sha256:5b9b8cb93f507e8dbaf22af6a2fd0ccbe8244bf30b1baad6b3954e935157ae3f"},
+ {file = "psutil-5.9.5-cp27-none-win_amd64.whl", hash = "sha256:8c5f7c5a052d1d567db4ddd231a9d27a74e8e4a9c3f44b1032762bd7b9fdcd42"},
+ {file = "psutil-5.9.5-cp36-abi3-macosx_10_9_x86_64.whl", hash = "sha256:3c6f686f4225553615612f6d9bc21f1c0e305f75d7d8454f9b46e901778e7217"},
+ {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7a7dd9997128a0d928ed4fb2c2d57e5102bb6089027939f3b722f3a210f9a8da"},
+ {file = "psutil-5.9.5-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:89518112647f1276b03ca97b65cc7f64ca587b1eb0278383017c2a0dcc26cbe4"},
+ {file = "psutil-5.9.5-cp36-abi3-win32.whl", hash = "sha256:104a5cc0e31baa2bcf67900be36acde157756b9c44017b86b2c049f11957887d"},
+ {file = "psutil-5.9.5-cp36-abi3-win_amd64.whl", hash = "sha256:b258c0c1c9d145a1d5ceffab1134441c4c5113b2417fafff7315a917a026c3c9"},
+ {file = "psutil-5.9.5-cp38-abi3-macosx_11_0_arm64.whl", hash = "sha256:c607bb3b57dc779d55e1554846352b4e358c10fff3abf3514a7a6601beebdb30"},
+ {file = "psutil-5.9.5.tar.gz", hash = "sha256:5410638e4df39c54d957fc51ce03048acd8e6d60abc0f5107af51e5fb566eb3c"},
+]
+
+[package.extras]
+test = ["enum34", "ipaddress", "mock", "pywin32", "wmi"]
+
+[[package]]
+name = "py-serializable"
+version = "0.11.1"
+description = "Library for serializing and deserializing Python Objects to and from JSON and XML."
+category = "dev"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "py-serializable-0.11.1.tar.gz", hash = "sha256:ba0e1287b9e4f645a5334f1913abd8e647e7250209f84f55dce3909498a6f586"},
+ {file = "py_serializable-0.11.1-py3-none-any.whl", hash = "sha256:79e21f0672822e6200b15f45ce9f636e8126466f62dbd7d488c67313c72b5c3e"},
+]
+
+[package.dependencies]
+defusedxml = ">=0.7.1,<0.8.0"
+
+[[package]]
+name = "pyarrow"
+version = "12.0.1"
+description = "Python library for Apache Arrow"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pyarrow-12.0.1-cp310-cp310-macosx_10_14_x86_64.whl", hash = "sha256:6d288029a94a9bb5407ceebdd7110ba398a00412c5b0155ee9813a40d246c5df"},
+ {file = "pyarrow-12.0.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:345e1828efdbd9aa4d4de7d5676778aba384a2c3add896d995b23d368e60e5af"},
+ {file = "pyarrow-12.0.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8d6009fdf8986332b2169314da482baed47ac053311c8934ac6651e614deacd6"},
+ {file = "pyarrow-12.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2d3c4cbbf81e6dd23fe921bc91dc4619ea3b79bc58ef10bce0f49bdafb103daf"},
+ {file = "pyarrow-12.0.1-cp310-cp310-win_amd64.whl", hash = "sha256:cdacf515ec276709ac8042c7d9bd5be83b4f5f39c6c037a17a60d7ebfd92c890"},
+ {file = "pyarrow-12.0.1-cp311-cp311-macosx_10_14_x86_64.whl", hash = "sha256:749be7fd2ff260683f9cc739cb862fb11be376de965a2a8ccbf2693b098db6c7"},
+ {file = "pyarrow-12.0.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:6895b5fb74289d055c43db3af0de6e16b07586c45763cb5e558d38b86a91e3a7"},
+ {file = "pyarrow-12.0.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1887bdae17ec3b4c046fcf19951e71b6a619f39fa674f9881216173566c8f718"},
+ {file = "pyarrow-12.0.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e2c9cb8eeabbadf5fcfc3d1ddea616c7ce893db2ce4dcef0ac13b099ad7ca082"},
+ {file = "pyarrow-12.0.1-cp311-cp311-win_amd64.whl", hash = "sha256:ce4aebdf412bd0eeb800d8e47db854f9f9f7e2f5a0220440acf219ddfddd4f63"},
+ {file = "pyarrow-12.0.1-cp37-cp37m-macosx_10_14_x86_64.whl", hash = "sha256:e0d8730c7f6e893f6db5d5b86eda42c0a130842d101992b581e2138e4d5663d3"},
+ {file = "pyarrow-12.0.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:43364daec02f69fec89d2315f7fbfbeec956e0d991cbbef471681bd77875c40f"},
+ {file = "pyarrow-12.0.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:051f9f5ccf585f12d7de836e50965b3c235542cc896959320d9776ab93f3b33d"},
+ {file = "pyarrow-12.0.1-cp37-cp37m-win_amd64.whl", hash = "sha256:be2757e9275875d2a9c6e6052ac7957fbbfc7bc7370e4a036a9b893e96fedaba"},
+ {file = "pyarrow-12.0.1-cp38-cp38-macosx_10_14_x86_64.whl", hash = "sha256:cf812306d66f40f69e684300f7af5111c11f6e0d89d6b733e05a3de44961529d"},
+ {file = "pyarrow-12.0.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:459a1c0ed2d68671188b2118c63bac91eaef6fc150c77ddd8a583e3c795737bf"},
+ {file = "pyarrow-12.0.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:85e705e33eaf666bbe508a16fd5ba27ca061e177916b7a317ba5a51bee43384c"},
+ {file = "pyarrow-12.0.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9120c3eb2b1f6f516a3b7a9714ed860882d9ef98c4b17edcdc91d95b7528db60"},
+ {file = "pyarrow-12.0.1-cp38-cp38-win_amd64.whl", hash = "sha256:c780f4dc40460015d80fcd6a6140de80b615349ed68ef9adb653fe351778c9b3"},
+ {file = "pyarrow-12.0.1-cp39-cp39-macosx_10_14_x86_64.whl", hash = "sha256:a3c63124fc26bf5f95f508f5d04e1ece8cc23a8b0af2a1e6ab2b1ec3fdc91b24"},
+ {file = "pyarrow-12.0.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:b13329f79fa4472324f8d32dc1b1216616d09bd1e77cfb13104dec5463632c36"},
+ {file = "pyarrow-12.0.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb656150d3d12ec1396f6dde542db1675a95c0cc8366d507347b0beed96e87ca"},
+ {file = "pyarrow-12.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6251e38470da97a5b2e00de5c6a049149f7b2bd62f12fa5dbb9ac674119ba71a"},
+ {file = "pyarrow-12.0.1-cp39-cp39-win_amd64.whl", hash = "sha256:3de26da901216149ce086920547dfff5cd22818c9eab67ebc41e863a5883bac7"},
+ {file = "pyarrow-12.0.1.tar.gz", hash = "sha256:cce317fc96e5b71107bf1f9f184d5e54e2bd14bbf3f9a3d62819961f0af86fec"},
+]
+
+[package.dependencies]
+numpy = ">=1.16.6"
+
+[[package]]
+name = "pycodestyle"
+version = "2.7.0"
+description = "Python style guide checker"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "pycodestyle-2.7.0-py2.py3-none-any.whl", hash = "sha256:514f76d918fcc0b55c6680472f0a37970994e07bbb80725808c17089be302068"},
+ {file = "pycodestyle-2.7.0.tar.gz", hash = "sha256:c389c1d06bf7904078ca03399a4816f974a1d590090fecea0c63ec26ebaf1cef"},
+]
+
+[[package]]
+name = "pycparser"
+version = "2.21"
+description = "C parser in Python"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"},
+ {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"},
+]
+
+[[package]]
+name = "pydub"
+version = "0.25.1"
+description = "Manipulate audio with an simple and easy high level interface"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pydub-0.25.1-py2.py3-none-any.whl", hash = "sha256:65617e33033874b59d87db603aa1ed450633288aefead953b30bded59cb599a6"},
+ {file = "pydub-0.25.1.tar.gz", hash = "sha256:980a33ce9949cab2a569606b65674d748ecbca4f0796887fd6f46173a7b0d30f"},
+]
+
+[[package]]
+name = "pyflakes"
+version = "2.3.1"
+description = "passive checker of Python programs"
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
+files = [
+ {file = "pyflakes-2.3.1-py2.py3-none-any.whl", hash = "sha256:7893783d01b8a89811dd72d7dfd4d84ff098e5eed95cfa8905b22bbffe52efc3"},
+ {file = "pyflakes-2.3.1.tar.gz", hash = "sha256:f5bc8ecabc05bb9d291eb5203d6810b49040f6ff446a756326104746cc00c1db"},
+]
+
+[[package]]
+name = "pygments"
+version = "2.15.1"
+description = "Pygments is a syntax highlighting package written in Python."
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "Pygments-2.15.1-py3-none-any.whl", hash = "sha256:db2db3deb4b4179f399a09054b023b6a586b76499d36965813c71aa8ed7b5fd1"},
+ {file = "Pygments-2.15.1.tar.gz", hash = "sha256:8ace4d3c1dd481894b2005f560ead0f9f19ee64fe983366be1a21e171d12775c"},
+]
+
+[package.extras]
+plugins = ["importlib-metadata"]
+
+[[package]]
+name = "pyjwt"
+version = "2.7.0"
+description = "JSON Web Token implementation in Python"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "PyJWT-2.7.0-py3-none-any.whl", hash = "sha256:ba2b425b15ad5ef12f200dc67dd56af4e26de2331f965c5439994dad075876e1"},
+ {file = "PyJWT-2.7.0.tar.gz", hash = "sha256:bd6ca4a3c4285c1a2d4349e5a035fdf8fb94e04ccd0fcbe6ba289dae9cc3e074"},
+]
+
+[package.dependencies]
+cryptography = {version = ">=3.4.0", optional = true, markers = "extra == \"crypto\""}
+
+[package.extras]
+crypto = ["cryptography (>=3.4.0)"]
+dev = ["coverage[toml] (==5.0.4)", "cryptography (>=3.4.0)", "pre-commit", "pytest (>=6.0.0,<7.0.0)", "sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"]
+docs = ["sphinx (>=4.5.0,<5.0.0)", "sphinx-rtd-theme", "zope.interface"]
+tests = ["coverage[toml] (==5.0.4)", "pytest (>=6.0.0,<7.0.0)"]
+
+[[package]]
+name = "pymongo"
+version = "3.13.0"
+description = "Python driver for MongoDB <http://www.mongodb.org>"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pymongo-3.13.0-cp27-cp27m-macosx_10_14_intel.whl", hash = "sha256:3ad3a3df830f7df7e0856c2bdb54d19f5bf188bd7420985e18643b8e4d2a075f"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_i686.whl", hash = "sha256:b96e0e9d2d48948240b510bac81614458fc10adcd3a93240c2fd96448b4efd35"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux1_x86_64.whl", hash = "sha256:9f592b202d77923498b32ddc5b376e5fa9ba280d3e16ed56cb8c932fe6d6a478"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:851f2bb52b5cb2f4711171ca925e0e05344a8452972a748a8a8ffdda1e1d72a7"},
+ {file = "pymongo-3.13.0-cp27-cp27m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:1c9d23f62a3fa7523d849c4942acc0d9ff7081ebc00c808ee7cfdc070df0687f"},
+ {file = "pymongo-3.13.0-cp27-cp27m-win32.whl", hash = "sha256:a17b81f22398e3e0f72bdf938e98c810286994b2bcc0a125cd5ad8fd4ea54ad7"},
+ {file = "pymongo-3.13.0-cp27-cp27m-win_amd64.whl", hash = "sha256:4f6dd55dab77adf60b445c11f426ee5cdfa1b86f6d54cb937bfcbf09572333ab"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_i686.whl", hash = "sha256:776f90bf2252f90a4ae838e7917638894c6356bef7265f424592e2fd1f577d05"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux1_x86_64.whl", hash = "sha256:50b99f4d3eee6f03778fe841d6f470e6c18e744dc665156da6da3bc6e65b398d"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:50a81b2d9f188c7909e0a1084fa969bb92a788076809c437ac1ae80393f46df9"},
+ {file = "pymongo-3.13.0-cp27-cp27mu-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c7c45a8a1a752002b0a7c81ab3a4c5e3b6f67f9826b16fbe3943f5329f565f24"},
+ {file = "pymongo-3.13.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:1037097708498bdc85f23c8798a5c46c7bce432d77d23608ff14e0d831f1a971"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux1_i686.whl", hash = "sha256:b5b733694e7df22d5c049581acfc487695a6ff813322318bed8dd66f79978636"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_aarch64.whl", hash = "sha256:d7c91747ec8dde51440dd594603158cc98abb3f7df84b2ed8a836f138285e4fb"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_i686.whl", hash = "sha256:f4175fcdddf764d371ee52ec4505a40facee2533e84abf2953cda86d050cfa1f"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_ppc64le.whl", hash = "sha256:93d4e9a02c17813b34e4bd9f6fbf07310c140c8f74341537c24d07c1cdeb24d1"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_s390x.whl", hash = "sha256:3b261d593f2563299062733ae003a925420a86ff4ddda68a69097d67204e43f3"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux2014_x86_64.whl", hash = "sha256:172db03182a22e9002157b262c1ea3b0045c73d4ff465adc152ce5b4b0e7b8d4"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09de3bfc995ae8cb955abb0c9ae963c134dba1b5622be3bcc527b89b0fd4091c"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:c0379447587ee4b8f983ba183202496e86c0358f47c45612619d634d1fcd82bd"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:30245a8747dc90019a3c9ad9df987e0280a3ea632ad36227cde7d1d8dcba0830"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65b6fddf6a7b91da044f202771a38e71bbb9bf42720a406b26b25fe2256e7102"},
+ {file = "pymongo-3.13.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5831a377d15a626fbec10890ffebc4c6abcd37e4126737932cd780a171eabdc1"},
+ {file = "pymongo-3.13.0-cp310-cp310-win32.whl", hash = "sha256:944249aa83dee314420c37d0f40c30a8f6dc4a3877566017b87062e53af449f4"},
+ {file = "pymongo-3.13.0-cp310-cp310-win_amd64.whl", hash = "sha256:ea8824ebc9a1a5c8269e8f1e3989b5a6bec876726e2f3c33ebd036cb488277f0"},
+ {file = "pymongo-3.13.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:bdd34c57b4da51a7961beb33645646d197e41f8517801dc76b37c1441e7a4e10"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:26f9cc42a162faa241c82e117ac85734ae9f14343dc2df1c90c6b2181f791b22"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4a82a1c10f5608e6494913faa169e213d703194bfca0aa710901f303be212414"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8927f22ef6a16229da7f18944deac8605bdc2c0858be5184259f2f7ce7fd4459"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e6f8191a282ef77e526f8f8f63753a437e4aa4bc78f5edd8b6b6ed0eaebd5363"},
+ {file = "pymongo-3.13.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4d9ed67c987bf9ac2ac684590ba3d2599cdfb0f331ee3db607f9684469b3b59d"},
+ {file = "pymongo-3.13.0-cp311-cp311-win32.whl", hash = "sha256:e8f6979664ff477cd61b06bf8aba206df7b2334209815ab3b1019931dab643d6"},
+ {file = "pymongo-3.13.0-cp311-cp311-win_amd64.whl", hash = "sha256:174fd1000e896d0dfbc7f6d7e6a1992a4868796c7dec31679e38218c78d6a942"},
+ {file = "pymongo-3.13.0-cp35-cp35m-macosx_10_6_intel.whl", hash = "sha256:d1ee773fb72ba024e7e3bb6ea8907fe52bccafcb5184aaced6bad995bd30ea20"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_i686.whl", hash = "sha256:28565e3dbd69fe5fe35a210067064dbb6ed5abe997079f653c19c873c3896fe6"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux1_x86_64.whl", hash = "sha256:5c1db7d366004d6c699eb08c716a63ae0a3e946d061cbebea65d7ce361950265"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e1956f3338c10308e2f99c2c9ff46ae412035cbcd7aaa76c39ccdb806854a247"},
+ {file = "pymongo-3.13.0-cp35-cp35m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:10f0fddc1d63ba3d4a4bffcc7720184c1b7efd570726ad5e2f55818da320239f"},
+ {file = "pymongo-3.13.0-cp35-cp35m-win32.whl", hash = "sha256:570ae3365b23d4fd8c669cb57613b1a90b2757e993588d3370ef90945dbeec4b"},
+ {file = "pymongo-3.13.0-cp35-cp35m-win_amd64.whl", hash = "sha256:79f777eaf3f5b2c6d81f9ef00d87837001d7063302503bbcbfdbf3e9bc27c96f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-macosx_10_6_intel.whl", hash = "sha256:d42eb29ba314adfd9c11234b4b646f61b0448bf9b00f14db4b317e6e4b947e77"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_i686.whl", hash = "sha256:e5e87c0eb774561c546f979342a8ff36ebee153c60a0b6c6b03ba989ceb9538c"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux1_x86_64.whl", hash = "sha256:0f2c5a5984599a88d087a15859860579b825098b473d8c843f1979a83d159f2e"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_aarch64.whl", hash = "sha256:59c98e86c5e861032b71e6e5b65f23e6afaacea6e82483b66f1191a5021a7b4f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_i686.whl", hash = "sha256:70b67390e27e58876853efbb87e43c85252de2515e2887f7dd901b4fa3d21973"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_ppc64le.whl", hash = "sha256:42ba8606492d76e6f9e4c7a458ed4bc712603be393259a52450345f0945da2cf"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_s390x.whl", hash = "sha256:0e5536994cf2d8488c6fd9dea71df3c4dbb3e0d2ba5e695da06d9142a29a0969"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux2014_x86_64.whl", hash = "sha256:fe8194f107f0fa3cabd14e9e809f174eca335993c1db72d1e74e0f496e7afe1f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d593d50815771f517d3ac4367ff716e3f3c78edae51d98e1e25791459f8848ff"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5136ebe8da6a1604998a8eb96be55935aa5f7129c41cc7bddc400d48e8df43be"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a424bdedfd84454d2905a861e0d4bb947cc5bd024fdeb3600c1a97d2be0f4255"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5161167b3840e9c84c80f2534ea6a099f51749d5673b662a3dd248be17c3208"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:644470442beaf969df99c4e00367a817eee05f0bba5d888f1ba6fe97b5e1c102"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:2406df90b2335371706c59b7d79e9633b81ed2a7ecd48c1faf8584552bdf2d90"},
+ {file = "pymongo-3.13.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:222591b828de10ac90064047b5d4916953f38c38b155009c4b8b5e0d33117c2b"},
+ {file = "pymongo-3.13.0-cp36-cp36m-win32.whl", hash = "sha256:7cb987b199fa223ad78eebaa9fbc183d5a5944bfe568a9d6f617316ca1c1f32f"},
+ {file = "pymongo-3.13.0-cp36-cp36m-win_amd64.whl", hash = "sha256:a6cbb73d9fc2282677e2b7a137d13da987bd0b13abd88ed27bba5534c226db06"},
+ {file = "pymongo-3.13.0-cp37-cp37m-macosx_10_6_intel.whl", hash = "sha256:b1223b826acbef07a7f5eb9bf37247b0b580119916dca9eae19d92b1290f5855"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_i686.whl", hash = "sha256:398fb86d374dc351a4abc2e24cd15e5e14b2127f6d90ce0df3fdf2adcc55ac1b"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux1_x86_64.whl", hash = "sha256:9c3d07ea19cd2856d9943dce37e75d69ecbb5baf93c3e4c82f73b6075c481292"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_aarch64.whl", hash = "sha256:2943d739715f265a2983ac43747595b6af3312d0a370614040959fd293763adf"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_i686.whl", hash = "sha256:c3b70ed82f20d18d22eafc9bda0ea656605071762f7d31f3c5afc35c59d3393b"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_ppc64le.whl", hash = "sha256:7ec2bb598847569ae34292f580842d37619eea3e546005042f485e15710180d5"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_s390x.whl", hash = "sha256:8cc37b437cba909bef06499dadd91a39c15c14225e8d8c7870020049f8a549fe"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux2014_x86_64.whl", hash = "sha256:65a063970e15a4f338f14b820561cf6cdaf2839691ac0adb2474ddff9d0b8b0b"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:02f0e1a75d3bc0e16c7e15daf9c56185642be055e425f3b34888fc6eb1b22401"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16e74b9c2aca2734c7f49f00fe68d6830a30d26df60e2ace7fe40ccb92087b94"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:24e954be35ad4537840f20bbc8d75320ae647d3cb4fab12cb8fcd2d55f408e76"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a149377d1ff766fd618500798d0d94637f66d0ae222bb6d28f41f3e15c626297"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61660710b054ae52c8fc10368e91d74719eb05554b631d7f8ca93d21d2bff2e6"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4bbc0d27dfef7689285e54f2e0a224f0c7cd9d5c46d2638fabad5500b951c92f"},
+ {file = "pymongo-3.13.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:9b2ed9c3b30f11cd4a3fbfc22167af7987b01b444215c2463265153fe7cf66d6"},
+ {file = "pymongo-3.13.0-cp37-cp37m-win32.whl", hash = "sha256:1c2c5e2b00e2fadcd590c0b2e293d71215e98ed1cb635cfca2be4998d197e534"},
+ {file = "pymongo-3.13.0-cp37-cp37m-win_amd64.whl", hash = "sha256:32eac95bbb030b2376ffd897376c6f870222a3457f01a9ce466b9057876132f8"},
+ {file = "pymongo-3.13.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a796ef39dadf9d73af05d24937644d386495e43a7d13617aa3651d836da542c8"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux1_i686.whl", hash = "sha256:b6793baf4639c72a500698a49e9250b293e17ae1faf11ac1699d8141194786fe"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux1_x86_64.whl", hash = "sha256:80d8576b04d0824f63bf803190359c0d3bcb6e7fa63fefbd4bc0ceaa7faae38c"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_aarch64.whl", hash = "sha256:db2e11507fe9cc2a722be21ccc62c1b1295398fe9724c1f14900cdc7166fc0d7"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_i686.whl", hash = "sha256:b01ce58eec5edeededf1992d2dce63fb8565e437be12d6f139d75b15614c4d08"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_ppc64le.whl", hash = "sha256:d1a19d6c5098f1f4e11430cd74621699453cbc534dd7ade9167e582f50814b19"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_s390x.whl", hash = "sha256:7219b1a726ced3bacecabef9bd114529bbb69477901373e800d7d0140baadc95"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux2014_x86_64.whl", hash = "sha256:2dae3b353a10c3767e0aa1c1492f2af388f1012b08117695ab3fd1f219e5814e"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:12721d926d43d33dd3318e58dce9b0250e8a9c6e1093fa8e09f4805193ff4b43"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6af0a4b17faf26779d5caee8542a4f2cba040cea27d3bffc476cbc6ccbd4c8ee"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:09b9d0f5a445c7e0ddcc021b09835aa6556f0166afc498f57dfdd72cdf6f02ad"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:db5b4f8ad8607a3d612da1d4c89a84e4cf5c88f98b46365820d9babe5884ba45"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:34dbf5fecf653c152edb75a35a8b15dfdc4549473484ee768aeb12c97983cead"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:34cd48df7e1fc69222f296d8f69e3957eb7c6b5aa0709d3467184880ed7538c0"},
+ {file = "pymongo-3.13.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:c8f755ff1f4ab4ca790d1d6d3229006100b301475948021b6b2757822e0d6c97"},
+ {file = "pymongo-3.13.0-cp38-cp38-win32.whl", hash = "sha256:b0746d0d4535f56bbaa63a8f6da362f330804d578e66e126b226eebe76c2bf00"},
+ {file = "pymongo-3.13.0-cp38-cp38-win_amd64.whl", hash = "sha256:8ad0515abb132f52ce9d8abd1a29681a1e65dba7b7fe13ea01e1a8db5715bf80"},
+ {file = "pymongo-3.13.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:3c5cb6c93c94df76a879bad4b89db0104b01806d17c2b803c1316ba50962b6d6"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux1_i686.whl", hash = "sha256:2e0854170813238f0c3131050c67cb1fb1ade75c93bf6cd156c1bd9a16095528"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux1_x86_64.whl", hash = "sha256:1410faa51ce835cc1234c99ec42e98ab4f3c6f50d92d86a2d4f6e11c97ee7a4e"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_aarch64.whl", hash = "sha256:d7910135f5de1c5c3578e61d6f4b087715b15e365f11d4fa51a9cee92988b2bd"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_i686.whl", hash = "sha256:028175dd8d2979a889153a2308e8e500b3df7d9e3fd1c33ca7fdeadf61cc87a2"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_ppc64le.whl", hash = "sha256:2bfc39276c0e6d07c95bd1088b5003f049e986e089509f7dbd68bb7a4b1e65ac"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_s390x.whl", hash = "sha256:4092b660ec720d44d3ca81074280dc25c7a3718df1b6c0fe9fe36ac6ed2833e4"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux2014_x86_64.whl", hash = "sha256:5bdeb71a610a7b801416268e500e716d0fe693fb10d809e17f0fb3dac5be5a34"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aa3bca8e76f5c00ed2bb4325e0e383a547d71595926d5275d7c88175aaf7435e"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7c7cab8155f430ca460a6fc7ae8a705b34f3e279a57adb5f900eb81943ec777c"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4a32f3dfcca4a4816373bdb6256c18c78974ebb3430e7da988516cd95b2bd6e4"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:30ed2788a6ec68743e2040ab1d16573d7d9f6e7333e45070ce9268cbc93d148c"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:21e61a536ffed84d10376c21c13a6ed1ebefb61989a844952547c229d6aeedf3"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0665412dce26b2318092a33bd2d2327d487c4490cfcde158d6946d39b1e28d78"},
+ {file = "pymongo-3.13.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.whl", hash = "sha256:64ed1a5ce5e5926727eb0f87c698c4d9a7a9f7b0953683a65e9ce2b7cc5f8e91"},
+ {file = "pymongo-3.13.0-cp39-cp39-win32.whl", hash = "sha256:7593cb1214185a0c5b43b96effc51ce82ddc933298ee36db7dc2bd45d61b4adc"},
+ {file = "pymongo-3.13.0-cp39-cp39-win_amd64.whl", hash = "sha256:3cfc9bc1e8b5667bc1f3dbe46d2f85b3f24ff7533893bdc1203058012db2c046"},
+ {file = "pymongo-3.13.0.tar.gz", hash = "sha256:e22d6cf5802cd09b674c307cc9e03870b8c37c503ebec3d25b86f2ce8c535dc7"},
+]
+
+[package.dependencies]
+dnspython = {version = ">=1.16.0,<1.17.0", optional = true, markers = "extra == \"srv\""}
+
+[package.extras]
+aws = ["pymongo-auth-aws (<2.0.0)"]
+encryption = ["pymongocrypt (>=1.1.0,<2.0.0)"]
+gssapi = ["pykerberos"]
+ocsp = ["certifi", "pyopenssl (>=17.2.0)", "requests (<3.0.0)", "service-identity (>=18.1.0)"]
+snappy = ["python-snappy"]
+srv = ["dnspython (>=1.16.0,<1.17.0)"]
+tls = ["ipaddress"]
+zstd = ["zstandard"]
+
+[[package]]
+name = "pyparsing"
+version = "3.1.0"
+description = "pyparsing module - Classes and methods to define and execute parsing grammars"
+category = "dev"
+optional = false
+python-versions = ">=3.6.8"
+files = [
+ {file = "pyparsing-3.1.0-py3-none-any.whl", hash = "sha256:d554a96d1a7d3ddaf7183104485bc19fd80543ad6ac5bdb6426719d766fb06c1"},
+ {file = "pyparsing-3.1.0.tar.gz", hash = "sha256:edb662d6fe322d6e990b1594b5feaeadf806803359e3d4d42f11e295e588f0ea"},
+]
+
+[package.extras]
+diagrams = ["jinja2", "railroad-diagrams"]
+
+[[package]]
+name = "pytest"
+version = "7.4.0"
+description = "pytest: simple powerful testing with Python"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "pytest-7.4.0-py3-none-any.whl", hash = "sha256:78bf16451a2eb8c7a2ea98e32dc119fd2aa758f1d5d66dbf0a59d69a3969df32"},
+ {file = "pytest-7.4.0.tar.gz", hash = "sha256:b4bf8c45bd59934ed84001ad51e11b4ee40d40a1229d2c79f9c592b0a3f6bd8a"},
+]
+
+[package.dependencies]
+colorama = {version = "*", markers = "sys_platform == \"win32\""}
+exceptiongroup = {version = ">=1.0.0rc8", markers = "python_version < \"3.11\""}
+iniconfig = "*"
+packaging = "*"
+pluggy = ">=0.12,<2.0"
+tomli = {version = ">=1.0.0", markers = "python_version < \"3.11\""}
+
+[package.extras]
+testing = ["argcomplete", "attrs (>=19.2.0)", "hypothesis (>=3.56)", "mock", "nose", "pygments (>=2.7.2)", "requests", "setuptools", "xmlschema"]
+
+[[package]]
+name = "pytest-cov"
+version = "2.12.1"
+description = "Pytest plugin for measuring coverage."
+category = "dev"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*"
+files = [
+ {file = "pytest-cov-2.12.1.tar.gz", hash = "sha256:261ceeb8c227b726249b376b8526b600f38667ee314f910353fa318caa01f4d7"},
+ {file = "pytest_cov-2.12.1-py2.py3-none-any.whl", hash = "sha256:261bb9e47e65bd099c89c3edf92972865210c36813f80ede5277dceb77a4a62a"},
+]
+
+[package.dependencies]
+coverage = ">=5.2.1"
+pytest = ">=4.6"
+toml = "*"
+
+[package.extras]
+testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtualenv"]
+
+[[package]]
+name = "pytest-httpserver"
+version = "1.0.8"
+description = "pytest-httpserver is a httpserver for pytest"
+category = "dev"
+optional = false
+python-versions = ">=3.8,<4.0"
+files = [
+ {file = "pytest_httpserver-1.0.8-py3-none-any.whl", hash = "sha256:24cd3d9f6a0b927c7bfc400d0b3fda7442721b8267ce29942bf307b190f0bb09"},
+ {file = "pytest_httpserver-1.0.8.tar.gz", hash = "sha256:e052f69bc8a9073db02484681e8e47004dd1fb3763b0ae833bd899e5895c559a"},
+]
+
+[package.dependencies]
+Werkzeug = ">=2.0.0"
+
+[[package]]
+name = "python-dateutil"
+version = "2.8.2"
+description = "Extensions to the standard Python datetime module"
+category = "main"
+optional = false
+python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7"
+files = [
+ {file = "python-dateutil-2.8.2.tar.gz", hash = "sha256:0123cacc1627ae19ddf3c27a5de5bd67ee4586fbdd6440d9748f8abb483d3e86"},
+ {file = "python_dateutil-2.8.2-py2.py3-none-any.whl", hash = "sha256:961d03dc3453ebbc59dbdea9e4e11c5651520a876d0f4db161e8674aae935da9"},
+]
+
+[package.dependencies]
+six = ">=1.5"
+
+[[package]]
+name = "python-dotenv"
+version = "1.0.0"
+description = "Read key-value pairs from a .env file and set them as environment variables"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "python-dotenv-1.0.0.tar.gz", hash = "sha256:a8df96034aae6d2d50a4ebe8216326c61c3eb64836776504fcca410e5937a3ba"},
+ {file = "python_dotenv-1.0.0-py3-none-any.whl", hash = "sha256:f5971a9226b701070a4bf2c38c89e5a3f0d64de8debda981d1db98583009122a"},
+]
+
+[package.extras]
+cli = ["click (>=5.0)"]
+
+[[package]]
+name = "pytz"
+version = "2020.5"
+description = "World timezone definitions, modern and historical"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "pytz-2020.5-py2.py3-none-any.whl", hash = "sha256:16962c5fb8db4a8f63a26646d8886e9d769b6c511543557bc84e9569fb9a9cb4"},
+ {file = "pytz-2020.5.tar.gz", hash = "sha256:180befebb1927b16f6b57101720075a984c019ac16b1b7575673bea42c6c3da5"},
+]
+
+[[package]]
+name = "pyyaml"
+version = "6.0"
+description = "YAML parser and emitter for Python"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:d4db7c7aef085872ef65a8fd7d6d09a14ae91f691dec3e87ee5ee0539d516f53"},
+ {file = "PyYAML-6.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9df7ed3b3d2e0ecfe09e14741b857df43adb5a3ddadc919a2d94fbdf78fea53c"},
+ {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77f396e6ef4c73fdc33a9157446466f1cff553d979bd00ecb64385760c6babdc"},
+ {file = "PyYAML-6.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a80a78046a72361de73f8f395f1f1e49f956c6be882eed58505a15f3e430962b"},
+ {file = "PyYAML-6.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:f84fbc98b019fef2ee9a1cb3ce93e3187a6df0b2538a651bfb890254ba9f90b5"},
+ {file = "PyYAML-6.0-cp310-cp310-win32.whl", hash = "sha256:2cd5df3de48857ed0544b34e2d40e9fac445930039f3cfe4bcc592a1f836d513"},
+ {file = "PyYAML-6.0-cp310-cp310-win_amd64.whl", hash = "sha256:daf496c58a8c52083df09b80c860005194014c3698698d1a57cbcfa182142a3a"},
+ {file = "PyYAML-6.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d4b0ba9512519522b118090257be113b9468d804b19d63c71dbcf4a48fa32358"},
+ {file = "PyYAML-6.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:81957921f441d50af23654aa6c5e5eaf9b06aba7f0a19c18a538dc7ef291c5a1"},
+ {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:afa17f5bc4d1b10afd4466fd3a44dc0e245382deca5b3c353d8b757f9e3ecb8d"},
+ {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dbad0e9d368bb989f4515da330b88a057617d16b6a8245084f1b05400f24609f"},
+ {file = "PyYAML-6.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:432557aa2c09802be39460360ddffd48156e30721f5e8d917f01d31694216782"},
+ {file = "PyYAML-6.0-cp311-cp311-win32.whl", hash = "sha256:bfaef573a63ba8923503d27530362590ff4f576c626d86a9fed95822a8255fd7"},
+ {file = "PyYAML-6.0-cp311-cp311-win_amd64.whl", hash = "sha256:01b45c0191e6d66c470b6cf1b9531a771a83c1c4208272ead47a3ae4f2f603bf"},
+ {file = "PyYAML-6.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:897b80890765f037df3403d22bab41627ca8811ae55e9a722fd0392850ec4d86"},
+ {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50602afada6d6cbfad699b0c7bb50d5ccffa7e46a3d738092afddc1f9758427f"},
+ {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:48c346915c114f5fdb3ead70312bd042a953a8ce5c7106d5bfb1a5254e47da92"},
+ {file = "PyYAML-6.0-cp36-cp36m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:98c4d36e99714e55cfbaaee6dd5badbc9a1ec339ebfc3b1f52e293aee6bb71a4"},
+ {file = "PyYAML-6.0-cp36-cp36m-win32.whl", hash = "sha256:0283c35a6a9fbf047493e3a0ce8d79ef5030852c51e9d911a27badfde0605293"},
+ {file = "PyYAML-6.0-cp36-cp36m-win_amd64.whl", hash = "sha256:07751360502caac1c067a8132d150cf3d61339af5691fe9e87803040dbc5db57"},
+ {file = "PyYAML-6.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:819b3830a1543db06c4d4b865e70ded25be52a2e0631ccd2f6a47a2822f2fd7c"},
+ {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:473f9edb243cb1935ab5a084eb238d842fb8f404ed2193a915d1784b5a6b5fc0"},
+ {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0ce82d761c532fe4ec3f87fc45688bdd3a4c1dc5e0b4a19814b9009a29baefd4"},
+ {file = "PyYAML-6.0-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:231710d57adfd809ef5d34183b8ed1eeae3f76459c18fb4a0b373ad56bedcdd9"},
+ {file = "PyYAML-6.0-cp37-cp37m-win32.whl", hash = "sha256:c5687b8d43cf58545ade1fe3e055f70eac7a5a1a0bf42824308d868289a95737"},
+ {file = "PyYAML-6.0-cp37-cp37m-win_amd64.whl", hash = "sha256:d15a181d1ecd0d4270dc32edb46f7cb7733c7c508857278d3d378d14d606db2d"},
+ {file = "PyYAML-6.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0b4624f379dab24d3725ffde76559cff63d9ec94e1736b556dacdfebe5ab6d4b"},
+ {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:213c60cd50106436cc818accf5baa1aba61c0189ff610f64f4a3e8c6726218ba"},
+ {file = "PyYAML-6.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9fa600030013c4de8165339db93d182b9431076eb98eb40ee068700c9c813e34"},
+ {file = "PyYAML-6.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:277a0ef2981ca40581a47093e9e2d13b3f1fbbeffae064c1d21bfceba2030287"},
+ {file = "PyYAML-6.0-cp38-cp38-win32.whl", hash = "sha256:d4eccecf9adf6fbcc6861a38015c2a64f38b9d94838ac1810a9023a0609e1b78"},
+ {file = "PyYAML-6.0-cp38-cp38-win_amd64.whl", hash = "sha256:1e4747bc279b4f613a09eb64bba2ba602d8a6664c6ce6396a4d0cd413a50ce07"},
+ {file = "PyYAML-6.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:055d937d65826939cb044fc8c9b08889e8c743fdc6a32b33e2390f66013e449b"},
+ {file = "PyYAML-6.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:e61ceaab6f49fb8bdfaa0f92c4b57bcfbea54c09277b1b4f7ac376bfb7a7c174"},
+ {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d67d839ede4ed1b28a4e8909735fc992a923cdb84e618544973d7dfc71540803"},
+ {file = "PyYAML-6.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba8c411ef271aa037d7357a2bc8f9ee8b58b9965831d9e51baf703280dc73d3"},
+ {file = "PyYAML-6.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:40527857252b61eacd1d9af500c3337ba8deb8fc298940291486c465c8b46ec0"},
+ {file = "PyYAML-6.0-cp39-cp39-win32.whl", hash = "sha256:b5b9eccad747aabaaffbc6064800670f0c297e52c12754eb1d976c57e4f74dcb"},
+ {file = "PyYAML-6.0-cp39-cp39-win_amd64.whl", hash = "sha256:b3d267842bf12586ba6c734f89d1f5b871df0273157918b0ccefa29deb05c21c"},
+ {file = "PyYAML-6.0.tar.gz", hash = "sha256:68fb519c14306fec9720a2a5b45bc9f0c8d1b9c72adf45c37baedfcd949c35a2"},
+]
+
+[[package]]
+name = "requests"
+version = "2.31.0"
+description = "Python HTTP for Humans."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "requests-2.31.0-py3-none-any.whl", hash = "sha256:58cd2187c01e70e6e26505bca751777aa9f2ee0b7f4300988b709f44e013003f"},
+ {file = "requests-2.31.0.tar.gz", hash = "sha256:942c5a758f98d790eaed1a29cb6eefc7ffb0d1cf7af05c3d2791656dbd6ad1e1"},
+]
+
+[package.dependencies]
+certifi = ">=2017.4.17"
+charset-normalizer = ">=2,<4"
+idna = ">=2.5,<4"
+urllib3 = ">=1.21.1,<3"
+
+[package.extras]
+socks = ["PySocks (>=1.5.6,!=1.5.7)"]
+use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"]
+
+[[package]]
+name = "rich"
+version = "13.4.2"
+description = "Render rich text, tables, progress bars, syntax highlighting, markdown and more to the terminal"
+category = "dev"
+optional = false
+python-versions = ">=3.7.0"
+files = [
+ {file = "rich-13.4.2-py3-none-any.whl", hash = "sha256:8f87bc7ee54675732fa66a05ebfe489e27264caeeff3728c945d25971b6485ec"},
+ {file = "rich-13.4.2.tar.gz", hash = "sha256:d653d6bccede5844304c605d5aac802c7cf9621efd700b46c7ec2b51ea914898"},
+]
+
+[package.dependencies]
+markdown-it-py = ">=2.2.0"
+pygments = ">=2.13.0,<3.0.0"
+
+[package.extras]
+jupyter = ["ipywidgets (>=7.5.1,<9)"]
+
+[[package]]
+name = "scikit-learn"
+version = "1.3.0"
+description = "A set of python modules for machine learning and data mining"
+category = "main"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "scikit-learn-1.3.0.tar.gz", hash = "sha256:8be549886f5eda46436b6e555b0e4873b4f10aa21c07df45c4bc1735afbccd7a"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:981287869e576d42c682cf7ca96af0c6ac544ed9316328fd0d9292795c742cf5"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:436aaaae2c916ad16631142488e4c82f4296af2404f480e031d866863425d2a2"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c7e28d8fa47a0b30ae1bd7a079519dd852764e31708a7804da6cb6f8b36e3630"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ae80c08834a473d08a204d966982a62e11c976228d306a2648c575e3ead12111"},
+ {file = "scikit_learn-1.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:552fd1b6ee22900cf1780d7386a554bb96949e9a359999177cf30211e6b20df6"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:79970a6d759eb00a62266a31e2637d07d2d28446fca8079cf9afa7c07b0427f8"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:850a00b559e636b23901aabbe79b73dc604b4e4248ba9e2d6e72f95063765603"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee04835fb016e8062ee9fe9074aef9b82e430504e420bff51e3e5fffe72750ca"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9d953531f5d9f00c90c34fa3b7d7cfb43ecff4c605dac9e4255a20b114a27369"},
+ {file = "scikit_learn-1.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:151ac2bf65ccf363664a689b8beafc9e6aae36263db114b4ca06fbbbf827444a"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:6a885a9edc9c0a341cab27ec4f8a6c58b35f3d449c9d2503a6fd23e06bbd4f6a"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-macosx_12_0_arm64.whl", hash = "sha256:9877af9c6d1b15486e18a94101b742e9d0d2f343d35a634e337411ddb57783f3"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c470f53cea065ff3d588050955c492793bb50c19a92923490d18fcb637f6383a"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fd6e2d7389542eae01077a1ee0318c4fec20c66c957f45c7aac0c6eb0fe3c612"},
+ {file = "scikit_learn-1.3.0-cp38-cp38-win_amd64.whl", hash = "sha256:3a11936adbc379a6061ea32fa03338d4ca7248d86dd507c81e13af428a5bc1db"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:998d38fcec96584deee1e79cd127469b3ad6fefd1ea6c2dfc54e8db367eb396b"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:ded35e810438a527e17623ac6deae3b360134345b7c598175ab7741720d7ffa7"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0e8102d5036e28d08ab47166b48c8d5e5810704daecf3a476a4282d562be9a28"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7617164951c422747e7c32be4afa15d75ad8044f42e7d70d3e2e0429a50e6718"},
+ {file = "scikit_learn-1.3.0-cp39-cp39-win_amd64.whl", hash = "sha256:1d54fb9e6038284548072df22fd34777e434153f7ffac72c8596f2d6987110dd"},
+]
+
+[package.dependencies]
+joblib = ">=1.1.1"
+numpy = ">=1.17.3"
+scipy = ">=1.5.0"
+threadpoolctl = ">=2.0.0"
+
+[package.extras]
+benchmark = ["matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "pandas (>=1.0.5)"]
+docs = ["Pillow (>=7.1.2)", "matplotlib (>=3.1.3)", "memory-profiler (>=0.57.0)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)", "sphinx (>=6.0.0)", "sphinx-copybutton (>=0.5.2)", "sphinx-gallery (>=0.10.1)", "sphinx-prompt (>=1.3.0)", "sphinxext-opengraph (>=0.4.2)"]
+examples = ["matplotlib (>=3.1.3)", "pandas (>=1.0.5)", "plotly (>=5.14.0)", "pooch (>=1.6.0)", "scikit-image (>=0.16.2)", "seaborn (>=0.9.0)"]
+tests = ["black (>=23.3.0)", "matplotlib (>=3.1.3)", "mypy (>=1.3)", "numpydoc (>=1.2.0)", "pandas (>=1.0.5)", "pooch (>=1.6.0)", "pyamg (>=4.0.0)", "pytest (>=7.1.2)", "pytest-cov (>=2.9.0)", "ruff (>=0.0.272)", "scikit-image (>=0.16.2)"]
+
+[[package]]
+name = "scipy"
+version = "1.11.1"
+description = "Fundamental algorithms for scientific computing in Python"
+category = "main"
+optional = false
+python-versions = "<3.13,>=3.9"
+files = [
+ {file = "scipy-1.11.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:aec8c62fbe52914f9cf28d846cf0401dd80ab80788bbab909434eb336ed07c04"},
+ {file = "scipy-1.11.1-cp310-cp310-macosx_12_0_arm64.whl", hash = "sha256:3b9963798df1d8a52db41a6fc0e6fa65b1c60e85d73da27ae8bb754de4792481"},
+ {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3e8eb42db36526b130dfbc417609498a6192381abc1975b91e3eb238e0b41c1a"},
+ {file = "scipy-1.11.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:366a6a937110d80dca4f63b3f5b00cc89d36f678b2d124a01067b154e692bab1"},
+ {file = "scipy-1.11.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:08d957ca82d3535b3b9ba6c8ff355d78fe975271874e2af267cb5add5bd78625"},
+ {file = "scipy-1.11.1-cp310-cp310-win_amd64.whl", hash = "sha256:e866514bc2d660608447b6ba95c8900d591f2865c07cca0aa4f7ff3c4ca70f30"},
+ {file = "scipy-1.11.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ba94eeef3c9caa4cea7b402a35bb02a5714ee1ee77eb98aca1eed4543beb0f4c"},
+ {file = "scipy-1.11.1-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:512fdc18c65f76dadaca139348e525646d440220d8d05f6d21965b8d4466bccd"},
+ {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cce154372f0ebe88556ed06d7b196e9c2e0c13080ecb58d0f35062dc7cc28b47"},
+ {file = "scipy-1.11.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b4bb943010203465ac81efa392e4645265077b4d9e99b66cf3ed33ae12254173"},
+ {file = "scipy-1.11.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:249cfa465c379c9bb2c20123001e151ff5e29b351cbb7f9c91587260602c58d0"},
+ {file = "scipy-1.11.1-cp311-cp311-win_amd64.whl", hash = "sha256:ffb28e3fa31b9c376d0fb1f74c1f13911c8c154a760312fbee87a21eb21efe31"},
+ {file = "scipy-1.11.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:39154437654260a52871dfde852adf1b93b1d1bc5dc0ffa70068f16ec0be2624"},
+ {file = "scipy-1.11.1-cp39-cp39-macosx_12_0_arm64.whl", hash = "sha256:b588311875c58d1acd4ef17c983b9f1ab5391755a47c3d70b6bd503a45bfaf71"},
+ {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d51565560565a0307ed06fa0ec4c6f21ff094947d4844d6068ed04400c72d0c3"},
+ {file = "scipy-1.11.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b41a0f322b4eb51b078cb3441e950ad661ede490c3aca66edef66f4b37ab1877"},
+ {file = "scipy-1.11.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:396fae3f8c12ad14c5f3eb40499fd06a6fef8393a6baa352a652ecd51e74e029"},
+ {file = "scipy-1.11.1-cp39-cp39-win_amd64.whl", hash = "sha256:be8c962a821957fdde8c4044efdab7a140c13294997a407eaee777acf63cbf0c"},
+ {file = "scipy-1.11.1.tar.gz", hash = "sha256:fb5b492fa035334fd249f0973cc79ecad8b09c604b42a127a677b45a9a3d4289"},
+]
+
+[package.dependencies]
+numpy = ">=1.21.6,<1.28.0"
+
+[package.extras]
+dev = ["click", "cython-lint (>=0.12.2)", "doit (>=0.36.0)", "mypy", "pycodestyle", "pydevtool", "rich-click", "ruff", "types-psutil", "typing_extensions"]
+doc = ["jupytext", "matplotlib (>2)", "myst-nb", "numpydoc", "pooch", "pydata-sphinx-theme (==0.9.0)", "sphinx (!=4.1.0)", "sphinx-design (>=0.2.0)"]
+test = ["asv", "gmpy2", "mpmath", "pooch", "pytest", "pytest-cov", "pytest-timeout", "pytest-xdist", "scikit-umfpack", "threadpoolctl"]
+
+[[package]]
+name = "setuptools"
+version = "68.0.0"
+description = "Easily download, build, install, upgrade, and uninstall Python packages"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "setuptools-68.0.0-py3-none-any.whl", hash = "sha256:11e52c67415a381d10d6b462ced9cfb97066179f0e871399e006c4ab101fc85f"},
+ {file = "setuptools-68.0.0.tar.gz", hash = "sha256:baf1fdb41c6da4cd2eae722e135500da913332ab3f2f5c7d33af9b492acb5235"},
+]
+
+[package.extras]
+docs = ["furo", "jaraco.packaging (>=9)", "jaraco.tidelift (>=1.4)", "pygments-github-lexers (==0.0.5)", "rst.linker (>=1.9)", "sphinx (>=3.5)", "sphinx-favicon", "sphinx-hoverxref (<2)", "sphinx-inline-tabs", "sphinx-lint", "sphinx-notfound-page (==0.8.3)", "sphinx-reredirects", "sphinxcontrib-towncrier"]
+testing = ["build[virtualenv]", "filelock (>=3.4.0)", "flake8-2020", "ini2toml[lite] (>=0.9)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pip (>=19.1)", "pip-run (>=8.8)", "pytest (>=6)", "pytest-black (>=0.3.7)", "pytest-checkdocs (>=2.4)", "pytest-cov", "pytest-enabler (>=1.3)", "pytest-mypy (>=0.9.1)", "pytest-perf", "pytest-ruff", "pytest-timeout", "pytest-xdist", "tomli-w (>=1.0.0)", "virtualenv (>=13.0.0)", "wheel"]
+testing-integration = ["build[virtualenv]", "filelock (>=3.4.0)", "jaraco.envs (>=2.2)", "jaraco.path (>=3.2.0)", "pytest", "pytest-enabler", "pytest-xdist", "tomli", "virtualenv (>=13.0.0)", "wheel"]
+
+[[package]]
+name = "six"
+version = "1.16.0"
+description = "Python 2 and 3 compatibility utilities"
+category = "main"
+optional = false
+python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*"
+files = [
+ {file = "six-1.16.0-py2.py3-none-any.whl", hash = "sha256:8abb2f1d86890a2dfb989f9a77cfcfd3e47c2a354b01111771326f8aa26e0254"},
+ {file = "six-1.16.0.tar.gz", hash = "sha256:1e61c37477a1626458e36f7b1d82aa5c9b094fa4802892072e49de9c60c4c926"},
+]
+
+[[package]]
+name = "smmap"
+version = "5.0.0"
+description = "A pure Python implementation of a sliding window memory map manager"
+category = "dev"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "smmap-5.0.0-py3-none-any.whl", hash = "sha256:2aba19d6a040e78d8b09de5c57e96207b09ed71d8e55ce0959eeee6c8e190d94"},
+ {file = "smmap-5.0.0.tar.gz", hash = "sha256:c840e62059cd3be204b0c9c9f74be2c09d5648eddd4580d9314c3ecde0b30936"},
+]
+
+[[package]]
+name = "sniffio"
+version = "1.3.0"
+description = "Sniff out which async library your code is running under"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "sniffio-1.3.0-py3-none-any.whl", hash = "sha256:eecefdce1e5bbfb7ad2eeaabf7c1eeb404d7757c379bd1f7e5cce9d8bf425384"},
+ {file = "sniffio-1.3.0.tar.gz", hash = "sha256:e60305c5e5d314f5389259b7f22aaa33d8f7dee49763119234af3755c55b9101"},
+]
+
+[[package]]
+name = "sortedcontainers"
+version = "2.4.0"
+description = "Sorted Containers -- Sorted List, Sorted Dict, Sorted Set"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "sortedcontainers-2.4.0-py2.py3-none-any.whl", hash = "sha256:a163dcaede0f1c021485e957a39245190e74249897e2ae4b2aa38595db237ee0"},
+ {file = "sortedcontainers-2.4.0.tar.gz", hash = "sha256:25caa5a06cc30b6b83d11423433f65d1f9d76c4c6a0c90e3379eaa43b9bfdb88"},
+]
+
+[[package]]
+name = "soundfile"
+version = "0.12.1"
+description = "An audio library based on libsndfile, CFFI and NumPy"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "soundfile-0.12.1-py2.py3-none-any.whl", hash = "sha256:828a79c2e75abab5359f780c81dccd4953c45a2c4cd4f05ba3e233ddf984b882"},
+ {file = "soundfile-0.12.1-py2.py3-none-macosx_10_9_x86_64.whl", hash = "sha256:d922be1563ce17a69582a352a86f28ed8c9f6a8bc951df63476ffc310c064bfa"},
+ {file = "soundfile-0.12.1-py2.py3-none-macosx_11_0_arm64.whl", hash = "sha256:bceaab5c4febb11ea0554566784bcf4bc2e3977b53946dda2b12804b4fe524a8"},
+ {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
+ {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_31_x86_64.whl", hash = "sha256:074247b771a181859d2bc1f98b5ebf6d5153d2c397b86ee9e29ba602a8dfe2a6"},
+ {file = "soundfile-0.12.1-py2.py3-none-win32.whl", hash = "sha256:59dfd88c79b48f441bbf6994142a19ab1de3b9bb7c12863402c2bc621e49091a"},
+ {file = "soundfile-0.12.1-py2.py3-none-win_amd64.whl", hash = "sha256:0d86924c00b62552b650ddd28af426e3ff2d4dc2e9047dae5b3d8452e0a49a77"},
+ {file = "soundfile-0.12.1.tar.gz", hash = "sha256:e8e1017b2cf1dda767aef19d2fd9ee5ebe07e050d430f77a0a7c66ba08b8cdae"},
+]
+
+[package.dependencies]
+cffi = ">=1.0"
+
+[package.extras]
+numpy = ["numpy"]
+
+[[package]]
+name = "soxr"
+version = "0.3.5"
+description = "High quality, one-dimensional sample-rate conversion library"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "soxr-0.3.5-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:21c3aa3b2e12351b4310eea9d56cf52ec0769e6832f911ee6ba32f85b7c92baa"},
+ {file = "soxr-0.3.5-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:ac3d7abc96082ff18a31fb1d678ddc0562f0c5e6d91f1cf0024b044989f63e93"},
+ {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:145e1e9d1b873a59ce0b5aa463ccacc40cf4bb74d9d8e6cef23433c752bfecea"},
+ {file = "soxr-0.3.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4a376b3678801ffc1d0b9ae918b958be29d5884ca1b4bbeab32e29c567723bb3"},
+ {file = "soxr-0.3.5-cp310-cp310-win32.whl", hash = "sha256:907e2eb176bdefec40cc8f6015b7cef7f3d525a34219b3580b603ee696cb25c6"},
+ {file = "soxr-0.3.5-cp310-cp310-win_amd64.whl", hash = "sha256:0a6dbf9c7b7a3642916aba264c1d0b872b2e173be56204ed1895dbe381a32077"},
+ {file = "soxr-0.3.5-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:22c08a41e8eee99241fc0e9afb510f9bc7ada4a149d469b8891b596281a27db3"},
+ {file = "soxr-0.3.5-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bdacbe4ce4a1001043f1f8f0744480e294f5c5106e7861fd7033a83a869ba371"},
+ {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4b9acd5c42159eac4a90807524d9aa450d6ea0c750df94455c151165896d922e"},
+ {file = "soxr-0.3.5-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:44b5d30f4e0d98b6d0034c00b04d5571ad070ce5cf3772f93193095b01b373de"},
+ {file = "soxr-0.3.5-cp311-cp311-win32.whl", hash = "sha256:677d5f44e85fdf0fdef33cd0e6087470732dd2e08fa73286c3659814110d1183"},
+ {file = "soxr-0.3.5-cp311-cp311-win_amd64.whl", hash = "sha256:a479984dd17bf0b50fb9fd659eba54a2dc59bf6eba9c29bb3a4a79ecec7dc9a4"},
+ {file = "soxr-0.3.5-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:a2eb4f273ca14d7cfa882b234a03497d0e5dfd6f769a488a0962fe500450838c"},
+ {file = "soxr-0.3.5-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:a254c5e1adddb1204d8f327158b6c11a854908a10b5782103f38a67156108334"},
+ {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5766727dfee4d3616edd2a866a9a0d2f272c01545bed165c5a2676fbfd278723"},
+ {file = "soxr-0.3.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2578664c6f94329685d864cdcae59794121bcbd808441572b2ffd01e7adc45dd"},
+ {file = "soxr-0.3.5-cp38-cp38-win32.whl", hash = "sha256:8a6f03804f48d986610eab8ca2b52e50b495f40ec13507741cd95f00ef7c2cb6"},
+ {file = "soxr-0.3.5-cp38-cp38-win_amd64.whl", hash = "sha256:592e9393e433501769a7e36b10460f4578c8e4ec3cddeec1aaaea4688e3558ef"},
+ {file = "soxr-0.3.5-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:93adbf04f51c7a5113059395633c2647f73bf195fa820256e1dd4da78af59275"},
+ {file = "soxr-0.3.5-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:37c4ec7ce275f284b0bf9741e5e6844a211ba1a850b2bf1c6a47769cdd3d109e"},
+ {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:18d5f3151fe4a88dfc37447bc6c397072aedcf36aeffb325cc817350ac5ad78e"},
+ {file = "soxr-0.3.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:549a8358ba3b99a75588453c96aaa802e0c84d40957bdbe1f820f14f83a052ca"},
+ {file = "soxr-0.3.5-cp39-cp39-win32.whl", hash = "sha256:799df1875803dc9c4a4d3a7c285b8c1cb34b40dc39dba7ac7bac85d072f936a5"},
+ {file = "soxr-0.3.5-cp39-cp39-win_amd64.whl", hash = "sha256:4dd3f61929eb304c109f1f3b6cc8243e3a1a46d636d5bd86b5a7f50609ecd7d6"},
+ {file = "soxr-0.3.5-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:028af32bd4ce4b4c8183bb36da99e23ae954a114034d74538b4cae1bf40a0555"},
+ {file = "soxr-0.3.5-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1299e2aae4d659e222bcbbaca69a51ee99571486070ed49a393725ea6010a8e9"},
+ {file = "soxr-0.3.5-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:162f4e8b9a014c6819b4db6def2d43f7f4d97432ae33f2edfc8e5d0c97cf1cb3"},
+ {file = "soxr-0.3.5.tar.gz", hash = "sha256:b6b60f6381c98249a2f2a594e9234b647b78856c76c060597d53ed27b6efd249"},
+]
+
+[package.dependencies]
+numpy = "*"
+
+[package.extras]
+docs = ["linkify-it-py", "myst-parser", "sphinx", "sphinx-book-theme"]
+test = ["pytest"]
+
+[[package]]
+name = "starlette"
+version = "0.28.0"
+description = "The little ASGI library that shines."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "starlette-0.28.0-py3-none-any.whl", hash = "sha256:e58b9fc402c579950260fbb6d57173395c4e62804c40d3ede7e9ef1074f0c579"},
+ {file = "starlette-0.28.0.tar.gz", hash = "sha256:7bf3da5e997e796cc202cef2bd3f96a7d9b1e1943203c2fe2b42e020bc658482"},
+]
+
+[package.dependencies]
+anyio = ">=3.4.0,<5"
+typing-extensions = {version = ">=3.10.0", markers = "python_version < \"3.10\""}
+
+[package.extras]
+full = ["httpx (>=0.22.0)", "itsdangerous", "jinja2", "python-multipart", "pyyaml"]
+
+[[package]]
+name = "starlette-prometheus"
+version = "0.9.0"
+description = "Prometheus integration for Starlette"
+category = "main"
+optional = false
+python-versions = ">=3.7,<4.0"
+files = [
+ {file = "starlette-prometheus-0.9.0.tar.gz", hash = "sha256:a52fb0f1df52b44a7a677a792759337ef0ce0d59ddf3e684a7d6459a93a90e99"},
+ {file = "starlette_prometheus-0.9.0-py3-none-any.whl", hash = "sha256:b4702e4ec67dce508d28551db0e45f12f58411afdb5d1078c92ff74331915381"},
+]
+
+[package.dependencies]
+prometheus_client = ">=0.12,<0.13"
+starlette = ">=0.12.2"
+
+[[package]]
+name = "stevedore"
+version = "5.1.0"
+description = "Manage dynamic plugins for Python applications"
+category = "dev"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "stevedore-5.1.0-py3-none-any.whl", hash = "sha256:8cc040628f3cea5d7128f2e76cf486b2251a4e543c7b938f58d9a377f6694a2d"},
+ {file = "stevedore-5.1.0.tar.gz", hash = "sha256:a54534acf9b89bc7ed264807013b505bf07f74dbe4bcfa37d32bd063870b087c"},
+]
+
+[package.dependencies]
+pbr = ">=2.0.0,<2.1.0 || >2.1.0"
+
+[[package]]
+name = "threadpoolctl"
+version = "3.1.0"
+description = "threadpoolctl"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"},
+ {file = "threadpoolctl-3.1.0.tar.gz", hash = "sha256:a335baacfaa4400ae1f0d8e3a58d6674d2f8828e3716bb2802c44955ad391380"},
+]
+
+[[package]]
+name = "toml"
+version = "0.10.2"
+description = "Python Library for Tom's Obvious, Minimal Language"
+category = "dev"
+optional = false
+python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*"
+files = [
+ {file = "toml-0.10.2-py2.py3-none-any.whl", hash = "sha256:806143ae5bfb6a3c6e736a764057db0e6a0e05e338b5630894a5f779cabb4f9b"},
+ {file = "toml-0.10.2.tar.gz", hash = "sha256:b3bda1d108d5dd99f4a20d24d9c348e91c4db7ab1b749200bded2f839ccbe68f"},
+]
+
+[[package]]
+name = "tomli"
+version = "2.0.1"
+description = "A lil' TOML parser"
+category = "dev"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "tomli-2.0.1-py3-none-any.whl", hash = "sha256:939de3e7a6161af0c887ef91b7d41a53e7c5a1ca976325f429cb46ea9bc30ecc"},
+ {file = "tomli-2.0.1.tar.gz", hash = "sha256:de526c12914f0c550d15924c62d72abc48d6fe7364aa87328337a31007fe8a4f"},
+]
+
+[[package]]
+name = "tqdm"
+version = "4.65.0"
+description = "Fast, Extensible Progress Meter"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "tqdm-4.65.0-py3-none-any.whl", hash = "sha256:c4f53a17fe37e132815abceec022631be8ffe1b9381c2e6e30aa70edc99e9671"},
+ {file = "tqdm-4.65.0.tar.gz", hash = "sha256:1871fb68a86b8fb3b59ca4cdd3dcccbc7e6d613eeed31f4c332531977b89beb5"},
+]
+
+[package.dependencies]
+colorama = {version = "*", markers = "platform_system == \"Windows\""}
+
+[package.extras]
+dev = ["py-make (>=0.1.0)", "twine", "wheel"]
+notebook = ["ipywidgets (>=6)"]
+slack = ["slack-sdk"]
+telegram = ["requests"]
+
+[[package]]
+name = "types-requests"
+version = "2.31.0.1"
+description = "Typing stubs for requests"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-requests-2.31.0.1.tar.gz", hash = "sha256:3de667cffa123ce698591de0ad7db034a5317457a596eb0b4944e5a9d9e8d1ac"},
+ {file = "types_requests-2.31.0.1-py3-none-any.whl", hash = "sha256:afb06ef8f25ba83d59a1d424bd7a5a939082f94b94e90ab5e6116bd2559deaa3"},
+]
+
+[package.dependencies]
+types-urllib3 = "*"
+
+[[package]]
+name = "types-urllib3"
+version = "1.26.25.13"
+description = "Typing stubs for urllib3"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-urllib3-1.26.25.13.tar.gz", hash = "sha256:3300538c9dc11dad32eae4827ac313f5d986b8b21494801f1bf97a1ac6c03ae5"},
+ {file = "types_urllib3-1.26.25.13-py3-none-any.whl", hash = "sha256:5dbd1d2bef14efee43f5318b5d36d805a489f6600252bb53626d4bfafd95e27c"},
+]
+
+[[package]]
+name = "typing-extensions"
+version = "4.7.1"
+description = "Backported and Experimental Type Hints for Python 3.7+"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "typing_extensions-4.7.1-py3-none-any.whl", hash = "sha256:440d5dd3af93b060174bf433bccd69b0babc3b15b1a8dca43789fd7f61514b36"},
+ {file = "typing_extensions-4.7.1.tar.gz", hash = "sha256:b75ddc264f0ba5615db7ba217daeb99701ad295353c45f9e95963337ceeeffb2"},
+]
+
+[[package]]
+name = "tzdata"
+version = "2023.3"
+description = "Provider of IANA time zone data"
+category = "main"
+optional = false
+python-versions = ">=2"
+files = [
+ {file = "tzdata-2023.3-py2.py3-none-any.whl", hash = "sha256:7e65763eef3120314099b6939b5546db7adce1e7d6f2e179e3df563c70511eda"},
+ {file = "tzdata-2023.3.tar.gz", hash = "sha256:11ef1e08e54acb0d4f95bdb1be05da659673de4acbd21bf9c69e94cc5e907a3a"},
+]
+
+[[package]]
+name = "urllib3"
+version = "2.0.3"
+description = "HTTP library with thread-safe connection pooling, file post, and more."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "urllib3-2.0.3-py3-none-any.whl", hash = "sha256:48e7fafa40319d358848e1bc6809b208340fafe2096f1725d05d67443d0483d1"},
+ {file = "urllib3-2.0.3.tar.gz", hash = "sha256:bee28b5e56addb8226c96f7f13ac28cb4c301dd5ea8a6ca179c0b9835e032825"},
+]
+
+[package.extras]
+brotli = ["brotli (>=1.0.9)", "brotlicffi (>=0.8.0)"]
+secure = ["certifi", "cryptography (>=1.9)", "idna (>=2.0.0)", "pyopenssl (>=17.1.0)", "urllib3-secure-extra"]
+socks = ["pysocks (>=1.5.6,!=1.5.7,<2.0)"]
+zstd = ["zstandard (>=0.18.0)"]
+
+[[package]]
+name = "webencodings"
+version = "0.5.1"
+description = "Character encoding aliases for legacy web content"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "webencodings-0.5.1-py2.py3-none-any.whl", hash = "sha256:a0af1213f3c2226497a97e2b3aa01a7e4bee4f403f95be16fc9acd2947514a78"},
+ {file = "webencodings-0.5.1.tar.gz", hash = "sha256:b36a1c245f2d304965eb4e0a82848379241dc04b865afcc4aab16748587e1923"},
+]
+
+[[package]]
+name = "werkzeug"
+version = "2.3.6"
+description = "The comprehensive WSGI web application library."
+category = "dev"
+optional = false
+python-versions = ">=3.8"
+files = [
+ {file = "Werkzeug-2.3.6-py3-none-any.whl", hash = "sha256:935539fa1413afbb9195b24880778422ed620c0fc09670945185cce4d91a8890"},
+ {file = "Werkzeug-2.3.6.tar.gz", hash = "sha256:98c774df2f91b05550078891dee5f0eb0cb797a522c757a2452b9cee5b202330"},
+]
+
+[package.dependencies]
+MarkupSafe = ">=2.1.1"
+
+[package.extras]
+watchdog = ["watchdog (>=2.3)"]
+
+[[package]]
+name = "xxhash"
+version = "3.2.0"
+description = "Python binding for xxHash"
+category = "main"
+optional = false
+python-versions = ">=3.6"
+files = [
+ {file = "xxhash-3.2.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:af44b9e59c4b2926a4e3c7f9d29949ff42fcea28637ff6b8182e654461932be8"},
+ {file = "xxhash-3.2.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:1bdd57973e2b802ef32553d7bebf9402dac1557874dbe5c908b499ea917662cd"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b7c9aa77bbce61a5e681bd39cb6a804338474dcc90abe3c543592aa5d6c9a9b"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:11bf87dc7bb8c3b0b5e24b7b941a9a19d8c1f88120b6a03a17264086bc8bb023"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2783d41487ce6d379fdfaa7332fca5187bf7010b9bddcf20cafba923bc1dc665"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:561076ca0dcef2fbc20b2bc2765bff099e002e96041ae9dbe910a863ca6ee3ea"},
+ {file = "xxhash-3.2.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3a26eeb4625a6e61cedc8c1b39b89327c9c7e1a8c2c4d786fe3f178eb839ede6"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d93a44d0104d1b9b10de4e7aadf747f6efc1d7ec5ed0aa3f233a720725dd31bd"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:89585adc73395a10306d2e2036e50d6c4ac0cf8dd47edf914c25488871b64f6d"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:a892b4b139126a86bfdcb97cd912a2f8c4e8623869c3ef7b50871451dd7afeb0"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e998efb190653f70e0f30d92b39fc645145369a4823bee46af8ddfc244aa969d"},
+ {file = "xxhash-3.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e8ed3bd2b8bb3277710843ca63e4f5c3ee6f8f80b083be5b19a7a9905420d11e"},
+ {file = "xxhash-3.2.0-cp310-cp310-win32.whl", hash = "sha256:20181cbaed033c72cb881b2a1d13c629cd1228f113046133469c9a48cfcbcd36"},
+ {file = "xxhash-3.2.0-cp310-cp310-win_amd64.whl", hash = "sha256:a0f7a16138279d707db778a63264d1d6016ac13ffd3f1e99f54b2855d6c0d8e1"},
+ {file = "xxhash-3.2.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:5daff3fb5bfef30bc5a2cb143810d376d43461445aa17aece7210de52adbe151"},
+ {file = "xxhash-3.2.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75bb5be3c5de702a547715f320ecf5c8014aeca750ed5147ca75389bd22e7343"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:01f36b671ff55cb1d5c2f6058b799b697fd0ae4b4582bba6ed0999678068172a"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4d4519123aac73c93159eb8f61db9682393862dd669e7eae034ecd0a35eadac"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:994e4741d5ed70fc2a335a91ef79343c6b1089d7dfe6e955dd06f8ffe82bede6"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:919bc1b010aa6ff0eb918838ff73a435aed9e9a19c3202b91acecd296bf75607"},
+ {file = "xxhash-3.2.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:17b65454c5accbb079c45eca546c27c4782f5175aa320758fafac896b1549d27"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:b0c094d5e65a46dbf3fe0928ff20873a747e6abfd2ed4b675beeb2750624bc2e"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f94163ebe2d5546e6a5977e96d83621f4689c1054053428cf8d4c28b10f92f69"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:cead7c0307977a00b3f784cff676e72c147adbcada19a2e6fc2ddf54f37cf387"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:a0e1bd0260c1da35c1883321ce2707ceea07127816ab625e1226ec95177b561a"},
+ {file = "xxhash-3.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:cc8878935671490efe9275fb4190a6062b73277bd273237179b9b5a2aa436153"},
+ {file = "xxhash-3.2.0-cp311-cp311-win32.whl", hash = "sha256:a433f6162b18d52f7068175d00bd5b1563b7405f926a48d888a97b90a160c40d"},
+ {file = "xxhash-3.2.0-cp311-cp311-win_amd64.whl", hash = "sha256:a32d546a1752e4ee7805d6db57944f7224afa7428d22867006b6486e4195c1f3"},
+ {file = "xxhash-3.2.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:82daaab720866bf690b20b49de5640b0c27e3b8eea2d08aa75bdca2b0f0cfb63"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3126df6520cbdbaddd87ce74794b2b6c45dd2cf6ac2b600a374b8cdb76a2548c"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e172c1ee40507ae3b8d220f4048aaca204f203e1e4197e8e652f5c814f61d1aa"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5384f1d9f30876f5d5b618464fb19ff7ce6c0fe4c690fbaafd1c52adc3aae807"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:26cb52174a7e96a17acad27a3ca65b24713610ac479c99ac9640843822d3bebf"},
+ {file = "xxhash-3.2.0-cp36-cp36m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fbcd613a5e76b1495fc24db9c37a6b7ee5f214fd85979187ec4e032abfc12ded"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_aarch64.whl", hash = "sha256:f988daf25f31726d5b9d0be6af636ca9000898f9ea43a57eac594daea25b0948"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_i686.whl", hash = "sha256:bbc30c98ab006ab9fc47e5ed439c00f706bc9d4441ff52693b8b6fea335163e0"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_ppc64le.whl", hash = "sha256:2408d49260b0a4a7cc6ba445aebf38e073aeaf482f8e32767ca477e32ccbbf9e"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_s390x.whl", hash = "sha256:3f4152fd0bf8b03b79f2f900fd6087a66866537e94b5a11fd0fd99ef7efe5c42"},
+ {file = "xxhash-3.2.0-cp36-cp36m-musllinux_1_1_x86_64.whl", hash = "sha256:0eea848758e4823a01abdbcccb021a03c1ee4100411cbeeb7a5c36a202a0c13c"},
+ {file = "xxhash-3.2.0-cp36-cp36m-win32.whl", hash = "sha256:77709139af5123c578ab06cf999429cdb9ab211047acd0c787e098dcb3f1cb4d"},
+ {file = "xxhash-3.2.0-cp36-cp36m-win_amd64.whl", hash = "sha256:91687671fd9d484a4e201ad266d366b695a45a1f2b41be93d116ba60f1b8f3b3"},
+ {file = "xxhash-3.2.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:e4af8bc5c3fcc2192c266421c6aa2daab1a18e002cb8e66ef672030e46ae25cf"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e8be562e2ce3e481d9209b6f254c3d7c5ff920eb256aba2380d2fb5ba75d4f87"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9eba0c7c12126b12f7fcbea5513f28c950d28f33d2a227f74b50b77789e478e8"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2198c4901a0223c48f6ec0a978b60bca4f4f7229a11ca4dc96ca325dd6a29115"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:50ce82a71b22a3069c02e914bf842118a53065e2ec1c6fb54786e03608ab89cc"},
+ {file = "xxhash-3.2.0-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b5019fb33711c30e54e4e57ae0ca70af9d35b589d385ac04acd6954452fa73bb"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:0d54ac023eef7e3ac9f0b8841ae8a376b933043bc2ad428121346c6fa61c491c"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c55fa832fc3fe64e0d29da5dc9b50ba66ca93312107cec2709300ea3d3bab5c7"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:f4ce006215497993ae77c612c1883ca4f3973899573ce0c52fee91f0d39c4561"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1afb9b9d27fd675b436cb110c15979976d92d761ad6e66799b83756402f3a974"},
+ {file = "xxhash-3.2.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:baa99cebf95c1885db21e119395f222a706a2bb75a545f0672880a442137725e"},
+ {file = "xxhash-3.2.0-cp37-cp37m-win32.whl", hash = "sha256:75aa692936942ccb2e8fd6a386c81c61630ac1b6d6e921698122db8a930579c3"},
+ {file = "xxhash-3.2.0-cp37-cp37m-win_amd64.whl", hash = "sha256:0a2cdfb5cae9fafb9f7b65fd52ecd60cf7d72c13bb2591ea59aaefa03d5a8827"},
+ {file = "xxhash-3.2.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:3a68d1e8a390b660d94b9360ae5baa8c21a101bd9c4790a8b30781bada9f1fc6"},
+ {file = "xxhash-3.2.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:ce7c3ce28f94302df95eaea7c9c1e2c974b6d15d78a0c82142a97939d7b6c082"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0dcb419bf7b0bc77d366e5005c25682249c5521a63fd36c51f584bd91bb13bd5"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ae521ed9287f86aac979eeac43af762f03d9d9797b2272185fb9ddd810391216"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0d16775094423088ffa357d09fbbb9ab48d2fb721d42c0856b801c86f616eec"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fe454aeab348c42f56d6f7434ff758a3ef90787ac81b9ad5a363cd61b90a1b0b"},
+ {file = "xxhash-3.2.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:052fd0efdd5525c2dbc61bebb423d92aa619c4905bba605afbf1e985a562a231"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:02badf3754e2133de254a4688798c4d80f0060635087abcb461415cb3eb82115"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:66b8a90b28c13c2aae7a71b32638ceb14cefc2a1c8cf23d8d50dfb64dfac7aaf"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:649cdf19df175925ad87289ead6f760cd840730ee85abc5eb43be326a0a24d97"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:4b948a03f89f5c72d69d40975af8af241111f0643228796558dc1cae8f5560b0"},
+ {file = "xxhash-3.2.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:49f51fab7b762da7c2cee0a3d575184d3b9be5e2f64f26cae2dd286258ac9b3c"},
+ {file = "xxhash-3.2.0-cp38-cp38-win32.whl", hash = "sha256:1a42994f0d42b55514785356722d9031f064fd34e495b3a589e96db68ee0179d"},
+ {file = "xxhash-3.2.0-cp38-cp38-win_amd64.whl", hash = "sha256:0a6d58ba5865475e53d6c2c4fa6a62e2721e7875e146e2681e5337a6948f12e7"},
+ {file = "xxhash-3.2.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:aabdbc082030f8df613e2d2ea1f974e7ad36a539bdfc40d36f34e55c7e4b8e94"},
+ {file = "xxhash-3.2.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:498843b66b9ca416e9d03037e5875c8d0c0ab9037527e22df3b39aa5163214cd"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a910b1193cd90af17228f5d6069816646df0148f14f53eefa6b2b11a1dedfcd0"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bb6d8ce31dc25faf4da92991320e211fa7f42de010ef51937b1dc565a4926501"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:883dc3d3942620f4c7dbc3fd6162f50a67f050b714e47da77444e3bcea7d91cc"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59dc8bfacf89b8f5be54d55bc3b4bd6d74d0c5320c8a63d2538ac7df5b96f1d5"},
+ {file = "xxhash-3.2.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:61e6aa1d30c2af692aa88c4dd48709426e8b37bff6a574ee2de677579c34a3d6"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:314ec0bd21f0ee8d30f2bd82ed3759314bd317ddbbd8555668f3d20ab7a8899a"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:dad638cde3a5357ad3163b80b3127df61fb5b5e34e9e05a87697144400ba03c7"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:eaa3ea15025b56076d806b248948612289b093e8dcda8d013776b3848dffff15"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:7deae3a312feb5c17c97cbf18129f83cbd3f1f9ec25b0f50e2bd9697befb22e7"},
+ {file = "xxhash-3.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:add774341c09853b1612c64a526032d95ab1683053325403e1afbe3ad2f374c5"},
+ {file = "xxhash-3.2.0-cp39-cp39-win32.whl", hash = "sha256:9b94749130ef3119375c599bfce82142c2500ef9ed3280089157ee37662a7137"},
+ {file = "xxhash-3.2.0-cp39-cp39-win_amd64.whl", hash = "sha256:e57d94a1552af67f67b27db5dba0b03783ea69d5ca2af2f40e098f0ba3ce3f5f"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-macosx_10_9_x86_64.whl", hash = "sha256:92fd765591c83e5c5f409b33eac1d3266c03d3d11c71a7dbade36d5cdee4fbc0"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8970f6a411a9839a02b23b7e90bbbba4a6de52ace009274998566dc43f36ca18"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c5f3e33fe6cbab481727f9aeb136a213aed7e33cd1ca27bd75e916ffacc18411"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:368265392cb696dd53907e2328b5a8c1bee81cf2142d0cc743caf1c1047abb36"},
+ {file = "xxhash-3.2.0-pp37-pypy37_pp73-win_amd64.whl", hash = "sha256:3b1f3c6d67fa9f49c4ff6b25ce0e7143bab88a5bc0f4116dd290c92337d0ecc7"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-macosx_10_9_x86_64.whl", hash = "sha256:c5e8db6e1ee7267b7c412ad0afd5863bf7a95286b8333a5958c8097c69f94cf5"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:761df3c7e2c5270088b691c5a8121004f84318177da1ca1db64222ec83c44871"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d2d15a707e7f689531eb4134eccb0f8bf3844bb8255ad50823aa39708d9e6755"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e6b2ba4ff53dd5f57d728095e3def7375eb19c90621ce3b41b256de84ec61cfd"},
+ {file = "xxhash-3.2.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:61b0bcf946fdfd8ab5f09179dc2b5c74d1ef47cedfc6ed0ec01fdf0ee8682dd3"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-macosx_10_9_x86_64.whl", hash = "sha256:f7b79f0f302396d8e0d444826ceb3d07b61977793886ebae04e82796c02e42dc"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e0773cd5c438ffcd5dbff91cdd503574f88a4b960e70cedeb67736583a17a918"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4ec1f57127879b419a2c8d2db9d9978eb26c61ae17e5972197830430ae78d25b"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3d4b15c00e807b1d3d0b612338c814739dec310b80fb069bd732b98ddc709ad7"},
+ {file = "xxhash-3.2.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:9d3f686e3d1c8900c5459eee02b60c7399e20ec5c6402364068a343c83a61d90"},
+ {file = "xxhash-3.2.0.tar.gz", hash = "sha256:1afd47af8955c5db730f630ad53ae798cf7fae0acb64cebb3cf94d35c47dd088"},
+]
+
+[[package]]
+name = "yarl"
+version = "1.9.2"
+description = "Yet another URL library"
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:8c2ad583743d16ddbdf6bb14b5cd76bf43b0d0006e918809d5d4ddf7bde8dd82"},
+ {file = "yarl-1.9.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:82aa6264b36c50acfb2424ad5ca537a2060ab6de158a5bd2a72a032cc75b9eb8"},
+ {file = "yarl-1.9.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c0c77533b5ed4bcc38e943178ccae29b9bcf48ffd1063f5821192f23a1bd27b9"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ee4afac41415d52d53a9833ebae7e32b344be72835bbb589018c9e938045a560"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9bf345c3a4f5ba7f766430f97f9cc1320786f19584acc7086491f45524a551ac"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2a96c19c52ff442a808c105901d0bdfd2e28575b3d5f82e2f5fd67e20dc5f4ea"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:891c0e3ec5ec881541f6c5113d8df0315ce5440e244a716b95f2525b7b9f3608"},
+ {file = "yarl-1.9.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c3a53ba34a636a256d767c086ceb111358876e1fb6b50dfc4d3f4951d40133d5"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:566185e8ebc0898b11f8026447eacd02e46226716229cea8db37496c8cdd26e0"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:2b0738fb871812722a0ac2154be1f049c6223b9f6f22eec352996b69775b36d4"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_ppc64le.whl", hash = "sha256:32f1d071b3f362c80f1a7d322bfd7b2d11e33d2adf395cc1dd4df36c9c243095"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_s390x.whl", hash = "sha256:e9fdc7ac0d42bc3ea78818557fab03af6181e076a2944f43c38684b4b6bed8e3"},
+ {file = "yarl-1.9.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:56ff08ab5df8429901ebdc5d15941b59f6253393cb5da07b4170beefcf1b2528"},
+ {file = "yarl-1.9.2-cp310-cp310-win32.whl", hash = "sha256:8ea48e0a2f931064469bdabca50c2f578b565fc446f302a79ba6cc0ee7f384d3"},
+ {file = "yarl-1.9.2-cp310-cp310-win_amd64.whl", hash = "sha256:50f33040f3836e912ed16d212f6cc1efb3231a8a60526a407aeb66c1c1956dde"},
+ {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:646d663eb2232d7909e6601f1a9107e66f9791f290a1b3dc7057818fe44fc2b6"},
+ {file = "yarl-1.9.2-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aff634b15beff8902d1f918012fc2a42e0dbae6f469fce134c8a0dc51ca423bb"},
+ {file = "yarl-1.9.2-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:a83503934c6273806aed765035716216cc9ab4e0364f7f066227e1aaea90b8d0"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b25322201585c69abc7b0e89e72790469f7dad90d26754717f3310bfe30331c2"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:22a94666751778629f1ec4280b08eb11815783c63f52092a5953faf73be24191"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8ec53a0ea2a80c5cd1ab397925f94bff59222aa3cf9c6da938ce05c9ec20428d"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:159d81f22d7a43e6eabc36d7194cb53f2f15f498dbbfa8edc8a3239350f59fe7"},
+ {file = "yarl-1.9.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:832b7e711027c114d79dffb92576acd1bd2decc467dec60e1cac96912602d0e6"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:95d2ecefbcf4e744ea952d073c6922e72ee650ffc79028eb1e320e732898d7e8"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:d4e2c6d555e77b37288eaf45b8f60f0737c9efa3452c6c44626a5455aeb250b9"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_ppc64le.whl", hash = "sha256:783185c75c12a017cc345015ea359cc801c3b29a2966c2655cd12b233bf5a2be"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_s390x.whl", hash = "sha256:b8cc1863402472f16c600e3e93d542b7e7542a540f95c30afd472e8e549fc3f7"},
+ {file = "yarl-1.9.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:822b30a0f22e588b32d3120f6d41e4ed021806418b4c9f0bc3048b8c8cb3f92a"},
+ {file = "yarl-1.9.2-cp311-cp311-win32.whl", hash = "sha256:a60347f234c2212a9f0361955007fcf4033a75bf600a33c88a0a8e91af77c0e8"},
+ {file = "yarl-1.9.2-cp311-cp311-win_amd64.whl", hash = "sha256:be6b3fdec5c62f2a67cb3f8c6dbf56bbf3f61c0f046f84645cd1ca73532ea051"},
+ {file = "yarl-1.9.2-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:38a3928ae37558bc1b559f67410df446d1fbfa87318b124bf5032c31e3447b74"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ac9bb4c5ce3975aeac288cfcb5061ce60e0d14d92209e780c93954076c7c4367"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3da8a678ca8b96c8606bbb8bfacd99a12ad5dd288bc6f7979baddd62f71c63ef"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:13414591ff516e04fcdee8dc051c13fd3db13b673c7a4cb1350e6b2ad9639ad3"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf74d08542c3a9ea97bb8f343d4fcbd4d8f91bba5ec9d5d7f792dbe727f88938"},
+ {file = "yarl-1.9.2-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6e7221580dc1db478464cfeef9b03b95c5852cc22894e418562997df0d074ccc"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:494053246b119b041960ddcd20fd76224149cfea8ed8777b687358727911dd33"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:52a25809fcbecfc63ac9ba0c0fb586f90837f5425edfd1ec9f3372b119585e45"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_ppc64le.whl", hash = "sha256:e65610c5792870d45d7b68c677681376fcf9cc1c289f23e8e8b39c1485384185"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_s390x.whl", hash = "sha256:1b1bba902cba32cdec51fca038fd53f8beee88b77efc373968d1ed021024cc04"},
+ {file = "yarl-1.9.2-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:662e6016409828ee910f5d9602a2729a8a57d74b163c89a837de3fea050c7582"},
+ {file = "yarl-1.9.2-cp37-cp37m-win32.whl", hash = "sha256:f364d3480bffd3aa566e886587eaca7c8c04d74f6e8933f3f2c996b7f09bee1b"},
+ {file = "yarl-1.9.2-cp37-cp37m-win_amd64.whl", hash = "sha256:6a5883464143ab3ae9ba68daae8e7c5c95b969462bbe42e2464d60e7e2698368"},
+ {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:5610f80cf43b6202e2c33ba3ec2ee0a2884f8f423c8f4f62906731d876ef4fac"},
+ {file = "yarl-1.9.2-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:b9a4e67ad7b646cd6f0938c7ebfd60e481b7410f574c560e455e938d2da8e0f4"},
+ {file = "yarl-1.9.2-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:83fcc480d7549ccebe9415d96d9263e2d4226798c37ebd18c930fce43dfb9574"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5fcd436ea16fee7d4207c045b1e340020e58a2597301cfbcfdbe5abd2356c2fb"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:84e0b1599334b1e1478db01b756e55937d4614f8654311eb26012091be109d59"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:3458a24e4ea3fd8930e934c129b676c27452e4ebda80fbe47b56d8c6c7a63a9e"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:838162460b3a08987546e881a2bfa573960bb559dfa739e7800ceeec92e64417"},
+ {file = "yarl-1.9.2-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f4e2d08f07a3d7d3e12549052eb5ad3eab1c349c53ac51c209a0e5991bbada78"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:de119f56f3c5f0e2fb4dee508531a32b069a5f2c6e827b272d1e0ff5ac040333"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:149ddea5abf329752ea5051b61bd6c1d979e13fbf122d3a1f9f0c8be6cb6f63c"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_ppc64le.whl", hash = "sha256:674ca19cbee4a82c9f54e0d1eee28116e63bc6fd1e96c43031d11cbab8b2afd5"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_s390x.whl", hash = "sha256:9b3152f2f5677b997ae6c804b73da05a39daa6a9e85a512e0e6823d81cdad7cc"},
+ {file = "yarl-1.9.2-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:5415d5a4b080dc9612b1b63cba008db84e908b95848369aa1da3686ae27b6d2b"},
+ {file = "yarl-1.9.2-cp38-cp38-win32.whl", hash = "sha256:f7a3d8146575e08c29ed1cd287068e6d02f1c7bdff8970db96683b9591b86ee7"},
+ {file = "yarl-1.9.2-cp38-cp38-win_amd64.whl", hash = "sha256:63c48f6cef34e6319a74c727376e95626f84ea091f92c0250a98e53e62c77c72"},
+ {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:75df5ef94c3fdc393c6b19d80e6ef1ecc9ae2f4263c09cacb178d871c02a5ba9"},
+ {file = "yarl-1.9.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:c027a6e96ef77d401d8d5a5c8d6bc478e8042f1e448272e8d9752cb0aff8b5c8"},
+ {file = "yarl-1.9.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f3b078dbe227f79be488ffcfc7a9edb3409d018e0952cf13f15fd6512847f3f7"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:59723a029760079b7d991a401386390c4be5bfec1e7dd83e25a6a0881859e716"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b03917871bf859a81ccb180c9a2e6c1e04d2f6a51d953e6a5cdd70c93d4e5a2a"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c1012fa63eb6c032f3ce5d2171c267992ae0c00b9e164efe4d73db818465fac3"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a74dcbfe780e62f4b5a062714576f16c2f3493a0394e555ab141bf0d746bb955"},
+ {file = "yarl-1.9.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8c56986609b057b4839968ba901944af91b8e92f1725d1a2d77cbac6972b9ed1"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:2c315df3293cd521033533d242d15eab26583360b58f7ee5d9565f15fee1bef4"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:b7232f8dfbd225d57340e441d8caf8652a6acd06b389ea2d3222b8bc89cbfca6"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_ppc64le.whl", hash = "sha256:53338749febd28935d55b41bf0bcc79d634881195a39f6b2f767870b72514caf"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_s390x.whl", hash = "sha256:066c163aec9d3d073dc9ffe5dd3ad05069bcb03fcaab8d221290ba99f9f69ee3"},
+ {file = "yarl-1.9.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:8288d7cd28f8119b07dd49b7230d6b4562f9b61ee9a4ab02221060d21136be80"},
+ {file = "yarl-1.9.2-cp39-cp39-win32.whl", hash = "sha256:b124e2a6d223b65ba8768d5706d103280914d61f5cae3afbc50fc3dfcc016623"},
+ {file = "yarl-1.9.2-cp39-cp39-win_amd64.whl", hash = "sha256:61016e7d582bc46a5378ffdd02cd0314fb8ba52f40f9cf4d9a5e7dbef88dee18"},
+ {file = "yarl-1.9.2.tar.gz", hash = "sha256:04ab9d4b9f587c06d801c2abfe9317b77cdf996c65a90d5e84ecc45010823571"},
+]
+
+[package.dependencies]
+idna = ">=2.0"
+multidict = ">=4.0"
+
+[metadata]
+lock-version = "2.0"
+python-versions = "3.9.15"
+content-hash = "9f4fbf1058f6ec31e437c39104086dac3ead6792fb7b266eca4f5594ea5cd24c"
diff --git a/libs/libapi/poetry.toml b/libs/libapi/poetry.toml
new file mode 100644
index 00000000..5fcef8cd
--- /dev/null
+++ b/libs/libapi/poetry.toml
@@ -0,0 +1,3 @@
+[virtualenvs]
+in-project = true
+prefer-active-python = true
diff --git a/libs/libapi/pyproject.toml b/libs/libapi/pyproject.toml
new file mode 100644
index 00000000..adc11924
--- /dev/null
+++ b/libs/libapi/pyproject.toml
@@ -0,0 +1,54 @@
+[tool.poetry]
+authors = ["The HuggingFace Authors."]
+description = "Library for the API services"
+name = "libapi"
+version = "0.1.0"
+license = "Apache-2.0"
+
+[tool.poetry.dependencies]
+cryptography = "^41.0.1"
+environs = "^9.5.0"
+libcommon = {path = "../../libs/libcommon", develop = true}
+orjson = "^3.8.6"
+pyjwt = { extras = ["crypto"], version = "^2.6.0" }
+python = "3.9.15"
+requests = "^2.28.2"
+starlette = "^0.28.0"
+starlette-prometheus = "^0.9.0"
+
+[tool.poetry.group.dev.dependencies]
+bandit = "^1.7.4"
+black = "^22.12.0"
+flake8 = "^3.9.2"
+isort = "^5.12.0"
+mypy = "^1.0.0"
+pip-audit = "^2.5.4"
+pytest = "^7.2.1"
+pytest-cov = "^2.12.1"
+pytest-httpserver = "^1.0.6"
+types-requests = "^2.28.11"
+
+[build-system]
+build-backend = "poetry.core.masonry.api"
+requires = ["poetry-core>=1.0.0"]
+
+[tool.pytest.ini_options]
+filterwarnings = ["ignore::DeprecationWarning"]
+[tool.coverage.run]
+source = ["libapi"]
+
+[tool.isort]
+profile = "black"
+
+[tool.black]
+line-length = 119
+preview = true
+
+[tool.mypy]
+strict = true
+
+[[tool.mypy.overrides]]
+module = [
+ "prometheus_client.*",
+]
+ignore_missing_imports = true
diff --git a/libs/libapi/src/libapi/__init__.py b/libs/libapi/src/libapi/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/libs/libapi/src/libapi/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/services/api/src/api/authentication.py b/libs/libapi/src/libapi/authentication.py
similarity index 98%
rename from services/api/src/api/authentication.py
rename to libs/libapi/src/libapi/authentication.py
index 1308657d..b423a07a 100644
--- a/services/api/src/api/authentication.py
+++ b/libs/libapi/src/libapi/authentication.py
@@ -13,2 +13 @@ from starlette.requests import Request
-from api.jwt_token import is_jwt_valid
-from api.utils import (
+from libapi.exceptions import (
@@ -18,0 +18 @@ from api.utils import (
+from libapi.jwt_token import is_jwt_valid
diff --git a/libs/libapi/src/libapi/config.py b/libs/libapi/src/libapi/config.py
new file mode 100644
index 00000000..a533253a
--- /dev/null
+++ b/libs/libapi/src/libapi/config.py
@@ -0,0 +1,67 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from dataclasses import dataclass
+from typing import Optional
+
+from environs import Env
+
+API_UVICORN_HOSTNAME = "localhost"
+API_UVICORN_NUM_WORKERS = 2
+API_UVICORN_PORT = 8000
+
+
+@dataclass(frozen=True)
+class UvicornConfig:
+ hostname: str = API_UVICORN_HOSTNAME
+ num_workers: int = API_UVICORN_NUM_WORKERS
+ port: int = API_UVICORN_PORT
+
+ @classmethod
+ def from_env(cls) -> "UvicornConfig":
+ env = Env(expand_vars=True)
+ with env.prefixed("API_UVICORN_"):
+ return cls(
+ hostname=env.str(name="HOSTNAME", default=API_UVICORN_HOSTNAME),
+ num_workers=env.int(name="NUM_WORKERS", default=API_UVICORN_NUM_WORKERS),
+ port=env.int(name="PORT", default=API_UVICORN_PORT),
+ )
+
+
+API_EXTERNAL_AUTH_URL = None
+API_HF_AUTH_PATH = "/api/datasets/%s/auth-check"
+API_HF_JWT_PUBLIC_KEY_URL = None
+API_HF_JWT_ALGORITHM = "EdDSA"
+API_HF_TIMEOUT_SECONDS = 0.2
+API_HF_WEBHOOK_SECRET = None
+API_MAX_AGE_LONG = 120 # 2 minutes
+API_MAX_AGE_SHORT = 10 # 10 seconds
+
+
+@dataclass(frozen=True)
+class ApiConfig:
+ external_auth_url: Optional[str] = API_EXTERNAL_AUTH_URL # not documented
+ hf_auth_path: str = API_HF_AUTH_PATH
+ hf_jwt_public_key_url: Optional[str] = API_HF_JWT_PUBLIC_KEY_URL
+ hf_jwt_algorithm: Optional[str] = API_HF_JWT_ALGORITHM
+ hf_timeout_seconds: Optional[float] = API_HF_TIMEOUT_SECONDS
+ hf_webhook_secret: Optional[str] = API_HF_WEBHOOK_SECRET
+ max_age_long: int = API_MAX_AGE_LONG
+ max_age_short: int = API_MAX_AGE_SHORT
+
+ @classmethod
+ def from_env(cls, hf_endpoint: str) -> "ApiConfig":
+ env = Env(expand_vars=True)
+ with env.prefixed("API_"):
+ hf_auth_path = env.str(name="HF_AUTH_PATH", default=API_HF_AUTH_PATH)
+ external_auth_url = None if hf_auth_path is None else f"{hf_endpoint}{hf_auth_path}"
+ return cls(
+ external_auth_url=external_auth_url,
+ hf_auth_path=hf_auth_path,
+ hf_jwt_public_key_url=env.str(name="HF_JWT_PUBLIC_KEY_URL", default=API_HF_JWT_PUBLIC_KEY_URL),
+ hf_jwt_algorithm=env.str(name="HF_JWT_ALGORITHM", default=API_HF_JWT_ALGORITHM),
+ hf_timeout_seconds=env.float(name="HF_TIMEOUT_SECONDS", default=API_HF_TIMEOUT_SECONDS),
+ hf_webhook_secret=env.str(name="HF_WEBHOOK_SECRET", default=API_HF_WEBHOOK_SECRET),
+ max_age_long=env.int(name="MAX_AGE_LONG", default=API_MAX_AGE_LONG),
+ max_age_short=env.int(name="MAX_AGE_SHORT", default=API_MAX_AGE_SHORT),
+ )
diff --git a/libs/libapi/src/libapi/exceptions.py b/libs/libapi/src/libapi/exceptions.py
new file mode 100644
index 00000000..b867f9de
--- /dev/null
+++ b/libs/libapi/src/libapi/exceptions.py
@@ -0,0 +1,106 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+from http import HTTPStatus
+from typing import Literal, Optional
+
+from libcommon.exceptions import CustomError
+
+ApiErrorCode = Literal[
+ "AuthCheckHubRequestError",
+ "ExternalAuthenticatedError",
+ "ExternalUnauthenticatedError",
+ "InvalidParameter",
+ "JWKError",
+ "MissingProcessingStepsError",
+ "MissingRequiredParameter",
+ "ResponseNotFound",
+ "ResponseNotReady",
+ "UnexpectedApiError",
+]
+
+
+class ApiError(CustomError):
+ """Base class for exceptions raised by an API service."""
+
+ def __init__(
+ self,
+ message: str,
+ status_code: HTTPStatus,
+ code: ApiErrorCode,
+ cause: Optional[BaseException] = None,
+ disclose_cause: bool = False,
+ ):
+ super().__init__(
+ message=message, status_code=status_code, code=code, cause=cause, disclose_cause=disclose_cause
+ )
+
+
+class AuthCheckHubRequestError(ApiError):
+ """Raised when the external authentication check failed or timed out."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(
+ message, HTTPStatus.INTERNAL_SERVER_ERROR, "AuthCheckHubRequestError", cause=cause, disclose_cause=False
+ )
+
+
+class ExternalAuthenticatedError(ApiError):
+ """Raised when the external authentication check failed while the user was authenticated.
+
+ Even if the external authentication server returns 403 in that case, we return 404 because
+ we don't know if the dataset exist or not. It's also coherent with how the Hugging Face Hub works."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.NOT_FOUND, "ExternalAuthenticatedError")
+
+
+class ExternalUnauthenticatedError(ApiError):
+ """Raised when the external authentication check failed while the user was unauthenticated."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.UNAUTHORIZED, "ExternalUnauthenticatedError")
+
+
+class InvalidParameterError(ApiError):
+ """Raised when a parameter has an invalid value."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "InvalidParameter")
+
+
+class JWKError(ApiError):
+ """Raised when the JWT key (JWK) could not be fetched or parsed."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "JWKError", cause=cause, disclose_cause=False)
+
+
+class MissingRequiredParameterError(ApiError):
+ """Raised when a required parameter is missing."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "MissingRequiredParameter")
+
+
+class ResponseNotFoundError(ApiError):
+ """Raised when the response has not been found."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.NOT_FOUND, "ResponseNotFound")
+
+
+class ResponseNotReadyError(ApiError):
+ """Raised when the response has not been processed yet."""
+
+ def __init__(self, message: str):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseNotReady")
+
+
+class UnexpectedApiError(ApiError):
+ """Raised when the server raised an unexpected error."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ logging.error(message, exc_info=cause)
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "UnexpectedApiError", cause)
diff --git a/services/api/src/api/jwt_token.py b/libs/libapi/src/libapi/jwt_token.py
similarity index 99%
rename from services/api/src/api/jwt_token.py
rename to libs/libapi/src/libapi/jwt_token.py
index 26d7ae1f..1fcf3707 100644
--- a/services/api/src/api/jwt_token.py
+++ b/libs/libapi/src/libapi/jwt_token.py
@@ -31 +31 @@ from jwt.algorithms import (
-from api.utils import JWKError
+from libapi.exceptions import JWKError
diff --git a/libs/libapi/src/libapi/py.typed b/libs/libapi/src/libapi/py.typed
new file mode 100644
index 00000000..e69de29b
diff --git a/libs/libapi/src/libapi/routes/__init__.py b/libs/libapi/src/libapi/routes/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/libs/libapi/src/libapi/routes/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/services/api/src/api/routes/healthcheck.py b/libs/libapi/src/libapi/routes/healthcheck.py
similarity index 100%
rename from services/api/src/api/routes/healthcheck.py
rename to libs/libapi/src/libapi/routes/healthcheck.py
diff --git a/services/api/src/api/routes/metrics.py b/libs/libapi/src/libapi/routes/metrics.py
similarity index 94%
rename from services/api/src/api/routes/metrics.py
rename to libs/libapi/src/libapi/routes/metrics.py
index ef637fe4..92ffc9bc 100644
--- a/services/api/src/api/routes/metrics.py
+++ b/libs/libapi/src/libapi/routes/metrics.py
@@ -11 +11 @@ from starlette.responses import Response
-from api.utils import Endpoint
+from libapi.utils import Endpoint
diff --git a/services/api/src/api/utils.py b/libs/libapi/src/libapi/utils.py
similarity index 54%
rename from services/api/src/api/utils.py
rename to libs/libapi/src/libapi/utils.py
index f6defeee..2f2480de 100644
--- a/services/api/src/api/utils.py
+++ b/libs/libapi/src/libapi/utils.py
@@ -4 +3,0 @@
-import logging
@@ -6 +5 @@ from http import HTTPStatus
-from typing import Any, Callable, Coroutine, List, Literal, Optional
+from typing import Any, Callable, Coroutine, List, Optional
@@ -16,95 +15 @@ from starlette.responses import JSONResponse, Response
-ApiErrorCode = Literal[
- "AuthCheckHubRequestError",
- "ExternalAuthenticatedError",
- "ExternalUnauthenticatedError",
- "InvalidParameter",
- "JWKError",
- "MissingRequiredParameter",
- "MissingProcessingStepsError",
- "ResponseNotFound",
- "ResponseNotReady",
- "UnexpectedError", # also in libcommon.exceptions
-]
-
-
-class ApiCustomError(CustomError):
- """Base class for exceptions in this module."""
-
- def __init__(
- self,
- message: str,
- status_code: HTTPStatus,
- code: ApiErrorCode,
- cause: Optional[BaseException] = None,
- disclose_cause: bool = False,
- ):
- super().__init__(message, status_code, str(code), cause, disclose_cause)
-
-
-class MissingRequiredParameterError(ApiCustomError):
- """Raised when a required parameter is missing."""
-
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "MissingRequiredParameter")
-
-
-class InvalidParameterError(ApiCustomError):
- """Raised when a parameter has an invalid value."""
-
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.UNPROCESSABLE_ENTITY, "InvalidParameter")
-
-
-class ResponseNotReadyError(ApiCustomError):
- """Raised when the response has not been processed yet."""
-
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "ResponseNotReady")
-
-
-class ResponseNotFoundError(ApiCustomError):
- """Raised when the response has not been found."""
-
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.NOT_FOUND, "ResponseNotFound")
-
-
-class UnexpectedError(ApiCustomError):
- """Raised when the server raised an unexpected error."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "UnexpectedError", cause)
- logging.error(message, exc_info=cause)
-
-
-class ExternalUnauthenticatedError(ApiCustomError):
- """Raised when the external authentication check failed while the user was unauthenticated."""
-
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.UNAUTHORIZED, "ExternalUnauthenticatedError")
-
-
-class ExternalAuthenticatedError(ApiCustomError):
- """Raised when the external authentication check failed while the user was authenticated.
-
- Even if the external authentication server returns 403 in that case, we return 404 because
- we don't know if the dataset exist or not. It's also coherent with how the Hugging Face Hub works."""
-
- def __init__(self, message: str):
- super().__init__(message, HTTPStatus.NOT_FOUND, "ExternalAuthenticatedError")
-
-
-class JWKError(ApiCustomError):
- """Raised when the JWT key (JWK) could not be fetched or parsed."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "JWKError", cause=cause, disclose_cause=False)
-
-
-class AuthCheckHubRequestError(ApiCustomError):
- """Raised when the external authentication check failed or timed out."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(
- message, HTTPStatus.INTERNAL_SERVER_ERROR, "AuthCheckHubRequestError", cause=cause, disclose_cause=False
- )
+from libapi.exceptions import ResponseNotFoundError, ResponseNotReadyError
@@ -154 +59 @@ def get_json_error_response(
-def get_json_api_error_response(error: ApiCustomError, max_age: int = 0, revision: Optional[str] = None) -> Response:
+def get_json_api_error_response(error: CustomError, max_age: int = 0, revision: Optional[str] = None) -> Response:
diff --git a/libs/libapi/tests/__init__.py b/libs/libapi/tests/__init__.py
new file mode 100644
index 00000000..fa0c50f2
--- /dev/null
+++ b/libs/libapi/tests/__init__.py
@@ -0,0 +1,2 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
diff --git a/libs/libapi/tests/conftest.py b/libs/libapi/tests/conftest.py
new file mode 100644
index 00000000..dcd155a6
--- /dev/null
+++ b/libs/libapi/tests/conftest.py
@@ -0,0 +1,36 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+from pytest import fixture
+
+from libapi.config import ApiConfig
+
+
+@fixture(scope="session")
+def hostname() -> str:
+ return "localhost"
+
+
+@fixture(scope="session")
+def port() -> str:
+ return "8888"
+
+
+@fixture(scope="session")
+def httpserver_listen_address(hostname: str, port: int) -> tuple[str, int]:
+ return (hostname, port)
+
+
+@fixture(scope="session")
+def hf_endpoint(hostname: str, port: int) -> str:
+ return f"http://{hostname}:{port}"
+
+
+@fixture(scope="session")
+def api_config(hf_endpoint: str) -> ApiConfig:
+ return ApiConfig.from_env(hf_endpoint=hf_endpoint)
+
+
+@fixture(scope="session")
+def hf_auth_path(api_config: ApiConfig) -> str:
+ return api_config.hf_auth_path
diff --git a/services/api/tests/test_authentication.py b/libs/libapi/tests/test_authentication.py
similarity index 78%
rename from services/api/tests/test_authentication.py
rename to libs/libapi/tests/test_authentication.py
index ab32c1dd..e0451dd6 100644
--- a/services/api/tests/test_authentication.py
+++ b/libs/libapi/tests/test_authentication.py
@@ -3,0 +4 @@
+import datetime
@@ -17,2 +18,2 @@ from werkzeug.wrappers import Response as WerkzeugResponse
-from api.authentication import auth_check
-from api.utils import (
+from libapi.authentication import auth_check
+from libapi.exceptions import (
@@ -24,8 +25,31 @@ from api.utils import (
-from .test_jwt_token import (
- algorithm_rs256,
- dataset_ok,
- payload_ok,
- private_key,
- public_key,
-)
-from .utils import auth_callback
+private_key = """-----BEGIN RSA PRIVATE KEY-----
+MIIBOQIBAAJAZTmplhS/Jd73ycVut7TglMObheQqXM7RZYlwazLU4wpfIVIwOh9I
+sCZGSgLyFq42KWIikKLEs/yqx3pRGfq+rwIDAQABAkAMyF9WCICq86Eu5bO5lynV
+H26AVfPTjHp87AI6R00C7p9n8hO/DhHaHpc3InOSsXsw9d2hmz37jwwBFiwMHMMh
+AiEAtbttHlIO+yO29oXw4P6+yO11lMy1UpT1sPVTnR9TXbUCIQCOl7Zuyy2ZY9ZW
+pDhW91x/14uXjnLXPypgY9bcfggJUwIhAJQG1LzrzjQWRUPMmgZKuhBkC3BmxhM8
+LlwzmCXVjEw5AiA7JnAFEb9+q82T71d3q/DxD0bWvb6hz5ASoBfXK2jGBQIgbaQp
+h4Tk6UJuj1xgKNs75Pk3pG2tj8AQiuBk3l62vRU=
+-----END RSA PRIVATE KEY-----"""
+public_key = """-----BEGIN PUBLIC KEY-----
+MFswDQYJKoZIhvcNAQEBBQADSgAwRwJAZTmplhS/Jd73ycVut7TglMObheQqXM7R
+ZYlwazLU4wpfIVIwOh9IsCZGSgLyFq42KWIikKLEs/yqx3pRGfq+rwIDAQAB
+-----END PUBLIC KEY-----"""
+
+dataset_ok = "dataset"
+exp_ok = datetime.datetime.now().timestamp() + 1000
+read_ok = True
+sub_ok = f"datasets/{dataset_ok}"
+payload_ok = {"sub": sub_ok, "read": read_ok, "exp": exp_ok}
+algorithm_rs256 = "RS256"
+
+
+def auth_callback(request: WerkzeugRequest) -> WerkzeugResponse:
+ # return 401 if a cookie has been provided, 404 if a token has been provided,
+ # and 200 if none has been provided
+ #
+ # caveat: the returned status codes don't simulate the reality
+ # they're just used to check every case
+ return WerkzeugResponse(
+ status=401 if request.headers.get("cookie") else 404 if request.headers.get("authorization") else 200
+ )
diff --git a/services/api/tests/test_jwt_token.py b/libs/libapi/tests/test_jwt_token.py
similarity index 98%
rename from services/api/tests/test_jwt_token.py
rename to libs/libapi/tests/test_jwt_token.py
index 4a9f6835..0060cfb0 100644
--- a/services/api/tests/test_jwt_token.py
+++ b/libs/libapi/tests/test_jwt_token.py
@@ -11 +11 @@ import pytest
-from api.jwt_token import is_jwt_valid, parse_jwt_public_key
+from libapi.jwt_token import is_jwt_valid, parse_jwt_public_key
diff --git a/services/api/Dockerfile b/services/api/Dockerfile
index 9a42a9d5..437a2367 100644
--- a/services/api/Dockerfile
+++ b/services/api/Dockerfile
@@ -28,0 +29 @@ COPY libs/libcommon ./libs/libcommon
+COPY libs/libapi ./libs/libapi
diff --git a/services/api/README.md b/services/api/README.md
index 3a147851..45239b60 100644
--- a/services/api/README.md
+++ b/services/api/README.md
@@ -3 +3 @@
-> API to get the first rows of 🤗 datasets
+> API on 🤗 datasets
@@ -11,21 +11 @@ The worker can be configured using environment variables. They are grouped by sc
-Set environment variables to configure the application (`API_` prefix):
-
-- `API_HF_AUTH_PATH`: the path of the external authentication service, on the hub (see `HF_ENDPOINT`). The string must contain `%s` which will be replaced with the dataset name. The external authentication service must return 200, 401, 403 or 404. Defaults to "/api/datasets/%s/auth-check".
-- `API_HF_JWT_PUBLIC_KEY_URL`: the URL where the "Hub JWT public key" is published. The "Hub JWT public key" must be in JWK format. It helps to decode a JWT sent by the Hugging Face Hub, for example, to bypass the external authentication check (JWT in the 'X-Api-Key' header). If not set, the JWT are ignored. Defaults to empty.
-- `API_HF_JWT_ALGORITHM`: the algorithm used to encode the JWT. Defaults to `"EdDSA"`.
-- `API_HF_TIMEOUT_SECONDS`: the timeout in seconds for the requests to the Hugging Face Hub. Defaults to `0.2` (200 ms).
-- `API_HF_WEBHOOK_SECRET`: a shared secret sent by the Hub in the "X-Webhook-Secret" header of POST requests sent to /webhook, to authenticate the originator and bypass some validation of the content (avoiding roundtrip to the Hub). If not set, all the validations are done. Defaults to empty.
-- `API_MAX_AGE_LONG`: number of seconds to set in the `max-age` header on data endpoints. Defaults to `120` (2 minutes).
-- `API_MAX_AGE_SHORT`: number of seconds to set in the `max-age` header on technical endpoints. Defaults to `10` (10 seconds).
-
-### Uvicorn
-
-The following environment variables are used to configure the Uvicorn server (`API_UVICORN_` prefix):
-
-- `API_UVICORN_HOSTNAME`: the hostname. Defaults to `"localhost"`.
-- `API_UVICORN_NUM_WORKERS`: the number of uvicorn workers. Defaults to `2`.
-- `API_UVICORN_PORT`: the port. Defaults to `8000`.
-
-### Prometheus
-
-- `PROMETHEUS_MULTIPROC_DIR`: the directory where the uvicorn workers share their prometheus metrics. See https://github.com/prometheus/client_python#multiprocess-mode-eg-gunicorn. Defaults to empty, in which case every worker manages its own metrics, and the /metrics endpoint returns the metrics of a random worker.
+See [../../libs/libapi/README.md](../../libs/libapi/README.md) for more information about the API configuration.
diff --git a/services/api/dev.Dockerfile b/services/api/dev.Dockerfile
index 7af4bd54..09196638 100644
--- a/services/api/dev.Dockerfile
+++ b/services/api/dev.Dockerfile
@@ -27,0 +28,2 @@ COPY libs/libcommon/pyproject.toml ./libs/libcommon/pyproject.toml
+COPY libs/libapi/poetry.lock ./libs/libapi/poetry.lock
+COPY libs/libapi/pyproject.toml ./libs/libapi/pyproject.toml
@@ -34,0 +37,3 @@ RUN mkdir ./libs/libcommon/src && mkdir ./libs/libcommon/src/libcommon && touch
+# Initialize an empty libapi
+# Mapping a volume to ./libs/libapi/src is required when running this image.
+RUN mkdir ./libs/libapi/src && mkdir ./libs/libapi/src/libapi && touch ./libs/libapi/src/libapi/__init__.py
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index d688e743..c60d328a 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
+# This file is automatically @generated by Poetry and should not be changed by hand.
@@ -1133,0 +1134,24 @@ test = ["pytest (>=7.2)", "pytest-cov (>=4.0)"]
+[[package]]
+name = "libapi"
+version = "0.1.0"
+description = "Library for the API services"
+category = "main"
+optional = false
+python-versions = "3.9.15"
+files = []
+develop = true
+
+[package.dependencies]
+cryptography = "^41.0.1"
+environs = "^9.5.0"
+libcommon = {path = "../../libs/libcommon", develop = true}
+orjson = "^3.8.6"
+pyjwt = {version = "^2.6.0", extras = ["crypto"]}
+requests = "^2.28.2"
+starlette = "^0.28.0"
+starlette-prometheus = "^0.9.0"
+
+[package.source]
+type = "directory"
+url = "../../libs/libapi"
+
@@ -1263,60 +1286,0 @@ testing = ["coverage", "pytest", "pytest-cov", "pytest-regressions"]
-[[package]]
-name = "markupsafe"
-version = "2.1.3"
-description = "Safely add untrusted strings to HTML/XML markup."
-category = "dev"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cd0f502fe016460680cd20aaa5a76d241d6f35a1c3350c474bac1273803893fa"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:e09031c87a1e51556fdcb46e5bd4f59dfb743061cf93c4d6831bf894f125eb57"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:68e78619a61ecf91e76aa3e6e8e33fc4894a2bebe93410754bd28fce0a8a4f9f"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:65c1a9bcdadc6c28eecee2c119465aebff8f7a584dd719facdd9e825ec61ab52"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:525808b8019e36eb524b8c68acdd63a37e75714eac50e988180b169d64480a00"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:962f82a3086483f5e5f64dbad880d31038b698494799b097bc59c2edf392fce6"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:aa7bd130efab1c280bed0f45501b7c8795f9fdbeb02e965371bbef3523627779"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:c9c804664ebe8f83a211cace637506669e7890fec1b4195b505c214e50dd4eb7"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-win32.whl", hash = "sha256:10bbfe99883db80bdbaff2dcf681dfc6533a614f700da1287707e8a5d78a8431"},
- {file = "MarkupSafe-2.1.3-cp310-cp310-win_amd64.whl", hash = "sha256:1577735524cdad32f9f694208aa75e422adba74f1baee7551620e43a3141f559"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:ad9e82fb8f09ade1c3e1b996a6337afac2b8b9e365f926f5a61aacc71adc5b3c"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:3c0fae6c3be832a0a0473ac912810b2877c8cb9d76ca48de1ed31e1c68386575"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b076b6226fb84157e3f7c971a47ff3a679d837cf338547532ab866c57930dbee"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bfce63a9e7834b12b87c64d6b155fdd9b3b96191b6bd334bf37db7ff1fe457f2"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:338ae27d6b8745585f87218a3f23f1512dbf52c26c28e322dbe54bcede54ccb9"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:e4dd52d80b8c83fdce44e12478ad2e85c64ea965e75d66dbeafb0a3e77308fcc"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:df0be2b576a7abbf737b1575f048c23fb1d769f267ec4358296f31c2479db8f9"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:5bbe06f8eeafd38e5d0a4894ffec89378b6c6a625ff57e3028921f8ff59318ac"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-win32.whl", hash = "sha256:dd15ff04ffd7e05ffcb7fe79f1b98041b8ea30ae9234aed2a9168b5797c3effb"},
- {file = "MarkupSafe-2.1.3-cp311-cp311-win_amd64.whl", hash = "sha256:134da1eca9ec0ae528110ccc9e48041e0828d79f24121a1a146161103c76e686"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8e254ae696c88d98da6555f5ace2279cf7cd5b3f52be2b5cf97feafe883b58d2"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cb0932dc158471523c9637e807d9bfb93e06a95cbf010f1a38b98623b929ef2b"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9402b03f1a1b4dc4c19845e5c749e3ab82d5078d16a2a4c2cd2df62d57bb0707"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ca379055a47383d02a5400cb0d110cef0a776fc644cda797db0c5696cfd7e18e"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_aarch64.whl", hash = "sha256:b7ff0f54cb4ff66dd38bebd335a38e2c22c41a8ee45aa608efc890ac3e3931bc"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c011a4149cfbcf9f03994ec2edffcb8b1dc2d2aede7ca243746df97a5d41ce48"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:56d9f2ecac662ca1611d183feb03a3fa4406469dafe241673d521dd5ae92a155"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-win32.whl", hash = "sha256:8758846a7e80910096950b67071243da3e5a20ed2546e6392603c096778d48e0"},
- {file = "MarkupSafe-2.1.3-cp37-cp37m-win_amd64.whl", hash = "sha256:787003c0ddb00500e49a10f2844fac87aa6ce977b90b0feaaf9de23c22508b24"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:2ef12179d3a291be237280175b542c07a36e7f60718296278d8593d21ca937d4"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:2c1b19b3aaacc6e57b7e25710ff571c24d6c3613a45e905b1fde04d691b98ee0"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8afafd99945ead6e075b973fefa56379c5b5c53fd8937dad92c662da5d8fd5ee"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8c41976a29d078bb235fea9b2ecd3da465df42a562910f9022f1a03107bd02be"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d080e0a5eb2529460b30190fcfcc4199bd7f827663f858a226a81bc27beaa97e"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:69c0f17e9f5a7afdf2cc9fb2d1ce6aabdb3bafb7f38017c0b77862bcec2bbad8"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:504b320cd4b7eff6f968eddf81127112db685e81f7e36e75f9f84f0df46041c3"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:42de32b22b6b804f42c5d98be4f7e5e977ecdd9ee9b660fda1a3edf03b11792d"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-win32.whl", hash = "sha256:ceb01949af7121f9fc39f7d27f91be8546f3fb112c608bc4029aef0bab86a2a5"},
- {file = "MarkupSafe-2.1.3-cp38-cp38-win_amd64.whl", hash = "sha256:1b40069d487e7edb2676d3fbdb2b0829ffa2cd63a2ec26c4938b2d34391b4ecc"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:8023faf4e01efadfa183e863fefde0046de576c6f14659e8782065bcece22198"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:6b2b56950d93e41f33b4223ead100ea0fe11f8e6ee5f641eb753ce4b77a7042b"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9dcdfd0eaf283af041973bff14a2e143b8bd64e069f4c383416ecd79a81aab58"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:05fb21170423db021895e1ea1e1f3ab3adb85d1c2333cbc2310f2a26bc77272e"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:282c2cb35b5b673bbcadb33a585408104df04f14b2d9b01d4c345a3b92861c2c"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:ab4a0df41e7c16a1392727727e7998a467472d0ad65f3ad5e6e765015df08636"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:7ef3cb2ebbf91e330e3bb937efada0edd9003683db6b57bb108c4001f37a02ea"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:0a4e4a1aff6c7ac4cd55792abf96c915634c2b97e3cc1c7129578aa68ebd754e"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-win32.whl", hash = "sha256:fec21693218efe39aa7f8599346e90c705afa52c5b31ae019b2e57e8f6542bb2"},
- {file = "MarkupSafe-2.1.3-cp39-cp39-win_amd64.whl", hash = "sha256:3fd4abcb888d15a94f32b75d8fd18ee162ca0c064f35b11134be77050296d6ba"},
- {file = "MarkupSafe-2.1.3.tar.gz", hash = "sha256:af598ed32d6ae86f1b747b82783958b1a4ab8f617b06fe68795c7f026abbdcad"},
-]
-
@@ -2528,15 +2491,0 @@ testing = ["fields", "hunter", "process-tests", "pytest-xdist", "six", "virtuale
-[[package]]
-name = "pytest-httpserver"
-version = "1.0.8"
-description = "pytest-httpserver is a httpserver for pytest"
-category = "dev"
-optional = false
-python-versions = ">=3.8,<4.0"
-files = [
- {file = "pytest_httpserver-1.0.8-py3-none-any.whl", hash = "sha256:24cd3d9f6a0b927c7bfc400d0b3fda7442721b8267ce29942bf307b190f0bb09"},
- {file = "pytest_httpserver-1.0.8.tar.gz", hash = "sha256:e052f69bc8a9073db02484681e8e47004dd1fb3763b0ae833bd899e5895c559a"},
-]
-
-[package.dependencies]
-Werkzeug = ">=2.0.0"
-
@@ -2851,0 +2801 @@ files = [
+ {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
@@ -2911 +2861 @@ name = "starlette"
-version = "0.27.0"
+version = "0.28.0"
@@ -2917,2 +2867,2 @@ files = [
- {file = "starlette-0.27.0-py3-none-any.whl", hash = "sha256:918416370e846586541235ccd38a474c08b80443ed31c578a418e2209b3eef91"},
- {file = "starlette-0.27.0.tar.gz", hash = "sha256:6a6b0d042acb8d469a01eba54e9cda6cbd24ac602c4cd016723117d6a7e73b75"},
+ {file = "starlette-0.28.0-py3-none-any.whl", hash = "sha256:e58b9fc402c579950260fbb6d57173395c4e62804c40d3ede7e9ef1074f0c579"},
+ {file = "starlette-0.28.0.tar.gz", hash = "sha256:7bf3da5e997e796cc202cef2bd3f96a7d9b1e1943203c2fe2b42e020bc658482"},
@@ -3195,18 +3144,0 @@ files = [
-[[package]]
-name = "werkzeug"
-version = "2.3.4"
-description = "The comprehensive WSGI web application library."
-category = "dev"
-optional = false
-python-versions = ">=3.8"
-files = [
- {file = "Werkzeug-2.3.4-py3-none-any.whl", hash = "sha256:48e5e61472fee0ddee27ebad085614ebedb7af41e88f687aaf881afb723a162f"},
- {file = "Werkzeug-2.3.4.tar.gz", hash = "sha256:1d5a58e0377d1fe39d061a5de4469e414e78ccb1e1e59c0f5ad6fa1c36c52b76"},
-]
-
-[package.dependencies]
-MarkupSafe = ">=2.1.1"
-
-[package.extras]
-watchdog = ["watchdog (>=2.3)"]
-
@@ -3412 +3344 @@ python-versions = "3.9.15"
-content-hash = "781ea6fdf22b868dd68ea1f37735c621a9064e2ee616b3e2538fa8924697b1b2"
+content-hash = "c46efd608bb9fde12d1d6aec4385e7b26da6d08011d84e287d5a1663f3c93217"
diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml
index 7fc589ed..46450593 100644
--- a/services/api/pyproject.toml
+++ b/services/api/pyproject.toml
@@ -9 +8,0 @@ license = "Apache-2.0"
-cryptography = "^41.0.1"
@@ -12,2 +11 @@ jsonschema = "^4.17.0"
-libcommon = {path = "../../libs/libcommon", develop = true}
-pyjwt = { extras = ["crypto"], version = "^2.6.0" }
+libapi = {path = "../../libs/libapi", develop = true}
@@ -16 +13,0 @@ python = "3.9.15"
-requests = "^2.28.2"
@@ -18,2 +14,0 @@ soundfile = ">=0.12.1"
-starlette = "^0.27.0"
-starlette-prometheus = "^0.9.0"
@@ -34 +28,0 @@ pytest-cov = "^2.12.1"
-pytest-httpserver = "^1.0.6"
@@ -36 +29,0 @@ types-psutil = "^5.9.5"
-types-requests = "^2.28.11"
@@ -37,0 +31 @@ types-jsonschema = "^4.17.0.4"
+types-requests = "^2.28.11"
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index a124b5bb..8f6992b2 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -4,0 +5,4 @@ import uvicorn
+from libapi.config import UvicornConfig
+from libapi.jwt_token import fetch_jwt_public_key
+from libapi.routes.healthcheck import healthcheck_endpoint
+from libapi.routes.metrics import create_metrics_endpoint
@@ -16,2 +20 @@ from starlette_prometheus import PrometheusMiddleware
-from api.config import AppConfig, EndpointConfig, UvicornConfig
-from api.jwt_token import fetch_jwt_public_key
+from api.config import AppConfig, EndpointConfig
@@ -19,2 +21,0 @@ from api.routes.endpoint import EndpointsDefinition, create_endpoint
-from api.routes.healthcheck import healthcheck_endpoint
-from api.routes.metrics import create_metrics_endpoint
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index e8518a51..28a62fc7 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -5 +5 @@ from dataclasses import dataclass, field
-from typing import List, Mapping, Optional
+from typing import List, Mapping
@@ -7 +7 @@ from typing import List, Mapping, Optional
-from environs import Env
+from libapi.config import ApiConfig
@@ -19,60 +18,0 @@ from libcommon.processing_graph import InputType
-API_UVICORN_HOSTNAME = "localhost"
-API_UVICORN_NUM_WORKERS = 2
-API_UVICORN_PORT = 8000
-
-
-@dataclass(frozen=True)
-class UvicornConfig:
- hostname: str = API_UVICORN_HOSTNAME
- num_workers: int = API_UVICORN_NUM_WORKERS
- port: int = API_UVICORN_PORT
-
- @classmethod
- def from_env(cls) -> "UvicornConfig":
- env = Env(expand_vars=True)
- with env.prefixed("API_UVICORN_"):
- return cls(
- hostname=env.str(name="HOSTNAME", default=API_UVICORN_HOSTNAME),
- num_workers=env.int(name="NUM_WORKERS", default=API_UVICORN_NUM_WORKERS),
- port=env.int(name="PORT", default=API_UVICORN_PORT),
- )
-
-
-API_EXTERNAL_AUTH_URL = None
-API_HF_AUTH_PATH = "/api/datasets/%s/auth-check"
-API_HF_JWT_PUBLIC_KEY_URL = None
-API_HF_JWT_ALGORITHM = "EdDSA"
-API_HF_TIMEOUT_SECONDS = 0.2
-API_HF_WEBHOOK_SECRET = None
-API_MAX_AGE_LONG = 120 # 2 minutes
-API_MAX_AGE_SHORT = 10 # 10 seconds
-
-
-@dataclass(frozen=True)
-class ApiConfig:
- external_auth_url: Optional[str] = API_EXTERNAL_AUTH_URL # not documented
- hf_auth_path: str = API_HF_AUTH_PATH
- hf_jwt_public_key_url: Optional[str] = API_HF_JWT_PUBLIC_KEY_URL
- hf_jwt_algorithm: Optional[str] = API_HF_JWT_ALGORITHM
- hf_timeout_seconds: Optional[float] = API_HF_TIMEOUT_SECONDS
- hf_webhook_secret: Optional[str] = API_HF_WEBHOOK_SECRET
- max_age_long: int = API_MAX_AGE_LONG
- max_age_short: int = API_MAX_AGE_SHORT
-
- @classmethod
- def from_env(cls, common_config: CommonConfig) -> "ApiConfig":
- env = Env(expand_vars=True)
- with env.prefixed("API_"):
- hf_auth_path = env.str(name="HF_AUTH_PATH", default=API_HF_AUTH_PATH)
- external_auth_url = None if hf_auth_path is None else f"{common_config.hf_endpoint}{hf_auth_path}"
- return cls(
- external_auth_url=external_auth_url,
- hf_auth_path=hf_auth_path,
- hf_jwt_public_key_url=env.str(name="HF_JWT_PUBLIC_KEY_URL", default=API_HF_JWT_PUBLIC_KEY_URL),
- hf_jwt_algorithm=env.str(name="HF_JWT_ALGORITHM", default=API_HF_JWT_ALGORITHM),
- hf_timeout_seconds=env.float(name="HF_TIMEOUT_SECONDS", default=API_HF_TIMEOUT_SECONDS),
- hf_webhook_secret=env.str(name="HF_WEBHOOK_SECRET", default=API_HF_WEBHOOK_SECRET),
- max_age_long=env.int(name="MAX_AGE_LONG", default=API_MAX_AGE_LONG),
- max_age_short=env.int(name="MAX_AGE_SHORT", default=API_MAX_AGE_SHORT),
- )
-
@@ -101 +41 @@ class AppConfig:
- api=ApiConfig.from_env(common_config=common_config),
+ api=ApiConfig.from_env(hf_endpoint=common_config.hf_endpoint),
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 71f6f00e..e01d22b4 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -8,0 +9,14 @@ from typing import List, Mapping, Optional, Tuple, TypedDict
+from libapi.authentication import auth_check
+from libapi.exceptions import (
+ ApiError,
+ MissingRequiredParameterError,
+ UnexpectedApiError,
+)
+from libapi.utils import (
+ Endpoint,
+ are_valid_parameters,
+ get_json_api_error_response,
+ get_json_error_response,
+ get_json_ok_response,
+ try_backfill_dataset,
+)
@@ -19 +32,0 @@ from starlette.responses import Response
-from api.authentication import auth_check
@@ -21,11 +33,0 @@ from api.config import EndpointConfig
-from api.utils import (
- ApiCustomError,
- Endpoint,
- MissingRequiredParameterError,
- UnexpectedError,
- are_valid_parameters,
- get_json_api_error_response,
- get_json_error_response,
- get_json_ok_response,
- try_backfill_dataset,
-)
@@ -319 +321 @@ def create_endpoint(
- error = e if isinstance(e, ApiCustomError) else UnexpectedError("Unexpected error.", e)
+ error = e if isinstance(e, ApiError) else UnexpectedApiError("Unexpected error.", e)
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index 18430cb7..2529260f 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -13,0 +14,14 @@ from fsspec.implementations.http import HTTPFileSystem
+from libapi.authentication import auth_check
+from libapi.exceptions import (
+ ApiError,
+ InvalidParameterError,
+ MissingRequiredParameterError,
+ UnexpectedApiError,
+)
+from libapi.utils import (
+ Endpoint,
+ are_valid_parameters,
+ get_json_api_error_response,
+ get_json_ok_response,
+ try_backfill_dataset,
+)
@@ -26,13 +39,0 @@ from starlette.responses import Response
-from api.authentication import auth_check
-from api.utils import (
- ApiCustomError,
- Endpoint,
- InvalidParameterError,
- MissingRequiredParameterError,
- UnexpectedError,
- are_valid_parameters,
- get_json_api_error_response,
- get_json_ok_response,
- try_backfill_dataset,
-)
-
@@ -381 +382 @@ def create_rows_endpoint(
- error = e if isinstance(e, ApiCustomError) else UnexpectedError("Unexpected error.", e)
+ error = e if isinstance(e, ApiError) else UnexpectedApiError("Unexpected error.", e)
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 029eaa97..10517797 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -7,0 +8,2 @@ from typing import List, Set, TypedDict
+from libapi.exceptions import UnexpectedApiError
+from libapi.utils import Endpoint, get_json_api_error_response, get_json_ok_response
@@ -14,7 +15,0 @@ from starlette.responses import Response
-from api.utils import (
- Endpoint,
- UnexpectedError,
- get_json_api_error_response,
- get_json_ok_response,
-)
-
@@ -79 +74,3 @@ def create_valid_endpoint(
- return get_json_api_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age_short)
+ return get_json_api_error_response(
+ UnexpectedApiError("Unexpected error.", e), max_age=max_age_short
+ )
diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py
index 96f2c810..5c43bef8 100644
--- a/services/api/src/api/routes/webhook.py
+++ b/services/api/src/api/routes/webhook.py
@@ -7,0 +8 @@ from jsonschema import ValidationError, validate
+from libapi.utils import Endpoint, get_response
@@ -16,2 +16,0 @@ from starlette.responses import Response
-from api.utils import Endpoint, get_response
-
diff --git a/services/api/tests/conftest.py b/services/api/tests/conftest.py
index 603e16bf..cbc59c89 100644
--- a/services/api/tests/conftest.py
+++ b/services/api/tests/conftest.py
@@ -6,0 +7 @@ from typing import Iterator
+from libapi.config import UvicornConfig
@@ -14 +15 @@ from pytest import MonkeyPatch, fixture
-from api.config import AppConfig, EndpointConfig, UvicornConfig
+from api.config import AppConfig, EndpointConfig
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index e691f92e..37f8f820 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -6,0 +7 @@ from unittest.mock import patch
+from libapi.exceptions import ResponseNotReadyError
@@ -15 +15,0 @@ from api.routes.endpoint import EndpointsDefinition, get_cache_entry_from_steps
-from api.utils import ResponseNotReadyError
diff --git a/services/api/tests/routes/test_rows.py b/services/api/tests/routes/test_rows.py
index 7613d52a..e8c944db 100644
--- a/services/api/tests/routes/test_rows.py
+++ b/services/api/tests/routes/test_rows.py
@@ -0,0 +1,3 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py
index 4827c75d..dd0f3a96 100644
--- a/services/api/tests/routes/test_valid.py
+++ b/services/api/tests/routes/test_valid.py
@@ -0,0 +1,3 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
diff --git a/services/api/tests/utils.py b/services/api/tests/utils.py
deleted file mode 100644
index b7cc894b..00000000
--- a/services/api/tests/utils.py
+++ /dev/null
@@ -1,16 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-from werkzeug.wrappers.request import Request
-from werkzeug.wrappers.response import Response
-
-
-def auth_callback(request: Request) -> Response:
- # return 401 if a cookie has been provided, 404 if a token has been provided,
- # and 200 if none has been provided
- #
- # caveat: the returned status codes don't simulate the reality
- # they're just used to check every case
- return Response(
- status=401 if request.headers.get("cookie") else 404 if request.headers.get("authorization") else 200
- )
diff --git a/tools/Python.mk b/tools/Python.mk
index 8632f448..91e40366 100644
--- a/tools/Python.mk
+++ b/tools/Python.mk
@@ -31 +31 @@ pip-audit:
- bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' |sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d')"
+ bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' |sed '/^torch @/d' | sed '/^libapi @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d')"
|
|
e1368b7dd349730d1e94a9bf94064919de2ee6d1
|
Quentin Lhoest
| 2023-07-04T12:30:09 |
optional num_bytes_original_files (#1478)
|
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py
index 5bdfcd35..ad008a4b 100644
--- a/services/worker/src/worker/dtos.py
+++ b/services/worker/src/worker/dtos.py
@@ -143 +143,3 @@ class ConfigSize(TypedDict):
- num_bytes_original_files: int
+ num_bytes_original_files: Optional[
+ int
+ ] # optional because partial parquet conversion can't provide the size of the original data
@@ -197 +199 @@ class DatasetSize(TypedDict):
- num_bytes_original_files: int
+ num_bytes_original_files: Optional[int]
diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py
index 9f47241b..785ea8f9 100644
--- a/services/worker/src/worker/job_runners/config/size.py
+++ b/services/worker/src/worker/job_runners/config/size.py
@@ -68 +68 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons
- "num_bytes_original_files": config_info["download_size"],
+ "num_bytes_original_files": config_info["download_size"] if not content["partial"] else None,
diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py
index 71217c1c..62c62704 100644
--- a/services/worker/src/worker/job_runners/dataset/size.py
+++ b/services/worker/src/worker/job_runners/dataset/size.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Tuple
+from typing import Optional, Tuple
@@ -93,0 +94,7 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
+ num_bytes_original_files: Optional[int] = 0
+ for config_size in config_sizes:
+ if num_bytes_original_files is not None and isinstance(config_size["num_bytes_original_files"], int):
+ num_bytes_original_files += config_size["num_bytes_original_files"]
+ else:
+ num_bytes_original_files = None
+ break
@@ -96 +103 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
- "num_bytes_original_files": sum(config_size["num_bytes_original_files"] for config_size in config_sizes),
+ "num_bytes_original_files": num_bytes_original_files,
|
|
00c0a16807736717bff563b7eb9db2e82e84374a
|
Quentin Lhoest
| 2023-07-04T09:38:22 |
unblock oscar (#1477)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index d787905d..953aeb10 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -100 +100 @@ parquetAndInfo:
- blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
+ blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
|
|
efb74b8330ebbe5a6832c566cd4f05873166e9ee
|
Quentin Lhoest
| 2023-07-04T09:38:14 |
More workers (#1476)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 1566f1fa..d787905d 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -124 +124 @@ cacheMaintenance:
- action: "backfill"
+ action: "skip"
@@ -129 +129 @@ cacheMaintenance:
- error_codes_to_retry: "DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError"
+ error_codes_to_retry: ""
@@ -272 +272 @@ workers:
- replicas: 20
+ replicas: 80
@@ -287 +287 @@ workers:
- replicas: 4
+ replicas: 12
|
|
10ecac78d4c640228d53ccddb83314c239252980
|
Quentin Lhoest
| 2023-07-03T18:31:46 |
fix stream_convert_to_parquet for GeneratorBasedBuilder (#1473)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 880b991a..cae1b12c 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -808,0 +809,3 @@ def stream_convert_to_parquet(builder: DatasetBuilder, max_dataset_size: int) ->
+ prepare_split_kwargs: dict[str, Any] = (
+ {"check_duplicate_keys": True} if isinstance(builder, datasets.builder.GeneratorBasedBuilder) else {}
+ )
@@ -813 +816,3 @@ def stream_convert_to_parquet(builder: DatasetBuilder, max_dataset_size: int) ->
- builder._prepare_split(split_generator=splits_generators[split], file_format="parquet")
+ builder._prepare_split(
+ split_generator=splits_generators[split], file_format="parquet", **prepare_split_kwargs
+ )
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index 6c4f72f2..a78ba301 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -840 +840 @@ def test_get_delete_operations(
-def test_stream_convert_to_parquet(
+def test_stream_convert_to_parquet_arrowbasedbuilder(
@@ -866,0 +867,38 @@ def test_stream_convert_to_parquet(
[email protected](
+ "max_dataset_size,expected_num_shards",
+ [
+ (1, 1),
+ (150, 19),
+ (300, 38),
+ (9999999, 1000),
+ ],
+)
+def test_stream_convert_to_parquet_generatorbasedbuilder(
+ max_dataset_size: int, expected_num_shards: int, tmp_path: Path
+) -> None:
+ num_rows = 1000
+
+ def long_generator() -> Iterator[Dict[str, int]]:
+ for i in range(num_rows):
+ yield {"foo": i}
+
+ cache_dir = str(tmp_path / "test_limit_parquet_writes_cache_dir")
+ builder = ParametrizedGeneratorBasedBuilder(generator=long_generator, cache_dir=cache_dir)
+ with patch("worker.job_runners.config.parquet_and_info.get_writer_batch_size", lambda ds_config_info: 1):
+ with patch.object(datasets.config, "MAX_SHARD_SIZE", 1):
+ parquet_operations, partial = stream_convert_to_parquet(builder, max_dataset_size=max_dataset_size)
+ num_shards = len(parquet_operations)
+ assert num_shards == expected_num_shards
+ assert partial == (expected_num_shards < num_rows)
+ assert all(isinstance(op.path_or_fileobj, str) for op in parquet_operations)
+ parquet_files = list_generated_parquet_files(builder, partial=partial)
+ assert len(parquet_files) == expected_num_shards
+ assert all(os.path.isfile(parquet_file.local_file) for parquet_file in parquet_files)
+ one_sample_max_size = 100
+ expected_max_dataset_size = max_dataset_size + one_sample_max_size
+ assert (
+ sum(pq.ParquetFile(parquet_file.local_file).read().nbytes for parquet_file in parquet_files)
+ < expected_max_dataset_size
+ )
+
+
|
|
9189407c5e423a1cdbeeae270fa5637439a3563a
|
Quentin Lhoest
| 2023-07-03T17:20:29 |
Add partial to subsequent parquet-and-info jobs (#1471)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
index c1f0b674..c4958fd5 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
@@ -20 +20 @@ class MigrationAddPartialToCacheResponse(Migration):
- " config-parquet-and-info"
+ " config-parquet-and-info and subsequent steps"
@@ -25 +25,12 @@ class MigrationAddPartialToCacheResponse(Migration):
- "kind": "config-parquet-and-info",
+ "kind": {
+ "$in": [
+ "config-parquet-and-info",
+ "config-parquet",
+ "dataset-parquet",
+ "config-parquet-metadata",
+ "config-info",
+ "dataset-info",
+ "config-size",
+ "dataset-size",
+ ]
+ },
@@ -36 +47,16 @@ class MigrationAddPartialToCacheResponse(Migration):
- {"kind": "config-parquet-and-info", "http_status": 200}, {"$unset": {"content.partial": ""}}
+ {
+ "kind": {
+ "$in": [
+ "config-parquet-and-info",
+ "config-parquet",
+ "dataset-parquet",
+ "config-parquet-metadata",
+ "config-info",
+ "dataset-info",
+ "config-size",
+ "dataset-size",
+ ]
+ },
+ "http_status": 200,
+ },
+ {"$unset": {"content.partial": ""}},
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
index 85e1a036..6f4eab65 100644
--- a/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
+++ b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
@@ -2,0 +3 @@
+from typing import Any, Dict, List
@@ -13 +14 @@ from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_con
-def assert_partial(dataset: str) -> None:
+def assert_partial(dataset: str, kind: str) -> None:
@@ -15 +16 @@ def assert_partial(dataset: str) -> None:
- entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset})
+ entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset, "kind": kind})
@@ -20 +21 @@ def assert_partial(dataset: str) -> None:
-def assert_unchanged(dataset: str) -> None:
+def assert_unchanged(dataset: str, kind: str) -> None:
@@ -22 +23 @@ def assert_unchanged(dataset: str) -> None:
- entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset})
+ entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset, "kind": kind})
@@ -27,5 +28,8 @@ def assert_unchanged(dataset: str) -> None:
-def test_cache_add_partial(mongo_host: str) -> None:
- with MongoResource(database="test_cache_add_partial", host=mongo_host, mongoengine_alias="cache"):
- db = get_db(CACHE_MONGOENGINE_ALIAS)
- db[CACHE_COLLECTION_RESPONSES].insert_many(
- [
+cache: List[Dict[str, Any]] = [
+ {
+ "config": "lhoestq--demo1",
+ "dataset": "lhoestq/demo1",
+ "kind": "config-parquet-and-info",
+ "split": None,
+ "content": {
+ "parquet_files": [
@@ -33 +36,0 @@ def test_cache_add_partial(mongo_host: str) -> None:
- "config": "lhoestq--demo1",
@@ -35,90 +38,5 @@ def test_cache_add_partial(mongo_host: str) -> None:
- "kind": "config-parquet-and-info",
- "split": None,
- "content": {
- "parquet_files": [
- {
- "dataset": "lhoestq/demo1",
- "config": "lhoestq--demo1",
- "split": "test",
- "url": "https://huggingface.co/.../csv-test.parquet",
- "filename": "csv-test.parquet",
- "size": 4415,
- },
- {
- "dataset": "lhoestq/demo1",
- "config": "lhoestq--demo1",
- "split": "train",
- "url": "https://huggingface.co/.../csv-train.parquet",
- "filename": "csv-train.parquet",
- "size": 5038,
- },
- ],
- "dataset_info": {
- "description": "",
- "citation": "",
- "homepage": "",
- "license": "",
- "features": {},
- "builder_name": "csv",
- "config_name": "lhoestq--demo1",
- "version": {},
- "splits": {},
- "download_checksums": {},
- "download_size": 2340,
- "dataset_size": 2464,
- "size_in_bytes": 4804,
- },
- },
- "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11",
- "details": None,
- "error_code": None,
- "http_status": 200,
- "job_runner_version": 3,
- "progress": 1.0,
- },
- {
- "config": "lhoestq--demo2",
- "dataset": "lhoestq/demo2",
- "kind": "config-parquet-and-info",
- "split": None,
- "content": {
- "parquet_files": [
- {
- "dataset": "lhoestq/demo2",
- "config": "lhoestq--demo2",
- "split": "test",
- "url": "https://huggingface.co/.../csv-test.parquet",
- "filename": "csv-test.parquet",
- "size": 4415,
- },
- {
- "dataset": "lhoestq/demo2",
- "config": "lhoestq--demo2",
- "split": "train",
- "url": "https://huggingface.co/.../csv-train.parquet",
- "filename": "csv-train.parquet",
- "size": 5038,
- },
- ],
- "dataset_info": {
- "description": "",
- "citation": "",
- "homepage": "",
- "license": "",
- "features": {},
- "builder_name": "csv",
- "config_name": "lhoestq--demo2",
- "version": {},
- "splits": {},
- "download_checksums": {},
- "download_size": 2340,
- "dataset_size": 2464,
- "size_in_bytes": 4804,
- },
- },
- "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11",
- "details": None,
- "error_code": None,
- "http_status": 200,
- "job_runner_version": 3,
- "progress": 1.0,
+ "config": "lhoestq--demo1",
+ "split": "test",
+ "url": "https://huggingface.co/.../csv-test.parquet",
+ "filename": "csv-test.parquet",
+ "size": 4415,
@@ -127,46 +45,6 @@ def test_cache_add_partial(mongo_host: str) -> None:
- "config": "lhoestq--error",
- "dataset": "lhoestq/error",
- "kind": "config-parquet-and-info",
- "split": None,
- "content": {"error": "Streaming is not supported for lhoestq/error"},
- "dataset_git_revision": "ec3c8d414af3dfe600399f5e6ef2c682938676f3",
- "details": {
- "error": "Streaming is not supported for lhoestq/error",
- "cause_exception": "TypeError",
- "cause_message": "Streaming is not supported for lhoestq/error",
- "cause_traceback": [
- "Traceback (most recent call last):\n",
- (
- ' File "/src/services/worker/src/worker/job_manager.py", line 163, in process\n '
- " job_result = self.job_runner.compute()\n"
- ),
- (
- ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
- " 932, in compute\n compute_config_parquet_and_info_response(\n"
- ),
- (
- ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
- " 825, in compute_config_parquet_and_info_response\n raise_if_not_supported(\n"
- ),
- (
- ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
- " 367, in raise_if_not_supported\n raise_if_too_big_from_external_data_files(\n"
- ),
- (
- ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
- " 447, in raise_if_too_big_from_external_data_files\n "
- " builder._split_generators(mock_dl_manager)\n"
- ),
- (
- " File"
- ' "/tmp/modules-cache/datasets_modules/.../error.py",'
- ' line 190, in _split_generators\n raise TypeError("Streaming is not supported for'
- ' lhoestq/error")\n'
- ),
- "TypeError: Streaming is not supported for lhoestq/error\n",
- ],
- },
- "error_code": "UnexpectedError",
- "http_status": 500,
- "job_runner_version": 3,
- "progress": None,
+ "dataset": "lhoestq/demo1",
+ "config": "lhoestq--demo1",
+ "split": "train",
+ "url": "https://huggingface.co/.../csv-train.parquet",
+ "filename": "csv-train.parquet",
+ "size": 5038,
@@ -174,2 +52,99 @@ def test_cache_add_partial(mongo_host: str) -> None:
- ]
- )
+ ],
+ "dataset_info": {
+ "description": "",
+ "citation": "",
+ "homepage": "",
+ "license": "",
+ "features": {},
+ "builder_name": "csv",
+ "config_name": "lhoestq--demo1",
+ "version": {},
+ "splits": {},
+ "download_checksums": {},
+ "download_size": 2340,
+ "dataset_size": 2464,
+ "size_in_bytes": 4804,
+ },
+ },
+ "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11",
+ "details": None,
+ "error_code": None,
+ "http_status": 200,
+ "job_runner_version": 3,
+ "progress": 1.0,
+ },
+ {
+ "config": "lhoestq--error",
+ "dataset": "lhoestq/error",
+ "kind": "config-parquet-and-info",
+ "split": None,
+ "content": {"error": "Streaming is not supported for lhoestq/error"},
+ "dataset_git_revision": "ec3c8d414af3dfe600399f5e6ef2c682938676f3",
+ "details": {
+ "error": "Streaming is not supported for lhoestq/error",
+ "cause_exception": "TypeError",
+ "cause_message": "Streaming is not supported for lhoestq/error",
+ "cause_traceback": [
+ "Traceback (most recent call last):\n",
+ (
+ ' File "/src/services/worker/src/worker/job_manager.py", line 163, in process\n '
+ " job_result = self.job_runner.compute()\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 932, in compute\n compute_config_parquet_and_info_response(\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 825, in compute_config_parquet_and_info_response\n raise_if_not_supported(\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 367, in raise_if_not_supported\n raise_if_too_big_from_external_data_files(\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 447, in raise_if_too_big_from_external_data_files\n "
+ " builder._split_generators(mock_dl_manager)\n"
+ ),
+ (
+ " File"
+ ' "/tmp/modules-cache/datasets_modules/.../error.py",'
+ ' line 190, in _split_generators\n raise TypeError("Streaming is not supported for'
+ ' lhoestq/error")\n'
+ ),
+ "TypeError: Streaming is not supported for lhoestq/error\n",
+ ],
+ },
+ "error_code": "UnexpectedError",
+ "http_status": 500,
+ "job_runner_version": 3,
+ "progress": None,
+ },
+]
+cache2: List[Dict[str, Any]] = [
+ {
+ "config": "lhoestq--demo2",
+ "dataset": "lhoestq/demo2",
+ "kind": kind,
+ "split": None,
+ "content": {},
+ "http_status": 200,
+ }
+ for kind in [
+ "config-parquet-and-info",
+ "config-parquet",
+ "dataset-parquet",
+ "config-parquet-metadata",
+ "config-info",
+ "dataset-info",
+ "config-size",
+ "dataset-size",
+ ]
+]
+
+
+def test_cache_add_partial(mongo_host: str) -> None:
+ with MongoResource(database="test_cache_add_partial", host=mongo_host, mongoengine_alias="cache"):
+ db = get_db(CACHE_MONGOENGINE_ALIAS)
+ db[CACHE_COLLECTION_RESPONSES].insert_many(cache + cache2)
@@ -183,3 +158,13 @@ def test_cache_add_partial(mongo_host: str) -> None:
- assert_partial("lhoestq/demo1")
- assert_partial("lhoestq/demo2")
- assert_unchanged("lhoestq/error")
+ assert_partial("lhoestq/demo1", kind="config-parquet-and-info")
+ assert_unchanged("lhoestq/error", kind="config-parquet-and-info")
+ for kind in [
+ "config-parquet-and-info",
+ "config-parquet",
+ "dataset-parquet",
+ "config-parquet-metadata",
+ "config-info",
+ "dataset-info",
+ "config-size",
+ "dataset-size",
+ ]:
+ assert_partial("lhoestq/demo2", kind=kind)
@@ -188,3 +173,13 @@ def test_cache_add_partial(mongo_host: str) -> None:
- assert_unchanged("lhoestq/demo1")
- assert_unchanged("lhoestq/demo2")
- assert_unchanged("lhoestq/error")
+ assert_unchanged("lhoestq/demo1", kind="config-parquet-and-info")
+ assert_unchanged("lhoestq/error", kind="config-parquet-and-info")
+ for kind in [
+ "config-parquet-and-info",
+ "config-parquet",
+ "dataset-parquet",
+ "config-parquet-metadata",
+ "config-info",
+ "dataset-info",
+ "config-size",
+ "dataset-size",
+ ]:
+ assert_unchanged("lhoestq/demo2", kind=kind)
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py
index ca029d38..5bdfcd35 100644
--- a/services/worker/src/worker/dtos.py
+++ b/services/worker/src/worker/dtos.py
@@ -112,0 +113 @@ class ConfigInfoResponse(TypedDict):
+ partial: bool
@@ -130,0 +132 @@ class ConfigParquetMetadataResponse(TypedDict):
+ partial: bool
@@ -134,0 +137 @@ class ConfigParquetResponse(TypedDict):
+ partial: bool
@@ -160,0 +164 @@ class ConfigSizeResponse(TypedDict):
+ partial: bool
@@ -175,0 +180 @@ class DatasetInfoResponse(TypedDict):
+ partial: bool
@@ -186,0 +192 @@ class DatasetParquetResponse(TypedDict):
+ partial: bool
@@ -206,0 +213 @@ class DatasetSizeResponse(TypedDict):
+ partial: bool
diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py
index 33d68219..ca254aff 100644
--- a/services/worker/src/worker/job_runners/config/info.py
+++ b/services/worker/src/worker/job_runners/config/info.py
@@ -33,0 +34 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons
+ partial = content["partial"]
@@ -45 +46 @@ def compute_config_info_response(dataset: str, config: str) -> ConfigInfoRespons
- return ConfigInfoResponse(dataset_info=config_info)
+ return ConfigInfoResponse(dataset_info=config_info, partial=partial)
diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py
index a6fc6833..fa614c97 100644
--- a/services/worker/src/worker/job_runners/config/parquet.py
+++ b/services/worker/src/worker/job_runners/config/parquet.py
@@ -44,0 +45 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse
+ partial = content["partial"]
@@ -47 +48 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse
- return ConfigParquetResponse(parquet_files=parquet_files)
+ return ConfigParquetResponse(parquet_files=parquet_files, partial=partial)
diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py
index 55de2fe3..0129a70e 100644
--- a/services/worker/src/worker/job_runners/config/parquet_metadata.py
+++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py
@@ -3,0 +4 @@
+import functools
@@ -5 +5,0 @@ import logging
-from functools import partial
@@ -71,0 +72 @@ def compute_parquet_metadata_response(
+ partial = config_parquet_best_response.response["content"]["partial"]
@@ -80 +81,5 @@ def compute_parquet_metadata_response(
- partial(get_parquet_file, fs=fs, hf_token=hf_token), source_urls, desc=desc, unit="pq", disable=True
+ functools.partial(get_parquet_file, fs=fs, hf_token=hf_token),
+ source_urls,
+ desc=desc,
+ unit="pq",
+ disable=True,
@@ -108 +113 @@ def compute_parquet_metadata_response(
- return ConfigParquetMetadataResponse(parquet_files_metadata=parquet_files_metadata)
+ return ConfigParquetMetadataResponse(parquet_files_metadata=parquet_files_metadata, partial=partial)
diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py
index 2610f994..9f47241b 100644
--- a/services/worker/src/worker/job_runners/config/size.py
+++ b/services/worker/src/worker/job_runners/config/size.py
@@ -76,0 +77 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons
+ partial = content["partial"]
@@ -85 +86,2 @@ def compute_config_size_response(dataset: str, config: str) -> ConfigSizeRespons
- }
+ },
+ "partial": partial,
diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py
index 1e8802c9..852af3e9 100644
--- a/services/worker/src/worker/job_runners/dataset/info.py
+++ b/services/worker/src/worker/job_runners/dataset/info.py
@@ -48,0 +49 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl
+ partial = False
@@ -76,0 +78 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl
+ partial = partial or config_response["content"]["partial"]
@@ -83 +85 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl
- return DatasetInfoResponse(dataset_info=config_infos, pending=pending, failed=failed), progress
+ return DatasetInfoResponse(dataset_info=config_infos, pending=pending, failed=failed, partial=partial), progress
diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py
index 9425c249..4b04f122 100644
--- a/services/worker/src/worker/job_runners/dataset/parquet.py
+++ b/services/worker/src/worker/job_runners/dataset/parquet.py
@@ -52,0 +53 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa
+ partial = False
@@ -84 +85,3 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa
- config_parquet_content = ConfigParquetResponse(parquet_files=response["content"]["parquet_files"])
+ config_parquet_content = ConfigParquetResponse(
+ parquet_files=response["content"]["parquet_files"], partial=response["content"]["partial"]
+ )
@@ -85,0 +89 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa
+ partial = partial or config_parquet_content["partial"]
@@ -92,5 +96 @@ def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, floa
- DatasetParquetResponse(
- parquet_files=parquet_files,
- pending=pending,
- failed=failed,
- ),
+ DatasetParquetResponse(parquet_files=parquet_files, pending=pending, failed=failed, partial=partial),
diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py
index 7019388d..71217c1c 100644
--- a/services/worker/src/worker/job_runners/dataset/size.py
+++ b/services/worker/src/worker/job_runners/dataset/size.py
@@ -55,0 +56 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
+ partial = False
@@ -87 +88,3 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
- config_size_content = ConfigSizeResponse(size=response["content"]["size"])
+ config_size_content = ConfigSizeResponse(
+ size=response["content"]["size"], partial=response["content"]["partial"]
+ )
@@ -89,0 +93 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
+ partial = partial or config_size_content["partial"]
@@ -111,0 +116 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
+ "partial": partial,
diff --git a/services/worker/tests/job_runners/config/test_info.py b/services/worker/tests/job_runners/config/test_info.py
index e0aac1b4..49be6df2 100644
--- a/services/worker/tests/job_runners/config/test_info.py
+++ b/services/worker/tests/job_runners/config/test_info.py
@@ -181,4 +181 @@ def get_job_runner(
- {
- "parquet_files": PARQUET_FILES,
- "dataset_info": CONFIG_INFO_1,
- },
+ {"parquet_files": PARQUET_FILES, "dataset_info": CONFIG_INFO_1, "partial": False},
@@ -186 +183 @@ def get_job_runner(
- {"dataset_info": CONFIG_INFO_1},
+ {"dataset_info": CONFIG_INFO_1, "partial": False},
diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py
index c2e9f007..1189453f 100644
--- a/services/worker/tests/job_runners/config/test_parquet.py
+++ b/services/worker/tests/job_runners/config/test_parquet.py
@@ -96 +96,2 @@ def get_job_runner(
- ]
+ ],
+ partial=False,
@@ -179 +180,2 @@ def get_job_runner(
- ]
+ ],
+ partial=False,
diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py
index 96f1cfbc..571a8f8e 100644
--- a/services/worker/tests/job_runners/config/test_parquet_metadata.py
+++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py
@@ -105,0 +106 @@ def get_job_runner(
+ partial=False,
@@ -130 +131,2 @@ def get_job_runner(
- ]
+ ],
+ partial=False,
diff --git a/services/worker/tests/job_runners/config/test_size.py b/services/worker/tests/job_runners/config/test_size.py
index b55b679f..df89b2e2 100644
--- a/services/worker/tests/job_runners/config/test_size.py
+++ b/services/worker/tests/job_runners/config/test_size.py
@@ -122,0 +123 @@ def get_job_runner(
+ "partial": False,
@@ -156 +157,2 @@ def get_job_runner(
- }
+ },
+ "partial": False,
diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py
index 26aeb369..0d6b8158 100644
--- a/services/worker/tests/job_runners/dataset/test_info.py
+++ b/services/worker/tests/job_runners/dataset/test_info.py
@@ -49 +49 @@ UPSTREAM_RESPONSE_CONFIG_INFO_1: UpstreamResponse = UpstreamResponse(
- content={"dataset_info": CONFIG_INFO_1},
+ content={"dataset_info": CONFIG_INFO_1, "partial": False},
@@ -57 +57 @@ UPSTREAM_RESPONSE_CONFIG_INFO_2: UpstreamResponse = UpstreamResponse(
- content={"dataset_info": CONFIG_INFO_2},
+ content={"dataset_info": CONFIG_INFO_2, "partial": False},
@@ -64,0 +65 @@ EXPECTED_OK = (
+ "partial": False,
@@ -82,0 +84 @@ EXPECTED_PARTIAL_PENDING = (
+ "partial": False,
@@ -100,0 +103 @@ EXPECTED_PARTIAL_FAILED = (
+ "partial": False,
diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py
index 8f63b188..28c04d3f 100644
--- a/services/worker/tests/job_runners/dataset/test_parquet.py
+++ b/services/worker/tests/job_runners/dataset/test_parquet.py
@@ -100 +100,2 @@ def get_job_runner(
- ]
+ ],
+ partial=False,
@@ -118 +119,2 @@ def get_job_runner(
- ]
+ ],
+ partial=False,
@@ -133,0 +136 @@ def get_job_runner(
+ partial=False,
diff --git a/services/worker/tests/job_runners/dataset/test_size.py b/services/worker/tests/job_runners/dataset/test_size.py
index 49490e44..76326d1d 100644
--- a/services/worker/tests/job_runners/dataset/test_size.py
+++ b/services/worker/tests/job_runners/dataset/test_size.py
@@ -120 +120,2 @@ def get_job_runner(
- }
+ },
+ "partial": False,
@@ -159 +160,2 @@ def get_job_runner(
- }
+ },
+ "partial": False,
@@ -233,0 +236 @@ def get_job_runner(
+ "partial": False,
|
|
cb4142773a2b46122049c34435dfb063ec656869
|
Polina Kazakova
| 2023-07-03T17:11:05 |
Rename `/dataset-info` endpoint to `/info` (#1468)
|
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml
index b4dbd129..bd3bb312 100644
--- a/docs/source/_toctree.yml
+++ b/docs/source/_toctree.yml
@@ -14,0 +15,2 @@
+ - local: info
+ title: Get dataset information
diff --git a/docs/source/info.mdx b/docs/source/info.mdx
new file mode 100644
index 00000000..dd9bc8c6
--- /dev/null
+++ b/docs/source/info.mdx
@@ -0,0 +1,139 @@
+# Get dataset information
+
+Datasets Server provides an `/info` endpoint for exploring the general information about dataset, including such fields as description, citation, homepage, license and features.
+
+The `/info` endpoint accepts two query parameters:
+
+- `dataset`: the dataset name
+- `config`: the configuration name
+
+<inferencesnippet>
+<python>
+```python
+import requests
+headers = {"Authorization": f"Bearer {API_TOKEN}"}
+API_URL = "https://datasets-server.huggingface.co/info?dataset=duorc&config=SelfRC"
+def query():
+ response = requests.get(API_URL, headers=headers)
+ return response.json()
+data = query()
+```
+</python>
+<js>
+```js
+import fetch from "node-fetch";
+async function query(data) {
+ const response = await fetch(
+ "https://datasets-server.huggingface.co/info?dataset=duorc&config=SelfRC",
+ {
+ headers: { Authorization: `Bearer ${API_TOKEN}` },
+ method: "GET"
+ }
+ );
+ const result = await response.json();
+ return result;
+}
+query().then((response) => {
+ console.log(JSON.stringify(response));
+});
+```
+</js>
+<curl>
+```curl
+curl https://datasets-server.huggingface.co/info?dataset=duorc&config=SelfRC \
+ -X GET \
+ -H "Authorization: Bearer ${API_TOKEN}"
+```
+</curl>
+</inferencesnippet>
+
+The endpoint response is a JSON with the `dataset_info` key. Its structure and content correspond to [DatasetInfo](https://huggingface.co/docs/datasets/package_reference/main_classes#datasets.DatasetInfo) object of the `datasets` library.
+
+```json
+{
+ "dataset_info": {
+ "description": "DuoRC contains 186,089 unique question-answer pairs created from a collection of 7680 pairs of movie plots where each pair in the collection reflects two versions of the same movie.\n",
+ "citation": "@inproceedings{DuoRC,\nauthor = { Amrita Saha and Rahul Aralikatte and Mitesh M. Khapra and Karthik Sankaranarayanan},title = {{DuoRC: Towards Complex Language Understanding with Paraphrased Reading Comprehension}},\nbooktitle = {Meeting of the Association for Computational Linguistics (ACL)},\nyear = {2018}\n}\n",
+ "homepage": "https://duorc.github.io/",
+ "license": "https://raw.githubusercontent.com/duorc/duorc/master/LICENSE",
+ "features": {
+ "plot_id": {
+ "dtype": "string",
+ "_type": "Value"
+ },
+ "plot": {
+ "dtype": "string",
+ "_type": "Value"
+ },
+ "title": {
+ "dtype": "string",
+ "_type": "Value"
+ },
+ "question_id": {
+ "dtype": "string",
+ "_type": "Value"
+ },
+ "question": {
+ "dtype": "string",
+ "_type": "Value"
+ },
+ "answers": {
+ "feature": {
+ "dtype": "string",
+ "_type": "Value"
+ },
+ "_type": "Sequence"
+ },
+ "no_answer": {
+ "dtype": "bool",
+ "_type": "Value"
+ }
+ },
+ "builder_name": "duorc",
+ "config_name": "SelfRC",
+ "version": {
+ "version_str": "1.0.0",
+ "major": 1,
+ "minor": 0,
+ "patch": 0
+ },
+ "splits": {
+ "train": {
+ "name": "train",
+ "num_bytes": 239852729,
+ "num_examples": 60721,
+ "dataset_name": "duorc"
+ },
+ "validation": {
+ "name": "validation",
+ "num_bytes": 51662519,
+ "num_examples": 12961,
+ "dataset_name": "duorc"
+ },
+ "test": {
+ "name": "test",
+ "num_bytes": 49142710,
+ "num_examples": 12559,
+ "dataset_name": "duorc"
+ }
+ },
+ "download_checksums": {
+ "https://raw.githubusercontent.com/duorc/duorc/master/dataset/SelfRC_train.json": {
+ "num_bytes": 24388192,
+ "checksum": null
+ },
+ "https://raw.githubusercontent.com/duorc/duorc/master/dataset/SelfRC_dev.json": {
+ "num_bytes": 5051240,
+ "checksum": null
+ },
+ "https://raw.githubusercontent.com/duorc/duorc/master/dataset/SelfRC_test.json": {
+ "num_bytes": 5023228,
+ "checksum": null
+ }
+ },
+ "download_size": 34462660,
+ "dataset_size": 340657958,
+ "size_in_bytes": 375120618
+ }
+}
+```
\ No newline at end of file
diff --git a/e2e/tests/test_11_api.py b/e2e/tests/test_11_api.py
index 7732ea1f..48e60d68 100644
--- a/e2e/tests/test_11_api.py
+++ b/e2e/tests/test_11_api.py
@@ -58,2 +58,2 @@ def test_auth_e2e(
- ("/dataset-info", "dataset"),
- ("/dataset-info", "config"),
+ ("/info", "dataset"),
+ ("/info", "config"),
diff --git a/services/api/src/api/config.py b/services/api/src/api/config.py
index 5bb0f533..e8518a51 100644
--- a/services/api/src/api/config.py
+++ b/services/api/src/api/config.py
@@ -138 +138 @@ class EndpointConfig:
- "/dataset-info": {"dataset": ["dataset-info"], "config": ["config-info"]},
+ "/info": {"dataset": ["dataset-info"], "config": ["config-info"]},
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index a3a98220..e691f92e 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -65 +65 @@ def test_endpoints_definition() -> None:
- dataset_info = definition["/dataset-info"]
+ dataset_info = definition["/info"]
@@ -90,0 +91,4 @@ def test_endpoints_definition() -> None:
+ # assert old endpoints don't exist
+ with raises(KeyError):
+ _ = definition["/dataset-info"]
+
|
|
793ce6a17d88315e6cc09075278606926257d526
|
Sylvain Lesage
| 2023-07-03T15:55:04 |
feat: 🎸 backfill cache entries older than 90 days (#1469)
|
diff --git a/chart/templates/_envCache.tpl b/chart/templates/_envCache.tpl
index 4701fbfd..f4974fd6 100644
--- a/chart/templates/_envCache.tpl
+++ b/chart/templates/_envCache.tpl
@@ -4,0 +5,2 @@
+- name: CACHE_MAX_DAYS
+ value: {{ .Values.cache.maxDays | quote }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 98cc505b..212ccc51 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -114,0 +115,2 @@ cache:
+ # Maximum number of days to keep the cached API responses
+ maxDays: 90
diff --git a/jobs/cache_maintenance/src/cache_maintenance/backfill.py b/jobs/cache_maintenance/src/cache_maintenance/backfill.py
index de3bedf1..a64bbad6 100644
--- a/jobs/cache_maintenance/src/cache_maintenance/backfill.py
+++ b/jobs/cache_maintenance/src/cache_maintenance/backfill.py
@@ -15,0 +16 @@ def backfill_cache(
+ cache_max_days: int,
@@ -50,0 +52 @@ def backfill_cache(
+ cache_max_days=cache_max_days,
diff --git a/jobs/cache_maintenance/src/cache_maintenance/main.py b/jobs/cache_maintenance/src/cache_maintenance/main.py
index e9acb982..a6926853 100644
--- a/jobs/cache_maintenance/src/cache_maintenance/main.py
+++ b/jobs/cache_maintenance/src/cache_maintenance/main.py
@@ -64,0 +65 @@ def run_job() -> None:
+ cache_max_days=job_config.cache.max_days,
diff --git a/libs/libcommon/README.md b/libs/libcommon/README.md
index 833ec48b..1b16130a 100644
--- a/libs/libcommon/README.md
+++ b/libs/libcommon/README.md
@@ -28,0 +29 @@ Set environment variables to configure the storage of precomputed API responses
+- `CACHE_MAX_DAYS`: maximum number of days to keep the cache entries. Defaults to `90`.
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 99c6bebc..3a713a1f 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -178,0 +179 @@ class LogConfig:
+CACHE_MAX_DAYS = 90 # 3 months
@@ -184,0 +186 @@ class CacheConfig:
+ max_days: int = CACHE_MAX_DAYS
@@ -192,0 +195 @@ class CacheConfig:
+ max_days=env.int(name="MAX_DAYS", default=CACHE_MAX_DAYS),
diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py
index 0797eef8..4a0b4603 100644
--- a/libs/libcommon/src/libcommon/operations.py
+++ b/libs/libcommon/src/libcommon/operations.py
@@ -15,0 +16 @@ def backfill_dataset(
+ cache_max_days: int,
@@ -24,0 +26 @@ def backfill_dataset(
+ cache_max_days (int): the number of days to keep the cache
@@ -31 +33 @@ def backfill_dataset(
- revision=revision, priority=priority, error_codes_to_retry=[]
+ revision=revision, priority=priority, error_codes_to_retry=[], cache_max_days=cache_max_days
diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py
index 1bbab28c..1a5669b0 100644
--- a/libs/libcommon/src/libcommon/orchestrator.py
+++ b/libs/libcommon/src/libcommon/orchestrator.py
@@ -33,0 +34 @@ class CacheStatus:
+ cache_is_old: Dict[str, ArtifactState] = field(default_factory=dict)
@@ -42,0 +44 @@ class CacheStatus:
+ "cache_is_old": sorted(self.cache_is_old.keys()),
@@ -309,0 +312 @@ class DatasetBackfillPlan(Plan):
+ cache_max_days: maximum number of days to keep the cache
@@ -317,0 +321 @@ class DatasetBackfillPlan(Plan):
+ cache_max_days: int
@@ -478,0 +483,5 @@ class DatasetBackfillPlan(Plan):
+ # is an old entry?
+ if artifact_state.cache_state.is_old(days=self.cache_max_days):
+ cache_status.cache_is_old[artifact_state.id] = artifact_state
+ continue
+
@@ -537,0 +547 @@ class DatasetBackfillPlan(Plan):
+ + list(self.cache_status.cache_is_old.values())
@@ -573 +583,3 @@ class DatasetOrchestrator:
- def set_revision(self, revision: str, priority: Priority, error_codes_to_retry: List[str]) -> None:
+ def set_revision(
+ self, revision: str, priority: Priority, error_codes_to_retry: List[str], cache_max_days: int
+ ) -> None:
@@ -583,0 +596 @@ class DatasetOrchestrator:
+ cache_max_days (int): The maximum number of days for which the cache is considered valid.
@@ -614,0 +628 @@ class DatasetOrchestrator:
+ cache_max_days=cache_max_days,
@@ -723 +737,3 @@ class DatasetOrchestrator:
- def backfill(self, revision: str, priority: Priority, error_codes_to_retry: Optional[List[str]] = None) -> int:
+ def backfill(
+ self, revision: str, priority: Priority, cache_max_days: int, error_codes_to_retry: Optional[List[str]] = None
+ ) -> int:
@@ -729,0 +746 @@ class DatasetOrchestrator:
+ cache_max_days (int): The maximum number of days to keep the cache.
@@ -752,0 +770 @@ class DatasetOrchestrator:
+ cache_max_days=cache_max_days,
diff --git a/libs/libcommon/src/libcommon/state.py b/libs/libcommon/src/libcommon/state.py
index b047fbe5..7a7f250e 100644
--- a/libs/libcommon/src/libcommon/state.py
+++ b/libs/libcommon/src/libcommon/state.py
@@ -12,0 +13 @@ from libcommon.simple_cache import CacheEntryMetadata, fetch_names
+from libcommon.utils import get_datetime
@@ -91,0 +93,7 @@ class CacheState:
+ def is_old(self, days: int) -> bool:
+ if self.cache_entry_metadata is None or days <= 0:
+ return False
+ return self.cache_entry_metadata["updated_at"] < get_datetime(days).replace(tzinfo=None)
+ # ^ we remove the timezone to avoid comparing timezone-aware and timezone-naive datetimes
+ # could be done better, but we don't need more precision
+
diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py
index b921ea78..ac79f831 100644
--- a/libs/libcommon/src/libcommon/utils.py
+++ b/libs/libcommon/src/libcommon/utils.py
@@ -7 +7 @@ import mimetypes
-from datetime import datetime, timezone
+from datetime import datetime, timedelta, timezone
@@ -91,2 +91,5 @@ def orjson_dumps(content: Any) -> bytes:
-def get_datetime() -> datetime:
- return datetime.now(timezone.utc)
+def get_datetime(days: Optional[int] = None) -> datetime:
+ date = datetime.now(timezone.utc)
+ if days is not None:
+ date = date - timedelta(days=days)
+ return date
diff --git a/libs/libcommon/tests/test_backfill.py b/libs/libcommon/tests/test_backfill.py
index 404f35a0..d69c231a 100644
--- a/libs/libcommon/tests/test_backfill.py
+++ b/libs/libcommon/tests/test_backfill.py
@@ -12 +12 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.utils import Priority, Status
+from libcommon.utils import Priority, Status, get_datetime
@@ -86,0 +87 @@ def test_initial_state(
+ "cache_is_old": [],
@@ -121,0 +123 @@ def test_da_is_computed(
+ "cache_is_old": [],
@@ -155,0 +158 @@ def test_ca_1_is_computed(
+ "cache_is_old": [],
@@ -199,0 +203 @@ def test_plan_one_job_creation_and_termination(
+ "cache_is_old": [],
@@ -218,0 +223 @@ def test_plan_one_job_creation_and_termination(
+ "cache_is_old": [],
@@ -237,0 +243 @@ def test_plan_one_job_creation_and_termination(
+ "cache_is_old": [],
@@ -282,0 +289 @@ def test_plan_all_job_creation_and_termination(processing_graph: ProcessingGraph
+ "cache_is_old": [],
@@ -299,0 +307 @@ def test_plan_all_job_creation_and_termination(processing_graph: ProcessingGraph
+ "cache_is_old": [],
@@ -343,0 +352 @@ def test_plan_compute_all(processing_graph: ProcessingGraph, up_to_date: List[st
+ "cache_is_old": [],
@@ -383,0 +393 @@ def test_plan_retry_error_and_outdated_by_parent(
+ "cache_is_old": [],
@@ -394,0 +405,29 @@ def test_plan_retry_error_and_outdated_by_parent(
[email protected](
+ "days_ago,is_old",
+ [(10, False), (30, True)],
+)
+def test_plan_old(days_ago: int, is_old: bool) -> None:
+ compute_all(processing_graph=PROCESSING_GRAPH_ONE_STEP)
+
+ CACHE_MAX_DAYS = 20
+ put_cache(step=STEP_DA, dataset=DATASET_NAME, revision=REVISION_NAME, updated_at=get_datetime(days_ago))
+
+ dataset_backfill_plan = get_dataset_backfill_plan(
+ processing_graph=PROCESSING_GRAPH_ONE_STEP, cache_max_days=CACHE_MAX_DAYS
+ )
+ assert_dataset_backfill_plan(
+ dataset_backfill_plan=dataset_backfill_plan,
+ cache_status={
+ "cache_has_different_git_revision": [],
+ "cache_is_old": [ARTIFACT_DA] if is_old else [],
+ "cache_is_outdated_by_parent": [],
+ "cache_is_empty": [],
+ "cache_is_error_to_retry": [],
+ "cache_is_job_runner_obsolete": [],
+ "up_to_date": [] if is_old else [ARTIFACT_DA],
+ },
+ queue_status={"in_process": []},
+ tasks=["CreateJobs,1"] if is_old else [],
+ )
+
+
@@ -428,0 +468 @@ def test_plan_outdated_by_parent(
+ "cache_is_old": [],
@@ -472,0 +513 @@ def test_plan_job_runner_version_and_outdated_by_parent(
+ "cache_is_old": [],
@@ -516,0 +558 @@ def test_plan_git_revision_and_outdated_by_parent(
+ "cache_is_old": [],
@@ -562,0 +605 @@ def test_plan_fan_in_updated(
+ "cache_is_old": [],
@@ -652,0 +696 @@ def test_plan_incoherent_state(
+ "cache_is_old": [],
@@ -669,0 +714 @@ def test_plan_incoherent_state(
+ "cache_is_old": [],
@@ -795,0 +841 @@ def test_delete_jobs(
+ "cache_is_old": [],
@@ -827,0 +874 @@ def test_multiple_revisions() -> None:
+ "cache_is_old": [],
@@ -848,0 +896 @@ def test_multiple_revisions() -> None:
+ "cache_is_old": [],
@@ -866,0 +915 @@ def test_multiple_revisions() -> None:
+ "cache_is_old": [],
@@ -884,0 +934 @@ def test_multiple_revisions() -> None:
+ "cache_is_old": [],
diff --git a/libs/libcommon/tests/test_backfill_on_real_graph.py b/libs/libcommon/tests/test_backfill_on_real_graph.py
index dfc46497..5cb1cf36 100644
--- a/libs/libcommon/tests/test_backfill_on_real_graph.py
+++ b/libs/libcommon/tests/test_backfill_on_real_graph.py
@@ -49,0 +50 @@ def test_plan_job_creation_and_termination() -> None:
+ "cache_is_old": [],
@@ -81,0 +83 @@ def test_plan_job_creation_and_termination() -> None:
+ "cache_is_old": [],
@@ -137,0 +140 @@ def test_plan_job_creation_and_termination() -> None:
+ "cache_is_old": [],
diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py
index 3131bc1a..32a687c4 100644
--- a/libs/libcommon/tests/test_orchestrator.py
+++ b/libs/libcommon/tests/test_orchestrator.py
@@ -40,0 +41,2 @@ from .utils import (
+CACHE_MAX_DAYS = 90
+
@@ -205 +207,3 @@ def test_set_revision(
- dataset_orchestrator.set_revision(revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[])
+ dataset_orchestrator.set_revision(
+ revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[], cache_max_days=CACHE_MAX_DAYS
+ )
@@ -239 +243,3 @@ def test_set_revision_handle_existing_jobs(
- dataset_orchestrator.set_revision(revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[])
+ dataset_orchestrator.set_revision(
+ revision=REVISION_NAME, priority=Priority.NORMAL, error_codes_to_retry=[], cache_max_days=CACHE_MAX_DAYS
+ )
diff --git a/libs/libcommon/tests/utils.py b/libs/libcommon/tests/utils.py
index f6bd9197..e4addd22 100644
--- a/libs/libcommon/tests/utils.py
+++ b/libs/libcommon/tests/utils.py
@@ -3,0 +4 @@
+from datetime import datetime
@@ -28,0 +30 @@ SPLIT_NAMES_CONTENT = {
+CACHE_MAX_DAYS = 90
@@ -211,0 +214 @@ def get_dataset_backfill_plan(
+ cache_max_days: Optional[int] = None,
@@ -217,0 +221 @@ def get_dataset_backfill_plan(
+ cache_max_days=CACHE_MAX_DAYS if cache_max_days is None else cache_max_days,
@@ -265,0 +270 @@ def put_cache(
+ updated_at: Optional[datetime] = None,
@@ -298,0 +304 @@ def put_cache(
+ updated_at=updated_at,
diff --git a/services/admin/src/admin/app.py b/services/admin/src/admin/app.py
index 120725b5..33855962 100644
--- a/services/admin/src/admin/app.py
+++ b/services/admin/src/admin/app.py
@@ -87,0 +88 @@ def create_app() -> Starlette:
+ cache_max_days=app_config.cache.max_days,
@@ -99,0 +101 @@ def create_app() -> Starlette:
+ cache_max_days=app_config.cache.max_days,
diff --git a/services/admin/src/admin/routes/dataset_backfill.py b/services/admin/src/admin/routes/dataset_backfill.py
index 3604391f..7df678fa 100644
--- a/services/admin/src/admin/routes/dataset_backfill.py
+++ b/services/admin/src/admin/routes/dataset_backfill.py
@@ -28,0 +29 @@ def create_dataset_backfill_endpoint(
+ cache_max_days: int,
@@ -53 +54,3 @@ def create_dataset_backfill_endpoint(
- dataset_orchestrator.backfill(revision=dataset_git_revision, priority=Priority.LOW)
+ dataset_orchestrator.backfill(
+ revision=dataset_git_revision, priority=Priority.LOW, cache_max_days=cache_max_days
+ )
diff --git a/services/admin/src/admin/routes/dataset_backfill_plan.py b/services/admin/src/admin/routes/dataset_backfill_plan.py
index ea9d071b..908a8b48 100644
--- a/services/admin/src/admin/routes/dataset_backfill_plan.py
+++ b/services/admin/src/admin/routes/dataset_backfill_plan.py
@@ -28,0 +29 @@ def create_dataset_backfill_plan_endpoint(
+ cache_max_days: int,
@@ -55,0 +57 @@ def create_dataset_backfill_plan_endpoint(
+ cache_max_days=cache_max_days,
diff --git a/services/admin/src/admin/routes/dataset_state.py b/services/admin/src/admin/routes/dataset_state.py
deleted file mode 100644
index 7daae82e..00000000
--- a/services/admin/src/admin/routes/dataset_state.py
+++ /dev/null
@@ -1,63 +0,0 @@
-# SPDX-License-Identifier: Apache-2.0
-# Copyright 2022 The HuggingFace Authors.
-
-import logging
-from typing import Optional
-
-from libcommon.dataset import get_dataset_git_revision
-from libcommon.orchestrator import DatasetBackfillPlan
-from libcommon.processing_graph import ProcessingGraph
-from starlette.requests import Request
-from starlette.responses import Response
-
-from admin.authentication import auth_check
-from admin.utils import (
- AdminCustomError,
- Endpoint,
- MissingRequiredParameterError,
- UnexpectedError,
- are_valid_parameters,
- get_json_admin_error_response,
- get_json_ok_response,
-)
-
-
-def create_dataset_backfill_plan_endpoint(
- processing_graph: ProcessingGraph,
- max_age: int,
- hf_endpoint: str,
- external_auth_url: Optional[str] = None,
- organization: Optional[str] = None,
- hf_token: Optional[str] = None,
- hf_timeout_seconds: Optional[float] = None,
-) -> Endpoint:
- async def dataset_backfill_plan_endpoint(request: Request) -> Response:
- try:
- dataset = request.query_params.get("dataset")
- if not are_valid_parameters([dataset]) or not dataset:
- raise MissingRequiredParameterError("Parameter 'dataset' is required")
- logging.info(f"/dataset-backfill-plan, dataset={dataset}")
-
- # if auth_check fails, it will raise an exception that will be caught below
- auth_check(
- external_auth_url=external_auth_url,
- request=request,
- organization=organization,
- hf_timeout_seconds=hf_timeout_seconds,
- )
-
- dataset_git_revision = get_dataset_git_revision(
- dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, hf_timeout_seconds=hf_timeout_seconds
- )
- dataset_backfill_plan = DatasetBackfillPlan(
- dataset=dataset,
- processing_graph=processing_graph,
- revision=dataset_git_revision,
- )
- return get_json_ok_response(dataset_backfill_plan.as_response(), max_age=max_age)
- except AdminCustomError as e:
- return get_json_admin_error_response(e, max_age=max_age)
- except Exception as e:
- return get_json_admin_error_response(UnexpectedError("Unexpected error.", e), max_age=max_age)
-
- return dataset_backfill_plan_endpoint
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 13479c33..a124b5bb 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -82,0 +83 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ cache_max_days=app_config.cache.max_days,
@@ -102 +103,3 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- processing_graph=processing_graph, hf_webhook_secret=app_config.api.hf_webhook_secret
+ processing_graph=processing_graph,
+ hf_webhook_secret=app_config.api.hf_webhook_secret,
+ cache_max_days=app_config.cache.max_days,
@@ -121,0 +125 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ cache_max_days=app_config.cache.max_days,
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 4ddb8bc5..71f6f00e 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -63,0 +64 @@ def get_cache_entry_from_steps(
+ cache_max_days: int,
@@ -88,0 +90 @@ def get_cache_entry_from_steps(
+ cache_max_days=cache_max_days,
@@ -223,0 +226 @@ def create_endpoint(
+ cache_max_days: int,
@@ -296,0 +300 @@ def create_endpoint(
+ cache_max_days=cache_max_days,
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index d0689d8e..18430cb7 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -256,0 +257 @@ def create_rows_endpoint(
+ cache_max_days: int,
@@ -331,0 +333 @@ def create_rows_endpoint(
+ cache_max_days=cache_max_days,
diff --git a/services/api/src/api/routes/webhook.py b/services/api/src/api/routes/webhook.py
index ecb375a9..96f2c810 100644
--- a/services/api/src/api/routes/webhook.py
+++ b/services/api/src/api/routes/webhook.py
@@ -65,0 +66 @@ def process_payload(
+ cache_max_days: int,
@@ -84 +85,5 @@ def process_payload(
- dataset=dataset, revision=revision, processing_graph=processing_graph, priority=Priority.NORMAL
+ dataset=dataset,
+ revision=revision,
+ processing_graph=processing_graph,
+ priority=Priority.NORMAL,
+ cache_max_days=cache_max_days,
@@ -90 +95,5 @@ def process_payload(
- dataset=moved_to, revision=revision, processing_graph=processing_graph, priority=Priority.NORMAL
+ dataset=moved_to,
+ revision=revision,
+ processing_graph=processing_graph,
+ priority=Priority.NORMAL,
+ cache_max_days=cache_max_days,
@@ -95 +104,3 @@ def process_payload(
-def create_webhook_endpoint(processing_graph: ProcessingGraph, hf_webhook_secret: Optional[str] = None) -> Endpoint:
+def create_webhook_endpoint(
+ processing_graph: ProcessingGraph, cache_max_days: int, hf_webhook_secret: Optional[str] = None
+) -> Endpoint:
@@ -130 +141,6 @@ def create_webhook_endpoint(processing_graph: ProcessingGraph, hf_webhook_secret
- process_payload(processing_graph=processing_graph, payload=payload, trust_sender=trust_sender)
+ process_payload(
+ processing_graph=processing_graph,
+ payload=payload,
+ trust_sender=trust_sender,
+ cache_max_days=cache_max_days,
+ )
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index 079529ee..f6defeee 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -175,0 +176 @@ def try_backfill_dataset(
+ cache_max_days: int,
@@ -195 +196,3 @@ def try_backfill_dataset(
- dataset_orchestrator.set_revision(revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[])
+ dataset_orchestrator.set_revision(
+ revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[], cache_max_days=cache_max_days
+ )
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index 69e0dfd2..a3a98220 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -16,0 +17,2 @@ from api.utils import ResponseNotReadyError
+CACHE_MAX_DAYS = 90
+
@@ -123,6 +125,7 @@ def test_get_cache_entry_from_steps() -> None:
- [step_without_error, step_with_error],
- dataset,
- config,
- None,
- processing_graph,
- app_config.common.hf_endpoint,
+ processing_steps=[step_without_error, step_with_error],
+ dataset=dataset,
+ config=config,
+ split=None,
+ processing_graph=processing_graph,
+ hf_endpoint=app_config.common.hf_endpoint,
+ cache_max_days=CACHE_MAX_DAYS,
@@ -135,6 +138,7 @@ def test_get_cache_entry_from_steps() -> None:
- [step_with_error, step_without_error],
- dataset,
- config,
- None,
- processing_graph,
- app_config.common.hf_endpoint,
+ processing_steps=[step_with_error, step_without_error],
+ dataset=dataset,
+ config=config,
+ split=None,
+ processing_graph=processing_graph,
+ hf_endpoint=app_config.common.hf_endpoint,
+ cache_max_days=CACHE_MAX_DAYS,
@@ -147 +151,7 @@ def test_get_cache_entry_from_steps() -> None:
- [step_with_error, step_with_error], dataset, config, None, processing_graph, app_config.common.hf_endpoint
+ processing_steps=[step_with_error, step_with_error],
+ dataset=dataset,
+ config=config,
+ split=None,
+ processing_graph=processing_graph,
+ hf_endpoint=app_config.common.hf_endpoint,
+ cache_max_days=CACHE_MAX_DAYS,
@@ -160 +170,7 @@ def test_get_cache_entry_from_steps() -> None:
- [non_existent_step], dataset, None, None, processing_graph, app_config.common.hf_endpoint
+ processing_steps=[non_existent_step],
+ dataset=dataset,
+ config=None,
+ split=None,
+ processing_graph=processing_graph,
+ hf_endpoint=app_config.common.hf_endpoint,
+ cache_max_days=CACHE_MAX_DAYS,
diff --git a/tools/docker-compose-base.yml b/tools/docker-compose-base.yml
index 0fdef1d0..19ab22aa 100644
--- a/tools/docker-compose-base.yml
+++ b/tools/docker-compose-base.yml
@@ -12,0 +13 @@ services:
+ CACHE_MAX_DAYS: ${CACHE_MAX_DAYS-90}
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
index b9ca8e9e..59ed778e 100644
--- a/tools/docker-compose-dev-base.yml
+++ b/tools/docker-compose-dev-base.yml
@@ -12,0 +13 @@ services:
+ CACHE_MAX_DAYS: ${CACHE_MAX_DAYS-90}
|
|
288a5de35f9f6c7cd05cbbb25bea5f38bd459992
|
Quentin Lhoest
| 2023-07-03T15:40:31 |
Stream convert to parquet (#1448)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index b0dc9cf6..1566f1fa 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -124 +124 @@ cacheMaintenance:
- action: "skip"
+ action: "backfill"
@@ -129 +129 @@ cacheMaintenance:
- error_codes_to_retry: ""
+ error_codes_to_retry: "DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError"
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 6d18f4da..eaf51193 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -49,0 +50,3 @@ from mongodb_migration.migrations._20230622131500_lock_add_owner import (
+from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_config_parquet_and_info import (
+ MigrationAddPartialToCacheResponse,
+)
@@ -237,0 +241,3 @@ class MigrationsCollector:
+ MigrationAddPartialToCacheResponse(
+ version="20230703110100", description="add 'partial' field to config-parquet-and-info"
+ ),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
new file mode 100644
index 00000000..c1f0b674
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
@@ -0,0 +1,42 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS
+from libcommon.simple_cache import CachedResponseDocument
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import Migration
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationAddPartialToCacheResponse(Migration):
+ def up(self) -> None:
+ # See https://docs.mongoengine.org/guide/migration.html#example-1-addition-of-a-field
+ logging.info(
+ "If missing, add the partial field with the default value (false) to the cached results of"
+ " config-parquet-and-info"
+ )
+ db = get_db(CACHE_MONGOENGINE_ALIAS)
+ db[CACHE_COLLECTION_RESPONSES].update_many(
+ {
+ "kind": "config-parquet-and-info",
+ "http_status": 200,
+ "content.partial": {"$exists": False},
+ },
+ {"$set": {"content.partial": False}},
+ )
+
+ def down(self) -> None:
+ logging.info("Remove the partial field from all the cached results")
+ db = get_db(CACHE_MONGOENGINE_ALIAS)
+ db[CACHE_COLLECTION_RESPONSES].update_many(
+ {"kind": "config-parquet-and-info", "http_status": 200}, {"$unset": {"content.partial": ""}}
+ )
+
+ def validate(self) -> None:
+ logging.info("Ensure that a random selection of cached results have the 'partial' field")
+
+ check_documents(DocCls=CachedResponseDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
new file mode 100644
index 00000000..85e1a036
--- /dev/null
+++ b/jobs/mongodb_migration/tests/migrations/test_20230703110100_cache_add_partial_field_in_config_parquet_and_info.py
@@ -0,0 +1,192 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_ALIAS
+from libcommon.resources import MongoResource
+from mongoengine.connection import get_db
+
+from mongodb_migration.migrations._20230703110100_cache_add_partial_field_in_config_parquet_and_info import (
+ MigrationAddPartialToCacheResponse,
+)
+
+
+def assert_partial(dataset: str) -> None:
+ db = get_db(CACHE_MONGOENGINE_ALIAS)
+ entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset})
+ assert entry is not None
+ assert entry["content"]["partial"] is False
+
+
+def assert_unchanged(dataset: str) -> None:
+ db = get_db(CACHE_MONGOENGINE_ALIAS)
+ entry = db[CACHE_COLLECTION_RESPONSES].find_one({"dataset": dataset})
+ assert entry is not None
+ assert "partial" not in entry["content"]
+
+
+def test_cache_add_partial(mongo_host: str) -> None:
+ with MongoResource(database="test_cache_add_partial", host=mongo_host, mongoengine_alias="cache"):
+ db = get_db(CACHE_MONGOENGINE_ALIAS)
+ db[CACHE_COLLECTION_RESPONSES].insert_many(
+ [
+ {
+ "config": "lhoestq--demo1",
+ "dataset": "lhoestq/demo1",
+ "kind": "config-parquet-and-info",
+ "split": None,
+ "content": {
+ "parquet_files": [
+ {
+ "dataset": "lhoestq/demo1",
+ "config": "lhoestq--demo1",
+ "split": "test",
+ "url": "https://huggingface.co/.../csv-test.parquet",
+ "filename": "csv-test.parquet",
+ "size": 4415,
+ },
+ {
+ "dataset": "lhoestq/demo1",
+ "config": "lhoestq--demo1",
+ "split": "train",
+ "url": "https://huggingface.co/.../csv-train.parquet",
+ "filename": "csv-train.parquet",
+ "size": 5038,
+ },
+ ],
+ "dataset_info": {
+ "description": "",
+ "citation": "",
+ "homepage": "",
+ "license": "",
+ "features": {},
+ "builder_name": "csv",
+ "config_name": "lhoestq--demo1",
+ "version": {},
+ "splits": {},
+ "download_checksums": {},
+ "download_size": 2340,
+ "dataset_size": 2464,
+ "size_in_bytes": 4804,
+ },
+ },
+ "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11",
+ "details": None,
+ "error_code": None,
+ "http_status": 200,
+ "job_runner_version": 3,
+ "progress": 1.0,
+ },
+ {
+ "config": "lhoestq--demo2",
+ "dataset": "lhoestq/demo2",
+ "kind": "config-parquet-and-info",
+ "split": None,
+ "content": {
+ "parquet_files": [
+ {
+ "dataset": "lhoestq/demo2",
+ "config": "lhoestq--demo2",
+ "split": "test",
+ "url": "https://huggingface.co/.../csv-test.parquet",
+ "filename": "csv-test.parquet",
+ "size": 4415,
+ },
+ {
+ "dataset": "lhoestq/demo2",
+ "config": "lhoestq--demo2",
+ "split": "train",
+ "url": "https://huggingface.co/.../csv-train.parquet",
+ "filename": "csv-train.parquet",
+ "size": 5038,
+ },
+ ],
+ "dataset_info": {
+ "description": "",
+ "citation": "",
+ "homepage": "",
+ "license": "",
+ "features": {},
+ "builder_name": "csv",
+ "config_name": "lhoestq--demo2",
+ "version": {},
+ "splits": {},
+ "download_checksums": {},
+ "download_size": 2340,
+ "dataset_size": 2464,
+ "size_in_bytes": 4804,
+ },
+ },
+ "dataset_git_revision": "87ecf163bedca9d80598b528940a9c4f99e14c11",
+ "details": None,
+ "error_code": None,
+ "http_status": 200,
+ "job_runner_version": 3,
+ "progress": 1.0,
+ },
+ {
+ "config": "lhoestq--error",
+ "dataset": "lhoestq/error",
+ "kind": "config-parquet-and-info",
+ "split": None,
+ "content": {"error": "Streaming is not supported for lhoestq/error"},
+ "dataset_git_revision": "ec3c8d414af3dfe600399f5e6ef2c682938676f3",
+ "details": {
+ "error": "Streaming is not supported for lhoestq/error",
+ "cause_exception": "TypeError",
+ "cause_message": "Streaming is not supported for lhoestq/error",
+ "cause_traceback": [
+ "Traceback (most recent call last):\n",
+ (
+ ' File "/src/services/worker/src/worker/job_manager.py", line 163, in process\n '
+ " job_result = self.job_runner.compute()\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 932, in compute\n compute_config_parquet_and_info_response(\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 825, in compute_config_parquet_and_info_response\n raise_if_not_supported(\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 367, in raise_if_not_supported\n raise_if_too_big_from_external_data_files(\n"
+ ),
+ (
+ ' File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line'
+ " 447, in raise_if_too_big_from_external_data_files\n "
+ " builder._split_generators(mock_dl_manager)\n"
+ ),
+ (
+ " File"
+ ' "/tmp/modules-cache/datasets_modules/.../error.py",'
+ ' line 190, in _split_generators\n raise TypeError("Streaming is not supported for'
+ ' lhoestq/error")\n'
+ ),
+ "TypeError: Streaming is not supported for lhoestq/error\n",
+ ],
+ },
+ "error_code": "UnexpectedError",
+ "http_status": 500,
+ "job_runner_version": 3,
+ "progress": None,
+ },
+ ]
+ )
+
+ migration = MigrationAddPartialToCacheResponse(
+ version="20230703110100",
+ description="add partial field to config-parquet-and-info",
+ )
+ migration.up()
+
+ assert_partial("lhoestq/demo1")
+ assert_partial("lhoestq/demo2")
+ assert_unchanged("lhoestq/error")
+
+ migration.down()
+ assert_unchanged("lhoestq/demo1")
+ assert_unchanged("lhoestq/demo2")
+ assert_unchanged("lhoestq/error")
+
+ db[CACHE_COLLECTION_RESPONSES].drop()
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py
index 46e66228..623f85a4 100644
--- a/libs/libcommon/src/libcommon/exceptions.py
+++ b/libs/libcommon/src/libcommon/exceptions.py
@@ -86,4 +85,0 @@ CacheableErrorCode = Literal[
- "DatasetTooBigFromDatasetsError",
- "DatasetTooBigFromHubError",
- "DatasetWithTooBigExternalFilesError",
- "DatasetWithTooManyExternalFilesError",
@@ -225,21 +220,0 @@ class DatasetRevisionNotFoundError(CacheableError):
-class DatasetTooBigFromDatasetsError(CacheableError):
- """Raised when the dataset size (sum of config sizes given by the datasets library) is too big."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromDatasetsError", cause, False)
-
-
-class DatasetTooBigFromHubError(CacheableError):
- """Raised when the dataset size (sum of files on the Hub) is too big."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetTooBigFromHubError", cause, False)
-
-
-class DatasetWithTooBigExternalFilesError(CacheableError):
- """Raised when the dataset size (sum of config sizes given by the datasets library) is too big."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooBigExternalFilesError", cause, True)
-
-
@@ -253,7 +227,0 @@ class DatasetWithTooManyConfigsError(CacheableError):
-class DatasetWithTooManyExternalFilesError(CacheableError):
- """Raised when the number of external data files of a dataset is too big."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyExternalFilesError", cause, True)
-
-
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py
index 17a6a41d..ca029d38 100644
--- a/services/worker/src/worker/dtos.py
+++ b/services/worker/src/worker/dtos.py
@@ -117,0 +118 @@ class ConfigParquetAndInfoResponse(TypedDict):
+ partial: bool
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 14da1e9e..880b991a 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -4 +4 @@
-import glob
+import functools
@@ -5,0 +6 @@ import logging
+import os
@@ -7 +8 @@ import re
-from functools import partial
+from contextlib import ExitStack
@@ -10 +11,14 @@ from pathlib import Path
-from typing import Any, List, Optional, Set, Tuple
+from types import TracebackType
+from typing import (
+ Any,
+ Callable,
+ Generator,
+ List,
+ Optional,
+ Set,
+ Tuple,
+ Type,
+ TypeVar,
+ Union,
+)
+from unittest.mock import patch
@@ -20,0 +35 @@ from datasets import DownloadConfig, Features, load_dataset_builder
+from datasets.arrow_writer import ParquetWriter
@@ -56,4 +70,0 @@ from libcommon.exceptions import (
- DatasetTooBigFromDatasetsError,
- DatasetTooBigFromHubError,
- DatasetWithTooBigExternalFilesError,
- DatasetWithTooManyExternalFilesError,
@@ -91,0 +103,8 @@ MAX_OPERATIONS_PER_COMMIT = 500
+T = TypeVar("T")
+
+
+def repo_file_rfilename_sort_key(repo_file: RepoFile) -> str:
+ if not isinstance(repo_file.rfilename, str): # check type for mypy
+ raise ValueError(f"Expected a string for repo_file.rfilename, but got a '{type(repo_file.rfilename)}'.")
+ return repo_file.rfilename
+
@@ -94 +113,3 @@ class ParquetFile:
- def __init__(self, local_file: str, local_dir: str, config: str):
+ def __init__(
+ self, local_file: str, local_dir: str, config: str, split: str, shard_idx: int, partial: bool = False
+ ):
@@ -96,0 +118,5 @@ class ParquetFile:
+ if shard_idx >= MAX_FILES_PER_DIRECTORY:
+ raise DatasetWithTooManyParquetFilesError(
+ "The dataset has too many parquet files and can't be uploaded in the parquet directory "
+ f"because it exceeds the maximum number of files per directory ({MAX_FILES_PER_DIRECTORY})."
+ )
@@ -99,0 +126,3 @@ class ParquetFile:
+ self.split = split
+ self.shard_idx = shard_idx
+ self.partial = partial
@@ -103 +132,5 @@ class ParquetFile:
- return f'{self.config}/{self.local_file.removeprefix(f"{self.local_dir}/")}'
+ if self.partial:
+ # Using 4 digits is ok since MAX_FILES_PER_DIRECTORY == 10_000
+ return f"{self.config}/partial/{self.split}/{self.shard_idx:04d}.parquet"
+ else:
+ return f'{self.config}/{self.local_file.removeprefix(f"{self.local_dir}/")}'
@@ -106 +139,2 @@ class ParquetFile:
-p = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>\w+(\.\w+)*?)(-[0-9]{5}-of-[0-9]{5})?.parquet")
+filename_pattern = re.compile("^[0-9]{4}\\.parquet$")
+legacy_filename_pattern = re.compile(r"(?P<builder>[\w-]+?)-(?P<split>\w+(\.\w+)*?)(-[0-9]{5}-of-[0-9]{5})?.parquet")
@@ -109,0 +144,12 @@ def parse_repo_filename(filename: str) -> Tuple[str, str]:
+ if not filename_pattern.match(os.path.basename(filename)):
+ return parse_legacy_repo_filename(filename)
+ parts = filename.split("/")
+ if len(parts) == 4 and parts[1] == "partial":
+ parts.pop(1)
+ if len(parts) != 3:
+ raise ValueError(f"Invalid filename: {filename}")
+ config, split, _ = parts
+ return config, split
+
+
+def parse_legacy_repo_filename(filename: str) -> Tuple[str, str]:
@@ -114 +160 @@ def parse_repo_filename(filename: str) -> Tuple[str, str]:
- m = p.match(fname)
+ m = legacy_filename_pattern.match(fname)
@@ -186 +232 @@ def is_parquet_builder_with_hub_files(builder: DatasetBuilder, hf_endpoint: str)
-def raise_if_too_big_from_hub(
+def _is_too_big_from_hub(
@@ -189 +235 @@ def raise_if_too_big_from_hub(
-) -> None:
+) -> bool:
@@ -199,6 +244,0 @@ def raise_if_too_big_from_hub(
- Returns:
- `None`
- Raises the following errors:
- - [`libcommon.exceptions.DatasetTooBigFromHubError`]
- If the dataset is too big to be converted to parquet, as measured by the sum of the repository
- files sizes given by the Hub.
@@ -207,5 +247 @@ def raise_if_too_big_from_hub(
- if dataset_size > max_dataset_size:
- raise DatasetTooBigFromHubError(
- f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes. "
- f"Current size of files on the hub is {dataset_size} bytes."
- )
+ return bool(dataset_size > max_dataset_size)
@@ -214 +250 @@ def raise_if_too_big_from_hub(
-def raise_if_too_big_from_datasets(
+def _is_too_big_from_datasets(
@@ -217 +253 @@ def raise_if_too_big_from_datasets(
-) -> None:
+) -> bool:
@@ -227,8 +262,0 @@ def raise_if_too_big_from_datasets(
- Returns:
- `None`
- Raises the following errors:
- - [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError)
- If the datasets.config.HF_ENDPOINT is not set to the expected value
- - [`libcommon.exceptions.DatasetTooBigFromDatasetsError`]
- If the dataset is too big to be converted to parquet, as measured by the sum of the configs
- sizes given by the datasets library.
@@ -237,6 +265 @@ def raise_if_too_big_from_datasets(
- if dataset_size > max_dataset_size:
- raise DatasetTooBigFromDatasetsError(
- f"The dataset is too big to be converted to Parquet. The size of the dataset ({dataset_size} B, as given"
- f" per the datasets library) exceeds the maximum supported size ({max_dataset_size} B). Please report the"
- " issue."
- )
+ return bool(dataset_size > max_dataset_size)
@@ -285 +308 @@ def raise_if_requires_manual_download(
-def raise_if_not_supported(
+def is_dataset_too_big(
@@ -292 +315 @@ def raise_if_not_supported(
-) -> None:
+) -> bool:
@@ -294,4 +317,4 @@ def raise_if_not_supported(
- Raise an error if the dataset is not supported:
- - if the dataset is in the list of blocked datasets
- - if the dataset cannot be accessed (does not exist, private)
- - if the dataset is too big, and not in the list of supported datasets
+ Check:
+ - the size of the dataset repository
+ - the size in dataset info
+ - the size and number of external files
@@ -320,14 +342,0 @@ def raise_if_not_supported(
- - [`libcommon.exceptions.DatasetManualDownloadError`]:
- If the dataset requires manual download.
- - [`libcommon.exceptions.DatasetRevisionNotFoundError`]
- If the revision does not exist or cannot be accessed using the token.
- - [`libcommon.exceptions.DatasetTooBigFromDatasetsError`]
- If the dataset is too big to be converted to parquet, as measured by the sum of the configs
- sizes given by the datasets library.
- - [`libcommon.exceptions.DatasetTooBigFromHubError`]
- If the dataset is too big to be converted to parquet, as measured by the sum of the repository
- files sizes given by the Hub.
- - [`libcommon.exceptions.DatasetWithTooManyExternalFilesError`]
- If the dataset has too many external files to be converted to parquet
- - [`libcommon.exceptions.DatasetWithTooBigExternalFilesError`]
- If the dataset is too big external files be converted to parquet
@@ -352,15 +361,12 @@ def raise_if_not_supported(
- raise_if_requires_manual_download(
- builder=builder,
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- )
- raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size)
- raise_if_too_big_from_external_data_files(
- builder=builder,
- max_dataset_size=max_dataset_size,
- max_external_data_files=max_external_data_files,
- hf_token=hf_token,
- )
- raise_if_too_big_from_datasets(
- builder.info,
- max_dataset_size=max_dataset_size,
+ return (
+ _is_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size)
+ or _is_too_big_from_datasets(
+ builder.info,
+ max_dataset_size=max_dataset_size,
+ )
+ or _is_too_big_from_external_data_files(
+ builder=builder,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ hf_token=hf_token,
+ )
@@ -426 +432 @@ class _MockStreamingDownloadManager(StreamingDownloadManager): # type: ignore
-def raise_if_too_big_from_external_data_files(
+def _is_too_big_from_external_data_files(
@@ -428 +434 @@ def raise_if_too_big_from_external_data_files(
-) -> None:
+) -> bool:
@@ -432 +438 @@ def raise_if_too_big_from_external_data_files(
- return
+ return False
@@ -488,4 +494 @@ def raise_if_too_big_from_external_data_files(
- raise DatasetWithTooManyExternalFilesError(
- f"The conversion to parquet is limited to datasets with less than {max_external_data_files} files. "
- f"However it uses {len(ext_data_files)} data files."
- )
+ return True
@@ -496 +499 @@ def raise_if_too_big_from_external_data_files(
- get_size = partial(_request_size, hf_token=hf_token)
+ get_size = functools.partial(_request_size, hf_token=hf_token)
@@ -500,6 +503,2 @@ def raise_if_too_big_from_external_data_files(
- if total_size > max_dataset_size:
- raise DatasetWithTooBigExternalFilesError(
- f"The conversion to parquet is limited to datasets under {max_dataset_size} bytes."
- f" However {i + 1} data files of {len(ext_data_files)} are already bigger than"
- f" {total_size} bytes."
- )
+ return total_size > max_dataset_size
+ return False
@@ -542,0 +542 @@ def raise_if_too_big_from_external_data_files(
+ return False
@@ -648 +648 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None:
- partial(retry_get_parquet_file_and_size, fs=fs, hf_token=hf_token),
+ functools.partial(retry_get_parquet_file_and_size, fs=fs, hf_token=hf_token),
@@ -668,0 +669,160 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None:
+class limit_parquet_writes:
+ """
+ Context manager that limits the number of bytes a `DatasetBuilder` can write to parquet.
+
+ It works by monitoring the calls to `pq.ParquetWriter.write_table` and stopping
+ the `GeneratorBasedBuilder._generate_examples` and `ArrowBasedBuilder._generate_tables`
+ generators once we reach the maximum number of bytes.
+
+ Since the generator is stopped after we reach the maximum number of bytes, the actual
+ number of bytes generated might be slightly higher than the requested limit.
+
+ Example of usage:
+
+ ```python
+ builder = load_dataset_builder("squad")
+ max_dataset_size = 10_000_000
+ with limit_parquet_writes(builder, max_dataset_size=max_dataset_size) as limiter:
+ builder.download_and_prepare(file_format="parquet")
+ assert builder.info.dataset_size == limiter.total_bytes < max_dataset_size + epsilon
+ ```
+
+ The limiter is usually used with a `StreamingDownloadManager` to not have to download
+ the full dataset:
+
+ ```python
+ builder = load_dataset_builder("squad")
+ max_dataset_size = 10_000_000
+ dl_manager = StreamingDownloadManager(...)
+ for split_generator in builder._split_generators(dl_manager):
+ with limit_parquet_writes(builder, max_dataset_size=max_dataset_size):
+ builder._prepare_split(split_generator=split_generator, file_format="parquet")
+ ```
+ """
+
+ def __init__(
+ self,
+ builder: Union[datasets.builder.GeneratorBasedBuilder, datasets.builder.ArrowBasedBuilder],
+ max_dataset_size: int,
+ ) -> None:
+ self.total_bytes = 0
+ self.builder = builder
+ self.max_dataset_size = max_dataset_size
+ self.exit_stack = ExitStack()
+
+ def __enter__(self) -> "limit_parquet_writes":
+ limiter = self
+
+ class _TrackedParquetWriter(pq.ParquetWriter): # type: ignore
+ """Count on-the-fly how many bytes are written"""
+
+ def track_write_table(self, pa_table: pa.Table) -> None:
+ limiter.total_bytes += pa_table.nbytes
+
+ def write_table(self, pa_table: pa.Table, row_group_size: Optional[int] = None) -> None:
+ self.track_write_table(pa_table)
+ super().write_table(pa_table, row_group_size=row_group_size)
+
+ def limited_generator(
+ generator: Callable[..., Generator[T, None, None]]
+ ) -> Callable[..., Generator[T, None, None]]:
+ """Stop the underlying generator once we reach the maximum dataset size"""
+
+ @functools.wraps(generator)
+ def wrapped(*args: Any, **kwargs: Any) -> Generator[T, None, None]:
+ for item in generator(*args, **kwargs):
+ if limiter.total_bytes < limiter.max_dataset_size:
+ yield item
+ else:
+ break
+
+ return wrapped
+
+ self.exit_stack.enter_context(patch.object(ParquetWriter, "_WRITER_CLASS", _TrackedParquetWriter))
+ if isinstance(self.builder, datasets.builder.GeneratorBasedBuilder):
+ self.exit_stack.enter_context(
+ patch.object(self.builder, "_generate_examples", limited_generator(self.builder._generate_examples))
+ )
+ else:
+ self.exit_stack.enter_context(
+ patch.object(self.builder, "_generate_tables", limited_generator(self.builder._generate_tables))
+ )
+ return self
+
+ def __exit__(
+ self,
+ exc_type: Optional[Type[BaseException]],
+ exc_value: Optional[BaseException],
+ traceback: Optional[TracebackType],
+ ) -> None:
+ return self.exit_stack.close()
+
+
+def list_generated_parquet_files(builder: DatasetBuilder, partial: bool = False) -> List[ParquetFile]:
+ """List the parquet files generated by `builder.download_and_prepare` in the `builder.cache_dir`."""
+ if not builder.info.splits:
+ raise EmptyDatasetError("No split found after generating parquet files")
+ split_dict = builder.info.splits
+ local_parquet_files: List[ParquetFile] = []
+ for split, split_info in split_dict.items():
+ # We know the `datasets` library uses a template for the shards names:
+ # - {builder.name}-{split}.parquet if there is only one shard
+ # - {builder.name}-{split}-{shard_idx:05d}-of-{num_shards:05d}.parquet otherwise
+ num_shards = len(split_info.shard_lengths) if isinstance(split_info.shard_lengths, list) else 1
+ fname_prefix = f"{builder.name}-{split}"
+ local_parquet_files.extend(
+ [
+ ParquetFile(
+ local_file=os.path.join(
+ builder.cache_dir,
+ fname_prefix
+ + (f"-{shard_idx:05d}-of-{num_shards:05d}.parquet" if num_shards > 1 else ".parquet"),
+ ),
+ local_dir=builder.cache_dir,
+ config=builder.config.name,
+ split=split,
+ shard_idx=shard_idx,
+ partial=partial,
+ )
+ for shard_idx in range(num_shards)
+ ]
+ )
+ return local_parquet_files
+
+
+def stream_convert_to_parquet(builder: DatasetBuilder, max_dataset_size: int) -> Tuple[List[CommitOperationAdd], bool]:
+ """Stream and prepare the dataset as parquet files and fills the builder info."""
+ writer_batch_size = get_writer_batch_size(builder.info)
+ if writer_batch_size is not None and (
+ builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size
+ ):
+ builder._writer_batch_size = writer_batch_size
+ dl_manager = StreamingDownloadManager(
+ base_path=builder.base_path,
+ download_config=DownloadConfig(use_auth_token=builder.use_auth_token, storage_options=builder.storage_options),
+ dataset_name=builder.name,
+ data_dir=builder.config.data_dir,
+ )
+ os.makedirs(builder.cache_dir, exist_ok=True)
+ split_dict = SplitDict(dataset_name=builder.name)
+ splits_generators = {sg.name: sg for sg in builder._split_generators(dl_manager)}
+ partial = False
+ for split in splits_generators:
+ split_dict.add(splits_generators[split].split_info)
+ with limit_parquet_writes(builder, max_dataset_size=max_dataset_size) as limiter:
+ builder._prepare_split(split_generator=splits_generators[split], file_format="parquet")
+ partial = partial or limiter.total_bytes >= max_dataset_size
+ builder.info.splits = split_dict
+ builder.info.dataset_size = sum(split.num_bytes for split in builder.info.splits.values())
+ builder.info.download_size = None
+ builder.info.size_in_bytes = None
+
+ # send the files to the target revision
+ local_parquet_files = list_generated_parquet_files(builder, partial=partial)
+ parquet_operations: List[CommitOperationAdd] = [
+ CommitOperationAdd(path_in_repo=parquet_file.path_in_repo, path_or_fileobj=parquet_file.local_file)
+ for parquet_file in local_parquet_files
+ ]
+ return parquet_operations, partial
+
+
@@ -670 +830 @@ def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]:
- """Download and prepare the dataset as parquet files and fills the builder info"""
+ """Download and prepare the dataset as parquet files and fills the builder info."""
@@ -680,4 +840 @@ def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]:
- local_parquet_files = [
- ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=builder.config.name)
- for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet")
- ]
+ local_parquet_files = list_generated_parquet_files(builder)
@@ -1016,0 +1174 @@ def compute_config_parquet_and_info_response(
+ partial = False
@@ -1020,0 +1179,5 @@ def compute_config_parquet_and_info_response(
+ raise_if_requires_manual_download(
+ builder=builder,
+ hf_endpoint=hf_endpoint,
+ hf_token=hf_token,
+ )
@@ -1024,10 +1187,11 @@ def compute_config_parquet_and_info_response(
- if dataset not in supported_datasets:
- raise_if_not_supported(
- dataset_info=dataset_info,
- builder=builder,
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- max_dataset_size=max_dataset_size,
- max_external_data_files=max_external_data_files,
- )
- parquet_operations = convert_to_parquet(builder)
+ if is_dataset_too_big(
+ dataset_info=dataset_info,
+ builder=builder,
+ hf_endpoint=hf_endpoint,
+ hf_token=hf_token,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ ):
+ parquet_operations, partial = stream_convert_to_parquet(builder, max_dataset_size=max_dataset_size)
+ else:
+ parquet_operations = convert_to_parquet(builder)
@@ -1071,0 +1236 @@ def compute_config_parquet_and_info_response(
+ repo_files.sort(key=repo_file_rfilename_sort_key)
@@ -1088,0 +1254 @@ def compute_config_parquet_and_info_response(
+ partial=partial,
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 11e9d9a0..1ceb456a 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -374,0 +375,18 @@ def create_dataset_info_response_for_csv(dataset: str, config: str) -> Any:
+def create_dataset_info_response_for_partially_generated_big_csv(config: str) -> Any:
+ # Dataset is partially converted to parquet: the first 10KB instead of the full 5MB
+ # Missing fields:
+ # - download_size: not applicable, because the dataset is generated using partially downloaded files
+ return {
+ "description": "",
+ "citation": "",
+ "homepage": "",
+ "license": "",
+ "features": BIG_cols,
+ "builder_name": "csv",
+ "config_name": config,
+ "version": {"version_str": "0.0.0", "major": 0, "minor": 0, "patch": 0},
+ "splits": {"train": {"name": "train", "num_bytes": 12380, "num_examples": 10, "dataset_name": "csv"}},
+ "dataset_size": 12380,
+ }
+
+
@@ -419 +437,3 @@ def create_parquet_and_info_response(
- dataset: str, data_type: Literal["csv", "audio", "big_parquet", "big_parquet_no_info"]
+ dataset: str,
+ data_type: Literal["csv", "big-csv", "audio", "big_parquet", "big_parquet_no_info"],
+ partial: bool = False,
@@ -423,2 +443,13 @@ def create_parquet_and_info_response(
- filename = "csv-train.parquet" if data_type == "csv" else "parquet-train.parquet"
- size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE if data_type == "audio" else BIG_PARQUET_FILE
+ if partial:
+ filename = "0000.parquet"
+ else:
+ filename = "csv-train.parquet" if "csv" in data_type else "parquet-train.parquet"
+ size = (
+ CSV_PARQUET_SIZE
+ if data_type == "csv"
+ else PARTIAL_CSV_PARQUET_SIZE
+ if data_type == "big-csv"
+ else AUDIO_PARQUET_SIZE
+ if data_type == "audio"
+ else BIG_PARQUET_FILE
+ )
@@ -427,0 +459,2 @@ def create_parquet_and_info_response(
+ else create_dataset_info_response_for_partially_generated_big_csv(config)
+ if data_type == "big-csv"
@@ -441 +474,3 @@ def create_parquet_and_info_response(
- repo_id=f"datasets/{dataset}", revision="refs%2Fconvert%2Fparquet", filename=f"{config}/{filename}"
+ repo_id=f"datasets/{dataset}",
+ revision="refs%2Fconvert%2Fparquet",
+ filename=f"{config}/partial/{split}/{filename}" if partial else f"{config}/{filename}",
@@ -447,0 +483 @@ def create_parquet_and_info_response(
+ "partial": partial,
@@ -451,0 +488 @@ CSV_PARQUET_SIZE = 1_866
+PARTIAL_CSV_PARQUET_SIZE = 8_188
@@ -748 +785,3 @@ def hub_reponses_big_csv(hub_public_big_csv: str) -> HubDatasetTest:
- "parquet_and_info_response": None,
+ "parquet_and_info_response": create_parquet_and_info_response(
+ dataset=hub_public_big_csv, data_type="big-csv", partial=True
+ ),
diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py
index c7db50e4..c2e9f007 100644
--- a/services/worker/tests/job_runners/config/test_parquet.py
+++ b/services/worker/tests/job_runners/config/test_parquet.py
@@ -84,0 +85 @@ def get_job_runner(
+ partial=False,
@@ -148,0 +150 @@ def get_job_runner(
+ partial=False,
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index ec3bc9c2..6c4f72f2 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -4,0 +5 @@ import io
+import os
@@ -11 +12,2 @@ from pathlib import Path
-from typing import Any, Callable, Iterator, List, Optional, Set, TypedDict
+from typing import Any, Callable, Dict, Iterator, List, Optional, Set, TypedDict
+from unittest.mock import patch
@@ -13,0 +16 @@ import datasets.builder
+import datasets.config
@@ -15,0 +19 @@ import pandas as pd
+import pyarrow.parquet as pq
@@ -18,0 +23,3 @@ from datasets import Audio, Features, Image, Value, load_dataset_builder
+from datasets.packaged_modules.generator.generator import (
+ Generator as ParametrizedGeneratorBasedBuilder,
+)
@@ -25,4 +31,0 @@ from libcommon.exceptions import (
- DatasetTooBigFromDatasetsError,
- DatasetTooBigFromHubError,
- DatasetWithTooBigExternalFilesError,
- DatasetWithTooManyExternalFilesError,
@@ -40,0 +44,3 @@ from worker.job_runners.config.parquet_and_info import (
+ _is_too_big_from_datasets,
+ _is_too_big_from_external_data_files,
+ _is_too_big_from_hub,
@@ -43,0 +50,2 @@ from worker.job_runners.config.parquet_and_info import (
+ limit_parquet_writes,
+ list_generated_parquet_files,
@@ -47,3 +55 @@ from worker.job_runners.config.parquet_and_info import (
- raise_if_too_big_from_datasets,
- raise_if_too_big_from_external_data_files,
- raise_if_too_big_from_hub,
+ stream_convert_to_parquet,
@@ -112 +118 @@ def assert_content_is_equal(content: Any, expected: Any) -> None:
- assert set(content) == {"parquet_files", "dataset_info"}, content
+ assert set(content) == {"parquet_files", "dataset_info", "partial"}, content
@@ -120,0 +127 @@ def assert_content_is_equal(content: Any, expected: Any) -> None:
+ assert content["partial"] == expected["partial"], content
@@ -241 +248 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_
- "name,raises",
+ "name,expected",
@@ -244 +251 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_
-def test_raise_if_too_big_from_hub(
+def test__is_too_big_from_hub(
@@ -248 +255 @@ def test_raise_if_too_big_from_hub(
- raises: bool,
+ expected: bool,
@@ -259,9 +266,4 @@ def test_raise_if_too_big_from_hub(
- if raises:
- with pytest.raises(DatasetTooBigFromHubError):
- raise_if_too_big_from_hub(
- dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size
- )
- else:
- raise_if_too_big_from_hub(
- dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size
- )
+ assert (
+ _is_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=app_config.parquet_and_info.max_dataset_size)
+ == expected
+ )
@@ -271 +273 @@ def test_raise_if_too_big_from_hub(
- "name,raises",
+ "name,expected",
@@ -274 +276 @@ def test_raise_if_too_big_from_hub(
-def test_raise_if_too_big_from_datasets(
+def test__is_too_big_from_datasets(
@@ -278 +280 @@ def test_raise_if_too_big_from_datasets(
- raises: bool,
+ expected: bool,
@@ -283,8 +285,2 @@ def test_raise_if_too_big_from_datasets(
- if raises:
- with pytest.raises(DatasetTooBigFromDatasetsError):
- raise_if_too_big_from_datasets(
- info=builder.info,
- max_dataset_size=app_config.parquet_and_info.max_dataset_size,
- )
- else:
- raise_if_too_big_from_datasets(
+ assert (
+ _is_too_big_from_datasets(
@@ -293,0 +290,2 @@ def test_raise_if_too_big_from_datasets(
+ == expected
+ )
@@ -297 +295 @@ def test_raise_if_too_big_from_datasets(
- "max_dataset_size,max_external_data_files,raises",
+ "max_dataset_size,max_external_data_files,expected",
@@ -303 +301 @@ def test_raise_if_too_big_from_datasets(
-def test_raise_if_too_big_external_files(
+def test__is_too_big_external_files(
@@ -305 +303 @@ def test_raise_if_too_big_external_files(
- raises: bool,
+ expected: bool,
@@ -312,10 +310,2 @@ def test_raise_if_too_big_external_files(
- if raises:
- with pytest.raises(DatasetWithTooBigExternalFilesError):
- raise_if_too_big_from_external_data_files(
- builder=external_files_dataset_builder,
- hf_token=app_config.common.hf_token,
- max_dataset_size=max_dataset_size,
- max_external_data_files=max_external_data_files,
- )
- else:
- raise_if_too_big_from_external_data_files(
+ assert (
+ _is_too_big_from_external_data_files(
@@ -326,0 +317,2 @@ def test_raise_if_too_big_external_files(
+ == expected
+ )
@@ -330 +322 @@ def test_raise_if_too_big_external_files(
- "max_dataset_size,max_external_data_files,raises",
+ "max_dataset_size,max_external_data_files,expected",
@@ -338 +330 @@ def test_raise_if_too_many_external_files(
- raises: bool,
+ expected: bool,
@@ -345,10 +337,2 @@ def test_raise_if_too_many_external_files(
- if raises:
- with pytest.raises(DatasetWithTooManyExternalFilesError):
- raise_if_too_big_from_external_data_files(
- builder=external_files_dataset_builder,
- hf_token=app_config.common.hf_token,
- max_dataset_size=max_dataset_size,
- max_external_data_files=max_external_data_files,
- )
- else:
- raise_if_too_big_from_external_data_files(
+ assert (
+ _is_too_big_from_external_data_files(
@@ -359,0 +344,2 @@ def test_raise_if_too_many_external_files(
+ == expected
+ )
@@ -387 +373 @@ def test_supported_if_big_parquet(
-def test_not_supported_if_big_non_parquet(
+def test_partially_converted_if_big_non_parquet(
@@ -403,3 +389,14 @@ def test_not_supported_if_big_non_parquet(
- with pytest.raises(CustomError) as e:
- job_runner.compute()
- assert e.typename == "DatasetTooBigFromHubError"
+ from datasets.packaged_modules.csv.csv import CsvConfig
+
+ # Set a small chunk size to yield more than one Arrow Table in _generate_tables
+ # to be able to stop the generation mid-way.
+ with patch.object(CsvConfig, "pd_read_csv_kwargs", {"chunksize": 10}):
+ response = job_runner.compute()
+ assert response
+ content = response.content
+ assert content
+ assert len(content["parquet_files"]) == 1
+ assert_content_is_equal(content, hub_reponses_big_csv["parquet_and_info_response"])
+ # dataset is partially generated
+ assert content["parquet_files"][0]["size"] < app_config.parquet_and_info.max_dataset_size
+ assert content["parquet_files"][0]["url"].endswith("/partial/train/0000.parquet")
@@ -597,0 +595,3 @@ def test_previous_step_error(
+ ("config/partial/split/0000.parquet", "split", "config", False),
+ ("config/partial/split.with.dots/0000.parquet", "split.with.dots", "config", False),
+ ("config/partial/toomanyzeros/00000.parquet", "toomanyzeros", "config", True),
@@ -828,0 +829,57 @@ def test_get_delete_operations(
+
+
[email protected](
+ "max_dataset_size,expected_num_shards",
+ [
+ (1, 1),
+ (150, 2),
+ (300, 4),
+ (9999999, 10),
+ ],
+)
+def test_stream_convert_to_parquet(
+ csv_path: str, max_dataset_size: int, expected_num_shards: int, tmp_path: Path
+) -> None:
+ num_data_files = 10
+ builder = load_dataset_builder(
+ "csv",
+ data_files={"train": [csv_path] * num_data_files},
+ cache_dir=str(tmp_path / f"test_stream_convert_to_parquet-{max_dataset_size=}"),
+ )
+ with patch("worker.job_runners.config.parquet_and_info.get_writer_batch_size", lambda ds_config_info: 1):
+ with patch.object(datasets.config, "MAX_SHARD_SIZE", 1):
+ parquet_operations, partial = stream_convert_to_parquet(builder, max_dataset_size=max_dataset_size)
+ num_shards = len(parquet_operations)
+ assert num_shards == expected_num_shards
+ assert partial == (expected_num_shards < num_data_files)
+ assert all(isinstance(op.path_or_fileobj, str) for op in parquet_operations)
+ parquet_files = list_generated_parquet_files(builder, partial=partial)
+ assert len(parquet_files) == expected_num_shards
+ assert all(os.path.isfile(parquet_file.local_file) for parquet_file in parquet_files)
+ one_sample_max_size = 100
+ expected_max_dataset_size = max_dataset_size + one_sample_max_size
+ assert (
+ sum(pq.ParquetFile(parquet_file.local_file).read().nbytes for parquet_file in parquet_files)
+ < expected_max_dataset_size
+ )
+
+
+def test_limit_parquet_writes(tmp_path: Path) -> None:
+ num_examples = 0
+
+ def long_generator() -> Iterator[Dict[str, int]]:
+ nonlocal num_examples
+ for i in range(10_000_000):
+ yield {"foo": i}
+ num_examples += 1
+
+ one_sample_size = 8
+ max_dataset_size = 50_000
+ expected_max_dataset_size = max_dataset_size + datasets.config.DEFAULT_MAX_BATCH_SIZE * one_sample_size
+ expected_max_num_examples = 1 + max_dataset_size // one_sample_size + datasets.config.DEFAULT_MAX_BATCH_SIZE
+ cache_dir = str(tmp_path / "test_limit_parquet_writes_cache_dir")
+ builder = ParametrizedGeneratorBasedBuilder(generator=long_generator, cache_dir=cache_dir)
+ with limit_parquet_writes(builder, max_dataset_size=max_dataset_size) as limiter:
+ builder.download_and_prepare(file_format="parquet")
+ assert builder.info.dataset_size == limiter.total_bytes <= expected_max_dataset_size
+ assert builder.info.splits["train"].num_examples == num_examples < expected_max_num_examples
|
|
2996e47b6f4abca258b1a59d942dd1019aac2c60
|
Sylvain Lesage
| 2023-07-03T15:26:54 |
feat: 🎸 use Normal priority only for API (#1470)
|
diff --git a/libs/libcommon/src/libcommon/operations.py b/libs/libcommon/src/libcommon/operations.py
index 37ee3ee2..0797eef8 100644
--- a/libs/libcommon/src/libcommon/operations.py
+++ b/libs/libcommon/src/libcommon/operations.py
@@ -16 +16 @@ def backfill_dataset(
- priority: Priority = Priority.NORMAL,
+ priority: Priority = Priority.LOW,
@@ -25 +25 @@ def backfill_dataset(
- priority (Priority, optional): The priority of the job. Defaults to Priority.NORMAL.
+ priority (Priority, optional): The priority of the job. Defaults to Priority.LOW.
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 40898dcf..57031369 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -129 +129 @@ class JobDocument(Document):
- priority (`Priority`, optional): The priority of the job. Defaults to Priority.NORMAL.
+ priority (`Priority`, optional): The priority of the job. Defaults to Priority.LOW.
@@ -169 +169 @@ class JobDocument(Document):
- priority = EnumField(Priority, default=Priority.NORMAL)
+ priority = EnumField(Priority, default=Priority.LOW)
@@ -350 +350 @@ class Queue:
- priority: Priority = Priority.NORMAL,
+ priority: Priority = Priority.LOW,
@@ -363 +363 @@ class Queue:
- priority (`Priority`, optional): The priority of the job. Defaults to Priority.NORMAL.
+ priority (`Priority`, optional): The priority of the job. Defaults to Priority.LOW.
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index fc81b103..792905ee 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -130,2 +130,2 @@ def test_priority_logic_creation_order() -> None:
- check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL)
- check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL)
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.LOW)
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.LOW)
@@ -143,2 +143,2 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None:
- check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL)
- check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.NORMAL)
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.LOW)
+ check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.LOW)
@@ -146 +146 @@ def test_priority_logic_started_jobs_per_dataset_order() -> None:
- check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL)
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.LOW)
@@ -161,6 +161,2 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None:
- check_job(
- queue=queue, expected_dataset="org1/dataset1", expected_split="split1", expected_priority=Priority.NORMAL
- )
- check_job(
- queue=queue, expected_dataset="org2/dataset2", expected_split="split1", expected_priority=Priority.NORMAL
- )
+ check_job(queue=queue, expected_dataset="org1/dataset1", expected_split="split1", expected_priority=Priority.LOW)
+ check_job(queue=queue, expected_dataset="org2/dataset2", expected_split="split1", expected_priority=Priority.LOW)
@@ -168,6 +164,2 @@ def test_priority_logic_started_jobs_per_namespace_order() -> None:
- check_job(
- queue=queue, expected_dataset="no_org_dataset3", expected_split="split1", expected_priority=Priority.NORMAL
- )
- check_job(
- queue=queue, expected_dataset="org1/dataset2", expected_split="split1", expected_priority=Priority.NORMAL
- )
+ check_job(queue=queue, expected_dataset="no_org_dataset3", expected_split="split1", expected_priority=Priority.LOW)
+ check_job(queue=queue, expected_dataset="org1/dataset2", expected_split="split1", expected_priority=Priority.LOW)
@@ -188 +179,0 @@ def test_priority_logic_priority_order() -> None:
- priority=Priority.LOW,
@@ -190 +181,8 @@ def test_priority_logic_priority_order() -> None:
- queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1")
+ queue.add_job(
+ job_type=test_type,
+ dataset="dataset2",
+ revision=test_revision,
+ config="config",
+ split="split1",
+ priority=Priority.NORMAL,
+ )
diff --git a/services/admin/src/admin/routes/dataset_backfill.py b/services/admin/src/admin/routes/dataset_backfill.py
index 4d388344..3604391f 100644
--- a/services/admin/src/admin/routes/dataset_backfill.py
+++ b/services/admin/src/admin/routes/dataset_backfill.py
@@ -53 +53 @@ def create_dataset_backfill_endpoint(
- dataset_orchestrator.backfill(revision=dataset_git_revision, priority=Priority.NORMAL)
+ dataset_orchestrator.backfill(revision=dataset_git_revision, priority=Priority.LOW)
diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py
index 439f4dde..3f37f937 100644
--- a/services/admin/src/admin/routes/force_refresh.py
+++ b/services/admin/src/admin/routes/force_refresh.py
@@ -10,0 +11 @@ from libcommon.queue import Queue
+from libcommon.utils import Priority
@@ -62 +63,8 @@ def create_force_refresh_endpoint(
- Queue().add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split)
+ Queue().add_job(
+ job_type=job_type,
+ dataset=dataset,
+ revision=revision,
+ config=config,
+ split=split,
+ priority=Priority.LOW,
+ )
|
|
dc604553180cd0e30ce2cce7a30e9d1b6f67fe65
|
Andrea Francis Soria Jimenez
| 2023-07-03T12:56:22 |
Adding retry to create duckdb index commit (#1466)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 8e82aa94..14da1e9e 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -80 +80,7 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas
-from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url, retry
+from worker.utils import (
+ HF_HUB_HTTP_ERROR_RETRY_SLEEPS,
+ LOCK_GIT_BRANCH_RETRY_SLEEPS,
+ create_branch,
+ hf_hub_url,
+ retry,
+)
@@ -753 +758,0 @@ def create_commits(
- sleeps = [1, 1, 1, 10, 10, 10]
@@ -757 +762 @@ def create_commits(
- retry_create_commit = retry(on=[HfHubHTTPError], sleeps=sleeps)(hf_api.create_commit)
+ retry_create_commit = retry(on=[HfHubHTTPError], sleeps=HF_HUB_HTTP_ERROR_RETRY_SLEEPS)(hf_api.create_commit)
@@ -772 +777 @@ def create_commits(
- f" {len(sleeps)} attempts)."
+ f" {len(HF_HUB_HTTP_ERROR_RETRY_SLEEPS)} attempts)."
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py
index 1ed21f5d..97fbb673 100644
--- a/services/worker/src/worker/job_runners/split/duckdb_index.py
+++ b/services/worker/src/worker/job_runners/split/duckdb_index.py
@@ -15 +15 @@ from huggingface_hub.hf_api import HfApi
-from huggingface_hub.utils._errors import RepositoryNotFoundError
+from huggingface_hub.utils._errors import HfHubHTTPError, RepositoryNotFoundError
@@ -19,0 +20 @@ from libcommon.exceptions import (
+ CreateCommitError,
@@ -38 +39,7 @@ from worker.job_runners.split.split_job_runner import SplitJobRunnerWithCache
-from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url
+from worker.utils import (
+ HF_HUB_HTTP_ERROR_RETRY_SLEEPS,
+ LOCK_GIT_BRANCH_RETRY_SLEEPS,
+ create_branch,
+ hf_hub_url,
+ retry,
+)
@@ -187,7 +194,2 @@ def compute_index_rows(
- committer_hf_api.create_commit(
- repo_id=dataset,
- repo_type=DATASET_TYPE,
- revision=target_revision,
- operations=delete_operations + add_operations,
- commit_message=commit_message,
- parent_commit=target_dataset_info.sha,
+ retry_create_commit = retry(on=[HfHubHTTPError], sleeps=HF_HUB_HTTP_ERROR_RETRY_SLEEPS)(
+ committer_hf_api.create_commit
@@ -194,0 +197,20 @@ def compute_index_rows(
+ try:
+ retry_create_commit(
+ repo_id=dataset,
+ repo_type=DATASET_TYPE,
+ revision=target_revision,
+ operations=delete_operations + add_operations,
+ commit_message=commit_message,
+ parent_commit=target_dataset_info.sha,
+ )
+ except RuntimeError as e:
+ if e.__cause__ and isinstance(e.__cause__, HfHubHTTPError):
+ raise CreateCommitError(
+ message=(
+ f"Commit {commit_message} could not be created on the Hub (after"
+ f" {len(HF_HUB_HTTP_ERROR_RETRY_SLEEPS)} attempts)."
+ ),
+ cause=e.__cause__,
+ ) from e.__cause__
+ raise e
+
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index c3b985d8..56308fc3 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -338,0 +339 @@ DATASET_TYPE = "dataset"
+HF_HUB_HTTP_ERROR_RETRY_SLEEPS = [1, 1, 1, 10, 10, 10]
|
|
812df7632bd074bf1fd46ec2e25e35ee5e2492ca
|
geethika-123
| 2023-07-01T15:49:32 |
Rename classes to indicate inheritance (#1428)
|
diff --git a/jobs/cache_maintenance/src/cache_maintenance/metrics.py b/jobs/cache_maintenance/src/cache_maintenance/metrics.py
index 2eeb4f0b..2aae8f11 100644
--- a/jobs/cache_maintenance/src/cache_maintenance/metrics.py
+++ b/jobs/cache_maintenance/src/cache_maintenance/metrics.py
@@ -6 +6 @@ import logging
-from libcommon.metrics import CacheTotalMetric, JobTotalMetric
+from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument
@@ -17 +17 @@ def collect_metrics(processing_graph: ProcessingGraph) -> None:
- JobTotalMetric.objects(queue=processing_step.job_type, status=status).upsert_one(total=total)
+ JobTotalMetricDocument.objects(queue=processing_step.job_type, status=status).upsert_one(total=total)
@@ -21 +21 @@ def collect_metrics(processing_graph: ProcessingGraph) -> None:
- CacheTotalMetric.objects(
+ CacheTotalMetricDocument.objects(
diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py
index 7f8cf4bd..aed47435 100644
--- a/jobs/cache_maintenance/tests/test_collect_metrics.py
+++ b/jobs/cache_maintenance/tests/test_collect_metrics.py
@@ -6 +6 @@ from http import HTTPStatus
-from libcommon.metrics import CacheTotalMetric, JobTotalMetric
+from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument
@@ -41 +41 @@ def test_collect_metrics() -> None:
- cache_metrics = CacheTotalMetric.objects()
+ cache_metrics = CacheTotalMetricDocument.objects()
@@ -45 +45 @@ def test_collect_metrics() -> None:
- job_metrics = JobTotalMetric.objects()
+ job_metrics = JobTotalMetricDocument.objects()
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py
index 46cc56f5..a87c27df 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20221117223000_cache_generic_response.py
@@ -7 +7 @@ import logging
-from libcommon.simple_cache import CachedResponse
+from libcommon.simple_cache import CachedResponseDocument
@@ -95 +95 @@ class MigrationMoveToGenericCachedResponse(Migration):
- check_documents(DocCls=CachedResponse, sample_size=10)
+ check_documents(DocCls=CachedResponseDocument, sample_size=10)
@@ -106 +106 @@ class MigrationMoveToGenericCachedResponse(Migration):
- cached_responses_count = CachedResponse.objects.count()
+ cached_responses_count = CachedResponseDocument.objects.count()
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py
index 28e4ddf5..b7cc73e2 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230126164900_queue_job_add_priority.py
@@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS
-from libcommon.queue import Job
+from libcommon.queue import JobDocument
@@ -30 +30 @@ class MigrationAddPriorityToJob(Migration):
- check_documents(DocCls=Job, sample_size=10)
+ check_documents(DocCls=JobDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py
index a30d5236..d63eaea5 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309123100_cache_add_progress.py
@@ -7 +7 @@ from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_AL
-from libcommon.simple_cache import CachedResponse
+from libcommon.simple_cache import CachedResponseDocument
@@ -30 +30 @@ class MigrationAddProgressToCacheResponse(Migration):
- check_documents(DocCls=CachedResponse, sample_size=10)
+ check_documents(DocCls=CachedResponseDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py
index e052a46c..6ea4e74e 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230309141600_cache_add_job_runner_version.py
@@ -7 +7 @@ from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_AL
-from libcommon.simple_cache import CachedResponse
+from libcommon.simple_cache import CachedResponseDocument
@@ -46 +46 @@ class MigrationAddJobRunnerVersionToCacheResponse(Migration):
- check_documents(DocCls=CachedResponse, sample_size=10)
+ check_documents(DocCls=CachedResponseDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py
index 976b3e7d..28a9b9b1 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230313164200_cache_remove_worker_version.py
@@ -7 +7 @@ from libcommon.constants import CACHE_COLLECTION_RESPONSES, CACHE_MONGOENGINE_AL
-from libcommon.simple_cache import CachedResponse
+from libcommon.simple_cache import CachedResponseDocument
@@ -27 +27 @@ class MigrationRemoveWorkerVersionFromCachedResponse(Migration):
- check_documents(DocCls=CachedResponse, sample_size=10)
+ check_documents(DocCls=CachedResponseDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py
index 6628d45b..c08d0590 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511100600_queue_remove_force.py
@@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS
-from libcommon.queue import Job
+from libcommon.queue import JobDocument
@@ -27 +27 @@ class MigrationRemoveForceFromJob(Migration):
- check_documents(DocCls=Job, sample_size=10)
+ check_documents(DocCls=JobDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py
index 4f6dea35..fb2cec30 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230511110700_queue_delete_skipped_jobs.py
@@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS
-from libcommon.queue import Job
+from libcommon.queue import JobDocument
@@ -31 +31 @@ class MigrationDeleteSkippedJobs(Migration):
- if not isinstance(doc, Job):
+ if not isinstance(doc, JobDocument):
@@ -36 +36 @@ class MigrationDeleteSkippedJobs(Migration):
- check_documents(DocCls=Job, sample_size=10, custom_validation=custom_validation)
+ check_documents(DocCls=JobDocument, sample_size=10, custom_validation=custom_validation)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py
index c1cee0a3..89fdd00e 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230516101500_queue_job_add_revision.py
@@ -7 +7 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS
-from libcommon.queue import Job
+from libcommon.queue import JobDocument
@@ -32 +32 @@ class MigrationQueueAddRevisionToJob(Migration):
- check_documents(DocCls=Job, sample_size=10)
+ check_documents(DocCls=JobDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py b/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py
index ce0ca4b5..b16aff74 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/renaming_migrations.py
@@ -7,2 +7,2 @@ from typing import Optional
-from libcommon.queue import Job
-from libcommon.simple_cache import CachedResponse
+from libcommon.queue import JobDocument
+from libcommon.simple_cache import CachedResponseDocument
@@ -47 +47 @@ class CacheRenamingMigration(CacheMigration):
- check_documents(DocCls=CachedResponse, sample_size=10)
+ check_documents(DocCls=CachedResponseDocument, sample_size=10)
@@ -114 +114 @@ class QueueRenamingMigration(QueueMigration):
- check_documents(DocCls=Job, sample_size=10)
+ check_documents(DocCls=JobDocument, sample_size=10)
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py
index b64bf12e..7fced24f 100644
--- a/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py
+++ b/jobs/mongodb_migration/tests/migrations/test_20230511100700_queue_delete_indexes_with_force.py
@@ -5 +5 @@ from libcommon.constants import QUEUE_COLLECTION_JOBS, QUEUE_MONGOENGINE_ALIAS
-from libcommon.queue import Job
+from libcommon.queue import JobDocument
@@ -21 +21 @@ def test_queue_delete_indexes_with_force(mongo_host: str) -> None:
- Job(
+ JobDocument(
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index 09cb6301..aa8b7027 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -13 +13 @@ from libcommon.constants import (
-from libcommon.queue import Job
+from libcommon.queue import JobDocument
@@ -121 +121 @@ def test_queue_delete_ttl_index(mongo_host: str) -> None:
- Job(
+ JobDocument(
diff --git a/libs/libcommon/src/libcommon/metrics.py b/libs/libcommon/src/libcommon/metrics.py
index 8c4343ce..b0833d61 100644
--- a/libs/libcommon/src/libcommon/metrics.py
+++ b/libs/libcommon/src/libcommon/metrics.py
@@ -39 +39 @@ class QuerySetManager(Generic[U]):
-class JobTotalMetric(Document):
+class JobTotalMetricDocument(Document):
@@ -60 +60 @@ class JobTotalMetric(Document):
- objects = QuerySetManager["JobTotalMetric"]()
+ objects = QuerySetManager["JobTotalMetricDocument"]()
@@ -63 +63 @@ class JobTotalMetric(Document):
-class CacheTotalMetric(Document):
+class CacheTotalMetricDocument(Document):
@@ -86 +86 @@ class CacheTotalMetric(Document):
- objects = QuerySetManager["CacheTotalMetric"]()
+ objects = QuerySetManager["CacheTotalMetricDocument"]()
@@ -91,2 +91,2 @@ def _clean_metrics_database() -> None:
- CacheTotalMetric.drop_collection() # type: ignore
- JobTotalMetric.drop_collection() # type: ignore
+ CacheTotalMetricDocument.drop_collection() # type: ignore
+ JobTotalMetricDocument.drop_collection() # type: ignore
diff --git a/libs/libcommon/src/libcommon/prometheus.py b/libs/libcommon/src/libcommon/prometheus.py
index 44a1c88a..27fa882a 100644
--- a/libs/libcommon/src/libcommon/prometheus.py
+++ b/libs/libcommon/src/libcommon/prometheus.py
@@ -19 +19 @@ from psutil import disk_usage
-from libcommon.metrics import CacheTotalMetric, JobTotalMetric
+from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument
@@ -67 +67 @@ def update_queue_jobs_total() -> None:
- for job_metric in JobTotalMetric.objects():
+ for job_metric in JobTotalMetricDocument.objects():
@@ -72 +72 @@ def update_responses_in_cache_total() -> None:
- for cache_metric in CacheTotalMetric.objects():
+ for cache_metric in CacheTotalMetricDocument.objects():
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 6aa66820..40898dcf 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -117 +117 @@ class NoWaitingJobError(Exception):
-class Job(Document):
+class JobDocument(Document):
@@ -193 +193 @@ class Job(Document):
- objects = QuerySetManager["Job"]()
+ objects = QuerySetManager["JobDocument"]()
@@ -211 +211 @@ class Job(Document):
- def get(cls, job_id: str) -> "Job":
+ def get(cls, job_id: str) -> "JobDocument":
@@ -351 +351 @@ class Queue:
- ) -> Job:
+ ) -> JobDocument:
@@ -367 +367 @@ class Queue:
- return Job(
+ return JobDocument(
@@ -393 +393 @@ class Queue:
- Job(
+ JobDocument(
@@ -412 +412 @@ class Queue:
- job_ids = Job.objects.insert(jobs, load_bulk=False)
+ job_ids = JobDocument.objects.insert(jobs, load_bulk=False)
@@ -429 +429 @@ class Queue:
- existing = Job.objects(pk__in=job_ids)
+ existing = JobDocument.objects(pk__in=job_ids)
@@ -440 +440 @@ class Queue:
- ) -> Job:
+ ) -> JobDocument:
@@ -467 +467 @@ class Queue:
- started_jobs = Job.objects(status=Status.STARTED, **filters)
+ started_jobs = JobDocument.objects(status=Status.STARTED, **filters)
@@ -473 +473 @@ class Queue:
- Job.objects(
+ JobDocument.objects(
@@ -506 +506 @@ class Queue:
- Job.objects(
+ JobDocument.objects(
@@ -524 +524 @@ class Queue:
- ) -> Job:
+ ) -> JobDocument:
@@ -549 +549 @@ class Queue:
- def _start_newest_job_and_cancel_others(self, job: Job) -> Job:
+ def _start_newest_job_and_cancel_others(self, job: JobDocument) -> JobDocument:
@@ -570 +570 @@ class Queue:
- waiting_jobs = Job.objects(
+ waiting_jobs = JobDocument.objects(
@@ -644 +644 @@ class Queue:
- def get_job_with_id(self, job_id: str) -> Job:
+ def get_job_with_id(self, job_id: str) -> JobDocument:
@@ -655 +655 @@ class Queue:
- return Job.objects(pk=job_id).get()
+ return JobDocument.objects(pk=job_id).get()
@@ -671 +671 @@ class Queue:
- def _get_started_job(self, job_id: str) -> Job:
+ def _get_started_job(self, job_id: str) -> JobDocument:
@@ -681 +681 @@ class Queue:
- job = Job.objects(pk=job_id).get()
+ job = JobDocument.objects(pk=job_id).get()
@@ -751 +751 @@ class Queue:
- Job.objects(
+ JobDocument.objects(
@@ -764 +764 @@ class Queue:
- for job in Job.objects(type=job_type, status=Status.STARTED.value):
+ for job in JobDocument.objects(type=job_type, status=Status.STARTED.value):
@@ -807 +807 @@ class Queue:
- for job in Job.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters)
+ for job in JobDocument.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters)
@@ -815 +815 @@ class Queue:
- return Job.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters).count() > 0
+ return JobDocument.objects(dataset=dataset, status__in=[Status.WAITING, Status.STARTED], **filters).count() > 0
@@ -827 +827 @@ class Queue:
- return Job.objects(type=job_type, status=status.value).count()
+ return JobDocument.objects(type=job_type, status=status.value).count()
@@ -856 +856 @@ class Queue:
- return [d.to_dict() for d in Job.objects(type=job_type, status=status.value)]
+ return [d.to_dict() for d in JobDocument.objects(type=job_type, status=status.value)]
@@ -875 +875 @@ class Queue:
- for d in Job.objects(
+ for d in JobDocument.objects(
@@ -899 +899 @@ class Queue:
- started_jobs = Job.objects(status=Status.STARTED)
+ started_jobs = JobDocument.objects(status=Status.STARTED)
@@ -923 +923 @@ def _clean_queue_database() -> None:
- Job.drop_collection() # type: ignore
+ JobDocument.drop_collection() # type: ignore
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index 096b8b2d..2d2d897a 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -70 +70 @@ class SplitFullName(NamedTuple):
-class CachedResponse(Document):
+class CachedResponseDocument(Document):
@@ -116 +116 @@ class CachedResponse(Document):
- objects = QuerySetManager["CachedResponse"]()
+ objects = QuerySetManager["CachedResponseDocument"]()
@@ -123,2 +123,2 @@ class CachedResponse(Document):
-CachedResponse.config.required = False # type: ignore
-CachedResponse.split.required = False # type: ignore
+CachedResponseDocument.config.required = False # type: ignore
+CachedResponseDocument.split.required = False # type: ignore
@@ -146 +146 @@ def upsert_response(
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split).upsert_one(
+ CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split).upsert_one(
@@ -188 +188 @@ def delete_response(
- return CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split).delete()
+ return CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split).delete()
@@ -192 +192 @@ def delete_dataset_responses(dataset: str) -> Optional[int]:
- return CachedResponse.objects(dataset=dataset).delete()
+ return CachedResponseDocument.objects(dataset=dataset).delete()
@@ -209 +209 @@ def get_response_without_content(
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split)
@@ -240 +240 @@ def get_response_metadata(
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split)
@@ -300 +300 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split)
@@ -322 +322 @@ def get_response_with_details(
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ CachedResponseDocument.objects(kind=kind, dataset=dataset, config=config, split=split)
@@ -444 +444 @@ def get_valid_datasets(kind: str) -> Set[str]:
- return set(CachedResponse.objects(kind=kind, http_status=HTTPStatus.OK).distinct("dataset"))
+ return set(CachedResponseDocument.objects(kind=kind, http_status=HTTPStatus.OK).distinct("dataset"))
@@ -448 +448 @@ def is_valid_for_kinds(dataset: str, kinds: List[str]) -> bool:
- return CachedResponse.objects(dataset=dataset, kind__in=kinds, http_status=HTTPStatus.OK).count() > 0
+ return CachedResponseDocument.objects(dataset=dataset, kind__in=kinds, http_status=HTTPStatus.OK).count() > 0
@@ -466 +466 @@ def get_responses_count_by_kind_status_and_error_code() -> List[CountEntry]:
- entries = CachedResponse.objects().only("kind", "http_status", "error_code")
+ entries = CachedResponseDocument.objects().only("kind", "http_status", "error_code")
@@ -535 +535 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo
- queryset = CachedResponse.objects(kind=kind)
+ queryset = CachedResponseDocument.objects(kind=kind)
@@ -538 +538 @@ def get_cache_reports(kind: str, cursor: Optional[str], limit: int) -> CacheRepo
- queryset = CachedResponse.objects(kind=kind, id__gt=ObjectId(cursor))
+ queryset = CachedResponseDocument.objects(kind=kind, id__gt=ObjectId(cursor))
@@ -566 +566 @@ def get_outdated_split_full_names_for_step(kind: str, current_version: int) -> L
- responses = CachedResponse.objects(kind=kind, job_runner_version__lt=current_version).only(
+ responses = CachedResponseDocument.objects(kind=kind, job_runner_version__lt=current_version).only(
@@ -575 +575 @@ def get_dataset_responses_without_content_for_kind(kind: str, dataset: str) -> L
- responses = CachedResponse.objects(kind=kind, dataset=dataset).exclude("content")
+ responses = CachedResponseDocument.objects(kind=kind, dataset=dataset).exclude("content")
@@ -628 +628 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int)
- queryset = CachedResponse.objects(kind=kind)
+ queryset = CachedResponseDocument.objects(kind=kind)
@@ -631 +631 @@ def get_cache_reports_with_content(kind: str, cursor: Optional[str], limit: int)
- queryset = CachedResponse.objects(kind=kind, id__gt=ObjectId(cursor))
+ queryset = CachedResponseDocument.objects(kind=kind, id__gt=ObjectId(cursor))
@@ -706 +706 @@ def get_cache_entries_df(dataset: str, cache_kinds: Optional[List[str]] = None)
- for response in CachedResponse.objects(dataset=dataset, **filters).only(
+ for response in CachedResponseDocument.objects(dataset=dataset, **filters).only(
@@ -723 +723 @@ def has_some_cache(dataset: str) -> bool:
- return CachedResponse.objects(dataset=dataset).count() > 0
+ return CachedResponseDocument.objects(dataset=dataset).count() > 0
@@ -760 +760 @@ def _clean_cache_database() -> None:
- CachedResponse.drop_collection() # type: ignore
+ CachedResponseDocument.drop_collection() # type: ignore
diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py
index 82848306..3131bc1a 100644
--- a/libs/libcommon/tests/test_orchestrator.py
+++ b/libs/libcommon/tests/test_orchestrator.py
@@ -11 +11 @@ from libcommon.processing_graph import Artifact, ProcessingGraph
-from libcommon.queue import Job, Queue
+from libcommon.queue import JobDocument, Queue
@@ -13 +13 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.simple_cache import CachedResponse, upsert_response_params
+from libcommon.simple_cache import CachedResponseDocument, upsert_response_params
@@ -169 +169 @@ def test_finish_job(
- assert Job.objects(dataset=DATASET_NAME).count() == 1 + len(artifacts_to_create)
+ assert JobDocument.objects(dataset=DATASET_NAME).count() == 1 + len(artifacts_to_create)
@@ -171 +171 @@ def test_finish_job(
- done_job = Job.objects(dataset=DATASET_NAME, status=Status.SUCCESS)
+ done_job = JobDocument.objects(dataset=DATASET_NAME, status=Status.SUCCESS)
@@ -174 +174 @@ def test_finish_job(
- waiting_jobs = Job.objects(dataset=DATASET_NAME, status=Status.WAITING)
+ waiting_jobs = JobDocument.objects(dataset=DATASET_NAME, status=Status.WAITING)
@@ -178,2 +178,2 @@ def test_finish_job(
- assert CachedResponse.objects(dataset=DATASET_NAME).count() == 1
- cached_response = CachedResponse.objects(dataset=DATASET_NAME).first()
+ assert CachedResponseDocument.objects(dataset=DATASET_NAME).count() == 1
+ cached_response = CachedResponseDocument.objects(dataset=DATASET_NAME).first()
diff --git a/libs/libcommon/tests/test_prometheus.py b/libs/libcommon/tests/test_prometheus.py
index 50e9d0ff..59c6a524 100644
--- a/libs/libcommon/tests/test_prometheus.py
+++ b/libs/libcommon/tests/test_prometheus.py
@@ -10 +10 @@ import pytest
-from libcommon.metrics import CacheTotalMetric, JobTotalMetric
+from libcommon.metrics import CacheTotalMetricDocument, JobTotalMetricDocument
@@ -151 +151 @@ def test_cache_metrics(metrics_mongo_resource: MetricsMongoResource) -> None:
- collection = CacheTotalMetric._get_collection()
+ collection = CacheTotalMetricDocument._get_collection()
@@ -184 +184 @@ def test_queue_metrics(metrics_mongo_resource: MetricsMongoResource) -> None:
- collection = JobTotalMetric._get_collection()
+ collection = JobTotalMetricDocument._get_collection()
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 34626226..fc81b103 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -18 +18 @@ from libcommon.constants import QUEUE_TTL_SECONDS
-from libcommon.queue import EmptyQueueError, Job, Lock, Queue, lock
+from libcommon.queue import EmptyQueueError, JobDocument, Lock, Queue, lock
@@ -320 +320 @@ def test_has_ttl_index_on_finished_at_field() -> None:
- for name, value in Job._get_collection().index_information().items()
+ for name, value in JobDocument._get_collection().index_information().items()
@@ -326 +326 @@ def test_has_ttl_index_on_finished_at_field() -> None:
- assert Job._get_collection().index_information()[ttl_index_name]["expireAfterSeconds"] == QUEUE_TTL_SECONDS
+ assert JobDocument._get_collection().index_information()[ttl_index_name]["expireAfterSeconds"] == QUEUE_TTL_SECONDS
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index 75f9845d..0a71e793 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -15 +15 @@ from libcommon.simple_cache import (
- CachedResponse,
+ CachedResponseDocument,
@@ -56 +56 @@ def test_insert_null_values() -> None:
- CachedResponse.objects(kind=kind, dataset=dataset_a, config=config, split=split).upsert_one(
+ CachedResponseDocument.objects(kind=kind, dataset=dataset_a, config=config, split=split).upsert_one(
@@ -60,2 +60,2 @@ def test_insert_null_values() -> None:
- assert CachedResponse.objects.count() == 1
- cached_response = CachedResponse.objects.get()
+ assert CachedResponseDocument.objects.count() == 1
+ cached_response = CachedResponseDocument.objects.get()
@@ -67 +67 @@ def test_insert_null_values() -> None:
- CachedResponse(
+ CachedResponseDocument(
@@ -70,2 +70,2 @@ def test_insert_null_values() -> None:
- assert CachedResponse.objects.count() == 2
- cached_response = CachedResponse.objects(dataset=dataset_b).get()
+ assert CachedResponseDocument.objects.count() == 2
+ cached_response = CachedResponseDocument.objects(dataset=dataset_b).get()
@@ -76 +76 @@ def test_insert_null_values() -> None:
- coll = CachedResponse._get_collection()
+ coll = CachedResponseDocument._get_collection()
@@ -87,2 +87,2 @@ def test_insert_null_values() -> None:
- assert CachedResponse.objects.count() == 3
- cached_response = CachedResponse.objects(dataset=dataset_c).get()
+ assert CachedResponseDocument.objects.count() == 3
+ cached_response = CachedResponseDocument.objects(dataset=dataset_c).get()
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index d83cbf26..3e30ca26 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -15 +15 @@ from libcommon.processing_graph import ProcessingGraph
-from libcommon.queue import Job, JobDoesNotExistError, Queue
+from libcommon.queue import JobDocument, JobDoesNotExistError, Queue
@@ -17 +17 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.simple_cache import CachedResponse
+from libcommon.simple_cache import CachedResponseDocument
@@ -86 +86 @@ def start_worker_loop_with_long_job() -> None:
- current_job = Job.objects(pk=current_job_info["job_id"]).get()
+ current_job = JobDocument.objects(pk=current_job_info["job_id"]).get()
@@ -107 +107 @@ def set_worker_state(worker_state_file_path: str) -> Iterator[WorkerState]:
-def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
+def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[JobDocument]:
@@ -112 +112 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I
- Job.get(job_id=job_info["job_id"]).delete()
+ JobDocument.get(job_id=job_info["job_id"]).delete()
@@ -116 +116 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I
- job = Job(
+ job = JobDocument(
@@ -136 +136,3 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I
-def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
+def set_long_running_job_in_queue(
+ app_config: AppConfig, queue_mongo_resource: QueueMongoResource
+) -> Iterator[JobDocument]:
@@ -141 +143 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q
- Job.get(job_id=job_info["job_id"]).delete()
+ JobDocument.get(job_id=job_info["job_id"]).delete()
@@ -146 +148 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q
- job = Job(
+ job = JobDocument(
@@ -167 +169 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q
-def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[Job]:
+def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterator[JobDocument]:
@@ -172 +174 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato
- Job.get(job_id=job_info["job_id"]).delete()
+ JobDocument.get(job_id=job_info["job_id"]).delete()
@@ -176 +178 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato
- job = Job(
+ job = JobDocument(
@@ -234 +236 @@ def test_executor_heartbeat(
- set_just_started_job_in_queue: Job,
+ set_just_started_job_in_queue: JobDocument,
@@ -248,3 +250,3 @@ def test_executor_kill_zombies(
- set_just_started_job_in_queue: Job,
- set_long_running_job_in_queue: Job,
- set_zombie_job_in_queue: Job,
+ set_just_started_job_in_queue: JobDocument,
+ set_long_running_job_in_queue: JobDocument,
+ set_zombie_job_in_queue: JobDocument,
@@ -259,3 +261,3 @@ def test_executor_kill_zombies(
- assert Job.objects(pk=zombie.pk).get().status in [Status.ERROR, Status.CANCELLED, Status.SUCCESS]
- assert Job.objects(pk=normal_job.pk).get().status == Status.STARTED
- response = CachedResponse.objects()[0]
+ assert JobDocument.objects(pk=zombie.pk).get().status in [Status.ERROR, Status.CANCELLED, Status.SUCCESS]
+ assert JobDocument.objects(pk=normal_job.pk).get().status == Status.STARTED
+ response = CachedResponseDocument.objects()[0]
@@ -273 +275 @@ def test_executor_kill_zombies(
- CachedResponse.objects().delete()
+ CachedResponseDocument.objects().delete()
@@ -279,2 +281,2 @@ def test_executor_start(
- set_just_started_job_in_queue: Job,
- set_zombie_job_in_queue: Job,
+ set_just_started_job_in_queue: JobDocument,
+ set_zombie_job_in_queue: JobDocument,
@@ -299 +301 @@ def test_executor_start(
- assert Job.objects(pk=set_just_started_job_in_queue.pk).get().last_heartbeat is not None
+ assert JobDocument.objects(pk=set_just_started_job_in_queue.pk).get().last_heartbeat is not None
@@ -301 +303,5 @@ def test_executor_start(
- assert Job.objects(pk=set_zombie_job_in_queue.pk).get().status in [Status.ERROR, Status.CANCELLED, Status.SUCCESS]
+ assert JobDocument.objects(pk=set_zombie_job_in_queue.pk).get().status in [
+ Status.ERROR,
+ Status.CANCELLED,
+ Status.SUCCESS,
+ ]
@@ -328,2 +334,2 @@ def test_executor_stops_on_long_job(
- set_long_running_job_in_queue: Job,
- set_just_started_job_in_queue: Job,
+ set_long_running_job_in_queue: JobDocument,
+ set_just_started_job_in_queue: JobDocument,
@@ -350 +356 @@ def test_executor_stops_on_long_job(
- responses = CachedResponse.objects()
+ responses = CachedResponseDocument.objects()
@@ -368 +374 @@ def test_executor_stops_on_long_job(
- CachedResponse.objects().delete()
+ CachedResponseDocument.objects().delete()
diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py
index 0e912bdb..a6e54257 100644
--- a/services/worker/tests/test_job_manager.py
+++ b/services/worker/tests/test_job_manager.py
@@ -8 +8 @@ from libcommon.processing_graph import ProcessingGraph, ProcessingStep
-from libcommon.queue import Job, Queue
+from libcommon.queue import JobDocument, Queue
@@ -10 +10 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.simple_cache import CachedResponse, get_response, upsert_response
+from libcommon.simple_cache import CachedResponseDocument, get_response, upsert_response
@@ -127 +127 @@ def test_backfill(priority: Priority, app_config: AppConfig) -> None:
- assert Job.objects().count() == 0
+ assert JobDocument.objects().count() == 0
@@ -197 +197 @@ def test_job_runner_set_crashed(
- assert Job.objects().count() == 0
+ assert JobDocument.objects().count() == 0
@@ -219 +219 @@ def test_job_runner_set_crashed(
- response = CachedResponse.objects()[0]
+ response = CachedResponseDocument.objects()[0]
|
|
e26ef0fd08298579346325193ba2d2405b93f43c
|
Sylvain Lesage
| 2023-06-30T20:02:12 |
Ensure parquet shards are sorted (#1465)
|
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json
index 9c3ed661..832fd7a7 100644
--- a/chart/static-files/openapi.json
+++ b/chart/static-files/openapi.json
@@ -2786 +2786 @@
- "description": "The dataset is converted to the parquet format. The endpoint gives the list of the parquet files.",
+ "description": "The dataset is converted to the parquet format. The endpoint gives the list of the parquet files. The same split can have multiple parquet files, called shards. They are sorted by their shard index.",
diff --git a/docs/source/parquet.mdx b/docs/source/parquet.mdx
index a789265e..927a5a1e 100644
--- a/docs/source/parquet.mdx
+++ b/docs/source/parquet.mdx
@@ -116 +116 @@ The endpoint also gives the filename and size of each file:
-Big datasets are partitioned into Parquet files (shards) of about 500MB each. The filename contains the name of the dataset, the split, the shard index, and the total number of shards (`dataset-name-train-0000-of-0004.parquet`). For example, the `train` split of the [`amazon_polarity`](https://datasets-server.huggingface.co/parquet?dataset=amazon_polarity) dataset is partitioned into 4 shards:
+Big datasets are partitioned into Parquet files (shards) of about 500MB each. The filename contains the name of the dataset, the split, the shard index, and the total number of shards (`dataset-name-train-0000-of-0004.parquet`). For a given split, the elements in the list are sorted by their shard index, in ascending order. For example, the `train` split of the [`amazon_polarity`](https://datasets-server.huggingface.co/parquet?dataset=amazon_polarity) dataset is partitioned into 4 shards:
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
index 7f57d086..5bb181a5 100644
--- a/libs/libcommon/src/libcommon/constants.py
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -20,4 +19,0 @@ DEFAULT_JOB_RUNNER_VERSION = 1
-PROCESSING_STEP_DATASET_CONFIG_NAMES_VERSION = 1
-PROCESSING_STEP_CONFIG_PARQUET_VERSION = 4
-PROCESSING_STEP_CONFIG_PARQUET_METADATA_VERSION = 1
-PROCESSING_STEP_CONFIG_SIZE_VERSION = 2
@@ -25,0 +22,7 @@ PROCESSING_STEP_CONFIG_OPT_IN_OUT_URLS_COUNT_VERSION = 3
+PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 3
+PROCESSING_STEP_CONFIG_PARQUET_METADATA_VERSION = 1
+PROCESSING_STEP_CONFIG_PARQUET_VERSION = 5
+PROCESSING_STEP_CONFIG_SIZE_VERSION = 2
+PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_INFO_VERSION = 3
+PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_STREAMING_VERSION = 3
+PROCESSING_STEP_DATASET_CONFIG_NAMES_VERSION = 1
@@ -30,0 +34,2 @@ PROCESSING_STEP_DATASET_SIZE_VERSION = 2
+PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 3
+PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION = 1
@@ -32 +36,0 @@ PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_PARQUET_VERSION = 2
-PROCESSING_STEP_CONFIG_PARQUET_AND_INFO_VERSION = 3
@@ -34,3 +38 @@ PROCESSING_STEP_SPLIT_FIRST_ROWS_FROM_STREAMING_VERSION = 3
-PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_INFO_VERSION = 3
-PROCESSING_STEP_CONFIG_SPLIT_NAMES_FROM_STREAMING_VERSION = 3
-PROCESSING_STEP_DATASET_SPLIT_NAMES_VERSION = 3
+PROCESSING_STEP_SPLIT_IMAGE_URL_COLUMNS_VERSION = 1
@@ -39,2 +40,0 @@ PROCESSING_STEP_SPLIT_OPT_IN_OUT_URLS_SCAN_VERSION = 4
-PROCESSING_STEP_SPLIT_IMAGE_URL_COLUMNS_VERSION = 1
-PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION = 1
diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py
index 572df22c..a6fc6833 100644
--- a/services/worker/src/worker/job_runners/config/parquet.py
+++ b/services/worker/src/worker/job_runners/config/parquet.py
@@ -38,2 +37,0 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse
- if "parquet_files" not in content:
- raise PreviousStepFormatError("Previous step did not return the expected content: 'parquet_files'.")
@@ -41 +39,8 @@ def compute_parquet_response(dataset: str, config: str) -> ConfigParquetResponse
- parquet_files = [parquet_file for parquet_file in content["parquet_files"] if parquet_file.get("config") == config]
+ try:
+ parquet_files = [
+ parquet_file for parquet_file in content["parquet_files"] if parquet_file.get("config") == config
+ ]
+ # sort by filename, which ensures the shards are in order: 00000, 00001, 00002, ...
+ parquet_files.sort(key=lambda x: x["filename"]) # type: ignore
+ except KeyError as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content: 'parquet_files'.", e) from e
diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py
index 000388ff..9425c249 100644
--- a/services/worker/src/worker/job_runners/dataset/parquet.py
+++ b/services/worker/src/worker/job_runners/dataset/parquet.py
@@ -26 +26 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float]:
+def compute_parquet_response(dataset: str) -> Tuple[DatasetParquetResponse, float]:
@@ -111 +111 @@ class DatasetParquetJobRunner(DatasetJobRunner):
- response_content, progress = compute_sizes_response(dataset=self.dataset)
+ response_content, progress = compute_parquet_response(dataset=self.dataset)
diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py
index e314000a..c7db50e4 100644
--- a/services/worker/tests/job_runners/config/test_parquet.py
+++ b/services/worker/tests/job_runners/config/test_parquet.py
@@ -116,0 +117,64 @@ def get_job_runner(
+ (
+ "shards_order",
+ "config_1",
+ HTTPStatus.OK,
+ ConfigParquetAndInfoResponse(
+ parquet_files=[
+ SplitHubFile(
+ dataset="ok",
+ config="config_1",
+ split="train",
+ url="url2",
+ filename="parquet-train-00001-of-05534.parquet",
+ size=0,
+ ),
+ SplitHubFile(
+ dataset="ok",
+ config="config_1",
+ split="train",
+ url="url1",
+ filename="parquet-train-00000-of-05534.parquet",
+ size=0,
+ ),
+ SplitHubFile(
+ dataset="ok",
+ config="config_1",
+ split="test",
+ url="url2",
+ filename="parquet-test-00000-of-00001.parquet",
+ size=0,
+ ),
+ ],
+ dataset_info={"description": "value", "dataset_size": 10},
+ ),
+ None,
+ ConfigParquetResponse(
+ parquet_files=[
+ SplitHubFile(
+ dataset="ok",
+ config="config_1",
+ split="test",
+ url="url2",
+ filename="parquet-test-00000-of-00001.parquet",
+ size=0,
+ ),
+ SplitHubFile(
+ dataset="ok",
+ config="config_1",
+ split="train",
+ url="url1",
+ filename="parquet-train-00000-of-05534.parquet",
+ size=0,
+ ),
+ SplitHubFile(
+ dataset="ok",
+ config="config_1",
+ split="train",
+ url="url2",
+ filename="parquet-train-00001-of-05534.parquet",
+ size=0,
+ ),
+ ]
+ ),
+ False,
+ ),
|
|
f9d5c9ff873ab93369ec605644b5e72f595ac6e2
|
Andrea Francis Soria Jimenez
| 2023-06-29T19:06:02 |
Disable backfill (#1461)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 62a4355f..b0dc9cf6 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -124 +124 @@ cacheMaintenance:
- action: "backfill"
+ action: "skip"
@@ -129 +129 @@ cacheMaintenance:
- error_codes_to_retry: "AskAccessHubRequestError,CachedResponseNotFound,DatasetInfoHubRequestError,DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError,ExternalFilesSizeRequestConnectionError,ExternalFilesSizeRequestTimeoutError,FileSystemError,GatedExtraFieldsError,JobManagerExceededMaximumDurationError,JobRunnerCrashedError,JobRunnerExceededMaximumDurationError,PreviousStepFormatError,PreviousStepStatusError"
+ error_codes_to_retry: ""
|
|
e6aae54dc8591379eeca3f36f9068f7d24abe729
|
Andrea Francis Soria Jimenez
| 2023-06-29T15:54:38 |
Adding fix (#1459)
|
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py
index 01874c4c..1ed21f5d 100644
--- a/services/worker/src/worker/job_runners/split/duckdb_index.py
+++ b/services/worker/src/worker/job_runners/split/duckdb_index.py
@@ -45 +45 @@ CREATE_SEQUENCE_COMMAND = "CREATE OR REPLACE SEQUENCE serial START 1;"
-CREATE_INDEX_COMMAND = "PRAGMA create_fts_index('data', '__hf_index_id', '*', overwrite=1);"
+CREATE_INDEX_COMMAND = "PRAGMA create_fts_index('data', '__hf_index_id', {columns}, overwrite=1);"
@@ -111 +111 @@ def compute_index_rows(
- column_names = ",".join(list(features.keys()))
+ column_names = ",".join('"' + column + '"' for column in list(features.keys()))
@@ -152,2 +152,3 @@ def compute_index_rows(
- logging.debug(CREATE_INDEX_COMMAND)
- con.sql(CREATE_INDEX_COMMAND)
+ create_index_sql = CREATE_INDEX_COMMAND.format(columns=column_names)
+ logging.debug(create_index_sql)
+ con.sql(create_index_sql)
diff --git a/services/worker/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py
index 6e987e20..7e7a28ff 100644
--- a/services/worker/tests/fixtures/datasets.py
+++ b/services/worker/tests/fixtures/datasets.py
@@ -158 +158,8 @@ def datasets() -> Mapping[str, Dataset]:
- ]
+ ],
+ "column with spaces": [
+ "a",
+ "b",
+ "c",
+ "d",
+ "e",
+ ],
|
|
7a83fc631a34627264f28a4af2b5e28549550e56
|
Sylvain Lesage
| 2023-06-29T15:31:04 |
feat: 🎸 backfill the datasets (#1460)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index b0dc9cf6..62a4355f 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -124 +124 @@ cacheMaintenance:
- action: "skip"
+ action: "backfill"
@@ -129 +129 @@ cacheMaintenance:
- error_codes_to_retry: ""
+ error_codes_to_retry: "AskAccessHubRequestError,CachedResponseNotFound,DatasetInfoHubRequestError,DatasetTooBigFromDatasetsError,DatasetTooBigFromHubError,DatasetWithTooBigExternalFilesError,DatasetWithTooManyExternalFilesError,ExternalFilesSizeRequestConnectionError,ExternalFilesSizeRequestTimeoutError,FileSystemError,GatedExtraFieldsError,JobManagerExceededMaximumDurationError,JobRunnerCrashedError,JobRunnerExceededMaximumDurationError,PreviousStepFormatError,PreviousStepStatusError"
|
|
e8a5be6f6b43860ae9613d18a64c0062216266c0
|
Sylvain Lesage
| 2023-06-29T14:13:14 |
Replace valid with preview and viewer (#1452)
|
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json
index 59a058d9..9c3ed661 100644
--- a/chart/static-files/openapi.json
+++ b/chart/static-files/openapi.json
@@ -898,4 +897,0 @@
- "preview": {
- "type": "array",
- "items": { "type": "string" }
- },
@@ -906 +902 @@
- "valid": {
+ "preview": {
@@ -908,2 +904 @@
- "items": { "type": "string" },
- "deprecated": true
+ "items": { "type": "string" }
@@ -915 +910 @@
- "required": ["valid"],
+ "required": ["preview", "viewer"],
@@ -917 +912,4 @@
- "valid": {
+ "viewer": {
+ "type": "boolean"
+ },
+ "preview": {
@@ -3190,7 +3187,0 @@
- ],
- "valid": [
- "0n1xus/codexglue",
- "0n1xus/pytorrent-standalone",
- "0x7194633/rupile",
- "51la5/keyword-extraction",
- "AHussain0418/day2_data"
@@ -3303 +3294,9 @@
- "valid": true
+ "viewer": true,
+ "preview": true
+ }
+ },
+ "preview": {
+ "summary": "dataset with only preview",
+ "value": {
+ "viewer": false,
+ "preview": true
@@ -3309 +3308,2 @@
- "valid": false
+ "viewer": false,
+ "preview": false
diff --git a/docs/source/valid.mdx b/docs/source/valid.mdx
index 3acd4e3d..4bbc0a67 100644
--- a/docs/source/valid.mdx
+++ b/docs/source/valid.mdx
@@ -5,2 +5,2 @@ Before you download a dataset from the Hub, it is helpful to know which datasets
-* `/valid` returns a list of all the datasets that work without any errors.
-* `/is-valid` checks if a specific dataset works without any errors.
+- `/valid` returns a list of all the datasets that work without any errors.
+- `/is-valid` checks if a specific dataset works without any errors.
@@ -11,5 +11,6 @@ The API endpoints will return an error for datasets that cannot be loaded with t
- Currently, only {" "}
- <a href="https://huggingface.co/docs/datasets/stream">streamable</a> datasets
- are supported so Datasets Server can extract the 100 first rows without downloading the
- whole dataset. This is especially useful for previewing large datasets where downloading
- the whole dataset may take hours!
+ The largest datasets are partially supported by Datasets Server. If they are{" "}
+ <a href="https://huggingface.co/docs/datasets/stream">streamable</a>, Datasets Server can
+ extract the first 100 rows without downloading the whole dataset. This is
+ especially useful for previewing large datasets where downloading the whole
+ dataset may take hours! See the <code>preview</code> field in the response of{" "}
+ <code>/valid</code> to view these partially supported datasets.
@@ -61 +62,4 @@ curl https://datasets-server.huggingface.co/valid \
-The endpoint response is a JSON containing a list valid datasets nested under the `valid` key:
+The endpoint response is a JSON containing lists of datasets nested under the following keys:
+
+- `viewer`: the dataset is fully-supported and the Dataset Viewer is working on the Hub dataset page. It also means the dataset has been auto-converted to Parquet.
+- `preview`: the dataset is partially supported and the Dataset Viewer on the Hub dataset page shows a preview of the first 100 rows obtained by streaming.
@@ -65,6 +69,14 @@ The endpoint response is a JSON containing a list valid datasets nested under th
- "valid": [
- "0n1xus/codexglue",
- "0n1xus/pytorrent-standalone",
- "0x7194633/rupile",
- "51la5/keyword-extraction",
- ...
+ "viewer": [
+ "0-hero/OIG-small-chip2",
+ "000alen/semantic","04-07-22/wep-probes",
+ "0721boy/nva-pic",
+ "0Tick/Danbooru-Random-Posts-Scrape",
+ "0Tick/E621-Random-PostsTag-Scrape",
+ "0n1xus/codexglue"
+ "..."
+ ],
+ "preview": [
+ "0x7194633/GCRL-flibusta",
+ "0xJustin/Dungeons_and_Diffusion_uncropped",
+ "0xaryan/music-classifier","13GP/training",
+ "..."
@@ -122 +134,13 @@ The response looks like this if a dataset is valid:
-{"valid": true}
+{
+ "viewer": true,
+ "preview": true
+}
+```
+
+If only the first rows of a dataset are available, then the response looks like:
+
+```json
+{
+ "viewer": false,
+ "preview": true
+}
@@ -125 +149 @@ The response looks like this if a dataset is valid:
-If a dataset is not valid, then the response looks like:
+Finally, if the dataset is not valid at all, then the response is:
@@ -128 +152,4 @@ If a dataset is not valid, then the response looks like:
-{"valid": false}
+{
+ "viewer": false,
+ "preview": false
+}
@@ -131,0 +159 @@ Some cases where a dataset is not valid are:
+
@@ -138 +166,2 @@ Some cases where a dataset is not valid are:
- Remember if a dataset is <a href="./quick_start#gated-datasets">gated</a>, you'll need to provide your user token to submit a successful query!
+ Remember if a dataset is <a href="./quick_start#gated-datasets">gated</a>,
+ you'll need to provide your user token to submit a successful query!
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index b36d4535..99c6bebc 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -322,0 +323 @@ class ProcessingGraphConfig:
+ # special case: triggered by all the steps that have "enables_preview" or "enables_viewer"
@@ -324 +325 @@ class ProcessingGraphConfig:
- "dataset-split-names",
+ "config-size",
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
index 66a609b6..7f57d086 100644
--- a/libs/libcommon/src/libcommon/constants.py
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -27 +27 @@ PROCESSING_STEP_DATASET_INFO_VERSION = 2
-PROCESSING_STEP_DATASET_IS_VALID_VERSION = 3
+PROCESSING_STEP_DATASET_IS_VALID_VERSION = 4
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index 13582b42..096b8b2d 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -447,11 +447,2 @@ def get_valid_datasets(kind: str) -> Set[str]:
-def get_validity_by_kind(dataset: str, kinds: Optional[List[str]] = None) -> Mapping[str, bool]:
- # TODO: rework with aggregate
- entries = (
- CachedResponse.objects(dataset=dataset)
- if kinds is None
- else CachedResponse.objects(dataset=dataset, kind__in=kinds)
- ).only("kind", "http_status")
- return {
- str(kind): entries(kind=kind, http_status=HTTPStatus.OK).first() is not None
- for kind in sorted(entries.distinct("kind"))
- }
+def is_valid_for_kinds(dataset: str, kinds: List[str]) -> bool:
+ return CachedResponse.objects(dataset=dataset, kind__in=kinds, http_status=HTTPStatus.OK).count() > 0
diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py
index ea672c6f..0a15fda6 100644
--- a/libs/libcommon/tests/test_processing_graph.py
+++ b/libs/libcommon/tests/test_processing_graph.py
@@ -115 +115 @@ def graph() -> ProcessingGraph:
- ["dataset-is-valid"],
+ [],
@@ -182 +182 @@ def graph() -> ProcessingGraph:
- ["dataset-size"],
+ ["dataset-is-valid", "dataset-size"],
@@ -196 +196 @@ def graph() -> ProcessingGraph:
- "dataset-split-names",
+ "config-size",
@@ -203 +202,0 @@ def graph() -> ProcessingGraph:
- "dataset-split-names",
@@ -205,0 +205 @@ def graph() -> ProcessingGraph:
+ "config-size",
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index e3e9d851..75f9845d 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -34 +34 @@ from libcommon.simple_cache import (
- get_validity_by_kind,
+ is_valid_for_kinds,
@@ -270,2 +270,2 @@ def test_get_valid_dataset_names_only_invalid_responses() -> None:
-def test_get_validity_by_kind_empty() -> None:
- assert not get_validity_by_kind(dataset="dataset")
+def test_is_valid_for_kinds_empty() -> None:
+ assert not is_valid_for_kinds(dataset="dataset", kinds=[])
@@ -274 +274 @@ def test_get_validity_by_kind_empty() -> None:
-def test_get_validity_by_kind_two_valid_datasets() -> None:
+def test_is_valid_for_kinds_two_valid_datasets() -> None:
@@ -281,5 +281,4 @@ def test_get_validity_by_kind_two_valid_datasets() -> None:
- assert get_validity_by_kind(dataset=dataset_a) == {kind: True}
- assert get_validity_by_kind(dataset=dataset_b) == {kind: True}
- assert get_validity_by_kind(dataset=dataset_b, kinds=[kind]) == {kind: True}
- assert not get_validity_by_kind(dataset=dataset_b, kinds=[other_kind])
- assert get_validity_by_kind(dataset=dataset_b, kinds=[kind, other_kind]) == {kind: True}
+ assert is_valid_for_kinds(dataset=dataset_a, kinds=[kind])
+ assert is_valid_for_kinds(dataset=dataset_b, kinds=[kind])
+ assert not is_valid_for_kinds(dataset=dataset_b, kinds=[other_kind])
+ assert is_valid_for_kinds(dataset=dataset_b, kinds=[kind, other_kind])
@@ -288 +287 @@ def test_get_validity_by_kind_two_valid_datasets() -> None:
-def test_get_validity_by_kind_two_valid_kinds() -> None:
+def test_is_valid_for_kinds_two_valid_kinds() -> None:
@@ -294 +293 @@ def test_get_validity_by_kind_two_valid_kinds() -> None:
- assert get_validity_by_kind(dataset=dataset) == {kind_a: True, kind_b: True}
+ assert is_valid_for_kinds(dataset=dataset, kinds=[kind_a, kind_b])
@@ -297 +296 @@ def test_get_validity_by_kind_two_valid_kinds() -> None:
-def test_get_validity_by_kind_at_least_one_valid_response() -> None:
+def test_is_valid_for_kinds_at_least_one_valid_response() -> None:
@@ -306 +305 @@ def test_get_validity_by_kind_at_least_one_valid_response() -> None:
- assert get_validity_by_kind(dataset=dataset) == {kind: True}
+ assert is_valid_for_kinds(dataset=dataset, kinds=[kind])
@@ -309 +308 @@ def test_get_validity_by_kind_at_least_one_valid_response() -> None:
-def test_get_validity_by_kind_only_invalid_responses() -> None:
+def test_is_valid_for_kinds_only_invalid_responses() -> None:
@@ -320 +319 @@ def test_get_validity_by_kind_only_invalid_responses() -> None:
- assert get_validity_by_kind(dataset=dataset) == {kind: False}
+ assert not is_valid_for_kinds(dataset=dataset, kinds=[kind])
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 606353e4..029eaa97 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -23,2 +22,0 @@ class ValidContent(TypedDict):
- valid: List[str]
- preview: List[str]
@@ -25,0 +24 @@ class ValidContent(TypedDict):
+ preview: List[str]
@@ -40 +38,0 @@ class ValidDatasets:
- _valid_set = set.union(_viewer_set, _preview_set)
@@ -42,2 +39,0 @@ class ValidDatasets:
- valid=sorted(_valid_set),
- preview=sorted(_preview_set),
@@ -44,0 +41 @@ class ValidDatasets:
+ preview=sorted(_preview_set),
diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py
index 38fb758f..4827c75d 100644
--- a/services/api/tests/routes/test_valid.py
+++ b/services/api/tests/routes/test_valid.py
@@ -35 +35 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) ->
- assert valid_datasets.content == {"valid": [], "preview": [], "viewer": []}
+ assert valid_datasets.content == {"preview": [], "viewer": []}
@@ -39 +39 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) ->
- "processing_graph_specification,expected_preview,expected_viewer,expected_valid",
+ "processing_graph_specification,expected_preview,expected_viewer",
@@ -41,6 +41,6 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) ->
- ({step_1: {}, step_2: {}}, [], [], []),
- ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset"], [], ["dataset"]),
- ({step_1: {}, step_2: {"enables_preview": True}}, [], [], []),
- ({step_1: {"enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]),
- ({step_1: {}, step_2: {"enables_viewer": True}}, [], [], []),
- ({step_1: {"enables_preview": True, "enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]),
+ ({step_1: {}, step_2: {}}, [], []),
+ ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset"], []),
+ ({step_1: {}, step_2: {"enables_preview": True}}, [], []),
+ ({step_1: {"enables_viewer": True}, step_2: {}}, [], ["dataset"]),
+ ({step_1: {}, step_2: {"enables_viewer": True}}, [], []),
+ ({step_1: {"enables_preview": True, "enables_viewer": True}, step_2: {}}, [], ["dataset"]),
@@ -53 +52,0 @@ def test_one_dataset(
- expected_valid: List[str],
@@ -61 +59,0 @@ def test_one_dataset(
- "valid": expected_valid,
@@ -68 +66 @@ def test_one_dataset(
- "processing_graph_specification,expected_preview,expected_viewer,expected_valid",
+ "processing_graph_specification,expected_preview,expected_viewer",
@@ -70,3 +68,3 @@ def test_one_dataset(
- ({step_1: {}, step_2: {}}, [], [], []),
- ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset1"], [], ["dataset1"]),
- ({step_1: {}, step_2: {"enables_preview": True}}, ["dataset2"], [], ["dataset2"]),
+ ({step_1: {}, step_2: {}}, [], []),
+ ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset1"], []),
+ ({step_1: {}, step_2: {"enables_preview": True}}, ["dataset2"], []),
@@ -77 +74,0 @@ def test_one_dataset(
- ["dataset1", "dataset2"],
@@ -83 +79,0 @@ def test_one_dataset(
- ["dataset1", "dataset2"],
@@ -91 +86,0 @@ def test_two_datasets(
- expected_valid: List[str],
@@ -108 +102,0 @@ def test_two_datasets(
- "valid": expected_valid,
@@ -115 +109 @@ def test_two_datasets(
- "processing_graph_specification,expected_preview,expected_viewer,expected_valid",
+ "processing_graph_specification,expected_preview,expected_viewer",
@@ -125 +118,0 @@ def test_two_datasets(
- [],
@@ -135 +127,0 @@ def test_two_datasets(
- ["dataset"],
@@ -149 +140,0 @@ def test_two_datasets(
- ["dataset"],
@@ -159 +149,0 @@ def test_two_datasets(
- ["dataset"],
@@ -173 +162,0 @@ def test_two_datasets(
- ["dataset"],
@@ -187 +175,0 @@ def test_two_datasets(
- ["dataset"],
@@ -195 +182,0 @@ def test_three_steps(
- expected_valid: List[str],
@@ -224 +210,0 @@ def test_three_steps(
- "valid": expected_valid,
@@ -241,4 +226,0 @@ def test_errors() -> None:
- "valid": [
- dataset_a,
- dataset_b,
- ],
diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py
index a883d39b..3e270714 100644
--- a/services/api/tests/test_app.py
+++ b/services/api/tests/test_app.py
@@ -51 +50,0 @@ def test_get_valid_datasets(client: TestClient) -> None:
- assert "valid" in response.json()
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py
index 5eb630d1..17a6a41d 100644
--- a/services/worker/src/worker/dtos.py
+++ b/services/worker/src/worker/dtos.py
@@ -178 +178,2 @@ class DatasetIsValidResponse(TypedDict):
- valid: bool
+ preview: bool
+ viewer: bool
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 87c48d90..493d7ba3 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -184,0 +185 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ processing_graph=self.processing_graph,
diff --git a/services/worker/src/worker/job_runners/dataset/is_valid.py b/services/worker/src/worker/job_runners/dataset/is_valid.py
index cb45bf8f..caf65b0a 100644
--- a/services/worker/src/worker/job_runners/dataset/is_valid.py
+++ b/services/worker/src/worker/job_runners/dataset/is_valid.py
@@ -5 +4,0 @@ import logging
-from typing import Tuple
@@ -8 +7,3 @@ from libcommon.constants import PROCESSING_STEP_DATASET_IS_VALID_VERSION
-from libcommon.simple_cache import get_validity_by_kind
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
+from libcommon.simple_cache import is_valid_for_kinds
+from libcommon.utils import JobInfo
@@ -10 +11,2 @@ from libcommon.simple_cache import get_validity_by_kind
-from worker.dtos import DatasetIsValidResponse, JobResult
+from worker.config import AppConfig
+from worker.dtos import CompleteJobResult, DatasetIsValidResponse, JobResult
@@ -13,2 +14,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-SPLIT_KINDS = ["config-split-names-from-streaming", "config-split-names-from-info"]
-FIRST_ROWS_KINDS = ["split-first-rows-from-streaming", "split-first-rows-from-parquet"]
@@ -16,2 +16 @@ FIRST_ROWS_KINDS = ["split-first-rows-from-streaming", "split-first-rows-from-pa
-
-def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, float]:
+def compute_is_valid_response(dataset: str, processing_graph: ProcessingGraph) -> DatasetIsValidResponse:
@@ -21,3 +20,4 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo
- A dataset is valid if:
- - /splits is valid for at least one of the configs
- - /first-rows is valid for at least one of the splits
+
+ A dataset is valid if at least one response of any of the artifacts for any of the
+ steps (for viewer and preview) is valid.
+ The deprecated `valid` field is an "or" of the `preview` and `viewer` fields.
@@ -28,0 +29,3 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo
+ processing_graph (`ProcessingGraph`):
+ The processing graph. In particular, it must provide the list of
+ processing steps that enable the viewer and the preview.
@@ -30 +33 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo
- `DatasetIsValidResponse`: An object with the is_valid_response.
+ `DatasetIsValidResponse`: The response (viewer, preview).
@@ -34,3 +37,5 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo
- validity_by_kind = get_validity_by_kind(dataset=dataset, kinds=SPLIT_KINDS + FIRST_ROWS_KINDS)
- is_valid = any(validity_by_kind[kind] for kind in SPLIT_KINDS if kind in validity_by_kind) and any(
- validity_by_kind[kind] for kind in FIRST_ROWS_KINDS if kind in validity_by_kind
+ viewer = is_valid_for_kinds(
+ dataset=dataset, kinds=[step.cache_kind for step in processing_graph.get_processing_steps_enables_viewer()]
+ )
+ preview = is_valid_for_kinds(
+ dataset=dataset, kinds=[step.cache_kind for step in processing_graph.get_processing_steps_enables_preview()]
@@ -39 +44 @@ def compute_is_valid_response(dataset: str) -> Tuple[DatasetIsValidResponse, flo
- return (DatasetIsValidResponse({"valid": is_valid}), 1.0)
+ return DatasetIsValidResponse({"viewer": viewer, "preview": preview})
@@ -50,0 +56,14 @@ class DatasetIsValidJobRunner(DatasetJobRunner):
+ def __init__(
+ self,
+ job_info: JobInfo,
+ app_config: AppConfig,
+ processing_step: ProcessingStep,
+ processing_graph: ProcessingGraph,
+ ) -> None:
+ super().__init__(
+ job_info=job_info,
+ app_config=app_config,
+ processing_step=processing_step,
+ )
+ self.processing_graph = processing_graph
+
@@ -54,2 +73,2 @@ class DatasetIsValidJobRunner(DatasetJobRunner):
- response_content, progress = compute_is_valid_response(dataset=self.dataset)
- return JobResult(response_content, progress=progress)
+ response_content = compute_is_valid_response(dataset=self.dataset, processing_graph=self.processing_graph)
+ return CompleteJobResult(response_content)
diff --git a/services/worker/tests/job_runners/dataset/test_is_valid.py b/services/worker/tests/job_runners/dataset/test_is_valid.py
index b03d2647..a6c64001 100644
--- a/services/worker/tests/job_runners/dataset/test_is_valid.py
+++ b/services/worker/tests/job_runners/dataset/test_is_valid.py
@@ -26,0 +27 @@ GetJobRunner = Callable[[str, AppConfig], DatasetIsValidJobRunner]
+DATASET = "dataset"
@@ -28,2 +29,2 @@ GetJobRunner = Callable[[str, AppConfig], DatasetIsValidJobRunner]
-UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING: UpstreamResponse = UpstreamResponse(
- kind="config-split-names-from-streaming", dataset="dataset_ok", config=None, http_status=HTTPStatus.OK, content={}
+UPSTREAM_RESPONSE_CONFIG_SIZE: UpstreamResponse = UpstreamResponse(
+ kind="config-size", dataset=DATASET, config="config", http_status=HTTPStatus.OK, content={}
@@ -31,2 +32,7 @@ UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING: UpstreamResponse = UpstreamRespons
-UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO: UpstreamResponse = UpstreamResponse(
- kind="config-split-names-from-info", dataset="dataset_ok", config=None, http_status=HTTPStatus.OK, content={}
+UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET: UpstreamResponse = UpstreamResponse(
+ kind="split-first-rows-from-parquet",
+ dataset=DATASET,
+ config="config",
+ split="split",
+ http_status=HTTPStatus.OK,
+ content={},
@@ -36 +42 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING: UpstreamResponse = UpstreamRe
- dataset="dataset_ok",
+ dataset=DATASET,
@@ -37,0 +44 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING: UpstreamResponse = UpstreamRe
+ split="split",
@@ -41,2 +48,2 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING: UpstreamResponse = UpstreamRe
-UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET: UpstreamResponse = UpstreamResponse(
- kind="split-first-rows-from-parquet", dataset="dataset_ok", config="config", http_status=HTTPStatus.OK, content={}
+UPSTREAM_RESPONSE_CONFIG_SIZE_ERROR: UpstreamResponse = UpstreamResponse(
+ kind="config-size", dataset=DATASET, config="config", http_status=HTTPStatus.INTERNAL_SERVER_ERROR, content={}
@@ -44,4 +51,5 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET: UpstreamResponse = UpstreamResp
-UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO_ERROR: UpstreamResponse = UpstreamResponse(
- kind="config-split-names-from-info",
- dataset="dataset_ok",
- config=None,
+UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR: UpstreamResponse = UpstreamResponse(
+ kind="split-first-rows-from-parquet",
+ dataset=DATASET,
+ config="config",
+ split="split",
@@ -53 +61 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR: UpstreamResponse = Upst
- dataset="dataset_ok",
+ dataset=DATASET,
@@ -54,0 +63 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR: UpstreamResponse = Upst
+ split="split",
@@ -58,2 +67,2 @@ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR: UpstreamResponse = Upst
-EXPECTED_OK = (
- {"valid": True},
+EXPECTED_ERROR = (
+ {"viewer": False, "preview": False},
@@ -62,2 +71,10 @@ EXPECTED_OK = (
-EXPECTED_ERROR = (
- {"valid": False},
+EXPECTED_VIEWER_OK = (
+ {"viewer": True, "preview": False},
+ 1.0,
+)
+EXPECTED_PREVIEW_OK = (
+ {"viewer": False, "preview": True},
+ 1.0,
+)
+EXPECTED_BOTH_OK = (
+ {"viewer": True, "preview": True},
@@ -78,8 +95 @@ def get_job_runner(
- processing_graph = ProcessingGraph(
- {
- processing_step_name: {
- "input_type": "dataset",
- "job_runner_version": DatasetIsValidJobRunner.get_job_runner_version(),
- }
- }
- )
+ processing_graph = ProcessingGraph(app_config.processing_graph.specification)
@@ -99,0 +110 @@ def get_job_runner(
+ processing_graph=processing_graph,
@@ -106 +117 @@ def get_job_runner(
- "dataset,upstream_responses,expected_error_code,expected,should_raise",
+ "upstream_responses,expected",
@@ -109 +119,0 @@ def get_job_runner(
- "dataset_ok",
@@ -111 +121,2 @@ def get_job_runner(
- UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING,
+ UPSTREAM_RESPONSE_CONFIG_SIZE,
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET,
@@ -114,3 +125 @@ def get_job_runner(
- None,
- EXPECTED_OK,
- False,
+ EXPECTED_BOTH_OK,
@@ -119 +127,0 @@ def get_job_runner(
- "dataset_ok",
@@ -121,2 +129,2 @@ def get_job_runner(
- UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING,
- UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO,
+ UPSTREAM_RESPONSE_CONFIG_SIZE_ERROR,
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET,
@@ -123,0 +132,5 @@ def get_job_runner(
+ ],
+ EXPECTED_PREVIEW_OK,
+ ),
+ (
+ [
@@ -124,0 +138 @@ def get_job_runner(
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING,
@@ -126,3 +140 @@ def get_job_runner(
- None,
- EXPECTED_OK,
- False,
+ EXPECTED_PREVIEW_OK,
@@ -130,3 +141,0 @@ def get_job_runner(
- ("dataset_ok", [], None, EXPECTED_ERROR, False),
- ("dataset_ok", [UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO], None, EXPECTED_ERROR, False),
- ("dataset_ok", [UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING], None, EXPECTED_ERROR, False),
@@ -134,5 +143,6 @@ def get_job_runner(
- "dataset_ok",
- [UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO_ERROR, UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING],
- None,
- EXPECTED_ERROR,
- False,
+ [
+ UPSTREAM_RESPONSE_CONFIG_SIZE,
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR,
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING,
+ ],
+ EXPECTED_BOTH_OK,
@@ -141,5 +151,6 @@ def get_job_runner(
- "dataset_ok",
- [UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO, UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR],
- None,
- EXPECTED_ERROR,
- False,
+ [
+ UPSTREAM_RESPONSE_CONFIG_SIZE,
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR,
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_STREAMING_ERROR,
+ ],
+ EXPECTED_VIEWER_OK,
@@ -148 +158,0 @@ def get_job_runner(
- "dataset_ok",
@@ -150,3 +160,8 @@ def get_job_runner(
- UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_STREAMING,
- UPSTREAM_RESPONSE_SPLIT_NAMES_FROM_DATASET_INFO_ERROR,
- UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET,
+ UPSTREAM_RESPONSE_CONFIG_SIZE,
+ ],
+ EXPECTED_VIEWER_OK,
+ ),
+ (
+ [
+ UPSTREAM_RESPONSE_CONFIG_SIZE_ERROR,
+ UPSTREAM_RESPONSE_SPLIT_FIRST_ROWS_FROM_PARQUET_ERROR,
@@ -155,3 +170,5 @@ def get_job_runner(
- None,
- EXPECTED_OK,
- False,
+ EXPECTED_ERROR,
+ ),
+ (
+ [],
+ EXPECTED_ERROR,
@@ -164 +180,0 @@ def test_compute(
- dataset: str,
@@ -166 +181,0 @@ def test_compute(
- expected_error_code: str,
@@ -168 +182,0 @@ def test_compute(
- should_raise: bool,
@@ -169,0 +184 @@ def test_compute(
+ dataset = DATASET
@@ -173,8 +188,3 @@ def test_compute(
- if should_raise:
- with pytest.raises(Exception) as e:
- job_runner.compute()
- assert e.typename == expected_error_code
- else:
- compute_result = job_runner.compute()
- assert compute_result.content == expected[0]
- assert compute_result.progress == expected[1]
+ compute_result = job_runner.compute()
+ assert compute_result.content == expected[0]
+ assert compute_result.progress == expected[1]
@@ -187 +197 @@ def test_doesnotexist(app_config: AppConfig, get_job_runner: GetJobRunner) -> No
- assert compute_result.content == {"valid": False}
+ assert compute_result.content == {"viewer": False, "preview": False}
diff --git a/services/worker/tests/job_runners/utils.py b/services/worker/tests/job_runners/utils.py
index 30e58fdf..04d80f76 100644
--- a/services/worker/tests/job_runners/utils.py
+++ b/services/worker/tests/job_runners/utils.py
@@ -8 +8 @@ from typing import Any, Mapping, Optional, TypedDict
-class UpstreamResponse(TypedDict):
+class _UpstreamResponse(TypedDict):
@@ -11 +10,0 @@ class UpstreamResponse(TypedDict):
- config: Optional[str]
@@ -13,0 +13,5 @@ class UpstreamResponse(TypedDict):
+
+
+class UpstreamResponse(_UpstreamResponse, total=False):
+ config: Optional[str]
+ split: Optional[str]
|
|
ee9072dedf7130cc64ffbbe608d9a055003d37e4
|
Sylvain Lesage
| 2023-06-29T07:27:39 |
feat: 🎸 reduce the number of workers back to 20 (#1457)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index a0f96bd5..b0dc9cf6 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -272 +272 @@ workers:
- replicas: 100
+ replicas: 20
|
|
0bb49fbb7fe6bec3237f34a760d1a4f02d85e01b
|
Andrea Francis Soria Jimenez
| 2023-06-28T20:42:34 |
Disable backfill (#1456)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 774978a2..a0f96bd5 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -124 +124 @@ cacheMaintenance:
- action: "backfill"
+ action: "skip"
@@ -129 +129 @@ cacheMaintenance:
- error_codes_to_retry: "CreateCommitError,LockedDatasetTimeoutError"
+ error_codes_to_retry: ""
|
|
f1c6f82d05510f06c666b182bddcb39c42fcdfec
|
Polina Kazakova
| 2023-06-28T18:33:53 |
Update quality target in Makefile for /chart (#1455)
|
diff --git a/chart/Makefile b/chart/Makefile
index 57716f53..779c0593 100644
--- a/chart/Makefile
+++ b/chart/Makefile
@@ -8 +8 @@ quality:
- helm lint --values env/dev.yaml
+ helm lint --values env/staging.yaml
|
|
7ab76da7a74c7f2ddc2656c3c3953ddb1b591d1d
|
Andrea Francis Soria Jimenez
| 2023-06-28T15:53:06 |
Enable backfill (#1454)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 22a72830..774978a2 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -124,2 +124,2 @@ cacheMaintenance:
- action: "skip"
- # ^ allowed values are {skip,backfill,upgrade}
+ action: "backfill"
+ # ^ allowed values are {backfill, collect-metrics,skip}
@@ -129 +129 @@ cacheMaintenance:
- error_codes_to_retry: ""
+ error_codes_to_retry: "CreateCommitError,LockedDatasetTimeoutError"
|
|
5d4d7bd731590ce56ac5448bce35cd5d8c581f7d
|
Andrea Francis Soria Jimenez
| 2023-06-28T15:31:04 |
Increase resources (#1453)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 96acf9f7..22a72830 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -272 +272 @@ workers:
- replicas: 20
+ replicas: 100
@@ -283 +283 @@ workers:
- workerJobTypesBlocked: "dataset-config-names,config-split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan"
+ workerJobTypesBlocked: "dataset-config-names,config-split-names-from-streaming,config-parquet-and-info,split-first-rows-from-parquet,split-first-rows-from-streaming,split-opt-in-out-urls-scan,split-duckdb-index"
|
|
2f6dbe2423a22c7749f34d73f3e54e42dd93f228
|
Andrea Francis Soria Jimenez
| 2023-06-28T14:50:36 |
Change commiter key (#1451)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 1a35e169..732dcf41 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.14.0
+version: 1.14.1
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
index 68b16da1..18e1497c 100644
--- a/chart/templates/_envWorker.tpl
+++ b/chart/templates/_envWorker.tpl
@@ -98 +98 @@
- key: HF_TOKEN
+ key: PARQUET_CONVERTER_HF_TOKEN
|
|
d9c9d42d9a9a3666272c45279a171041a43deab4
|
Andrea Francis Soria Jimenez
| 2023-06-28T14:18:17 |
Adding debug logs for split-duckdb-index (#1449)
|
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py
index daaa0d25..01874c4c 100644
--- a/services/worker/src/worker/job_runners/split/duckdb_index.py
+++ b/services/worker/src/worker/job_runners/split/duckdb_index.py
@@ -163,0 +164 @@ def compute_index_rows(
+ logging.debug(f"try to create branch for {dataset=} with {target_revision=} on {hf_endpoint=}")
@@ -170,0 +172 @@ def compute_index_rows(
+ logging.debug(f"get dataset info for {dataset=} with {target_revision=}")
@@ -175,0 +178 @@ def compute_index_rows(
+ logging.debug(f"delete operations for {dataset=} {delete_operations=}")
@@ -180,0 +184 @@ def compute_index_rows(
+ logging.debug(f"add operations for {dataset=} {add_operations=}")
@@ -189,0 +194 @@ def compute_index_rows(
+ logging.debug(f"create commit {commit_message} for {dataset=} {add_operations=}")
@@ -192,0 +198 @@ def compute_index_rows(
+ logging.debug(f"dataset info for {dataset=} {target_dataset_info=}")
|
|
f17eb0f6371bd6df1dd5f79eb27340e049c46d54
|
Andrea Francis Soria Jimenez
| 2023-06-27T20:01:09 |
Adding other processing steps (#1441)
|
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 7d454058..b36d4535 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -362,0 +363,2 @@ class ProcessingGraphConfig:
+ "config-split-names-from-streaming",
+ "config-parquet-and-info",
diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py
index c75d2104..ea672c6f 100644
--- a/libs/libcommon/tests/test_processing_graph.py
+++ b/libs/libcommon/tests/test_processing_graph.py
@@ -85,0 +86 @@ def graph() -> ProcessingGraph:
+ "split-duckdb-index",
@@ -106,0 +108 @@ def graph() -> ProcessingGraph:
+ "split-duckdb-index",
@@ -298 +300 @@ def graph() -> ProcessingGraph:
- ["config-split-names-from-info"],
+ ["config-split-names-from-info", "config-split-names-from-streaming", "config-parquet-and-info"],
@@ -300,0 +303 @@ def graph() -> ProcessingGraph:
+ "config-split-names-from-streaming",
|
|
228334075f2368885e7a1ed1ae77573165fa12d1
|
Andrea Francis Soria Jimenez
| 2023-06-27T19:12:46 |
Try to fix Duckdb extensions (#1440)
|
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py
index 822ccb50..daaa0d25 100644
--- a/services/worker/src/worker/job_runners/split/duckdb_index.py
+++ b/services/worker/src/worker/job_runners/split/duckdb_index.py
@@ -129,0 +130,4 @@ def compute_index_rows(
+ # index all columns
+ db_path = duckdb_index_file_directory.resolve() / DUCKDB_DEFAULT_INDEX_FILENAME
+ con = duckdb.connect(str(db_path.resolve()))
+
@@ -132 +136 @@ def compute_index_rows(
- duckdb.execute(SET_EXTENSIONS_DIRECTORY_COMMAND.format(directory=extensions_directory))
+ con.execute(SET_EXTENSIONS_DIRECTORY_COMMAND.format(directory=extensions_directory))
@@ -134,4 +138,4 @@ def compute_index_rows(
- duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="httpfs"))
- duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="httpfs"))
- duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="fts"))
- duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="fts"))
+ con.execute(INSTALL_EXTENSION_COMMAND.format(extension="httpfs"))
+ con.execute(LOAD_EXTENSION_COMMAND.format(extension="httpfs"))
+ con.execute(INSTALL_EXTENSION_COMMAND.format(extension="fts"))
+ con.execute(LOAD_EXTENSION_COMMAND.format(extension="fts"))
@@ -139,4 +142,0 @@ def compute_index_rows(
- # index all columns
- db_path = duckdb_index_file_directory.resolve() / DUCKDB_DEFAULT_INDEX_FILENAME
-
- con = duckdb.connect(str(db_path.resolve()))
|
|
b2a91994683ef8107197d2f0e12cbcd17df42011
|
Andrea Francis Soria Jimenez
| 2023-06-27T16:45:36 |
Set Duckdb extensions install directory (#1439)
|
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
index 77a03377..68b16da1 100644
--- a/chart/templates/_envWorker.tpl
+++ b/chart/templates/_envWorker.tpl
@@ -110,0 +111,2 @@
+- name: DUCKDB_INDEX_EXTENSIONS_DIRECTORY
+ value: "/tmp/duckdb-extensions"
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 231ec1f5..7d454058 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -113,0 +114 @@ DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s"
+DUCKDB_INDEX_EXTENSIONS_DIRECTORY: Optional[str] = None
@@ -123,0 +125 @@ class DuckDbIndexConfig:
+ extensions_directory: Optional[str] = DUCKDB_INDEX_EXTENSIONS_DIRECTORY
@@ -137,0 +140 @@ class DuckDbIndexConfig:
+ extensions_directory=env.str(name="EXTENSIONS_DIRECTORY", default=DUCKDB_INDEX_EXTENSIONS_DIRECTORY),
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py
index b715de60..822ccb50 100644
--- a/services/worker/src/worker/job_runners/split/duckdb_index.py
+++ b/services/worker/src/worker/job_runners/split/duckdb_index.py
@@ -48,0 +49 @@ LOAD_EXTENSION_COMMAND = "LOAD '{extension}';"
+SET_EXTENSIONS_DIRECTORY_COMMAND = "SET extension_directory='{directory}';"
@@ -62,0 +64 @@ def compute_index_rows(
+ extensions_directory: Optional[str],
@@ -128,0 +131,3 @@ def compute_index_rows(
+ if extensions_directory is not None:
+ duckdb.execute(SET_EXTENSIONS_DIRECTORY_COMMAND.format(directory=extensions_directory))
+
@@ -259,0 +265 @@ class SplitDuckDbIndexJobRunner(SplitJobRunnerWithCache):
+ extensions_directory=self.duckdb_index_config.extensions_directory,
|
|
be96d0a6693b8405e95e850d5181246c476839db
|
Andrea Francis Soria Jimenez
| 2023-06-27T15:31:40 |
Increase chart version (#1438)
|
diff --git a/chart/Chart.yaml b/chart/Chart.yaml
index 13b7046f..1a35e169 100644
--- a/chart/Chart.yaml
+++ b/chart/Chart.yaml
@@ -21 +21 @@ type: application
-version: 1.13.0
+version: 1.14.0
|
|
b988e218f04a9340d88d1afc6dbd0176b943b92a
|
Andrea Francis Soria Jimenez
| 2023-06-27T13:41:16 |
feat: Index the (text) datasets contents to enable full-text search - DuckDB (#1296)
|
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json
index 11b6b1fe..59a058d9 100644
--- a/chart/static-files/openapi.json
+++ b/chart/static-files/openapi.json
@@ -928 +928 @@
- "items": { "$ref": "#/components/schemas/ParquetFileItem" }
+ "items": { "$ref": "#/components/schemas/SplitHubFile" }
@@ -932 +932 @@
- "ParquetFileItem": {
+ "SplitHubFile": {
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
index 6f2d1411..77a03377 100644
--- a/chart/templates/_envWorker.tpl
+++ b/chart/templates/_envWorker.tpl
@@ -89,0 +90,21 @@
+# specific to 'split-duckdb-index' job runner
+- name: DUCKDB_INDEX_COMMIT_MESSAGE
+ value: {{ .Values.duckDBIndex.commitMessage | quote }}
+- name: DUCKDB_INDEX_COMMITTER_HF_TOKEN
+ {{- if .Values.secrets.appParquetConverterHfToken.fromSecret }}
+ valueFrom:
+ secretKeyRef:
+ name: {{ .Values.secrets.appParquetConverterHfToken.secretName | quote }}
+ key: HF_TOKEN
+ optional: false
+ {{- else }}
+ value: {{ .Values.secrets.appParquetConverterHfToken.value }}
+ {{- end }}
+- name: DUCKDB_INDEX_TARGET_REVISION
+ value: {{ .Values.duckDBIndex.targetRevision | quote }}
+- name: DUCKDB_INDEX_URL_TEMPLATE
+ value: {{ .Values.duckDBIndex.urlTemplate | quote }}
+- name: DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES
+ value: {{ .Values.duckDBIndex.maxParquetSizeBytes | quote }}
+- name: DUCKDB_INDEX_STORAGE_DIRECTORY
+ value: {{ .Values.duckDBIndex.storageDirectory | quote }}
diff --git a/chart/templates/_helpers.tpl b/chart/templates/_helpers.tpl
index 12aeec4c..ac9370e6 100644
--- a/chart/templates/_helpers.tpl
+++ b/chart/templates/_helpers.tpl
@@ -171,0 +172,9 @@ The parquet-metadata/ subpath in the NFS
+{{/*
+The duckdb-index/ subpath in the NFS
+- in a subdirectory named as the chart (datasets-server/), and below it,
+- in a subdirectory named as the Release, so that Releases will not share the same dir
+*/}}
+{{- define "duckDBIndex.subpath" -}}
+{{- printf "%s/%s/%s/" .Chart.Name .Release.Name "duckdb-index" }}
+{{- end }}
+
diff --git a/chart/templates/_initContainerDuckDBIndex.tpl b/chart/templates/_initContainerDuckDBIndex.tpl
new file mode 100644
index 00000000..ed7cb43b
--- /dev/null
+++ b/chart/templates/_initContainerDuckDBIndex.tpl
@@ -0,0 +1,21 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+{{- define "initContainerDuckDBIndex" -}}
+- name: prepare-duckdb-index
+ image: ubuntu:focal
+ imagePullPolicy: {{ .Values.images.pullPolicy }}
+ command: ["/bin/sh", "-c"]
+ args:
+ - chown {{ .Values.uid }}:{{ .Values.gid }} /mounted-path;
+ volumeMounts:
+ - mountPath: /mounted-path
+ mountPropagation: None
+ name: data
+ subPath: "{{ include "duckDBIndex.subpath" . }}"
+ readOnly: false
+ securityContext:
+ runAsNonRoot: false
+ runAsUser: 0
+ runAsGroup: 0
+{{- end -}}
diff --git a/chart/templates/_volumeMountDuckDBIndex.tpl b/chart/templates/_volumeMountDuckDBIndex.tpl
new file mode 100644
index 00000000..01c37b89
--- /dev/null
+++ b/chart/templates/_volumeMountDuckDBIndex.tpl
@@ -0,0 +1,10 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+{{- define "volumeMountDuckDBIndexRW" -}}
+- mountPath: {{ .Values.duckDBIndex.storageDirectory | quote }}
+ mountPropagation: None
+ name: data
+ subPath: "{{ include "duckDBIndex.subpath" . }}"
+ readOnly: false
+{{- end -}}
diff --git a/chart/templates/worker/_container.tpl b/chart/templates/worker/_container.tpl
index 9f83bad8..f9b86817 100644
--- a/chart/templates/worker/_container.tpl
+++ b/chart/templates/worker/_container.tpl
@@ -26,0 +27 @@
+ {{ include "volumeMountDuckDBIndexRW" . | nindent 2 }}
diff --git a/chart/templates/worker/_deployment.yaml b/chart/templates/worker/_deployment.yaml
index e06d319c..03a70646 100644
--- a/chart/templates/worker/_deployment.yaml
+++ b/chart/templates/worker/_deployment.yaml
@@ -28,0 +29 @@ spec:
+ {{ include "initContainerDuckDBIndex" . | nindent 8 }}
diff --git a/chart/values.yaml b/chart/values.yaml
index 4a57fc7a..98cc505b 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -216,0 +217,11 @@ parquetMetadata:
+duckDBIndex:
+ # Directory on the shared storage (used temporarily to prepare the duckdb indexes before sending to the Hub)
+ storageDirectory: "/duckdb-index"
+ # the git commit message when the duckdb index file is uploaded to the Hub. Defaults to `Update duckdb index files`.
+ commitMessage: "Update duckdb index files"
+ # the git revision of the dataset where to store the duckdb index file. Defaults to `refs/convert/parquet`.
+ targetRevision: "refs/convert/parquet"
+ # the URL template to build the duckdb index file URL. Defaults to `/datasets/%s/resolve/%s/%s`.
+ urlTemplate: "/datasets/%s/resolve/%s/%s"
+ # the maximum size of the split parquets.
+ maxParquetSizeBytes: "100_000_000"
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index e4d63e70..231ec1f5 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -26,0 +27 @@ from libcommon.constants import (
+ PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION,
@@ -106,0 +108,33 @@ class ParquetMetadataConfig:
+DUCKDB_INDEX_STORAGE_DIRECTORY = None
+DUCKDB_INDEX_COMMIT_MESSAGE = "Update duckdb index file"
+DUCKDB_INDEX_COMMITTER_HF_TOKEN = None
+DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES = 100_000_000
+DUCKDB_INDEX_TARGET_REVISION = "refs/convert/parquet"
+DUCKDB_INDEX_URL_TEMPLATE = "/datasets/%s/resolve/%s/%s"
+
+
+@dataclass(frozen=True)
+class DuckDbIndexConfig:
+ storage_directory: Optional[str] = DUCKDB_INDEX_STORAGE_DIRECTORY
+ commit_message: str = DUCKDB_INDEX_COMMIT_MESSAGE
+ committer_hf_token: Optional[str] = DUCKDB_INDEX_COMMITTER_HF_TOKEN
+ target_revision: str = DUCKDB_INDEX_TARGET_REVISION
+ url_template: str = DUCKDB_INDEX_URL_TEMPLATE
+ max_parquet_size_bytes: int = DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES
+
+ @classmethod
+ def from_env(cls) -> "DuckDbIndexConfig":
+ env = Env(expand_vars=True)
+ with env.prefixed("DUCKDB_INDEX_"):
+ return cls(
+ storage_directory=env.str(name="STORAGE_DIRECTORY", default=DUCKDB_INDEX_STORAGE_DIRECTORY),
+ commit_message=env.str(name="COMMIT_MESSAGE", default=DUCKDB_INDEX_COMMIT_MESSAGE),
+ committer_hf_token=env.str(name="COMMITTER_HF_TOKEN", default=DUCKDB_INDEX_COMMITTER_HF_TOKEN),
+ target_revision=env.str(name="TARGET_REVISION", default=DUCKDB_INDEX_TARGET_REVISION),
+ url_template=env.str(name="URL_TEMPLATE", default=DUCKDB_INDEX_URL_TEMPLATE),
+ max_parquet_size_bytes=env.int(
+ name="MAX_PARQUET_SIZE_BYTES", default=DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES
+ ),
+ )
+
+
@@ -321,0 +356,7 @@ class ProcessingGraphConfig:
+ "split-duckdb-index": {
+ "input_type": "split",
+ "triggered_by": [
+ "config-split-names-from-info",
+ ],
+ "job_runner_version": PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION,
+ },
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
index cd41126f..66a609b6 100644
--- a/libs/libcommon/src/libcommon/constants.py
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -8,0 +9 @@ PARQUET_METADATA_CACHE_APPNAME = "datasets_server_parquet_metadata"
+DUCKDB_INDEX_CACHE_APPNAME = "datasets_server_duckdb_index"
@@ -38,0 +40 @@ PROCESSING_STEP_SPLIT_IMAGE_URL_COLUMNS_VERSION = 1
+PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION = 1
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py
index f9de8f3f..46e66228 100644
--- a/libs/libcommon/src/libcommon/exceptions.py
+++ b/libs/libcommon/src/libcommon/exceptions.py
@@ -75,0 +76 @@ CacheableErrorCode = Literal[
+ "CacheDirectoryNotInitializedError",
@@ -91,0 +93 @@ CacheableErrorCode = Literal[
+ "DuckDBIndexFileNotFoundError",
@@ -104,0 +107 @@ CacheableErrorCode = Literal[
+ "NoIndexableColumnsError",
@@ -114,0 +118 @@ CacheableErrorCode = Literal[
+ "SplitWithTooBigParquetError",
@@ -138,0 +143,7 @@ class CacheableError(CustomError):
+class CacheDirectoryNotInitializedError(CacheableError):
+ """Raised when the cache directory has not been initialized before job compute."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "CacheDirectoryNotInitializedError", cause, True)
+
+
@@ -234,0 +246,7 @@ class DatasetWithTooBigExternalFilesError(CacheableError):
+class DatasetWithTooManyConfigsError(CacheableError):
+ """Raised when the number of configs of a dataset exceeded the limit."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyConfigsError", cause, True)
+
+
@@ -249,2 +267,2 @@ class DatasetWithTooManyParquetFilesError(CacheableError):
-class LockedDatasetTimeoutError(CacheableError):
- """Raised when a dataset is locked by another job."""
+class DuckDBIndexFileNotFoundError(CacheableError):
+ """Raised when no duckdb index file was found for split."""
@@ -253 +271 @@ class LockedDatasetTimeoutError(CacheableError):
- super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "LockedDatasetTimeoutError", cause, True)
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "DuckDBIndexFileNotFoundError", cause, False)
@@ -357,0 +376,7 @@ class JobManagerExceededMaximumDurationError(CacheableError):
+class LockedDatasetTimeoutError(CacheableError):
+ """Raised when a dataset is locked by another job."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "LockedDatasetTimeoutError", cause, True)
+
+
@@ -371,0 +397,7 @@ class NormalRowsError(CacheableError):
+class NoIndexableColumnsError(CacheableError):
+ """Raised when split does not have string columns to index."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "NoIndexableColumnsError", cause, True)
+
+
@@ -452,0 +485,7 @@ class SplitNotFoundError(CacheableError):
+class SplitWithTooBigParquetError(CacheableError):
+ """Raised when the split parquet size (sum of parquet sizes given) is too big."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "SplitWithTooBigParquetError", cause, False)
+
+
@@ -499,7 +537,0 @@ class UnsupportedExternalFilesError(CacheableError):
-
-
-class DatasetWithTooManyConfigsError(CacheableError):
- """Raised when the number of configs of a dataset exceeded the limit."""
-
- def __init__(self, message: str, cause: Optional[BaseException] = None):
- super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyConfigsError", cause, True)
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py
index 0f8ca107..351cb65e 100644
--- a/libs/libcommon/src/libcommon/parquet_utils.py
+++ b/libs/libcommon/src/libcommon/parquet_utils.py
@@ -22,0 +23 @@ from libcommon.simple_cache import get_previous_step_or_raise
+from libcommon.utils import SplitHubFile
@@ -39,9 +39,0 @@ class FileSystemError(Exception):
-class ParquetFileItem(TypedDict):
- dataset: str
- config: str
- split: str
- url: str
- filename: str
- size: int
-
-
@@ -137 +129 @@ class ParquetIndexWithoutMetadata:
- parquet_file_items: List[ParquetFileItem],
+ parquet_file_items: List[SplitHubFile],
diff --git a/libs/libcommon/src/libcommon/storage.py b/libs/libcommon/src/libcommon/storage.py
index bbef1442..63d9c108 100644
--- a/libs/libcommon/src/libcommon/storage.py
+++ b/libs/libcommon/src/libcommon/storage.py
@@ -14,0 +15 @@ from libcommon.constants import (
+ DUCKDB_INDEX_CACHE_APPNAME,
@@ -83,0 +85,14 @@ def init_parquet_metadata_dir(directory: Optional[StrPath] = None) -> StrPath:
+def init_duckdb_index_cache_dir(directory: Optional[StrPath] = None) -> StrPath:
+ """Initialize the duckdb index directory.
+
+ If directory is None, it will be set to the default duckdb index location on the machine.
+
+ Args:
+ directory (Optional[Union[str, PathLike[str]]], optional): The directory to initialize. Defaults to None.
+
+ Returns:
+ Union[str, PathLike[str]]: The directory.
+ """
+ return init_dir(directory, appname=DUCKDB_INDEX_CACHE_APPNAME)
+
+
diff --git a/libs/libcommon/src/libcommon/utils.py b/libs/libcommon/src/libcommon/utils.py
index 301a7609..b921ea78 100644
--- a/libs/libcommon/src/libcommon/utils.py
+++ b/libs/libcommon/src/libcommon/utils.py
@@ -67,0 +68,9 @@ class JobResult(TypedDict):
+class SplitHubFile(TypedDict):
+ dataset: str
+ config: str
+ split: str
+ url: str
+ filename: str
+ size: int
+
+
diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py
index 49bd37fb..c75d2104 100644
--- a/libs/libcommon/tests/test_processing_graph.py
+++ b/libs/libcommon/tests/test_processing_graph.py
@@ -95,0 +96 @@ def graph() -> ProcessingGraph:
+ "split-duckdb-index",
@@ -102 +103,5 @@ def graph() -> ProcessingGraph:
- ["split-first-rows-from-streaming", "dataset-split-names", "config-opt-in-out-urls-count"],
+ [
+ "split-first-rows-from-streaming",
+ "dataset-split-names",
+ "config-opt-in-out-urls-count",
+ ],
@@ -289,0 +295,11 @@ def graph() -> ProcessingGraph:
+ (
+ "split-duckdb-index",
+ [],
+ ["config-split-names-from-info"],
+ [
+ "config-split-names-from-info",
+ "config-parquet-and-info",
+ "config-info",
+ "dataset-config-names",
+ ],
+ ),
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 0ece7030..f2363ed9 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -996,0 +997,62 @@ idna = ["idna (>=2.1)"]
+[[package]]
+name = "duckdb"
+version = "0.8.1"
+description = "DuckDB embedded database"
+category = "main"
+optional = false
+python-versions = "*"
+files = [
+ {file = "duckdb-0.8.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:14781d21580ee72aba1f5dcae7734674c9b6c078dd60470a08b2b420d15b996d"},
+ {file = "duckdb-0.8.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:f13bf7ab0e56ddd2014ef762ae4ee5ea4df5a69545ce1191b8d7df8118ba3167"},
+ {file = "duckdb-0.8.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:e4032042d8363e55365bbca3faafc6dc336ed2aad088f10ae1a534ebc5bcc181"},
+ {file = "duckdb-0.8.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:31a71bd8f0b0ca77c27fa89b99349ef22599ffefe1e7684ae2e1aa2904a08684"},
+ {file = "duckdb-0.8.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:24568d6e48f3dbbf4a933109e323507a46b9399ed24c5d4388c4987ddc694fd0"},
+ {file = "duckdb-0.8.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:297226c0dadaa07f7c5ae7cbdb9adba9567db7b16693dbd1b406b739ce0d7924"},
+ {file = "duckdb-0.8.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:5792cf777ece2c0591194006b4d3e531f720186102492872cb32ddb9363919cf"},
+ {file = "duckdb-0.8.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:12803f9f41582b68921d6b21f95ba7a51e1d8f36832b7d8006186f58c3d1b344"},
+ {file = "duckdb-0.8.1-cp310-cp310-win32.whl", hash = "sha256:d0953d5a2355ddc49095e7aef1392b7f59c5be5cec8cdc98b9d9dc1f01e7ce2b"},
+ {file = "duckdb-0.8.1-cp310-cp310-win_amd64.whl", hash = "sha256:6e6583c98a7d6637e83bcadfbd86e1f183917ea539f23b6b41178f32f813a5eb"},
+ {file = "duckdb-0.8.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:fad7ed0d4415f633d955ac24717fa13a500012b600751d4edb050b75fb940c25"},
+ {file = "duckdb-0.8.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:81ae602f34d38d9c48dd60f94b89f28df3ef346830978441b83c5b4eae131d08"},
+ {file = "duckdb-0.8.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:7d75cfe563aaa058d3b4ccaaa371c6271e00e3070df5de72361fd161b2fe6780"},
+ {file = "duckdb-0.8.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8dbb55e7a3336f2462e5e916fc128c47fe1c03b6208d6bd413ac11ed95132aa0"},
+ {file = "duckdb-0.8.1-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a6df53efd63b6fdf04657385a791a4e3c4fb94bfd5db181c4843e2c46b04fef5"},
+ {file = "duckdb-0.8.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1b188b80b70d1159b17c9baaf541c1799c1ce8b2af4add179a9eed8e2616be96"},
+ {file = "duckdb-0.8.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:5ad481ee353f31250b45d64b4a104e53b21415577943aa8f84d0af266dc9af85"},
+ {file = "duckdb-0.8.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d1d1b1729993611b1892509d21c21628917625cdbe824a61ce891baadf684b32"},
+ {file = "duckdb-0.8.1-cp311-cp311-win32.whl", hash = "sha256:2d8f9cc301e8455a4f89aa1088b8a2d628f0c1f158d4cf9bc78971ed88d82eea"},
+ {file = "duckdb-0.8.1-cp311-cp311-win_amd64.whl", hash = "sha256:07457a43605223f62d93d2a5a66b3f97731f79bbbe81fdd5b79954306122f612"},
+ {file = "duckdb-0.8.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:d2c8062c3e978dbcd80d712ca3e307de8a06bd4f343aa457d7dd7294692a3842"},
+ {file = "duckdb-0.8.1-cp36-cp36m-win32.whl", hash = "sha256:fad486c65ae944eae2de0d590a0a4fb91a9893df98411d66cab03359f9cba39b"},
+ {file = "duckdb-0.8.1-cp36-cp36m-win_amd64.whl", hash = "sha256:86fa4506622c52d2df93089c8e7075f1c4d0ba56f4bf27faebde8725355edf32"},
+ {file = "duckdb-0.8.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:60e07a62782f88420046e30cc0e3de842d0901c4fd5b8e4d28b73826ec0c3f5e"},
+ {file = "duckdb-0.8.1-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f18563675977f8cbf03748efee0165b4c8ef64e0cbe48366f78e2914d82138bb"},
+ {file = "duckdb-0.8.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:16e179443832bea8439ae4dff93cf1e42c545144ead7a4ef5f473e373eea925a"},
+ {file = "duckdb-0.8.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a413d5267cb41a1afe69d30dd6d4842c588256a6fed7554c7e07dad251ede095"},
+ {file = "duckdb-0.8.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:3784680df59eadd683b0a4c2375d451a64470ca54bd171c01e36951962b1d332"},
+ {file = "duckdb-0.8.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:67a1725c2b01f9b53571ecf3f92959b652f60156c1c48fb35798302e39b3c1a2"},
+ {file = "duckdb-0.8.1-cp37-cp37m-win32.whl", hash = "sha256:197d37e2588c5ad063e79819054eedb7550d43bf1a557d03ba8f8f67f71acc42"},
+ {file = "duckdb-0.8.1-cp37-cp37m-win_amd64.whl", hash = "sha256:3843feb79edf100800f5037c32d5d5a5474fb94b32ace66c707b96605e7c16b2"},
+ {file = "duckdb-0.8.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:624c889b0f2d656794757b3cc4fc58030d5e285f5ad2ef9fba1ea34a01dab7fb"},
+ {file = "duckdb-0.8.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:fcbe3742d77eb5add2d617d487266d825e663270ef90253366137a47eaab9448"},
+ {file = "duckdb-0.8.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:47516c9299d09e9dbba097b9fb339b389313c4941da5c54109df01df0f05e78c"},
+ {file = "duckdb-0.8.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cf1ba718b7522d34399446ebd5d4b9fcac0b56b6ac07bfebf618fd190ec37c1d"},
+ {file = "duckdb-0.8.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e36e35d38a9ae798fe8cf6a839e81494d5b634af89f4ec9483f4d0a313fc6bdb"},
+ {file = "duckdb-0.8.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:23493313f88ce6e708a512daacad13e83e6d1ea0be204b175df1348f7fc78671"},
+ {file = "duckdb-0.8.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:1fb9bf0b6f63616c8a4b9a6a32789045e98c108df100e6bac783dc1e36073737"},
+ {file = "duckdb-0.8.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:12fc13ecd5eddd28b203b9e3999040d3a7374a8f4b833b04bd26b8c5685c2635"},
+ {file = "duckdb-0.8.1-cp38-cp38-win32.whl", hash = "sha256:a12bf4b18306c9cb2c9ba50520317e6cf2de861f121d6f0678505fa83468c627"},
+ {file = "duckdb-0.8.1-cp38-cp38-win_amd64.whl", hash = "sha256:e4e809358b9559c00caac4233e0e2014f3f55cd753a31c4bcbbd1b55ad0d35e4"},
+ {file = "duckdb-0.8.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:7acedfc00d97fbdb8c3d120418c41ef3cb86ef59367f3a9a30dff24470d38680"},
+ {file = "duckdb-0.8.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:99bfe264059cdc1e318769103f656f98e819cd4e231cd76c1d1a0327f3e5cef8"},
+ {file = "duckdb-0.8.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:538b225f361066231bc6cd66c04a5561de3eea56115a5dd773e99e5d47eb1b89"},
+ {file = "duckdb-0.8.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ae0be3f71a18cd8492d05d0fc1bc67d01d5a9457b04822d025b0fc8ee6efe32e"},
+ {file = "duckdb-0.8.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cd82ba63b58672e46c8ec60bc9946aa4dd7b77f21c1ba09633d8847ad9eb0d7b"},
+ {file = "duckdb-0.8.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:780a34559aaec8354e83aa4b7b31b3555f1b2cf75728bf5ce11b89a950f5cdd9"},
+ {file = "duckdb-0.8.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:01f0d4e9f7103523672bda8d3f77f440b3e0155dd3b2f24997bc0c77f8deb460"},
+ {file = "duckdb-0.8.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:31f692decb98c2d57891da27180201d9e93bb470a3051fcf413e8da65bca37a5"},
+ {file = "duckdb-0.8.1-cp39-cp39-win32.whl", hash = "sha256:e7fe93449cd309bbc67d1bf6f6392a6118e94a9a4479ab8a80518742e855370a"},
+ {file = "duckdb-0.8.1-cp39-cp39-win_amd64.whl", hash = "sha256:81d670bc6807672f038332d9bf587037aabdd741b0810de191984325ed307abd"},
+ {file = "duckdb-0.8.1.tar.gz", hash = "sha256:a54d37f4abc2afc4f92314aaa56ecf215a411f40af4bffe1e86bd25e62aceee9"},
+]
+
@@ -5594 +5656 @@ python-versions = "3.9.15"
-content-hash = "109af95b92d54671ee4da9ddd322e3b03b5f9882ec38c9ace064afcbe37b5a2b"
+content-hash = "3aa60ce2866418d5594a71e79a63dbd8e2bd3991c079c53bc055a7c584b3f69e"
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index f5feb938..a2e6034d 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -14,0 +15 @@ conllu = "^4.5.2"
+duckdb = "^0.8.1"
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index bc3fd397..0a57557e 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -11,0 +12 @@ from libcommon.config import (
+ DuckDbIndexConfig,
@@ -251,0 +253 @@ class AppConfig:
+ duckdb_index: DuckDbIndexConfig = field(default_factory=DuckDbIndexConfig)
@@ -269,0 +272 @@ class AppConfig:
+ duckdb_index=DuckDbIndexConfig.from_env(),
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py
index ef4a4cc0..5eb630d1 100644
--- a/services/worker/src/worker/dtos.py
+++ b/services/worker/src/worker/dtos.py
@@ -6,0 +7,2 @@ from typing import Any, Dict, List, Mapping, Optional, TypedDict
+from libcommon.utils import SplitHubFile
+
@@ -113,6 +114,0 @@ class ConfigInfoResponse(TypedDict):
-class ParquetFileItem(SplitItem):
- url: str
- filename: str
- size: int
-
-
@@ -120 +116 @@ class ConfigParquetAndInfoResponse(TypedDict):
- parquet_files: List[ParquetFileItem]
+ parquet_files: List[SplitHubFile]
@@ -137 +133 @@ class ConfigParquetResponse(TypedDict):
- parquet_files: List[ParquetFileItem]
+ parquet_files: List[SplitHubFile]
@@ -186 +182 @@ class DatasetParquetResponse(TypedDict):
- parquet_files: List[ParquetFileItem]
+ parquet_files: List[SplitHubFile]
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 2352944c..87c48d90 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -36,0 +37 @@ from worker.job_runners.dataset.split_names import DatasetSplitNamesJobRunner
+from worker.job_runners.split.duckdb_index import SplitDuckDbIndexJobRunner
@@ -75,0 +77 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ duckdb_index_cache_directory: StrPath
@@ -217,0 +220,8 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ if job_type == SplitDuckDbIndexJobRunner.get_job_type():
+ return SplitDuckDbIndexJobRunner(
+ job_info=job_info,
+ app_config=self.app_config,
+ processing_step=processing_step,
+ duckdb_index_cache_directory=self.duckdb_index_cache_directory,
+ )
+
@@ -236,0 +247 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ SplitDuckDbIndexJobRunner.get_job_type(),
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 8df7ea7a..8e82aa94 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -11 +11 @@ from typing import Any, List, Optional, Set, Tuple
-from urllib.parse import quote, unquote
+from urllib.parse import unquote
@@ -74 +74 @@ from libcommon.simple_cache import get_previous_step_or_raise
-from libcommon.utils import JobInfo
+from libcommon.utils import JobInfo, SplitHubFile
@@ -78 +78 @@ from worker.config import AppConfig, ParquetAndInfoConfig
-from worker.dtos import CompleteJobResult, ConfigParquetAndInfoResponse, ParquetFileItem
+from worker.dtos import CompleteJobResult, ConfigParquetAndInfoResponse
@@ -80 +80 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas
-from worker.utils import retry
+from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url, retry
@@ -100,6 +99,0 @@ class ParquetFile:
-# TODO: use huggingface_hub's hf_hub_url after
-# https://github.com/huggingface/huggingface_hub/issues/1082
-def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url_template: str) -> str:
- return (hf_endpoint + url_template) % (repo_id, quote(revision, safe=""), filename)
-
-
@@ -128 +122 @@ def create_parquet_file_item(
-) -> ParquetFileItem:
+) -> SplitHubFile:
@@ -1037 +1030,0 @@ def compute_config_parquet_and_info_response(
- sleeps = [1, 1, 1, 1, 1, 10, 10, 10, 10, 100] * 3
@@ -1039 +1032,3 @@ def compute_config_parquet_and_info_response(
- with lock.git_branch(dataset=dataset, branch=target_revision, owner=job_id, sleeps=sleeps):
+ with lock.git_branch(
+ dataset=dataset, branch=target_revision, owner=job_id, sleeps=LOCK_GIT_BRANCH_RETRY_SLEEPS
+ ):
@@ -1042,2 +1037,5 @@ def compute_config_parquet_and_info_response(
- refs = retry(on=[requests.exceptions.ConnectionError], sleeps=[1, 1, 1, 10, 10])(hf_api.list_repo_refs)(
- repo_id=dataset, repo_type=DATASET_TYPE
+ create_branch(
+ dataset=dataset,
+ target_revision=target_revision,
+ hf_api=hf_api,
+ committer_hf_api=committer_hf_api,
@@ -1045,9 +1043 @@ def compute_config_parquet_and_info_response(
- if all(ref.ref != target_revision for ref in refs.converts):
- initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id
- committer_hf_api.create_branch(
- repo_id=dataset,
- branch=target_revision,
- repo_type=DATASET_TYPE,
- revision=initial_commit,
- exist_ok=True,
- )
+
diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py
index 29a34f9a..55de2fe3 100644
--- a/services/worker/src/worker/job_runners/config/parquet_metadata.py
+++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py
@@ -8 +7,0 @@ from typing import List, Optional
-from datasets.utils.file_utils import get_authentication_headers_for_url
@@ -19 +18 @@ from libcommon.storage import StrPath
-from libcommon.utils import JobInfo
+from libcommon.utils import JobInfo, SplitHubFile
@@ -28 +26,0 @@ from worker.dtos import (
- ParquetFileItem,
@@ -32,5 +30 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-
-
-def get_parquet_file(url: str, fs: HTTPFileSystem, hf_token: Optional[str]) -> ParquetFile:
- headers = get_authentication_headers_for_url(url, use_auth_token=hf_token)
- return ParquetFile(fs.open(url, headers=headers))
+from worker.utils import get_parquet_file
@@ -73 +67 @@ def compute_parquet_metadata_response(
- parquet_file_items: List[ParquetFileItem] = [
+ parquet_file_items: List[SplitHubFile] = [
diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py
index 9ba4eddc..000388ff 100644
--- a/services/worker/src/worker/job_runners/dataset/parquet.py
+++ b/services/worker/src/worker/job_runners/dataset/parquet.py
@@ -14,0 +15 @@ from libcommon.simple_cache import (
+from libcommon.utils import SplitHubFile
@@ -20 +20,0 @@ from worker.dtos import (
- ParquetFileItem,
@@ -49 +49 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float]
- parquet_files: list[ParquetFileItem] = []
+ parquet_files: list[SplitHubFile] = []
diff --git a/services/worker/src/worker/job_runners/split/duckdb_index.py b/services/worker/src/worker/job_runners/split/duckdb_index.py
new file mode 100644
index 00000000..b715de60
--- /dev/null
+++ b/services/worker/src/worker/job_runners/split/duckdb_index.py
@@ -0,0 +1,265 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+import logging
+from pathlib import Path
+from typing import List, Optional, Set
+
+import duckdb
+from huggingface_hub._commit_api import (
+ CommitOperation,
+ CommitOperationAdd,
+ CommitOperationDelete,
+)
+from huggingface_hub.hf_api import HfApi
+from huggingface_hub.utils._errors import RepositoryNotFoundError
+from libcommon.config import DuckDbIndexConfig
+from libcommon.constants import PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION
+from libcommon.exceptions import (
+ CacheDirectoryNotInitializedError,
+ DatasetNotFoundError,
+ DuckDBIndexFileNotFoundError,
+ LockedDatasetTimeoutError,
+ NoIndexableColumnsError,
+ ParquetResponseEmptyError,
+ PreviousStepFormatError,
+ SplitNotFoundError,
+ SplitWithTooBigParquetError,
+)
+from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import lock
+from libcommon.simple_cache import get_previous_step_or_raise
+from libcommon.storage import StrPath
+from libcommon.utils import JobInfo, SplitHubFile
+
+from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
+from worker.job_runners.split.split_job_runner import SplitJobRunnerWithCache
+from worker.utils import LOCK_GIT_BRANCH_RETRY_SLEEPS, create_branch, hf_hub_url
+
+DATASET_TYPE = "dataset"
+STRING_FEATURE_DTYPE = "string"
+VALUE_FEATURE_TYPE = "Value"
+DUCKDB_DEFAULT_INDEX_FILENAME = "index.duckdb"
+CREATE_SEQUENCE_COMMAND = "CREATE OR REPLACE SEQUENCE serial START 1;"
+CREATE_INDEX_COMMAND = "PRAGMA create_fts_index('data', '__hf_index_id', '*', overwrite=1);"
+CREATE_TABLE_COMMAND = "CREATE OR REPLACE TABLE data AS SELECT nextval('serial') AS __hf_index_id, {columns} FROM"
+INSTALL_EXTENSION_COMMAND = "INSTALL '{extension}';"
+LOAD_EXTENSION_COMMAND = "LOAD '{extension}';"
+
+
+def compute_index_rows(
+ job_id: str,
+ dataset: str,
+ config: str,
+ split: str,
+ duckdb_index_file_directory: Path,
+ target_revision: str,
+ hf_endpoint: str,
+ commit_message: str,
+ url_template: str,
+ hf_token: Optional[str],
+ max_parquet_size_bytes: int,
+ committer_hf_token: Optional[str],
+) -> SplitHubFile:
+ logging.info(f"get split-duckdb-index for dataset={dataset} config={config} split={split}")
+
+ # validate split
+ split_names_best_response = get_previous_step_or_raise(
+ kinds=["config-split-names-from-streaming", "config-split-names-from-info"], dataset=dataset, config=config
+ )
+ try:
+ splits_content = split_names_best_response.response["content"]["splits"]
+ except Exception as e:
+ raise PreviousStepFormatError("Previous step did not return the expected content.", e) from e
+
+ if split not in [split_item["split"] for split_item in splits_content]:
+ raise SplitNotFoundError(f"The split '{split}' does not exist for the config '{config}' of the dataset.")
+
+ # get parquet urls and dataset_info
+ config_parquet_and_info_step = "config-parquet-and-info"
+ parquet_and_info_best_response = get_previous_step_or_raise(
+ kinds=[config_parquet_and_info_step],
+ dataset=dataset,
+ config=config,
+ )
+ content_parquet_and_info = parquet_and_info_best_response.response["content"]
+ try:
+ split_parquet_files = [
+ parquet_file
+ for parquet_file in content_parquet_and_info["parquet_files"]
+ if parquet_file["config"] == config and parquet_file["split"] == split
+ ]
+
+ split_parquets_size = sum(parquet_file["size"] for parquet_file in split_parquet_files)
+
+ if split_parquets_size > max_parquet_size_bytes:
+ raise SplitWithTooBigParquetError(
+ f"The indexing is limited to split parquets under {max_parquet_size_bytes} bytes. "
+ f"Current size of sum of split parquets is {split_parquets_size} bytes."
+ )
+
+ parquet_urls = [parquet_file["url"] for parquet_file in split_parquet_files]
+
+ if not parquet_urls:
+ raise ParquetResponseEmptyError("No parquet files found.")
+
+ # get the features
+ features = content_parquet_and_info["dataset_info"]["features"]
+ column_names = ",".join(list(features.keys()))
+
+ # look for string columns
+ string_columns = [
+ column
+ for column, feature in features.items()
+ if "dtype" in feature
+ and "_type" in feature
+ and feature["dtype"] == STRING_FEATURE_DTYPE
+ and feature["_type"] == VALUE_FEATURE_TYPE
+ ]
+ if not string_columns:
+ raise NoIndexableColumnsError("No string columns available to index.")
+
+ except KeyError as e:
+ raise PreviousStepFormatError(
+ f"Previous step '{config_parquet_and_info_step}' did not return the expected content.", e
+ ) from e
+
+ # configure duckdb extensions
+ duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="httpfs"))
+ duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="httpfs"))
+ duckdb.execute(INSTALL_EXTENSION_COMMAND.format(extension="fts"))
+ duckdb.execute(LOAD_EXTENSION_COMMAND.format(extension="fts"))
+
+ # index all columns
+ db_path = duckdb_index_file_directory.resolve() / DUCKDB_DEFAULT_INDEX_FILENAME
+
+ con = duckdb.connect(str(db_path.resolve()))
+ logging.debug(CREATE_SEQUENCE_COMMAND)
+ con.sql(CREATE_SEQUENCE_COMMAND)
+
+ create_command_sql = f"{CREATE_TABLE_COMMAND.format(columns=column_names)} read_parquet({parquet_urls});"
+ logging.debug(create_command_sql)
+ con.sql(create_command_sql)
+
+ # TODO: by default, 'porter' stemmer is being used, use a specific one by dataset language in the future
+ # see https://duckdb.org/docs/extensions/full_text_search.html for more details about 'stemmer' parameter
+ logging.debug(CREATE_INDEX_COMMAND)
+ con.sql(CREATE_INDEX_COMMAND)
+ con.close()
+
+ hf_api = HfApi(endpoint=hf_endpoint, token=hf_token)
+ committer_hf_api = HfApi(endpoint=hf_endpoint, token=committer_hf_token)
+ index_file_location = f"{config}/{split}/{DUCKDB_DEFAULT_INDEX_FILENAME}"
+
+ try:
+ with lock.git_branch(
+ dataset=dataset, branch=target_revision, owner=job_id, sleeps=LOCK_GIT_BRANCH_RETRY_SLEEPS
+ ):
+ create_branch(
+ dataset=dataset,
+ target_revision=target_revision,
+ hf_api=hf_api,
+ committer_hf_api=committer_hf_api,
+ )
+
+ target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False)
+ all_repo_files: Set[str] = {f.rfilename for f in target_dataset_info.siblings}
+ delete_operations: List[CommitOperation] = []
+ if index_file_location in all_repo_files:
+ delete_operations.append(CommitOperationDelete(path_in_repo=index_file_location))
+
+ # send the files to the target revision
+ add_operations: List[CommitOperation] = [
+ CommitOperationAdd(path_in_repo=index_file_location, path_or_fileobj=db_path.resolve())
+ ]
+
+ committer_hf_api.create_commit(
+ repo_id=dataset,
+ repo_type=DATASET_TYPE,
+ revision=target_revision,
+ operations=delete_operations + add_operations,
+ commit_message=commit_message,
+ parent_commit=target_dataset_info.sha,
+ )
+
+ # call the API again to get the index file
+ target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True)
+ except TimeoutError as err:
+ raise LockedDatasetTimeoutError("the dataset is currently locked, please try again later.") from err
+ except RepositoryNotFoundError as err:
+ raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err
+
+ repo_files = [
+ repo_file for repo_file in target_dataset_info.siblings if repo_file.rfilename == index_file_location
+ ]
+
+ if not repo_files or len(repo_files) != 1:
+ logging.warning(f"Found {len(repo_files)} index files, should be only 1")
+ raise DuckDBIndexFileNotFoundError("No index file was found")
+
+ repo_file = repo_files[0]
+ if repo_file.size is None:
+ raise ValueError(f"Cannot get size of {repo_file.rfilename}")
+
+ return SplitHubFile(
+ dataset=dataset,
+ config=config,
+ split=split,
+ url=hf_hub_url(
+ repo_id=dataset,
+ filename=repo_file.rfilename,
+ hf_endpoint=hf_endpoint,
+ revision=target_revision,
+ url_template=url_template,
+ ),
+ filename=Path(repo_file.rfilename).name,
+ size=repo_file.size,
+ )
+
+
+class SplitDuckDbIndexJobRunner(SplitJobRunnerWithCache):
+ duckdb_index_config: DuckDbIndexConfig
+
+ def __init__(
+ self,
+ job_info: JobInfo,
+ app_config: AppConfig,
+ processing_step: ProcessingStep,
+ duckdb_index_cache_directory: StrPath,
+ ) -> None:
+ super().__init__(
+ job_info=job_info,
+ app_config=app_config,
+ processing_step=processing_step,
+ cache_directory=Path(duckdb_index_cache_directory),
+ )
+ self.duckdb_index_config = app_config.duckdb_index
+
+ @staticmethod
+ def get_job_type() -> str:
+ return "split-duckdb-index"
+
+ @staticmethod
+ def get_job_runner_version() -> int:
+ return PROCESSING_STEP_SPLIT_DUCKDB_INDEX_VERSION
+
+ def compute(self) -> CompleteJobResult:
+ if self.cache_subdirectory is None:
+ raise CacheDirectoryNotInitializedError("Cache directory has not been initialized.")
+ return CompleteJobResult(
+ compute_index_rows(
+ job_id=self.job_info["job_id"],
+ dataset=self.dataset,
+ config=self.config,
+ split=self.split,
+ duckdb_index_file_directory=self.cache_subdirectory,
+ hf_token=self.app_config.common.hf_token,
+ url_template=self.duckdb_index_config.url_template,
+ commit_message=self.duckdb_index_config.commit_message,
+ committer_hf_token=self.duckdb_index_config.committer_hf_token,
+ hf_endpoint=self.app_config.common.hf_endpoint,
+ target_revision=self.duckdb_index_config.target_revision,
+ max_parquet_size_bytes=self.duckdb_index_config.max_parquet_size_bytes,
+ )
+ )
diff --git a/services/worker/src/worker/main.py b/services/worker/src/worker/main.py
index da297ccc..4d207280 100644
--- a/services/worker/src/worker/main.py
+++ b/services/worker/src/worker/main.py
@@ -9 +9,5 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.storage import init_assets_dir, init_parquet_metadata_dir
+from libcommon.storage import (
+ init_assets_dir,
+ init_duckdb_index_cache_dir,
+ init_parquet_metadata_dir,
+)
@@ -29,0 +34 @@ if __name__ == "__main__":
+ duckdb_index_cache_directory = init_duckdb_index_cache_dir(directory=app_config.duckdb_index.storage_directory)
@@ -56,0 +62 @@ if __name__ == "__main__":
+ duckdb_index_cache_directory=duckdb_index_cache_directory,
diff --git a/services/worker/src/worker/start_worker_loop.py b/services/worker/src/worker/start_worker_loop.py
index 5aa4246d..d5e69a82 100644
--- a/services/worker/src/worker/start_worker_loop.py
+++ b/services/worker/src/worker/start_worker_loop.py
@@ -9 +9,5 @@ from libcommon.resources import CacheMongoResource, QueueMongoResource
-from libcommon.storage import init_assets_dir, init_parquet_metadata_dir
+from libcommon.storage import (
+ init_assets_dir,
+ init_duckdb_index_cache_dir,
+ init_parquet_metadata_dir,
+)
@@ -28,0 +33 @@ if __name__ == "__main__":
+ duckdb_index_cache_directory = init_duckdb_index_cache_dir(directory=app_config.duckdb_index.storage_directory)
@@ -55,0 +61 @@ if __name__ == "__main__":
+ duckdb_index_cache_directory=duckdb_index_cache_directory,
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index fab6b031..c3b985d8 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -20,0 +21 @@ from typing import (
+from urllib.parse import quote
@@ -22,0 +24 @@ import PIL
+import requests
@@ -31 +33,9 @@ from datasets import (
-from libcommon.exceptions import NormalRowsError, StreamingRowsError
+from datasets.utils.file_utils import get_authentication_headers_for_url
+from fsspec.implementations.http import HTTPFileSystem
+from huggingface_hub.hf_api import HfApi
+from huggingface_hub.utils._errors import RepositoryNotFoundError
+from libcommon.exceptions import (
+ DatasetNotFoundError,
+ NormalRowsError,
+ StreamingRowsError,
+)
@@ -32,0 +43 @@ from libcommon.utils import orjson_dumps
+from pyarrow.parquet import ParquetFile
@@ -312,0 +324,31 @@ def get_rows_or_raise(
+
+
+# TODO: use huggingface_hub's hf_hub_url after
+# https://github.com/huggingface/huggingface_hub/issues/1082
+def hf_hub_url(repo_id: str, filename: str, hf_endpoint: str, revision: str, url_template: str) -> str:
+ return (hf_endpoint + url_template) % (repo_id, quote(revision, safe=""), filename)
+
+
+def get_parquet_file(url: str, fs: HTTPFileSystem, hf_token: Optional[str]) -> ParquetFile:
+ headers = get_authentication_headers_for_url(url, use_auth_token=hf_token)
+ return ParquetFile(fs.open(url, headers=headers))
+
+
+DATASET_TYPE = "dataset"
+
+LIST_REPO_REFS_RETRY_SLEEPS = [1, 1, 1, 10, 10]
+LOCK_GIT_BRANCH_RETRY_SLEEPS = [1, 1, 1, 1, 1, 10, 10, 10, 10, 100] * 3
+
+
+def create_branch(dataset: str, target_revision: str, hf_api: HfApi, committer_hf_api: HfApi) -> None:
+ try:
+ refs = retry(on=[requests.exceptions.ConnectionError], sleeps=LIST_REPO_REFS_RETRY_SLEEPS)(
+ hf_api.list_repo_refs
+ )(repo_id=dataset, repo_type=DATASET_TYPE)
+ if all(ref.ref != target_revision for ref in refs.converts):
+ initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id
+ committer_hf_api.create_branch(
+ repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit, exist_ok=True
+ )
+ except RepositoryNotFoundError as err:
+ raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index 851a99c6..fab26487 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -11 +11,6 @@ from libcommon.simple_cache import _clean_cache_database
-from libcommon.storage import StrPath, init_assets_dir, init_parquet_metadata_dir
+from libcommon.storage import (
+ StrPath,
+ init_assets_dir,
+ init_duckdb_index_cache_dir,
+ init_parquet_metadata_dir,
+)
@@ -67,0 +73 @@ def set_env_vars(
+ mp.setenv("DUCKDB_INDEX_COMMITTER_HF_TOKEN", CI_PARQUET_CONVERTER_APP_TOKEN)
@@ -121,0 +128,5 @@ def parquet_metadata_directory(app_config: AppConfig) -> StrPath:
+@fixture
+def duckdb_index_cache_directory(app_config: AppConfig) -> StrPath:
+ return init_duckdb_index_cache_dir(app_config.duckdb_index.storage_directory)
+
+
diff --git a/services/worker/tests/fixtures/datasets.py b/services/worker/tests/fixtures/datasets.py
index 18edaad1..6e987e20 100644
--- a/services/worker/tests/fixtures/datasets.py
+++ b/services/worker/tests/fixtures/datasets.py
@@ -145,0 +146,17 @@ def datasets() -> Mapping[str, Dataset]:
+ "duckdb_index": Dataset.from_pandas(
+ pd.DataFrame(
+ {
+ "text": [
+ (
+ "Grand Moff Tarkin and Lord Vader are interrupted in their discussion by the buzz of the"
+ " comlink"
+ ),
+ "There goes another one.",
+ "Vader turns round and round in circles as his ship spins into space.",
+ "We count thirty Rebel ships, Lord Vader.",
+ "The wingman spots the pirateship coming at him and warns the Dark Lord",
+ ]
+ },
+ dtype=pd.StringDtype(storage="python"),
+ )
+ ),
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index f34ca587..11e9d9a0 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -283,0 +284,7 @@ def hub_public_spawning_opt_in_out(datasets: Mapping[str, Dataset]) -> Iterator[
[email protected](scope="session")
+def hub_public_duckdb_index(datasets: Mapping[str, Dataset]) -> Iterator[str]:
+ repo_id = create_hub_dataset_repo(prefix="duckdb_index", dataset=datasets["duckdb_index"])
+ yield repo_id
+ delete_hub_dataset_repo(repo_id=repo_id)
+
+
@@ -766,0 +774,13 @@ def hub_reponses_spawning_opt_in_out(hub_public_spawning_opt_in_out: str) -> Hub
+
+
[email protected]
+def hub_responses_duckdb_index(hub_public_duckdb_index: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_duckdb_index,
+ "config_names_response": create_config_names_response(hub_public_duckdb_index),
+ "splits_response": create_splits_response(hub_public_duckdb_index),
+ "first_rows_response": create_first_rows_response(hub_public_duckdb_index, TEXT_cols, TEXT_rows),
+ "parquet_and_info_response": create_parquet_and_info_response(
+ dataset=hub_public_duckdb_index, data_type="csv"
+ ),
+ }
diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py
index 031fee96..e314000a 100644
--- a/services/worker/tests/job_runners/config/test_parquet.py
+++ b/services/worker/tests/job_runners/config/test_parquet.py
@@ -12 +12 @@ from libcommon.simple_cache import CachedArtifactError, upsert_response
-from libcommon.utils import Priority
+from libcommon.utils import Priority, SplitHubFile
@@ -15,5 +15 @@ from worker.config import AppConfig
-from worker.dtos import (
- ConfigParquetAndInfoResponse,
- ConfigParquetResponse,
- ParquetFileItem,
-)
+from worker.dtos import ConfigParquetAndInfoResponse, ConfigParquetResponse
@@ -81 +77 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
@@ -84 +80 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
@@ -93 +89 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
@@ -96 +92 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py
index 613d616a..96f1cfbc 100644
--- a/services/worker/tests/job_runners/config/test_parquet_metadata.py
+++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py
@@ -22 +22 @@ from libcommon.storage import StrPath
-from libcommon.utils import Priority
+from libcommon.utils import Priority, SplitHubFile
@@ -28 +27,0 @@ from worker.dtos import (
- ParquetFileItem,
@@ -100 +99 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
@@ -103 +102 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py
index 61cfbfbc..8f63b188 100644
--- a/services/worker/tests/job_runners/dataset/test_parquet.py
+++ b/services/worker/tests/job_runners/dataset/test_parquet.py
@@ -12 +12 @@ from libcommon.simple_cache import CachedArtifactError, upsert_response
-from libcommon.utils import Priority
+from libcommon.utils import Priority, SplitHubFile
@@ -15 +15 @@ from worker.config import AppConfig
-from worker.dtos import ConfigParquetResponse, DatasetParquetResponse, ParquetFileItem
+from worker.dtos import ConfigParquetResponse, DatasetParquetResponse
@@ -92 +92 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
@@ -110 +110 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
@@ -125 +125 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
@@ -128 +128 @@ def get_job_runner(
- ParquetFileItem(
+ SplitHubFile(
diff --git a/services/worker/tests/job_runners/split/test_duckdb_index.py b/services/worker/tests/job_runners/split/test_duckdb_index.py
new file mode 100644
index 00000000..a8fe40cf
--- /dev/null
+++ b/services/worker/tests/job_runners/split/test_duckdb_index.py
@@ -0,0 +1,242 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+import os
+from dataclasses import replace
+from http import HTTPStatus
+from typing import Callable, Optional
+
+import duckdb
+import pytest
+import requests
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.simple_cache import upsert_response
+from libcommon.storage import StrPath
+from libcommon.utils import Priority
+
+from worker.config import AppConfig
+from worker.job_runners.config.parquet_and_info import ConfigParquetAndInfoJobRunner
+from worker.job_runners.split.duckdb_index import SplitDuckDbIndexJobRunner
+from worker.resources import LibrariesResource
+
+from ...fixtures.hub import HubDatasetTest
+
+GetJobRunner = Callable[[str, str, str, AppConfig], SplitDuckDbIndexJobRunner]
+
+GetParquetJobRunner = Callable[[str, str, AppConfig], ConfigParquetAndInfoJobRunner]
+
+
[email protected]
+def get_job_runner(
+ duckdb_index_cache_directory: StrPath,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ config: str,
+ split: str,
+ app_config: AppConfig,
+ ) -> SplitDuckDbIndexJobRunner:
+ processing_step_name = SplitDuckDbIndexJobRunner.get_job_type()
+ processing_graph = ProcessingGraph(
+ {
+ "dataset-step": {"input_type": "dataset"},
+ "config-parquet": {
+ "input_type": "config",
+ "triggered_by": "dataset-step",
+ "provides_config_parquet": True,
+ },
+ "config-split-names-from-streaming": {
+ "input_type": "config",
+ "triggered_by": "dataset-step",
+ },
+ processing_step_name: {
+ "input_type": "dataset",
+ "job_runner_version": SplitDuckDbIndexJobRunner.get_job_runner_version(),
+ "triggered_by": ["config-parquet", "config-split-names-from-streaming"],
+ },
+ }
+ )
+ return SplitDuckDbIndexJobRunner(
+ job_info={
+ "type": SplitDuckDbIndexJobRunner.get_job_type(),
+ "params": {
+ "dataset": dataset,
+ "revision": "revision",
+ "config": config,
+ "split": split,
+ },
+ "job_id": "job_id",
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=processing_graph.get_processing_step(processing_step_name),
+ duckdb_index_cache_directory=duckdb_index_cache_directory,
+ )
+
+ return _get_job_runner
+
+
[email protected]
+def get_parquet_job_runner(
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetParquetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ config: str,
+ app_config: AppConfig,
+ ) -> ConfigParquetAndInfoJobRunner:
+ processing_step_name = ConfigParquetAndInfoJobRunner.get_job_type()
+ processing_graph = ProcessingGraph(
+ {
+ "dataset-level": {"input_type": "dataset"},
+ processing_step_name: {
+ "input_type": "config",
+ "job_runner_version": ConfigParquetAndInfoJobRunner.get_job_runner_version(),
+ "triggered_by": "dataset-level",
+ },
+ }
+ )
+ return ConfigParquetAndInfoJobRunner(
+ job_info={
+ "type": ConfigParquetAndInfoJobRunner.get_job_type(),
+ "params": {
+ "dataset": dataset,
+ "revision": "revision",
+ "config": config,
+ "split": None,
+ },
+ "job_id": "job_id",
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=processing_graph.get_processing_step(processing_step_name),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_job_runner
+
+
[email protected](
+ "hub_dataset_name,max_parquet_size_bytes,expected_error_code",
+ [
+ ("duckdb_index", None, None),
+ ("duckdb_index", 1_000, "SplitWithTooBigParquetError"), # parquet size is 2812
+ ("public", None, "NoIndexableColumnsError"), # dataset does not have string columns to index
+ ],
+)
+def test_compute(
+ get_parquet_job_runner: GetParquetJobRunner,
+ get_job_runner: GetJobRunner,
+ app_config: AppConfig,
+ hub_responses_public: HubDatasetTest,
+ hub_responses_duckdb_index: HubDatasetTest,
+ hub_dataset_name: str,
+ max_parquet_size_bytes: Optional[int],
+ expected_error_code: str,
+) -> None:
+ hub_datasets = {"public": hub_responses_public, "duckdb_index": hub_responses_duckdb_index}
+ dataset = hub_datasets[hub_dataset_name]["name"]
+ config_names = hub_datasets[hub_dataset_name]["config_names_response"]
+ config = hub_datasets[hub_dataset_name]["config_names_response"]["config_names"][0]["config"]
+ splits_response = hub_datasets[hub_dataset_name]["splits_response"]
+ split = "train"
+
+ upsert_response(
+ "dataset-config-names",
+ dataset=dataset,
+ http_status=HTTPStatus.OK,
+ content=config_names,
+ )
+
+ upsert_response(
+ "config-split-names-from-streaming",
+ dataset=dataset,
+ config=config,
+ http_status=HTTPStatus.OK,
+ content=splits_response,
+ )
+
+ app_config = (
+ app_config
+ if max_parquet_size_bytes is None
+ else replace(
+ app_config, duckdb_index=replace(app_config.duckdb_index, max_parquet_size_bytes=max_parquet_size_bytes)
+ )
+ )
+
+ parquet_job_runner = get_parquet_job_runner(dataset, config, app_config)
+ parquet_response = parquet_job_runner.compute()
+ config_parquet = parquet_response.content
+
+ # simulate more than one parquet file to index
+ extra_parquet_file = config_parquet["parquet_files"][0]
+ config_parquet["parquet_files"].append(extra_parquet_file)
+
+ upsert_response(
+ "config-parquet-and-info",
+ dataset=dataset,
+ config=config,
+ http_status=HTTPStatus.OK,
+ content=config_parquet,
+ )
+
+ assert parquet_response
+ job_runner = get_job_runner(dataset, config, split, app_config)
+ job_runner.pre_compute()
+
+ if expected_error_code:
+ with pytest.raises(Exception) as e:
+ job_runner.compute()
+ assert e.typename == expected_error_code
+ else:
+ job_runner.pre_compute()
+ response = job_runner.compute()
+ assert response
+ content = response.content
+ url = content["url"]
+ file_name = content["filename"]
+ assert url is not None
+ assert file_name is not None
+ job_runner.post_compute()
+
+ # download locally duckdb index file
+ duckdb_file = requests.get(url)
+ with open(file_name, "wb") as f:
+ f.write(duckdb_file.content)
+
+ duckdb.execute("INSTALL 'fts';")
+ duckdb.execute("LOAD 'fts';")
+ con = duckdb.connect(file_name)
+
+ # validate number of inserted records
+ record_count = con.sql("SELECT COUNT(*) FROM data;").fetchall()
+ assert record_count is not None
+ assert isinstance(record_count, list)
+ assert record_count[0] == (10,) # dataset has 5 rows but since parquet file was duplicate it is 10
+
+ # perform a search to validate fts feature
+ query = "Lord Vader"
+ result = con.execute(
+ "SELECT text FROM data WHERE fts_main_data.match_bm25(__hf_index_id, ?) IS NOT NULL;",
+ [query],
+ )
+ rows = result.df()
+ assert rows is not None
+ assert (rows["text"].eq("Vader turns round and round in circles as his ship spins into space.")).any()
+ assert (rows["text"].eq("The wingman spots the pirateship coming at him and warns the Dark Lord")).any()
+ assert (rows["text"].eq("We count thirty Rebel ships, Lord Vader.")).any()
+ assert (
+ rows["text"].eq(
+ "Grand Moff Tarkin and Lord Vader are interrupted in their discussion by the buzz of the comlink"
+ )
+ ).any()
+ assert not (rows["text"].eq("There goes another one.")).any()
+
+ con.close()
+ os.remove(file_name)
+ job_runner.post_compute()
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index 87f44cd5..d83cbf26 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -201,0 +202 @@ def job_runner_factory(
+ duckdb_index_cache_directory: StrPath,
@@ -209,0 +211 @@ def job_runner_factory(
+ duckdb_index_cache_directory=duckdb_index_cache_directory,
diff --git a/services/worker/tests/test_job_runner_factory.py b/services/worker/tests/test_job_runner_factory.py
index 3ed3b0e7..e10bc8c0 100644
--- a/services/worker/tests/test_job_runner_factory.py
+++ b/services/worker/tests/test_job_runner_factory.py
@@ -41,0 +42 @@ def test_create_job_runner(
+ duckdb_index_cache_directory: StrPath,
@@ -50,0 +52 @@ def test_create_job_runner(
+ duckdb_index_cache_directory=duckdb_index_cache_directory,
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index d6b2bd92..37b1c87d 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -93,0 +94 @@ services:
+ - duckdb-index:${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index}:rw
@@ -114,0 +116,6 @@ services:
+ DUCKDB_INDEX_STORAGE_DIRECTORY: ${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index}
+ DUCKDB_INDEX_COMMIT_MESSAGE: ${DUCKDB_INDEX_COMMIT_MESSAGE-Update duckdb index file}
+ DUCKDB_INDEX_COMMITTER_HF_TOKEN: ${DUCKDB_INDEX_COMMITTER_HF_TOKEN-}
+ DUCKDB_INDEX_TARGET_REVISION: ${DUCKDB_INDEX_TARGET_REVISION-refs/convert/parquet}
+ DUCKDB_INDEX_URL_TEMPLATE: ${DUCKDB_INDEX_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
+ DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES: ${DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES-100_000_000}
@@ -147,0 +155 @@ volumes:
+ duckdb-index:
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index c50d781f..233e90f2 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -97,0 +98 @@ services:
+ - duckdb-index:${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index}:rw
@@ -118,0 +120,6 @@ services:
+ DUCKDB_INDEX_STORAGE_DIRECTORY: ${DUCKDB_INDEX_STORAGE_DIRECTORY-/duckdb-index}
+ DUCKDB_INDEX_COMMIT_MESSAGE: ${DUCKDB_INDEX_COMMIT_MESSAGE-Update duckdb index files}
+ DUCKDB_INDEX_COMMITTER_HF_TOKEN: ${DUCKDB_INDEX_COMMITTER_HF_TOKEN-}
+ DUCKDB_INDEX_TARGET_REVISION: ${DUCKDB_INDEX_TARGET_REVISION-refs/convert/parquet}
+ DUCKDB_INDEX_URL_TEMPLATE: ${DUCKDB_INDEX_URL_TEMPLATE-/datasets/%s/resolve/%s/%s}
+ DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES: ${DUCKDB_INDEX_MAX_PARQUET_SIZE_BYTES-100_000_000}
@@ -149,0 +157 @@ volumes:
+ duckdb-index:
|
|
04f53d629c65a6d280b4ea79daf87c5822fe5785
|
Sylvain Lesage
| 2023-06-27T13:04:00 |
refactor: 💡 remove dead code (#1435)
|
diff --git a/services/worker/src/worker/job_manager.py b/services/worker/src/worker/job_manager.py
index d4807508..5a5a4f63 100644
--- a/services/worker/src/worker/job_manager.py
+++ b/services/worker/src/worker/job_manager.py
@@ -30,3 +29,0 @@ from worker.job_runner import JobRunner
-# List of error codes that should trigger a retry.
-ERROR_CODES_TO_RETRY: list[str] = ["ClientConnectionError"]
-
|
|
1928052b47b1619a516f96ab5c1df781df4af36a
|
Quentin Lhoest
| 2023-06-27T12:12:19 |
unqote path and revision (#1434)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index c5c879be..8df7ea7a 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -11 +11 @@ from typing import Any, List, Optional, Set, Tuple
-from urllib.parse import quote
+from urllib.parse import quote, unquote
@@ -586,0 +587,2 @@ def copy_parquet_files(builder: DatasetBuilder) -> List[CommitOperationCopy]:
+ src_revision = unquote(src_revision)
+ src_path_in_repo = unquote(src_path_in_repo)
|
|
64e84e792c755525d3a6c48425567617931e001c
|
Quentin Lhoest
| 2023-06-26T17:23:12 |
Fix auth in rows again (#1430)
|
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py
index f95a63f8..0f8ca107 100644
--- a/libs/libcommon/src/libcommon/parquet_utils.py
+++ b/libs/libcommon/src/libcommon/parquet_utils.py
@@ -265,0 +266 @@ class ParquetIndexWithMetadata:
+ **self.httpfs.kwargs,
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index 177865f5..d0689d8e 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -274 +274 @@ def create_rows_endpoint(
- httpfs=HTTPFileSystem(storage_options={"headers": {"authorization": f"Bearer {hf_token}"}}),
+ httpfs=HTTPFileSystem(headers={"authorization": f"Bearer {hf_token}"}),
diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py
index 56a3815d..613d616a 100644
--- a/services/worker/tests/job_runners/config/test_parquet_metadata.py
+++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py
@@ -7 +7 @@ from pathlib import Path
-from typing import Any, Callable
+from typing import Any, Callable, Mapping, Optional
@@ -13 +13,3 @@ import pytest
-from fsspec.implementations.http import HTTPFileSystem
+from datasets import Dataset, Features
+from fsspec.implementations.http import HTTPFile, HTTPFileSystem
+from huggingface_hub import hf_hub_url
@@ -14,0 +17 @@ from libcommon.exceptions import PreviousStepFormatError
+from libcommon.parquet_utils import ParquetIndexWithMetadata
@@ -29,0 +33,3 @@ from worker.job_runners.config.parquet_metadata import ConfigParquetMetadataJobR
+from ...constants import CI_USER_TOKEN
+from ...fixtures.hub import hf_api
+
@@ -188,0 +195,71 @@ def test_compute(
+
+
+class AuthenticatedHTTPFile(HTTPFile): # type: ignore
+ last_url: Optional[str] = None
+
+ def __init__( # type: ignore
+ self,
+ fs,
+ url,
+ session=None,
+ block_size=None,
+ mode="rb",
+ cache_type="bytes",
+ cache_options=None,
+ size=None,
+ loop=None,
+ asynchronous=False,
+ **kwargs,
+ ) -> None:
+ super().__init__(
+ fs,
+ url,
+ session=session,
+ block_size=block_size,
+ mode=mode,
+ cache_type=cache_type,
+ cache_options=cache_options,
+ size=size,
+ loop=loop,
+ asynchronous=asynchronous,
+ **kwargs,
+ )
+ assert self.kwargs == {"headers": {"authorization": f"Bearer {CI_USER_TOKEN}"}}
+ AuthenticatedHTTPFile.last_url = url
+
+
+def test_ParquetIndexWithMetadata_query(
+ datasets: Mapping[str, Dataset], hub_public_big: str, tmp_path_factory: pytest.TempPathFactory
+) -> None:
+ ds = datasets["big"]
+ httpfs = HTTPFileSystem(headers={"authorization": f"Bearer {CI_USER_TOKEN}"})
+ filename = next(
+ iter(
+ repo_file
+ for repo_file in hf_api.list_repo_files(repo_id=hub_public_big, repo_type="dataset")
+ if repo_file.endswith(".parquet")
+ )
+ )
+ url = hf_hub_url(repo_id=hub_public_big, filename=filename, repo_type="dataset")
+ metadata_path = str(tmp_path_factory.mktemp("test_ParquetIndexWithMetadata_query") / "metadata.parquet")
+ with httpfs.open(url) as f:
+ num_bytes = f.size
+ pf = pq.ParquetFile(url, filesystem=httpfs)
+ num_rows = pf.metadata.num_rows
+ features = Features.from_arrow_schema(pf.schema_arrow)
+ pf.metadata.write_metadata_file(metadata_path)
+ index = ParquetIndexWithMetadata(
+ features=features,
+ supported_columns=list(features),
+ unsupported_columns=[],
+ parquet_files_urls=[url],
+ metadata_paths=[metadata_path],
+ num_rows=[num_rows],
+ num_bytes=[num_bytes],
+ httpfs=httpfs,
+ hf_token=CI_USER_TOKEN,
+ )
+ with patch("libcommon.parquet_utils.HTTPFile", AuthenticatedHTTPFile):
+ out = index.query(offset=0, length=2).to_pydict()
+ assert out == ds[:2]
+ assert AuthenticatedHTTPFile.last_url == url
|
|
a249894dd692ea6da078912aed9c462d3a99b51c
|
Sylvain Lesage
| 2023-06-26T16:17:02 |
feat: 🎸 don't insert a new lock when releasing (#1432)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index e3b7df0e..6aa66820 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -296 +295,0 @@ class lock(contextlib.AbstractContextManager["lock"]):
- upsert=True,
|
|
a0479f8a8ac7868c94344370a7965575872c05e3
|
Sylvain Lesage
| 2023-06-26T15:56:46 |
fix: 🐛 remove the "required" constraint on created_at in Lock (#1431)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index e7c92ebd..e3b7df0e 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -239 +239 @@ class Lock(Document):
- created_at = DateTimeField(required=True)
+ created_at = DateTimeField()
|
|
97cc53ff804ca606e05744eb699ae240b5842eff
|
Sylvain Lesage
| 2023-06-26T15:37:26 |
Ensure only one job is started for the same unicity_id (#1420)
|
diff --git a/jobs/cache_maintenance/tests/test_collect_metrics.py b/jobs/cache_maintenance/tests/test_collect_metrics.py
index eb58797e..7f8cf4bd 100644
--- a/jobs/cache_maintenance/tests/test_collect_metrics.py
+++ b/jobs/cache_maintenance/tests/test_collect_metrics.py
@@ -27 +27 @@ def test_collect_metrics() -> None:
- queue.upsert_job(
+ queue.add_job(
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 4b6dba13..6d18f4da 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -46,0 +47,3 @@ from mongodb_migration.migrations._20230516101600_queue_delete_index_without_rev
+from mongodb_migration.migrations._20230622131500_lock_add_owner import (
+ MigrationAddOwnerToQueueLock,
+)
@@ -231,0 +235,3 @@ class MigrationsCollector:
+ MigrationAddOwnerToQueueLock(
+ version="20230622131800", description="add 'owner' field copying the job_id value"
+ ),
diff --git a/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230622131500_lock_add_owner.py b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230622131500_lock_add_owner.py
new file mode 100644
index 00000000..e539be50
--- /dev/null
+++ b/jobs/mongodb_migration/src/mongodb_migration/migrations/_20230622131500_lock_add_owner.py
@@ -0,0 +1,32 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+import logging
+
+from libcommon.constants import QUEUE_COLLECTION_LOCKS, QUEUE_MONGOENGINE_ALIAS
+from libcommon.queue import Lock
+from mongoengine.connection import get_db
+
+from mongodb_migration.check import check_documents
+from mongodb_migration.migration import Migration
+
+
+# connection already occurred in the main.py (caveat: we use globals)
+class MigrationAddOwnerToQueueLock(Migration):
+ def up(self) -> None:
+ # See https://docs.mongoengine.org/guide/migration.html#example-1-addition-of-a-field
+ logging.info("If missing, add the owner field with the same value as the field job_id to the locks")
+ db = get_db(QUEUE_MONGOENGINE_ALIAS)
+ db[QUEUE_COLLECTION_LOCKS].update_many(
+ {"owner": {"$exists": False}}, [{"$set": {"owner": "$job_id"}}] # type: ignore
+ )
+
+ def down(self) -> None:
+ logging.info("Remove the owner field from all the locks")
+ db = get_db(QUEUE_MONGOENGINE_ALIAS)
+ db[QUEUE_COLLECTION_LOCKS].update_many({}, {"$unset": {"owner": ""}})
+
+ def validate(self) -> None:
+ logging.info("Ensure that a random selection of locks have the 'owner' field")
+
+ check_documents(DocCls=Lock, sample_size=10)
diff --git a/jobs/mongodb_migration/tests/migrations/test_20230622131500_lock_add_owner.py b/jobs/mongodb_migration/tests/migrations/test_20230622131500_lock_add_owner.py
new file mode 100644
index 00000000..f97639a9
--- /dev/null
+++ b/jobs/mongodb_migration/tests/migrations/test_20230622131500_lock_add_owner.py
@@ -0,0 +1,64 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from typing import Optional
+
+from libcommon.constants import QUEUE_COLLECTION_LOCKS, QUEUE_MONGOENGINE_ALIAS
+from libcommon.resources import MongoResource
+from mongoengine.connection import get_db
+
+from mongodb_migration.migrations._20230622131500_lock_add_owner import (
+ MigrationAddOwnerToQueueLock,
+)
+
+
+def assert_owner(key: str, owner: Optional[str]) -> None:
+ db = get_db(QUEUE_MONGOENGINE_ALIAS)
+ entry = db[QUEUE_COLLECTION_LOCKS].find_one({"key": key})
+ assert entry is not None
+ if owner is None:
+ assert "owner" not in entry or entry["owner"] is None
+ else:
+ assert entry["owner"] == owner
+
+
+def test_lock_add_owner(mongo_host: str) -> None:
+ with MongoResource(database="test_lock_add_owner", host=mongo_host, mongoengine_alias="queue"):
+ db = get_db(QUEUE_MONGOENGINE_ALIAS)
+ db[QUEUE_COLLECTION_LOCKS].insert_many(
+ [
+ {
+ "key": "key1",
+ "job_id": "job_id1",
+ "created_at": "2022-01-01T00:00:00.000000Z",
+ },
+ {
+ "key": "key2",
+ "job_id": None,
+ "created_at": "2022-01-01T00:00:00.000000Z",
+ },
+ {
+ "key": "key3",
+ "job_id": "job_id3",
+ "owner": "owner3",
+ "created_at": "2022-01-01T00:00:00.000000Z",
+ },
+ ]
+ )
+
+ migration = MigrationAddOwnerToQueueLock(
+ version="20230622131500",
+ description="add owner field to locks",
+ )
+ migration.up()
+
+ assert_owner("key1", "job_id1")
+ assert_owner("key2", None)
+ assert_owner("key3", "owner3")
+
+ migration.down()
+ assert_owner("key1", None)
+ assert_owner("key2", None)
+ assert_owner("key3", None)
+
+ db[QUEUE_COLLECTION_LOCKS].drop()
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 55aea4de..e7c92ebd 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -98,0 +99,12 @@ class JobDoesNotExistError(DoesNotExist):
+class AlreadyStartedJobError(Exception):
+ pass
+
+
+class LockTimeoutError(Exception):
+ pass
+
+
+class NoWaitingJobError(Exception):
+ pass
+
+
@@ -222 +234 @@ class Lock(Document):
- meta = {"collection": QUEUE_COLLECTION_LOCKS, "db_alias": QUEUE_MONGOENGINE_ALIAS, "indexes": [("key", "job_id")]}
+ meta = {"collection": QUEUE_COLLECTION_LOCKS, "db_alias": QUEUE_MONGOENGINE_ALIAS, "indexes": [("key", "owner")]}
@@ -224 +236,2 @@ class Lock(Document):
- job_id = StringField()
+ owner = StringField()
+ job_id = StringField() # deprecated
@@ -234,2 +247,3 @@ class lock(contextlib.AbstractContextManager["lock"]):
- Provides a simple way of inter-worker communication using a MongoDB lock.
- A lock is used to indicate another worker of your application that a resource
+ Provides a simple way of inter-applications communication using a MongoDB lock.
+
+ An example usage is to another worker of your application that a resource
@@ -242 +256 @@ class lock(contextlib.AbstractContextManager["lock"]):
- with lock(key=key, job_id=job.pk):
+ with lock(key=key, owner=job.pk):
@@ -251 +265 @@ class lock(contextlib.AbstractContextManager["lock"]):
- lock(key=key, job_id=job.pk).acquire()
+ lock(key=key, owner=job.pk).acquire()
@@ -259 +273 @@ class lock(contextlib.AbstractContextManager["lock"]):
- def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> None:
+ def __init__(self, key: str, owner: str, sleeps: Sequence[float] = _default_sleeps) -> None:
@@ -261 +275 @@ class lock(contextlib.AbstractContextManager["lock"]):
- self.job_id = job_id
+ self.owner = owner
@@ -267 +281 @@ class lock(contextlib.AbstractContextManager["lock"]):
- Lock.objects(key=self.key, job_id__in=[None, self.job_id]).update(
+ Lock.objects(key=self.key, owner__in=[None, self.owner]).update(
@@ -271 +285 @@ class lock(contextlib.AbstractContextManager["lock"]):
- job_id=self.job_id,
+ owner=self.owner,
@@ -276 +290 @@ class lock(contextlib.AbstractContextManager["lock"]):
- logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'")
+ logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for owner='{self.owner}'")
@@ -281 +295 @@ class lock(contextlib.AbstractContextManager["lock"]):
- Lock.objects(key=self.key, job_id=self.job_id).update(
+ Lock.objects(key=self.key, owner=self.owner).update(
@@ -285 +299 @@ class lock(contextlib.AbstractContextManager["lock"]):
- job_id=None,
+ owner=None,
@@ -300 +314 @@ class lock(contextlib.AbstractContextManager["lock"]):
- def git_branch(cls, dataset: str, branch: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> "lock":
+ def git_branch(cls, dataset: str, branch: str, owner: str, sleeps: Sequence[float] = _default_sleeps) -> "lock":
@@ -307 +321 @@ class lock(contextlib.AbstractContextManager["lock"]):
- job_id (`str`): the current job id that holds the lock
+ owner (`str`): the current job id that holds the lock
@@ -311 +325 @@ class lock(contextlib.AbstractContextManager["lock"]):
- return cls(key=key, job_id=job_id, sleeps=sleeps)
+ return cls(key=key, owner=owner, sleeps=sleeps)
@@ -330 +344 @@ class Queue:
- def _add_job(
+ def add_job(
@@ -341 +355,2 @@ class Queue:
- This method should not be called directly. Use `upsert_job` instead.
+ Note that the same "unicity_id" can have multiple jobs in the waiting state, with the same or different
+ revisions and or priorities.
@@ -366,38 +380,0 @@ class Queue:
- def upsert_job(
- self,
- job_type: str,
- dataset: str,
- revision: str,
- config: Optional[str] = None,
- split: Optional[str] = None,
- priority: Priority = Priority.NORMAL,
- ) -> Job:
- """Add, or update, a job to the queue in the waiting state.
-
- If jobs already exist with the same parameters in the waiting state, they are cancelled and replaced by a new
- one.
- Note that the new job inherits the highest priority of the previous waiting jobs.
-
- Args:
- job_type (`str`): The type of the job
- dataset (`str`): The dataset on which to apply the job.
- revision (`str`): The git revision of the dataset.
- config (`str`, optional): The config on which to apply the job.
- split (`str`, optional): The config on which to apply the job.
- priority (`Priority`, optional): The priority of the job. Defaults to Priority.NORMAL.
-
- Returns: the job
- """
- canceled_jobs = self.cancel_jobs(
- job_type=job_type,
- dataset=dataset,
- config=config,
- split=split,
- statuses_to_cancel=[Status.WAITING],
- )
- if any(job["priority"] == Priority.NORMAL for job in canceled_jobs):
- priority = Priority.NORMAL
- return self._add_job(
- job_type=job_type, dataset=dataset, revision=revision, config=config, split=split, priority=priority
- )
-
@@ -441,39 +417,0 @@ class Queue:
- def cancel_jobs(
- self,
- job_type: str,
- dataset: str,
- config: Optional[str] = None,
- split: Optional[str] = None,
- statuses_to_cancel: Optional[List[Status]] = None,
- ) -> List[JobDict]:
- """Cancel jobs from the queue.
-
- Note that the jobs for all the revisions are canceled.
-
- Returns the list of canceled jobs (as JobDict, before they are canceled, to be able to know their previous
- status)
-
- Args:
- job_type (`str`): The type of the job
- dataset (`str`): The dataset on which to apply the job.
- config (`str`, optional): The config on which to apply the job.
- split (`str`, optional): The config on which to apply the job.
- statuses_to_cancel (`list[Status]`, optional): The list of statuses to cancel. Defaults to
- [Status.WAITING, Status.STARTED].
-
- Returns:
- `list[JobDict]`: The list of canceled jobs
- """
- if statuses_to_cancel is None:
- statuses_to_cancel = [Status.WAITING, Status.STARTED]
- existing = Job.objects(
- type=job_type,
- dataset=dataset,
- config=config,
- split=split,
- status__in=statuses_to_cancel,
- )
- job_dicts = [job.to_dict() for job in existing]
- existing.update(finished_at=get_datetime(), status=Status.CANCELLED)
- return job_dicts
-
@@ -540 +478 @@ class Queue:
- .only("type", "dataset", "revision", "config", "split", "priority")
+ .only("type", "dataset", "revision", "config", "split", "priority", "unicity_id")
@@ -577 +515 @@ class Queue:
- .only("type", "dataset", "revision", "config", "split", "priority")
+ .only("type", "dataset", "revision", "config", "split", "priority", "unicity_id")
@@ -612 +550,15 @@ class Queue:
- def _start_job(self, job: Job) -> Job:
+ def _start_newest_job_and_cancel_others(self, job: Job) -> Job:
+ """Start a job (the newest one for unicity_id) and cancel the other ones.
+
+ A lock is used to ensure that the job is not started by another worker.
+
+ Args:
+ job: the job to start
+
+ Returns:
+ the started job
+
+ Raises:
+ AlreadyStartedJobError: if a started job already exist for the same unicity_id.
+ LockTimeoutError: if the lock could not be acquired after 20 retries.
+ """
@@ -614,2 +566,40 @@ class Queue:
- job.update(started_at=get_datetime(), status=Status.STARTED)
- return job
+ RETRIES = 20
+ try:
+ # retry for 2 seconds
+ with lock(key=job.unicity_id, owner=str(job.pk), sleeps=[0.1] * RETRIES):
+ # get all the pending jobs for the same unicity_id
+ waiting_jobs = Job.objects(
+ unicity_id=job.unicity_id, status__in=[Status.WAITING, Status.STARTED]
+ ).order_by("-created_at")
+ datetime = get_datetime()
+ # raise if any job has already been started for unicity_id
+ num_started_jobs = waiting_jobs(status=Status.STARTED).count()
+ if num_started_jobs > 0:
+ if num_started_jobs > 1:
+ logging.critical(f"job {job.unicity_id} has been started {num_started_jobs} times. Max is 1.")
+ raise AlreadyStartedJobError(f"job {job.unicity_id} has been started by another worker")
+ # get the most recent one
+ first_job = waiting_jobs.first()
+ if not first_job:
+ raise NoWaitingJobError(f"no waiting job could be found for {job.unicity_id}")
+ # start it
+ first_job.update(
+ started_at=datetime,
+ status=Status.STARTED,
+ write_concern={"w": "majority", "fsync": True},
+ read_concern={"level": "majority"},
+ )
+ # and cancel the other ones, if any
+ for job in waiting_jobs.skip(1):
+ # TODO: try to do a bulk update (not sure it's possible)
+ job.update(
+ finished_at=datetime,
+ status=Status.CANCELLED,
+ write_concern={"w": "majority", "fsync": True},
+ read_concern={"level": "majority"},
+ )
+ return first_job.reload()
+ except TimeoutError as err:
+ raise LockTimeoutError(
+ f"could not acquire the lock for job {job.unicity_id} after {RETRIES} retries."
+ ) from err
@@ -622 +612,2 @@ class Queue:
- The job is moved from the waiting state to the started state.
+ The job is moved from the waiting state to the started state. A lock is used to ensure that only one worker
+ can start a job at a time.
@@ -630,0 +622,2 @@ class Queue:
+ AlreadyStartedJobError: if a started job already exist for the same unicity_id
+ LockTimeoutError: if the lock cannot be acquired
@@ -633,0 +627 @@ class Queue:
+
@@ -640 +633,0 @@ class Queue:
- self._start_job(next_waiting_job)
@@ -649 +642,2 @@ class Queue:
- return next_waiting_job.info()
+ started_job = self._start_newest_job_and_cancel_others(job=next_waiting_job)
+ return started_job.info()
@@ -773 +767 @@ class Queue:
- self.upsert_job(
+ self.add_job(
diff --git a/libs/libcommon/tests/test_backfill.py b/libs/libcommon/tests/test_backfill.py
index 7e445658..404f35a0 100644
--- a/libs/libcommon/tests/test_backfill.py
+++ b/libs/libcommon/tests/test_backfill.py
@@ -770 +770 @@ def test_delete_jobs(
- job = queue._add_job(job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority)
+ job = queue.add_job(job_type=STEP_DA, dataset="dataset", revision="revision", priority=priority)
@@ -775 +775,3 @@ def test_delete_jobs(
- queue._start_job(job)
+ job.status = Status.STARTED
+ job.started_at = datetime.now()
+ job.save()
diff --git a/libs/libcommon/tests/test_orchestrator.py b/libs/libcommon/tests/test_orchestrator.py
index 82e476d5..82848306 100644
--- a/libs/libcommon/tests/test_orchestrator.py
+++ b/libs/libcommon/tests/test_orchestrator.py
@@ -145 +145 @@ def test_finish_job(
- Queue()._add_job(
+ Queue().add_job(
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 82b930fb..34626226 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -24 +24,3 @@ def get_old_datetime() -> datetime:
- return get_datetime() - timedelta(days=1)
+ # Beware: the TTL index is set to 10 minutes. So it will delete the finished jobs after 10 minutes.
+ # We have to use a datetime that is not older than 10 minutes.
+ return get_datetime() - timedelta(seconds=(QUEUE_TTL_SECONDS / 2))
@@ -32 +34 @@ def queue_mongo_resource_autouse(queue_mongo_resource: QueueMongoResource) -> Qu
-def test__add_job() -> None:
+def test_add_job() -> None:
@@ -39 +41 @@ def test__add_job() -> None:
- queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ job1 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
@@ -41 +43 @@ def test__add_job() -> None:
- queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ job2 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
@@ -43 +45 @@ def test__add_job() -> None:
- # get and start the first job
+ # get and start a job the second one should have been picked
@@ -44,0 +47 @@ def test__add_job() -> None:
+ assert job2.reload().status == Status.STARTED
@@ -49,0 +53,2 @@ def test__add_job() -> None:
+ # and the first job should have been cancelled
+ assert job1.reload().status == Status.CANCELLED
@@ -53 +58,2 @@ def test__add_job() -> None:
- queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ job3 = queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ assert job3.status == Status.WAITING
@@ -61,4 +67 @@ def test__add_job() -> None:
- # process the second job
- job_info = queue.start_job()
- queue.finish_job(job_id=job_info["job_id"], is_success=True)
- # and the third one
+ # process the third job
@@ -78,78 +80,0 @@ def test__add_job() -> None:
-def test_upsert_job() -> None:
- test_type = "test_type"
- test_dataset = "test_dataset"
- test_revision_1 = "test_revision_1"
- test_revision_2 = "test_revision_2"
- # get the queue
- queue = Queue()
- # upsert a job
- queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_1)
- # a second call creates a second waiting job, and the first one is cancelled
- queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_1)
- # a third call, with a different revision, creates a third waiting job, and the second one is cancelled
- # because the unicity_id is the same
- queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_2)
- assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2)
- # get and start the last job
- job_info = queue.start_job()
- assert job_info["type"] == test_type
- assert job_info["params"]["dataset"] == test_dataset
- assert job_info["params"]["revision"] == test_revision_2
- assert job_info["params"]["config"] is None
- assert job_info["params"]["split"] is None
- assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2)
- # adding the job while the first one has not finished yet adds a new waiting job
- queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision_2)
- with pytest.raises(EmptyQueueError):
- # but: it's not possible to start two jobs with the same arguments
- queue.start_job()
- # finish the first job
- queue.finish_job(job_id=job_info["job_id"], is_success=True)
- # the queue is not empty
- assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2)
- # process the second job
- job_info = queue.start_job()
- queue.finish_job(job_id=job_info["job_id"], is_success=True)
- # the queue is empty
- assert not queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2)
- with pytest.raises(EmptyQueueError):
- # an error is raised if we try to start a job
- queue.start_job()
-
-
[email protected](
- "statuses_to_cancel, expected_remaining_number",
- [
- (None, 0),
- ([Status.WAITING], 1),
- ([Status.WAITING, Status.STARTED], 0),
- ([Status.STARTED], 1),
- ([Status.SUCCESS], 2),
- ],
-)
-def test_cancel_jobs(statuses_to_cancel: Optional[List[Status]], expected_remaining_number: int) -> None:
- test_type = "test_type"
- test_dataset = "test_dataset"
- test_revision_1 = "test_revision_1"
- test_revision_2 = "test_revision_2"
- queue = Queue()
- queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision_1)
- queue._add_job(job_type=test_type, dataset=test_dataset, revision=test_revision_2)
- queue.start_job()
-
- canceled_job_dicts = queue.cancel_jobs(
- job_type=test_type, dataset=test_dataset, statuses_to_cancel=statuses_to_cancel
- )
- assert len(canceled_job_dicts) == 2 - expected_remaining_number
-
- if expected_remaining_number == 0:
- with pytest.raises(EmptyQueueError):
- queue.start_job()
- assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_1) == (
- statuses_to_cancel is not None and Status.STARTED not in statuses_to_cancel
- )
- assert queue.is_job_in_process(job_type=test_type, dataset=test_dataset, revision=test_revision_2) == (
- statuses_to_cancel is not None and Status.WAITING not in statuses_to_cancel
- )
-
-
@@ -173 +98 @@ def test_cancel_jobs_by_job_id(
- job = queue._add_job(job_type=test_type, dataset=job_id, revision="test_revision")
+ job = queue.add_job(job_type=test_type, dataset=job_id, revision="test_revision")
@@ -199 +124,13 @@ def check_job(queue: Queue, expected_dataset: str, expected_split: str, expected
-def test_priority_logic() -> None:
+def test_priority_logic_creation_order() -> None:
+ test_type = "test_type"
+ test_revision = "test_revision"
+ queue = Queue()
+ queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1")
+ queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2")
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL)
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL)
+ with pytest.raises(EmptyQueueError):
+ queue.start_job()
+
+
+def test_priority_logic_started_jobs_per_dataset_order() -> None:
@@ -203,3 +140,20 @@ def test_priority_logic() -> None:
- queue.upsert_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1")
- queue.upsert_job(
- job_type=test_type, dataset="dataset1/dataset", revision=test_revision, config="config", split="split1"
+ queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1")
+ queue.add_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2")
+ queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1")
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL)
+ check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.NORMAL)
+ # ^ before, even if the creation date is after, because the dataset is different and has no started job
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL)
+ with pytest.raises(EmptyQueueError):
+ queue.start_job()
+
+
+def test_priority_logic_started_jobs_per_namespace_order() -> None:
+ test_type = "test_type"
+ test_revision = "test_revision"
+ queue = Queue()
+ queue.add_job(job_type=test_type, dataset="org1/dataset1", revision=test_revision, config="config", split="split1")
+ queue.add_job(job_type=test_type, dataset="org1/dataset2", revision=test_revision, config="config", split="split1")
+ queue.add_job(job_type=test_type, dataset="org2/dataset2", revision=test_revision, config="config", split="split1")
+ queue.add_job(
+ job_type=test_type, dataset="no_org_dataset3", revision=test_revision, config="config", split="split1"
@@ -207,8 +161,2 @@ def test_priority_logic() -> None:
- queue.upsert_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split2")
- queue.upsert_job(
- job_type=test_type,
- dataset="dataset2",
- revision=test_revision,
- config="config",
- split="split1",
- priority=Priority.LOW,
+ check_job(
+ queue=queue, expected_dataset="org1/dataset1", expected_split="split1", expected_priority=Priority.NORMAL
@@ -216,7 +164,2 @@ def test_priority_logic() -> None:
- queue.upsert_job(
- job_type=test_type,
- dataset="dataset2/dataset",
- revision=test_revision,
- config="config",
- split="split1",
- priority=Priority.LOW,
+ check_job(
+ queue=queue, expected_dataset="org2/dataset2", expected_split="split1", expected_priority=Priority.NORMAL
@@ -224,9 +167,6 @@ def test_priority_logic() -> None:
- queue.upsert_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split2")
- queue.upsert_job(job_type=test_type, dataset="dataset3", revision=test_revision, config="config", split="split1")
- queue.upsert_job(
- job_type=test_type,
- dataset="dataset3",
- revision=test_revision,
- config="config",
- split="split1",
- priority=Priority.LOW,
+ # ^ before, even if the creation date is after, because the namespace is different and has no started job
+ check_job(
+ queue=queue, expected_dataset="no_org_dataset3", expected_split="split1", expected_priority=Priority.NORMAL
+ )
+ check_job(
+ queue=queue, expected_dataset="org1/dataset2", expected_split="split1", expected_priority=Priority.NORMAL
@@ -234,2 +174,9 @@ def test_priority_logic() -> None:
- queue.upsert_job(job_type=test_type, dataset="dataset1", revision=test_revision, config="config", split="split1")
- queue.upsert_job(
+ with pytest.raises(EmptyQueueError):
+ queue.start_job()
+
+
+def test_priority_logic_priority_order() -> None:
+ test_type = "test_type"
+ test_revision = "test_revision"
+ queue = Queue()
+ queue.add_job(
@@ -237 +184 @@ def test_priority_logic() -> None:
- dataset="dataset2",
+ dataset="dataset1",
@@ -243,16 +190,4 @@ def test_priority_logic() -> None:
- check_job(
- queue=queue, expected_dataset="dataset1/dataset", expected_split="split1", expected_priority=Priority.NORMAL
- )
- check_job(queue=queue, expected_dataset="dataset2", expected_split="split2", expected_priority=Priority.NORMAL)
- check_job(queue=queue, expected_dataset="dataset3", expected_split="split1", expected_priority=Priority.NORMAL)
- # ^ before the other "dataset3" jobs because its priority is higher (it inherited Priority.NORMAL in upsert_job)
- check_job(queue=queue, expected_dataset="dataset1", expected_split="split2", expected_priority=Priority.NORMAL)
- # ^ same namespace as dataset1/dataset, goes after namespaces without any started job
- check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.NORMAL)
- # ^ comes after the other "dataset1" jobs because the last upsert_job call moved its creation date
- check_job(
- queue=queue, expected_dataset="dataset2/dataset", expected_split="split1", expected_priority=Priority.LOW
- )
- # ^ comes after the other "dataset2" jobs because its priority is lower
- check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.LOW)
- # ^ the rest of the rules apply for Priority.LOW jobs
+ queue.add_job(job_type=test_type, dataset="dataset2", revision=test_revision, config="config", split="split1")
+ check_job(queue=queue, expected_dataset="dataset2", expected_split="split1", expected_priority=Priority.NORMAL)
+ # ^ before, even if the creation date is after, because the priority is higher
+ check_job(queue=queue, expected_dataset="dataset1", expected_split="split1", expected_priority=Priority.LOW)
@@ -284 +219 @@ def test_job_types_only(
- queue.upsert_job(job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None)
+ queue.add_job(job_type=job_type, dataset=test_dataset, revision=test_revision, config=None, split=None)
@@ -309 +244 @@ def test_count_by_status() -> None:
- queue.upsert_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
+ queue.add_job(job_type=test_type, dataset=test_dataset, revision=test_revision)
@@ -314 +249 @@ def test_count_by_status() -> None:
- queue.upsert_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision)
+ queue.add_job(job_type=test_other_type, dataset=test_dataset, revision=test_revision)
@@ -333 +268 @@ def test_get_dataset_pending_jobs_for_type() -> None:
- queue.upsert_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
+ queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
@@ -339 +274 @@ def test_get_dataset_pending_jobs_for_type() -> None:
- queue.upsert_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
+ queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
@@ -344 +279 @@ def test_get_dataset_pending_jobs_for_type() -> None:
- queue.upsert_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
+ queue.add_job(job_type=job_type, dataset=dataset, revision=test_revision, config=config, split=None)
@@ -356 +291 @@ def test_queue_heartbeat() -> None:
- job = queue.upsert_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1")
+ job = queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split1")
@@ -370 +305 @@ def test_queue_get_zombies() -> None:
- zombie = queue.upsert_job(
+ zombie = queue.add_job(
@@ -374 +309 @@ def test_queue_get_zombies() -> None:
- queue.upsert_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split2")
+ queue.add_job(job_type=job_type, dataset="dataset1", revision="revision", config="config", split="split2")
@@ -411 +346 @@ def locked_increment(tmp_file: Path) -> None:
- with lock(key="test_lock", job_id=str(os.getpid()), sleeps=sleeps):
+ with lock(key="test_lock", owner=str(os.getpid()), sleeps=sleeps):
@@ -434 +369 @@ def git_branch_locked_increment(tmp_file: Path) -> None:
- with lock.git_branch(dataset=dataset, branch=branch, job_id=str(os.getpid()), sleeps=sleeps):
+ with lock.git_branch(dataset=dataset, branch=branch, owner=str(os.getpid()), sleeps=sleeps):
@@ -452 +387 @@ def test_lock_git_branch(tmp_path_factory: pytest.TempPathFactory, queue_mongo_r
- assert Lock.objects().get().job_id is None
+ assert Lock.objects().get().owner is None
diff --git a/services/admin/src/admin/routes/force_refresh.py b/services/admin/src/admin/routes/force_refresh.py
index 47e48854..439f4dde 100644
--- a/services/admin/src/admin/routes/force_refresh.py
+++ b/services/admin/src/admin/routes/force_refresh.py
@@ -62 +62 @@ def create_force_refresh_endpoint(
- Queue().upsert_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split)
+ Queue().add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split)
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index ab566739..69e0dfd2 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -154 +154 @@ def test_get_cache_entry_from_steps() -> None:
- queue.upsert_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config)
+ queue.add_job(job_type="dataset-split-names", dataset=dataset, revision=revision, config=config)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 704e6a5a..c5c879be 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -1037 +1037 @@ def compute_config_parquet_and_info_response(
- with lock.git_branch(dataset=dataset, branch=target_revision, job_id=job_id, sleeps=sleeps):
+ with lock.git_branch(dataset=dataset, branch=target_revision, owner=job_id, sleeps=sleeps):
diff --git a/services/worker/src/worker/loop.py b/services/worker/src/worker/loop.py
index a1dc8d9f..c0303b21 100644
--- a/services/worker/src/worker/loop.py
+++ b/services/worker/src/worker/loop.py
@@ -14 +14,7 @@ from libcommon.processing_graph import ProcessingGraph
-from libcommon.queue import EmptyQueueError, Queue
+from libcommon.queue import (
+ AlreadyStartedJobError,
+ EmptyQueueError,
+ LockTimeoutError,
+ NoWaitingJobError,
+ Queue,
+)
@@ -129 +135 @@ class Loop:
- except EmptyQueueError:
+ except (EmptyQueueError, AlreadyStartedJobError, LockTimeoutError, NoWaitingJobError) as e:
@@ -131 +137 @@ class Loop:
- logging.debug("no job in the queue")
+ logging.debug(e)
diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py
index e4eb41dd..0e912bdb 100644
--- a/services/worker/tests/test_job_manager.py
+++ b/services/worker/tests/test_job_manager.py
@@ -128 +128 @@ def test_backfill(priority: Priority, app_config: AppConfig) -> None:
- queue.upsert_job(
+ queue.add_job(
@@ -198 +198 @@ def test_job_runner_set_crashed(
- queue.upsert_job(
+ queue.add_job(
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 35fc1c6a..2a766b92 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -72 +72 @@ def test_process_next_job(
- loop.queue.upsert_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split)
+ loop.queue.add_job(job_type=job_type, dataset=dataset, revision=revision, config=config, split=split)
|
|
b7a9721cf1120b11604bff978d8f28554c467e80
|
Albert Villanova del Moral
| 2023-06-26T13:01:25 |
Remove too restrictive __all__ definitions (#1423)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py
index 9be31dc5..7c7b13e0 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/database_migrations.py
@@ -7 +7 @@ from typing import Generic, Type, TypeVar
-from mongoengine import Document, DoesNotExist
+from mongoengine import Document
@@ -58,4 +57,0 @@ def _clean_maintenance_database() -> None:
-
-
-# explicit re-export
-__all__ = ["DoesNotExist"]
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 550b1a52..55aea4de 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -18,2 +18,2 @@ import pytz
-from mongoengine import Document, DoesNotExist
-from mongoengine.errors import NotUniqueError
+from mongoengine import Document
+from mongoengine.errors import DoesNotExist, NotUniqueError
@@ -94,0 +95,4 @@ class EmptyQueueError(Exception):
+class JobDoesNotExistError(DoesNotExist):
+ pass
+
+
@@ -193,0 +198,7 @@ class Job(Document):
+ @classmethod
+ def get(cls, job_id: str) -> "Job":
+ try:
+ return cls.objects(pk=job_id).get()
+ except DoesNotExist as e:
+ raise JobDoesNotExistError(f"Job does not exist: {job_id=}") from e
+
@@ -921,4 +931,0 @@ def _clean_queue_database() -> None:
-
-
-# explicit re-export
-__all__ = ["DoesNotExist"]
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index f537eb28..13582b42 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -25 +25,2 @@ from bson.errors import InvalidId
-from mongoengine import Document, DoesNotExist
+from mongoengine import Document
+from mongoengine.errors import DoesNotExist
@@ -125,0 +127,4 @@ CachedResponse.split.required = False # type: ignore
+class CacheEntryDoesNotExistError(DoesNotExist):
+ pass
+
+
@@ -198 +203 @@ class CacheEntryWithoutContent(TypedDict):
-# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them
+# Note: we let the exceptions throw: it's the responsibility of the caller to manage them
@@ -202,5 +207,8 @@ def get_response_without_content(
- response = (
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
- .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress")
- .get()
- )
+ try:
+ response = (
+ CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress")
+ .get()
+ )
+ except DoesNotExist as e:
+ raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e
@@ -226 +234 @@ class CacheEntryMetadata(CacheEntryWithoutContent):
-# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them
+# Note: we let the exceptions throw: it's the responsibility of the caller to manage them
@@ -230,5 +238,8 @@ def get_response_metadata(
- response = (
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
- .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "updated_at")
- .get()
- )
+ try:
+ response = (
+ CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ .only("http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "updated_at")
+ .get()
+ )
+ except DoesNotExist as e:
+ raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e
@@ -285 +296 @@ class CachedArtifactError(Exception):
-# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them
+# Note: we let the exceptions throw: it's the responsibility of the caller to manage them
@@ -287,5 +298,8 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O
- response = (
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
- .only("content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress")
- .get()
- )
+ try:
+ response = (
+ CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ .only("content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress")
+ .get()
+ )
+ except DoesNotExist as e:
+ raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e
@@ -302 +316 @@ def get_response(kind: str, dataset: str, config: Optional[str] = None, split: O
-# Note: we let the exceptions throw (ie DoesNotExist): it's the responsibility of the caller to manage them
+# Note: we let the exceptions throw: it's the responsibility of the caller to manage them
@@ -306,4 +320,13 @@ def get_response_with_details(
- response = (
- CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
- .only(
- "content", "http_status", "error_code", "job_runner_version", "dataset_git_revision", "progress", "details"
+ try:
+ response = (
+ CachedResponse.objects(kind=kind, dataset=dataset, config=config, split=split)
+ .only(
+ "content",
+ "http_status",
+ "error_code",
+ "job_runner_version",
+ "dataset_git_revision",
+ "progress",
+ "details",
+ )
+ .get()
@@ -311,2 +334,2 @@ def get_response_with_details(
- .get()
- )
+ except DoesNotExist as e:
+ raise CacheEntryDoesNotExistError(f"Cache entry does not exist: {kind=} {dataset=} {config=} {split=}") from e
@@ -332 +355 @@ def get_response_or_missing_error(
- except DoesNotExist:
+ except CacheEntryDoesNotExistError:
@@ -747,4 +769,0 @@ def _clean_cache_database() -> None:
-
-
-# explicit re-export
-__all__ = ["DoesNotExist"]
diff --git a/libs/libcommon/tests/test_simple_cache.py b/libs/libcommon/tests/test_simple_cache.py
index d78a0fa1..e3e9d851 100644
--- a/libs/libcommon/tests/test_simple_cache.py
+++ b/libs/libcommon/tests/test_simple_cache.py
@@ -15,0 +16 @@ from libcommon.simple_cache import (
+ CacheEntryDoesNotExistError,
@@ -18 +18,0 @@ from libcommon.simple_cache import (
- DoesNotExist,
@@ -141 +141 @@ def test_upsert_response(config: Optional[str], split: Optional[str]) -> None:
- with pytest.raises(DoesNotExist):
+ with pytest.raises(CacheEntryDoesNotExistError):
@@ -181 +181 @@ def test_delete_response() -> None:
- with pytest.raises(DoesNotExist):
+ with pytest.raises(CacheEntryDoesNotExistError):
@@ -200 +200 @@ def test_delete_dataset_responses() -> None:
- with pytest.raises(DoesNotExist):
+ with pytest.raises(CacheEntryDoesNotExistError):
@@ -202 +202 @@ def test_delete_dataset_responses() -> None:
- with pytest.raises(DoesNotExist):
+ with pytest.raises(CacheEntryDoesNotExistError):
diff --git a/services/worker/src/worker/job_manager.py b/services/worker/src/worker/job_manager.py
index 17ee0782..d4807508 100644
--- a/services/worker/src/worker/job_manager.py
+++ b/services/worker/src/worker/job_manager.py
@@ -22 +22 @@ from libcommon.simple_cache import (
- DoesNotExist,
+ CacheEntryDoesNotExistError,
@@ -146 +146 @@ class JobManager:
- except DoesNotExist:
+ except CacheEntryDoesNotExistError:
diff --git a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
index 524ff628..7de09d90 100644
--- a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
@@ -11 +11 @@ from libcommon.simple_cache import (
- DoesNotExist,
+ CacheEntryDoesNotExistError,
@@ -48 +48 @@ def compute_opt_in_out_urls_scan_response(dataset: str, config: str) -> Tuple[Op
- except DoesNotExist:
+ except CacheEntryDoesNotExistError:
diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py
index d5251d9b..1e8802c9 100644
--- a/services/worker/src/worker/job_runners/dataset/info.py
+++ b/services/worker/src/worker/job_runners/dataset/info.py
@@ -11 +11 @@ from libcommon.simple_cache import (
- DoesNotExist,
+ CacheEntryDoesNotExistError,
@@ -54 +54 @@ def compute_dataset_info_response(dataset: str) -> Tuple[DatasetInfoResponse, fl
- except DoesNotExist:
+ except CacheEntryDoesNotExistError:
diff --git a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
index 06533df7..4d1a1e46 100644
--- a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
@@ -11 +11 @@ from libcommon.simple_cache import (
- DoesNotExist,
+ CacheEntryDoesNotExistError,
@@ -42 +42 @@ def compute_opt_in_out_urls_count_response(dataset: str) -> Tuple[OptInOutUrlsCo
- except DoesNotExist:
+ except CacheEntryDoesNotExistError:
diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py
index 2ebb9bac..9ba4eddc 100644
--- a/services/worker/src/worker/job_runners/dataset/parquet.py
+++ b/services/worker/src/worker/job_runners/dataset/parquet.py
@@ -11 +11 @@ from libcommon.simple_cache import (
- DoesNotExist,
+ CacheEntryDoesNotExistError,
@@ -58 +58 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetParquetResponse, float]
- except DoesNotExist:
+ except CacheEntryDoesNotExistError:
diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py
index b48c0f6e..7019388d 100644
--- a/services/worker/src/worker/job_runners/dataset/size.py
+++ b/services/worker/src/worker/job_runners/dataset/size.py
@@ -11 +11 @@ from libcommon.simple_cache import (
- DoesNotExist,
+ CacheEntryDoesNotExistError,
@@ -61 +61 @@ def compute_sizes_response(dataset: str) -> Tuple[DatasetSizeResponse, float]:
- except DoesNotExist:
+ except CacheEntryDoesNotExistError:
diff --git a/services/worker/tests/test_executor.py b/services/worker/tests/test_executor.py
index a34fa1f3..87f44cd5 100644
--- a/services/worker/tests/test_executor.py
+++ b/services/worker/tests/test_executor.py
@@ -15 +15 @@ from libcommon.processing_graph import ProcessingGraph
-from libcommon.queue import DoesNotExist, Job, Queue
+from libcommon.queue import Job, JobDoesNotExistError, Queue
@@ -112,2 +112,2 @@ def set_just_started_job_in_queue(queue_mongo_resource: QueueMongoResource) -> I
- Job.objects(pk=job_info["job_id"]).get().delete()
- except DoesNotExist:
+ Job.get(job_id=job_info["job_id"]).delete()
+ except JobDoesNotExistError:
@@ -141,2 +141,2 @@ def set_long_running_job_in_queue(app_config: AppConfig, queue_mongo_resource: Q
- Job.objects(pk=job_info["job_id"]).get().delete()
- except DoesNotExist:
+ Job.get(job_id=job_info["job_id"]).delete()
+ except JobDoesNotExistError:
@@ -172,2 +172,2 @@ def set_zombie_job_in_queue(queue_mongo_resource: QueueMongoResource) -> Iterato
- Job.objects(pk=job_info["job_id"]).get().delete()
- except DoesNotExist:
+ Job.get(job_id=job_info["job_id"]).delete()
+ except JobDoesNotExistError:
|
|
1a549c7cf593ebcf18644a1668741b9425e61c76
|
Andrea Francis Soria Jimenez
| 2023-06-26T11:04:12 |
Move dtos to its own file (#1426)
|
diff --git a/services/worker/src/worker/dtos.py b/services/worker/src/worker/dtos.py
new file mode 100644
index 00000000..ef4a4cc0
--- /dev/null
+++ b/services/worker/src/worker/dtos.py
@@ -0,0 +1,208 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+from dataclasses import dataclass, field
+from typing import Any, Dict, List, Mapping, Optional, TypedDict
+
+
+class JobRunnerInfo(TypedDict):
+ job_type: str
+ job_runner_version: int
+
+
+@dataclass
+class JobResult:
+ content: Mapping[str, Any]
+ progress: float
+
+ def __post_init__(self) -> None:
+ if self.progress < 0.0 or self.progress > 1.0:
+ raise ValueError(f"Progress should be between 0 and 1, but got {self.progress}")
+
+
+@dataclass
+class CompleteJobResult(JobResult):
+ content: Mapping[str, Any]
+ progress: float = field(init=False, default=1.0)
+
+
+class DatasetItem(TypedDict):
+ dataset: str
+
+
+class ConfigItem(DatasetItem):
+ config: Optional[str]
+
+
+class SplitItem(ConfigItem):
+ split: Optional[str]
+
+
+class SplitsList(TypedDict):
+ splits: List[SplitItem]
+
+
+class FailedConfigItem(ConfigItem):
+ error: Mapping[str, Any]
+
+
+class DatasetSplitNamesResponse(TypedDict):
+ splits: List[SplitItem]
+ pending: List[ConfigItem]
+ failed: List[FailedConfigItem]
+
+
+class PreviousJob(SplitItem):
+ kind: str
+
+
+class FeatureItem(TypedDict):
+ feature_idx: int
+ name: str
+ type: Mapping[str, Any]
+
+
+class RowItem(TypedDict):
+ row_idx: int
+ row: Mapping[str, Any]
+ truncated_cells: List[str]
+
+
+class SplitFirstRowsResponse(SplitItem):
+ features: List[FeatureItem]
+ rows: List[RowItem]
+
+
+class OptUrl(TypedDict):
+ url: str
+ row_idx: int
+ column_name: str
+
+
+class OptInOutUrlsCountResponse(TypedDict):
+ urls_columns: List[str]
+ num_opt_in_urls: int
+ num_opt_out_urls: int
+ num_urls: int
+ num_scanned_rows: int
+ has_urls_columns: bool
+ full_scan: Optional[bool]
+
+
+class OptInOutUrlsScanResponse(OptInOutUrlsCountResponse):
+ opt_in_urls: List[OptUrl]
+ opt_out_urls: List[OptUrl]
+
+
+class ImageUrlColumnsResponse(TypedDict):
+ columns: List[str]
+
+
+Row = Mapping[str, Any]
+
+
+class RowsContent(TypedDict):
+ rows: List[Row]
+ all_fetched: bool
+
+
+class ConfigInfoResponse(TypedDict):
+ dataset_info: Dict[str, Any]
+
+
+class ParquetFileItem(SplitItem):
+ url: str
+ filename: str
+ size: int
+
+
+class ConfigParquetAndInfoResponse(TypedDict):
+ parquet_files: List[ParquetFileItem]
+ dataset_info: Dict[str, Any]
+
+
+class ParquetFileMetadataItem(SplitItem):
+ url: str
+ filename: str
+ size: int
+ num_rows: int
+ parquet_metadata_subpath: str
+
+
+class ConfigParquetMetadataResponse(TypedDict):
+ parquet_files_metadata: List[ParquetFileMetadataItem]
+
+
+class ConfigParquetResponse(TypedDict):
+ parquet_files: List[ParquetFileItem]
+
+
+class ConfigSize(TypedDict):
+ dataset: str
+ config: str
+ num_bytes_original_files: int
+ num_bytes_parquet_files: int
+ num_bytes_memory: int
+ num_rows: int
+ num_columns: int
+
+
+class SplitSize(SplitItem):
+ num_bytes_parquet_files: int
+ num_bytes_memory: int
+ num_rows: int
+ num_columns: int
+
+
+class ConfigSizeContent(TypedDict):
+ config: ConfigSize
+ splits: list[SplitSize]
+
+
+class ConfigSizeResponse(TypedDict):
+ size: ConfigSizeContent
+
+
+class ConfigNameItem(TypedDict):
+ dataset: str
+ config: str
+
+
+class DatasetConfigNamesResponse(TypedDict):
+ config_names: List[ConfigNameItem]
+
+
+class DatasetInfoResponse(TypedDict):
+ dataset_info: Dict[str, Any]
+ pending: List[PreviousJob]
+ failed: List[PreviousJob]
+
+
+class DatasetIsValidResponse(TypedDict):
+ valid: bool
+
+
+class DatasetParquetResponse(TypedDict):
+ parquet_files: List[ParquetFileItem]
+ pending: list[PreviousJob]
+ failed: list[PreviousJob]
+
+
+class DatasetSize(TypedDict):
+ dataset: str
+ num_bytes_original_files: int
+ num_bytes_parquet_files: int
+ num_bytes_memory: int
+ num_rows: int
+
+
+class DatasetSizeContent(TypedDict):
+ dataset: DatasetSize
+ configs: list[ConfigSize]
+ splits: list[SplitSize]
+
+
+class DatasetSizeResponse(TypedDict):
+ size: DatasetSizeContent
+ pending: list[PreviousJob]
+ failed: list[PreviousJob]
diff --git a/services/worker/src/worker/job_runner.py b/services/worker/src/worker/job_runner.py
index d4c9d839..8a99f94a 100644
--- a/services/worker/src/worker/job_runner.py
+++ b/services/worker/src/worker/job_runner.py
@@ -11 +11 @@ from worker.config import AppConfig
-from worker.utils import JobResult, JobRunnerInfo
+from worker.dtos import JobResult, JobRunnerInfo
diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py
index ae0e2760..33d68219 100644
--- a/services/worker/src/worker/job_runners/config/info.py
+++ b/services/worker/src/worker/job_runners/config/info.py
@@ -2 +1,0 @@ import logging
-from typing import Any, Dict, TypedDict
@@ -7,0 +7 @@ from libcommon.simple_cache import get_previous_step_or_raise
+from worker.dtos import CompleteJobResult, ConfigInfoResponse
@@ -9,5 +8,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import CompleteJobResult
-
-
-class ConfigInfoResponse(TypedDict):
- dataset_info: Dict[str, Any]
diff --git a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
index 9c9306c4..524ff628 100644
--- a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
@@ -15,0 +16 @@ from libcommon.simple_cache import (
+from worker.dtos import JobResult, OptInOutUrlsCountResponse
@@ -17 +17,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import JobResult, OptInOutUrlsCountResponse
diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py
index eae1d780..572df22c 100644
--- a/services/worker/src/worker/job_runners/config/parquet.py
+++ b/services/worker/src/worker/job_runners/config/parquet.py
@@ -5 +4,0 @@ import logging
-from typing import List, TypedDict
@@ -10,0 +10 @@ from libcommon.simple_cache import get_previous_step_or_raise
+from worker.dtos import CompleteJobResult, ConfigParquetResponse
@@ -12,6 +11,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.job_runners.config.parquet_and_info import ParquetFileItem
-from worker.utils import CompleteJobResult
-
-
-class ConfigParquetResponse(TypedDict):
- parquet_files: List[ParquetFileItem]
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 313166d5..704e6a5a 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -10 +10 @@ from pathlib import Path
-from typing import Any, Dict, List, Optional, Set, Tuple, TypedDict
+from typing import Any, List, Optional, Set, Tuple
@@ -77,0 +78 @@ from worker.config import AppConfig, ParquetAndInfoConfig
+from worker.dtos import CompleteJobResult, ConfigParquetAndInfoResponse, ParquetFileItem
@@ -79,16 +80 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas
-from worker.utils import CompleteJobResult, retry
-
-
-class ParquetFileItem(TypedDict):
- dataset: str
- config: str
- split: str
- url: str
- filename: str
- size: int
-
-
-class ConfigParquetAndInfoResponse(TypedDict):
- parquet_files: List[ParquetFileItem]
- dataset_info: Dict[str, Any]
-
+from worker.utils import retry
diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py
index e1cb6ef5..29a34f9a 100644
--- a/services/worker/src/worker/job_runners/config/parquet_metadata.py
+++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py
@@ -6 +6 @@ from functools import partial
-from typing import List, Optional, TypedDict
+from typing import List, Optional
@@ -24,0 +25,6 @@ from worker.config import AppConfig
+from worker.dtos import (
+ CompleteJobResult,
+ ConfigParquetMetadataResponse,
+ ParquetFileItem,
+ ParquetFileMetadataItem,
+)
@@ -26,17 +31,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.job_runners.config.parquet_and_info import ParquetFileItem
-from worker.utils import CompleteJobResult
-
-
-class ParquetFileMetadataItem(TypedDict):
- dataset: str
- config: str
- split: str
- url: str
- filename: str
- size: int
- num_rows: int
- parquet_metadata_subpath: str
-
-
-class ConfigParquetMetadataResponse(TypedDict):
- parquet_files_metadata: List[ParquetFileMetadataItem]
diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py
index e46b9108..2610f994 100644
--- a/services/worker/src/worker/job_runners/config/size.py
+++ b/services/worker/src/worker/job_runners/config/size.py
@@ -5 +4,0 @@ import logging
-from typing import TypedDict
@@ -10,0 +10 @@ from libcommon.simple_cache import get_previous_step_or_raise
+from worker.dtos import CompleteJobResult, ConfigSize, ConfigSizeResponse, SplitSize
@@ -12,30 +11,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import CompleteJobResult
-
-
-class ConfigSize(TypedDict):
- dataset: str
- config: str
- num_bytes_original_files: int
- num_bytes_parquet_files: int
- num_bytes_memory: int
- num_rows: int
- num_columns: int
-
-
-class SplitSize(TypedDict):
- dataset: str
- config: str
- split: str
- num_bytes_parquet_files: int
- num_bytes_memory: int
- num_rows: int
- num_columns: int
-
-
-class ConfigSizeContent(TypedDict):
- config: ConfigSize
- splits: list[SplitSize]
-
-
-class ConfigSizeResponse(TypedDict):
- size: ConfigSizeContent
diff --git a/services/worker/src/worker/job_runners/config/split_names_from_info.py b/services/worker/src/worker/job_runners/config/split_names_from_info.py
index 7037de05..6b97ce14 100644
--- a/services/worker/src/worker/job_runners/config/split_names_from_info.py
+++ b/services/worker/src/worker/job_runners/config/split_names_from_info.py
@@ -13,0 +14 @@ from libcommon.simple_cache import get_previous_step_or_raise
+from worker.dtos import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList
@@ -15 +15,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList
diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
index 6c4ddc10..8c300c26 100644
--- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
@@ -19,0 +20 @@ from libcommon.exceptions import (
+from worker.dtos import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList
@@ -21 +21,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatas
-from worker.utils import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList
diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py
index faf74109..51b7a4e1 100644
--- a/services/worker/src/worker/job_runners/dataset/config_names.py
+++ b/services/worker/src/worker/job_runners/dataset/config_names.py
@@ -5 +5 @@ import logging
-from typing import List, Optional, TypedDict, Union
+from typing import List, Optional, Union
@@ -16,0 +17 @@ from libcommon.exceptions import (
+from worker.dtos import CompleteJobResult, ConfigNameItem, DatasetConfigNamesResponse
@@ -20,10 +20,0 @@ from worker.job_runners.dataset.dataset_job_runner import (
-from worker.utils import CompleteJobResult
-
-
-class ConfigNameItem(TypedDict):
- dataset: str
- config: str
-
-
-class DatasetConfigNamesResponse(TypedDict):
- config_names: List[ConfigNameItem]
diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py
index ad0c1408..d5251d9b 100644
--- a/services/worker/src/worker/job_runners/dataset/info.py
+++ b/services/worker/src/worker/job_runners/dataset/info.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Any, Dict, List, Tuple, TypedDict
+from typing import Any, Dict, Tuple
@@ -15,0 +16 @@ from libcommon.simple_cache import (
+from worker.dtos import DatasetInfoResponse, JobResult, PreviousJob
@@ -17,7 +17,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult, PreviousJob
-
-
-class DatasetInfoResponse(TypedDict):
- dataset_info: Dict[str, Any]
- pending: List[PreviousJob]
- failed: List[PreviousJob]
diff --git a/services/worker/src/worker/job_runners/dataset/is_valid.py b/services/worker/src/worker/job_runners/dataset/is_valid.py
index 9023a1d9..cb45bf8f 100644
--- a/services/worker/src/worker/job_runners/dataset/is_valid.py
+++ b/services/worker/src/worker/job_runners/dataset/is_valid.py
@@ -5 +5 @@ import logging
-from typing import Tuple, TypedDict
+from typing import Tuple
@@ -9,0 +10 @@ from libcommon.simple_cache import get_validity_by_kind
+from worker.dtos import DatasetIsValidResponse, JobResult
@@ -11,6 +11,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult
-
-
-class DatasetIsValidResponse(TypedDict):
- valid: bool
-
diff --git a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
index 530d3145..06533df7 100644
--- a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
@@ -15,0 +16 @@ from libcommon.simple_cache import (
+from worker.dtos import JobResult, OptInOutUrlsCountResponse
@@ -17 +17,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult, OptInOutUrlsCountResponse
diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py
index 4a360950..2ebb9bac 100644
--- a/services/worker/src/worker/job_runners/dataset/parquet.py
+++ b/services/worker/src/worker/job_runners/dataset/parquet.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import List, Tuple, TypedDict
+from typing import Tuple
@@ -16,2 +16,7 @@ from libcommon.simple_cache import (
-from worker.job_runners.config.parquet import ConfigParquetResponse
-from worker.job_runners.config.parquet_and_info import ParquetFileItem
+from worker.dtos import (
+ ConfigParquetResponse,
+ DatasetParquetResponse,
+ JobResult,
+ ParquetFileItem,
+ PreviousJob,
+)
@@ -19,7 +23,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult, PreviousJob
-
-
-class DatasetParquetResponse(TypedDict):
- parquet_files: List[ParquetFileItem]
- pending: list[PreviousJob]
- failed: list[PreviousJob]
diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py
index 3b69e948..b48c0f6e 100644
--- a/services/worker/src/worker/job_runners/dataset/size.py
+++ b/services/worker/src/worker/job_runners/dataset/size.py
@@ -6 +6 @@ from http import HTTPStatus
-from typing import Tuple, TypedDict
+from typing import Tuple
@@ -16 +16,9 @@ from libcommon.simple_cache import (
-from worker.job_runners.config.size import ConfigSize, ConfigSizeResponse, SplitSize
+from worker.dtos import (
+ ConfigSize,
+ ConfigSizeResponse,
+ DatasetSize,
+ DatasetSizeResponse,
+ JobResult,
+ PreviousJob,
+ SplitSize,
+)
@@ -18,21 +25,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult, PreviousJob
-
-
-class DatasetSize(TypedDict):
- dataset: str
- num_bytes_original_files: int
- num_bytes_parquet_files: int
- num_bytes_memory: int
- num_rows: int
-
-
-class DatasetSizeContent(TypedDict):
- dataset: DatasetSize
- configs: list[ConfigSize]
- splits: list[SplitSize]
-
-
-class DatasetSizeResponse(TypedDict):
- size: DatasetSizeContent
- pending: list[PreviousJob]
- failed: list[PreviousJob]
diff --git a/services/worker/src/worker/job_runners/dataset/split_names.py b/services/worker/src/worker/job_runners/dataset/split_names.py
index 259e66d4..06aa9222 100644
--- a/services/worker/src/worker/job_runners/dataset/split_names.py
+++ b/services/worker/src/worker/job_runners/dataset/split_names.py
@@ -12,2 +12 @@ from libcommon.simple_cache import get_best_response, get_previous_step_or_raise
-from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import (
+from worker.dtos import (
@@ -19,0 +19 @@ from worker.utils import (
+from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
index 07bdb663..ba42e24e 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
@@ -25,2 +25 @@ from worker.config import AppConfig, FirstRowsConfig
-from worker.job_runners.split.split_job_runner import SplitJobRunner
-from worker.utils import (
+from worker.dtos import (
@@ -32,3 +30,0 @@ from worker.utils import (
- create_truncated_row_items,
- get_json_size,
- to_features_list,
@@ -35,0 +32,2 @@ from worker.utils import (
+from worker.job_runners.split.split_job_runner import SplitJobRunner
+from worker.utils import create_truncated_row_items, get_json_size, to_features_list
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
index 3dc01f41..a9694d04 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
@@ -35,0 +36 @@ from worker.config import AppConfig, FirstRowsConfig
+from worker.dtos import CompleteJobResult, JobRunnerInfo, Row, SplitFirstRowsResponse
@@ -38,4 +38,0 @@ from worker.utils import (
- CompleteJobResult,
- JobRunnerInfo,
- Row,
- SplitFirstRowsResponse,
diff --git a/services/worker/src/worker/job_runners/split/image_url_columns.py b/services/worker/src/worker/job_runners/split/image_url_columns.py
index c2b19607..84756946 100644
--- a/services/worker/src/worker/job_runners/split/image_url_columns.py
+++ b/services/worker/src/worker/job_runners/split/image_url_columns.py
@@ -11,2 +11 @@ from libcommon.utils import is_image_url
-from worker.job_runners.split.split_job_runner import SplitJobRunner
-from worker.utils import (
+from worker.dtos import (
@@ -16,0 +16 @@ from worker.utils import (
+from worker.job_runners.split.split_job_runner import SplitJobRunner
diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py
index af094266..0b602b84 100644
--- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py
@@ -9,0 +10 @@ from libcommon.simple_cache import get_previous_step_or_raise
+from worker.dtos import CompleteJobResult, OptInOutUrlsCountResponse
@@ -11 +11,0 @@ from worker.job_runners.split.split_job_runner import SplitJobRunner
-from worker.utils import CompleteJobResult, OptInOutUrlsCountResponse
diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
index dbb20277..2ccb4443 100644
--- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
@@ -24,0 +25 @@ from worker.config import AppConfig, OptInOutUrlsScanConfig
+from worker.dtos import CompleteJobResult, OptInOutUrlsScanResponse, OptUrl
@@ -26,6 +27 @@ from worker.job_runners.split.split_job_runner import SplitJobRunnerWithDatasets
-from worker.utils import (
- CompleteJobResult,
- OptInOutUrlsScanResponse,
- OptUrl,
- get_rows_or_raise,
-)
+from worker.utils import get_rows_or_raise
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index 982da72d..fab6b031 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -9 +8,0 @@ import warnings
-from dataclasses import dataclass, field
@@ -14 +12,0 @@ from typing import (
- Mapping,
@@ -19 +16,0 @@ from typing import (
- TypedDict,
@@ -36,0 +34,2 @@ from libcommon.utils import orjson_dumps
+from worker.dtos import FeatureItem, Row, RowItem, RowsContent
+
@@ -41,110 +39,0 @@ MAX_IMAGE_PIXELS = 10_000_000_000
-class JobRunnerInfo(TypedDict):
- job_type: str
- job_runner_version: int
-
-
-@dataclass
-class JobResult:
- content: Mapping[str, Any]
- progress: float
-
- def __post_init__(self) -> None:
- if self.progress < 0.0 or self.progress > 1.0:
- raise ValueError(f"Progress should be between 0 and 1, but got {self.progress}")
-
-
-@dataclass
-class CompleteJobResult(JobResult):
- content: Mapping[str, Any]
- progress: float = field(init=False, default=1.0)
-
-
-class DatasetItem(TypedDict):
- dataset: str
-
-
-class ConfigItem(DatasetItem):
- config: Optional[str]
-
-
-class SplitItem(ConfigItem):
- split: Optional[str]
-
-
-class SplitsList(TypedDict):
- splits: List[SplitItem]
-
-
-class FailedConfigItem(ConfigItem):
- error: Mapping[str, Any]
-
-
-class DatasetSplitNamesResponse(TypedDict):
- splits: List[SplitItem]
- pending: List[ConfigItem]
- failed: List[FailedConfigItem]
-
-
-class PreviousJob(TypedDict):
- kind: str
- dataset: str
- config: Optional[str]
- split: Optional[str]
-
-
-class FeatureItem(TypedDict):
- feature_idx: int
- name: str
- type: Mapping[str, Any]
-
-
-class RowItem(TypedDict):
- row_idx: int
- row: Mapping[str, Any]
- truncated_cells: List[str]
-
-
-class SplitFirstRowsResponse(TypedDict):
- dataset: str
- config: str
- split: str
- features: List[FeatureItem]
- rows: List[RowItem]
-
-
-class OptUrl(TypedDict):
- url: str
- row_idx: int
- column_name: str
-
-
-class OptInOutUrlsCountResponse(TypedDict):
- urls_columns: List[str]
- num_opt_in_urls: int
- num_opt_out_urls: int
- num_urls: int
- num_scanned_rows: int
- has_urls_columns: bool
- full_scan: Optional[bool]
-
-
-class OptInOutUrlsScanResponse(OptInOutUrlsCountResponse):
- opt_in_urls: List[OptUrl]
- opt_out_urls: List[OptUrl]
-
-
-class ImageUrlColumnsResponse(TypedDict):
- columns: List[str]
-
-
-Row = Mapping[str, Any]
-
-
-class RowsContent(TypedDict):
- rows: List[Row]
- all_fetched: bool
-
-
-# TODO: separate functions from common classes and named dicts otherwise this file will continue growing
-
-
diff --git a/services/worker/tests/job_runners/config/test_config_job_runner.py b/services/worker/tests/job_runners/config/test_config_job_runner.py
index 1e8fa750..01a5ce8e 100644
--- a/services/worker/tests/job_runners/config/test_config_job_runner.py
+++ b/services/worker/tests/job_runners/config/test_config_job_runner.py
@@ -11,0 +12 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -13 +13,0 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import CompleteJobResult
diff --git a/services/worker/tests/job_runners/config/test_parquet.py b/services/worker/tests/job_runners/config/test_parquet.py
index 4b58e5f7..031fee96 100644
--- a/services/worker/tests/job_runners/config/test_parquet.py
+++ b/services/worker/tests/job_runners/config/test_parquet.py
@@ -15,5 +15 @@ from worker.config import AppConfig
-from worker.job_runners.config.parquet import (
- ConfigParquetJobRunner,
- ConfigParquetResponse,
-)
-from worker.job_runners.config.parquet_and_info import (
+from worker.dtos import (
@@ -20,0 +17 @@ from worker.job_runners.config.parquet_and_info import (
+ ConfigParquetResponse,
@@ -22,0 +20 @@ from worker.job_runners.config.parquet_and_info import (
+from worker.job_runners.config.parquet import ConfigParquetJobRunner
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index e9631ea8..ec3bc9c2 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -36,0 +37 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -52 +52,0 @@ from worker.resources import LibrariesResource
-from worker.utils import CompleteJobResult
diff --git a/services/worker/tests/job_runners/config/test_parquet_metadata.py b/services/worker/tests/job_runners/config/test_parquet_metadata.py
index fadd6e71..56a3815d 100644
--- a/services/worker/tests/job_runners/config/test_parquet_metadata.py
+++ b/services/worker/tests/job_runners/config/test_parquet_metadata.py
@@ -22,4 +22 @@ from worker.config import AppConfig
-from worker.job_runners.config.parquet import ConfigParquetResponse
-from worker.job_runners.config.parquet_and_info import ParquetFileItem
-from worker.job_runners.config.parquet_metadata import (
- ConfigParquetMetadataJobRunner,
+from worker.dtos import (
@@ -26,0 +24,2 @@ from worker.job_runners.config.parquet_metadata import (
+ ConfigParquetResponse,
+ ParquetFileItem,
@@ -28,0 +28 @@ from worker.job_runners.config.parquet_metadata import (
+from worker.job_runners.config.parquet_metadata import ConfigParquetMetadataJobRunner
diff --git a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py
index 2d8a95d2..49f7fa06 100644
--- a/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py
+++ b/services/worker/tests/job_runners/dataset/test_dataset_job_runner.py
@@ -11,0 +12 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -13 +13,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import CompleteJobResult
diff --git a/services/worker/tests/job_runners/dataset/test_info.py b/services/worker/tests/job_runners/dataset/test_info.py
index 94c8d7d4..26aeb369 100644
--- a/services/worker/tests/job_runners/dataset/test_info.py
+++ b/services/worker/tests/job_runners/dataset/test_info.py
@@ -14,0 +15 @@ from worker.config import AppConfig
+from worker.dtos import PreviousJob
@@ -16 +16,0 @@ from worker.job_runners.dataset.info import DatasetInfoJobRunner
-from worker.utils import PreviousJob
diff --git a/services/worker/tests/job_runners/dataset/test_parquet.py b/services/worker/tests/job_runners/dataset/test_parquet.py
index c698848c..61cfbfbc 100644
--- a/services/worker/tests/job_runners/dataset/test_parquet.py
+++ b/services/worker/tests/job_runners/dataset/test_parquet.py
@@ -15,6 +15,2 @@ from worker.config import AppConfig
-from worker.job_runners.config.parquet import ConfigParquetResponse
-from worker.job_runners.config.parquet_and_info import ParquetFileItem
-from worker.job_runners.dataset.parquet import (
- DatasetParquetJobRunner,
- DatasetParquetResponse,
-)
+from worker.dtos import ConfigParquetResponse, DatasetParquetResponse, ParquetFileItem
+from worker.job_runners.dataset.parquet import DatasetParquetJobRunner
diff --git a/services/worker/tests/job_runners/split/test_image_url_columns.py b/services/worker/tests/job_runners/split/test_image_url_columns.py
index cf13a2d3..68fb6f01 100644
--- a/services/worker/tests/job_runners/split/test_image_url_columns.py
+++ b/services/worker/tests/job_runners/split/test_image_url_columns.py
@@ -17,0 +18 @@ from worker.config import AppConfig
+from worker.dtos import ImageUrlColumnsResponse
@@ -19 +19,0 @@ from worker.job_runners.split.image_url_columns import SplitImageUrlColumnsJobRu
-from worker.utils import ImageUrlColumnsResponse
diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
index 8c8d3d74..2de50d1d 100644
--- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
+++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
@@ -23,0 +24 @@ from worker.config import AppConfig
+from worker.dtos import ImageUrlColumnsResponse
@@ -29 +29,0 @@ from worker.resources import LibrariesResource
-from worker.utils import ImageUrlColumnsResponse
diff --git a/services/worker/tests/job_runners/split/test_split_job_runner.py b/services/worker/tests/job_runners/split/test_split_job_runner.py
index d944a8ea..80d88ba4 100644
--- a/services/worker/tests/job_runners/split/test_split_job_runner.py
+++ b/services/worker/tests/job_runners/split/test_split_job_runner.py
@@ -11,0 +12 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -13 +13,0 @@ from worker.job_runners.split.split_job_runner import SplitJobRunner
-from worker.utils import CompleteJobResult
diff --git a/services/worker/tests/job_runners/test__job_runner_with_cache.py b/services/worker/tests/job_runners/test__job_runner_with_cache.py
index db580ea6..11aa9fd1 100644
--- a/services/worker/tests/job_runners/test__job_runner_with_cache.py
+++ b/services/worker/tests/job_runners/test__job_runner_with_cache.py
@@ -13,0 +14 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -16 +16,0 @@ from worker.resources import LibrariesResource
-from worker.utils import CompleteJobResult
diff --git a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
index 30ea2795..2667da40 100644
--- a/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
+++ b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
@@ -13,0 +14 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -18 +18,0 @@ from worker.resources import LibrariesResource
-from worker.utils import CompleteJobResult
diff --git a/services/worker/tests/test_job_manager.py b/services/worker/tests/test_job_manager.py
index 93b8df65..e4eb41dd 100644
--- a/services/worker/tests/test_job_manager.py
+++ b/services/worker/tests/test_job_manager.py
@@ -13,0 +14 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -16 +16,0 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import CompleteJobResult
diff --git a/services/worker/tests/test_loop.py b/services/worker/tests/test_loop.py
index 4176c770..35fc1c6a 100644
--- a/services/worker/tests/test_loop.py
+++ b/services/worker/tests/test_loop.py
@@ -7,0 +8 @@ from worker.config import AppConfig
+from worker.dtos import CompleteJobResult
@@ -12 +12,0 @@ from worker.resources import LibrariesResource
-from worker.utils import CompleteJobResult
|
|
120c2547e31780e584d0a37ac4d9305f4b988195
|
Andrea Francis Soria Jimenez
| 2023-06-26T11:01:35 |
Create missing Jobs when /rows cache does not exists yet (#1425)
|
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py
index e0ddad9f..f95a63f8 100644
--- a/libs/libcommon/src/libcommon/parquet_utils.py
+++ b/libs/libcommon/src/libcommon/parquet_utils.py
@@ -20 +19,0 @@ from libcommon.constants import PARQUET_REVISION
-from libcommon.exceptions import UnexpectedError
@@ -399,12 +398,8 @@ class RowsIndex:
- try:
- result = get_previous_step_or_raise(
- kinds=cache_kinds,
- dataset=self.dataset,
- config=self.config,
- split=None,
- )
- self.revision = result.response["dataset_git_revision"]
- content = result.response["content"]
- except Exception as e:
- raise UnexpectedError("Could not get the list of parquet files to fetch the rows from.") from e
- # ^ TODO: improve the error, depending on the case
+ result = get_previous_step_or_raise(
+ kinds=cache_kinds,
+ dataset=self.dataset,
+ config=self.config,
+ split=None,
+ )
+ self.revision = result.response["dataset_git_revision"]
+ content = result.response["content"]
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 34d68dc4..13479c33 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -113,0 +114 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
+ hf_endpoint=app_config.common.hf_endpoint,
diff --git a/services/api/src/api/routes/endpoint.py b/services/api/src/api/routes/endpoint.py
index 8b8a4dab..4ddb8bc5 100644
--- a/services/api/src/api/routes/endpoint.py
+++ b/services/api/src/api/routes/endpoint.py
@@ -9,2 +8,0 @@ from typing import List, Mapping, Optional, Tuple, TypedDict
-from libcommon.dataset import get_dataset_git_revision
-from libcommon.orchestrator import DatasetOrchestrator
@@ -18 +15,0 @@ from libcommon.simple_cache import (
-from libcommon.utils import Priority
@@ -28,2 +24,0 @@ from api.utils import (
- ResponseNotFoundError,
- ResponseNotReadyError,
@@ -34,0 +30 @@ from api.utils import (
+ try_backfill_dataset,
@@ -86,30 +82,8 @@ def get_cache_entry_from_steps(
- dataset_orchestrator = DatasetOrchestrator(dataset=dataset, processing_graph=processing_graph)
- if not dataset_orchestrator.has_some_cache():
- # We have to check if the dataset exists and is supported
- try:
- revision = get_dataset_git_revision(
- dataset=dataset,
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- hf_timeout_seconds=hf_timeout_seconds,
- )
- except Exception as e:
- # The dataset is not supported
- raise ResponseNotFoundError("Not found.") from e
- # The dataset is supported, and the revision is known. We set the revision (it will create the jobs)
- # and tell the user to retry.
- dataset_orchestrator.set_revision(revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[])
- raise ResponseNotReadyError(
- "The server is busier than usual and the response is not ready yet. Please retry later."
- )
- elif dataset_orchestrator.has_pending_ancestor_jobs(
- processing_step_names=[processing_step.name for processing_step in processing_steps]
- ):
- # some jobs are still in progress, the cache entries could exist in the future
- raise ResponseNotReadyError(
- "The server is busier than usual and the response is not ready yet. Please retry later."
- )
- else:
- # no pending job: the cache entry will not be created
- raise ResponseNotFoundError("Not found.")
-
+ try_backfill_dataset(
+ processing_steps=processing_steps,
+ processing_graph=processing_graph,
+ dataset=dataset,
+ hf_endpoint=hf_endpoint,
+ hf_timeout_seconds=hf_timeout_seconds,
+ hf_token=hf_token,
+ )
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index 6f8d2d51..177865f5 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -16,0 +17 @@ from libcommon.prometheus import StepProfiler
+from libcommon.simple_cache import CachedArtifactError
@@ -34,0 +36 @@ from api.utils import (
+ try_backfill_dataset,
@@ -254,0 +257 @@ def create_rows_endpoint(
+ hf_endpoint: str,
@@ -309,6 +312,21 @@ def create_rows_endpoint(
- rows_index = indexer.get_rows_index(
- dataset=dataset,
- config=config,
- split=split,
- )
- revision = rows_index.revision
+ try:
+ rows_index = indexer.get_rows_index(
+ dataset=dataset,
+ config=config,
+ split=split,
+ )
+ revision = rows_index.revision
+ except CachedArtifactError:
+ config_parquet_processing_steps = processing_graph.get_config_parquet_processing_steps()
+ config_parquet_metadata_processing_steps = (
+ processing_graph.get_config_parquet_metadata_processing_steps()
+ )
+ try_backfill_dataset(
+ processing_steps=config_parquet_metadata_processing_steps
+ + config_parquet_processing_steps,
+ processing_graph=processing_graph,
+ dataset=dataset,
+ hf_endpoint=hf_endpoint,
+ hf_timeout_seconds=hf_timeout_seconds,
+ hf_token=hf_token,
+ )
diff --git a/services/api/src/api/utils.py b/services/api/src/api/utils.py
index 99fe6657..079529ee 100644
--- a/services/api/src/api/utils.py
+++ b/services/api/src/api/utils.py
@@ -7,0 +8 @@ from typing import Any, Callable, Coroutine, List, Literal, Optional
+from libcommon.dataset import get_dataset_git_revision
@@ -9 +10,3 @@ from libcommon.exceptions import CustomError
-from libcommon.utils import orjson_dumps
+from libcommon.orchestrator import DatasetOrchestrator
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
+from libcommon.utils import Priority, orjson_dumps
@@ -168,0 +172,39 @@ def are_valid_parameters(parameters: List[Any]) -> bool:
+def try_backfill_dataset(
+ processing_steps: List[ProcessingStep],
+ dataset: str,
+ processing_graph: ProcessingGraph,
+ hf_endpoint: str,
+ hf_token: Optional[str] = None,
+ hf_timeout_seconds: Optional[float] = None,
+) -> None:
+ dataset_orchestrator = DatasetOrchestrator(dataset=dataset, processing_graph=processing_graph)
+ if not dataset_orchestrator.has_some_cache():
+ # We have to check if the dataset exists and is supported
+ try:
+ revision = get_dataset_git_revision(
+ dataset=dataset,
+ hf_endpoint=hf_endpoint,
+ hf_token=hf_token,
+ hf_timeout_seconds=hf_timeout_seconds,
+ )
+ except Exception as e:
+ # The dataset is not supported
+ raise ResponseNotFoundError("Not found.") from e
+ # The dataset is supported, and the revision is known. We set the revision (it will create the jobs)
+ # and tell the user to retry.
+ dataset_orchestrator.set_revision(revision=revision, priority=Priority.NORMAL, error_codes_to_retry=[])
+ raise ResponseNotReadyError(
+ "The server is busier than usual and the response is not ready yet. Please retry later."
+ )
+ elif dataset_orchestrator.has_pending_ancestor_jobs(
+ processing_step_names=[processing_step.name for processing_step in processing_steps]
+ ):
+ # some jobs are still in progress, the cache entries could exist in the future
+ raise ResponseNotReadyError(
+ "The server is busier than usual and the response is not ready yet. Please retry later."
+ )
+ else:
+ # no pending job: the cache entry will not be created
+ raise ResponseNotFoundError("Not found.")
+
+
diff --git a/services/api/tests/routes/test_endpoint.py b/services/api/tests/routes/test_endpoint.py
index ebdd22a4..ab566739 100644
--- a/services/api/tests/routes/test_endpoint.py
+++ b/services/api/tests/routes/test_endpoint.py
@@ -156 +156 @@ def test_get_cache_entry_from_steps() -> None:
- with patch("api.routes.endpoint.get_dataset_git_revision", return_value=revision):
+ with patch("libcommon.dataset.get_dataset_git_revision", return_value=revision):
|
|
07887d39931663f1e4c1a78cd1140c6395a8d356
|
Quentin Lhoest
| 2023-06-26T10:40:24 |
Ignore duckdb files in parquet and info (#1429)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 1e1c4d9b..313166d5 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -798,0 +799,27 @@ def create_commits(
+def get_delete_operations(
+ parquet_operations: List[CommitOperationAdd], all_repo_files: Set[str], config_names: Set[str], config: str
+) -> List[CommitOperationDelete]:
+ # - get files that will be preserved in repo:
+ # 1. parquet files belonging to any other config (otherwise outdated files might be preserved)
+ # 2. duckdb files belonging to any config
+ # 3. .gitattributes
+ pattern_in_any_config_dir = re.compile(f"^({'|'.join(re.escape(conf) for conf in config_names)})/")
+ pattern_in_any_other_config_dir = re.compile(
+ f"^({'|'.join(re.escape(conf) for conf in config_names.difference({config}))})/"
+ )
+ files_to_ignore: Set[str] = {
+ file
+ for file in all_repo_files
+ if (pattern_in_any_other_config_dir.match(file) and file.endswith(".parquet"))
+ or (pattern_in_any_config_dir.match(file) and file.endswith(".duckdb"))
+ }.union({".gitattributes"})
+ # - get files to be deleted - all files except for:
+ # - the files to be preserved
+ # - parquet files obtained for current config at this processing step
+ files_to_add = [operation.path_in_repo for operation in parquet_operations]
+ files_to_delete = all_repo_files - set(files_to_add).union(files_to_ignore)
+ delete_operations = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete]
+ logging.debug(f"{delete_operations=}")
+ return delete_operations
+
+
@@ -850 +876,0 @@ def commit_parquet_conversion(
- # - get repo parquet files
@@ -852,18 +878,3 @@ def commit_parquet_conversion(
- repo_parquet_files: Set[str] = {file for file in all_repo_files if file.endswith(".parquet")}
- # - get files that will be preserved in repo: files belonging to other configs and .gitattributes
- # we exclude files of current config because otherwise outdated files might be preserved
- files_to_ignore: Set[str] = {
- file
- for other_config in config_names.difference({config})
- for file in repo_parquet_files
- if file.startswith(f"{other_config}/")
- }.union({".gitattributes"})
- # - get files to be deleted - all files except for:
- # - parquet files obtained for current config at this processing step,
- # - parquet files belonging to other existing configs
- # - .gitattributes
- files_to_add = [operation.path_in_repo for operation in parquet_operations]
- files_to_delete = all_repo_files - set(files_to_add).union(files_to_ignore)
- delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete]
- logging.debug(f"{delete_operations=}")
-
+ delete_operations = get_delete_operations(
+ parquet_operations=parquet_operations, all_repo_files=all_repo_files, config_names=config_names, config=config
+ )
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index 9de4ac77..e9631ea8 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -11 +11 @@ from pathlib import Path
-from typing import Any, Callable, Iterator, List, Optional, TypedDict
+from typing import Any, Callable, Iterator, List, Optional, Set, TypedDict
@@ -40,0 +41 @@ from worker.job_runners.config.parquet_and_info import (
+ get_delete_operations,
@@ -796,0 +798,31 @@ def test_concurrency(
+
+
[email protected](
+ "parquet_files,all_repo_files,config_names,config,deleted_files",
+ [
+ (
+ set(),
+ {"dummy", "c1/dummy", "c1/0.parquet", "c2/0.parquet", "c1/index.duckdb"},
+ {"c1", "c2"},
+ "c1",
+ {"dummy", "c1/dummy", "c1/0.parquet"},
+ ),
+ (
+ {"c1/0.parquet"},
+ {"dummy", "c1/dummy", "c1/0.parquet", "c2/0.parquet", "c1/index.duckdb"},
+ {"c1", "c2"},
+ "c1",
+ {"dummy", "c1/dummy"},
+ ),
+ ],
+)
+def test_get_delete_operations(
+ parquet_files: Set[str], all_repo_files: Set[str], config_names: Set[str], config: str, deleted_files: Set[str]
+) -> None:
+ parquet_operations = [
+ CommitOperationAdd(path_in_repo=path_in_repo, path_or_fileobj=b"") for path_in_repo in parquet_files
+ ]
+ delete_operations = get_delete_operations(
+ parquet_operations=parquet_operations, all_repo_files=all_repo_files, config_names=config_names, config=config
+ )
+ assert set(delete_operation.path_in_repo for delete_operation in delete_operations) == deleted_files
|
|
1d668b0ad9ea88fec9f37101a42b1934d61d726f
|
Quentin Lhoest
| 2023-06-26T08:21:49 |
Add auth in rows (#1427)
|
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index 2806e5b0..6f8d2d51 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -271 +271 @@ def create_rows_endpoint(
- httpfs=HTTPFileSystem(),
+ httpfs=HTTPFileSystem(storage_options={"headers": {"authorization": f"Bearer {hf_token}"}}),
|
|
ad4351dc9ee2995cf716aa83afb50f8f1fb305f1
|
Quentin Lhoest
| 2023-06-23T16:01:52 |
Fix regression: use parquet metadata when possible (#1424)
|
diff --git a/e2e/tests/test_11_api.py b/e2e/tests/test_11_api.py
index d6abeef3..7732ea1f 100644
--- a/e2e/tests/test_11_api.py
+++ b/e2e/tests/test_11_api.py
@@ -109 +109 @@ def test_rows_endpoint(
- limit = 10
+ length = 10
@@ -111 +111 @@ def test_rows_endpoint(
- relative_url=f"/rows?dataset={dataset}&config={config}&split={split}&offset={offset}&limit={limit}",
+ relative_url=f"/rows?dataset={dataset}&config={config}&split={split}&offset={offset}&length={length}",
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py
index 3a498888..e0ddad9f 100644
--- a/libs/libcommon/src/libcommon/parquet_utils.py
+++ b/libs/libcommon/src/libcommon/parquet_utils.py
@@ -7 +7 @@ from os import PathLike
-from typing import Callable, List, Literal, Optional, Tuple, TypedDict, Union, cast
+from typing import Callable, List, Literal, Optional, Tuple, TypedDict, Union
@@ -60,22 +59,0 @@ class ParquetFileMetadataItem(TypedDict):
-class HTTPFileSystemSession(object):
- _singleton: Optional["HTTPFileSystemSession"] = None
-
- @classmethod
- def get_instance(cls) -> "HTTPFileSystemSession":
- if cls._singleton is None:
- HTTPFileSystemSession()
- return cast("HTTPFileSystemSession", HTTPFileSystemSession._singleton)
-
- def __new__(cls) -> "HTTPFileSystemSession":
- if HTTPFileSystemSession._singleton is not None:
- raise RuntimeError(
- "cannot initialize another instance of HTTPFileSystemSession, use .get_instance() instead"
- )
- return super(HTTPFileSystemSession, cls).__new__(cls)
-
- def __init__(self) -> None:
- HTTPFileSystemSession._singleton = self
- self.httpfs = HTTPFileSystem()
- self.session = asyncio.run(self.httpfs.set_session())
-
-
@@ -236,0 +215 @@ class ParquetIndexWithMetadata:
+ httpfs: HTTPFileSystem
@@ -238,0 +218,6 @@ class ParquetIndexWithMetadata:
+ def __post_init__(self) -> None:
+ if self.httpfs._session is None:
+ self.httpfs_session = asyncio.run(self.httpfs.set_session())
+ else:
+ self.httpfs_session = self.httpfs._session
+
@@ -273,3 +257,0 @@ class ParquetIndexWithMetadata:
- httpFileSystemSession = HTTPFileSystemSession.get_instance()
- session = httpFileSystemSession.session
- httpfs = httpFileSystemSession.httpfs
@@ -278 +260,8 @@ class ParquetIndexWithMetadata:
- HTTPFile(httpfs, url, session=session, size=size, loop=httpfs.loop, cache_type=None),
+ HTTPFile(
+ self.httpfs,
+ url,
+ session=self.httpfs_session,
+ size=size,
+ loop=self.httpfs.loop,
+ cache_type=None,
+ ),
@@ -319,0 +309 @@ class ParquetIndexWithMetadata:
+ httpfs: HTTPFileSystem,
@@ -359,0 +350 @@ class ParquetIndexWithMetadata:
+ httpfs=httpfs,
@@ -370,0 +362 @@ class RowsIndex:
+ httpfs: HfFileSystem,
@@ -379,0 +372 @@ class RowsIndex:
+ self.httpfs = httpfs
@@ -403,2 +396,2 @@ class RowsIndex:
- cache_kinds = [step.cache_kind for step in config_parquet_processing_steps]
- cache_kinds.extend([step.cache_kind for step in config_parquet_metadata_processing_steps])
+ cache_kinds = [step.cache_kind for step in config_parquet_metadata_processing_steps]
+ cache_kinds.extend([step.cache_kind for step in config_parquet_processing_steps])
@@ -438,0 +432 @@ class RowsIndex:
+ httpfs=self.httpfs,
@@ -465,0 +460 @@ class Indexer:
+ httpfs: HTTPFileSystem,
@@ -471,0 +467 @@ class Indexer:
+ self.httpfs = httpfs
@@ -492,0 +489 @@ class Indexer:
+ httpfs=self.httpfs,
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index 618a6e02..2806e5b0 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -12,0 +13 @@ from datasets import Features
+from fsspec.implementations.http import HTTPFileSystem
@@ -269,0 +271 @@ def create_rows_endpoint(
+ httpfs=HTTPFileSystem(),
@@ -274,0 +277 @@ def create_rows_endpoint(
+ await indexer.httpfs.set_session()
diff --git a/services/api/tests/routes/test_rows.py b/services/api/tests/routes/test_rows.py
index 9dbad3ed..7613d52a 100644
--- a/services/api/tests/routes/test_rows.py
+++ b/services/api/tests/routes/test_rows.py
@@ -14,0 +15 @@ from fsspec import AbstractFileSystem
+from fsspec.implementations.http import HTTPFileSystem
@@ -247 +248,5 @@ def dataset_image_with_config_parquet() -> dict[str, Any]:
-def indexer(app_config: AppConfig, processing_graph: ProcessingGraph, parquet_metadata_directory: StrPath) -> Indexer:
+def indexer(
+ app_config: AppConfig,
+ processing_graph: ProcessingGraph,
+ parquet_metadata_directory: StrPath,
+) -> Indexer:
@@ -251,0 +257 @@ def indexer(app_config: AppConfig, processing_graph: ProcessingGraph, parquet_me
+ httpfs=HTTPFileSystem(),
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
index a4f64ece..07bdb663 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
@@ -7,0 +8 @@ from datasets import Audio, Features, Image
+from fsspec.implementations.http import HTTPFileSystem
@@ -198,0 +200 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner):
+ httpfs=HTTPFileSystem(),
|
|
bddd6c15995b6f72a93a8b1cd692aca58cc2a4c8
|
Quentin Lhoest
| 2023-06-23T12:02:30 |
update to datasets 2.13.1 (#1422)
|
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index d43c651e..0de28c38 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -529 +529 @@ name = "datasets"
-version = "2.13.0"
+version = "2.13.1"
@@ -535,2 +535,2 @@ files = [
- {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
- {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
@@ -1248 +1248 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
@@ -2754 +2753,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock
index 6b1a8a05..acc68895 100644
--- a/jobs/cache_maintenance/poetry.lock
+++ b/jobs/cache_maintenance/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -579 +579 @@ name = "datasets"
-version = "2.13.0"
+version = "2.13.1"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
- {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
@@ -1026 +1026 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
@@ -2566 +2565,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index 6b1a8a05..acc68895 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -579 +579 @@ name = "datasets"
-version = "2.13.0"
+version = "2.13.1"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
- {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
@@ -1026 +1026 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
@@ -2566 +2565,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index 99b16a6a..aa8af056 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -579 +579 @@ name = "datasets"
-version = "2.13.0"
+version = "2.13.1"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
- {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
@@ -2547 +2546,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
@@ -3015 +3014 @@ python-versions = "3.9.15"
-content-hash = "bfdd77727f041d88e26a4da8c23a6b8434f20c6747f068f1ca3c11d841e38aa1"
+content-hash = "c796cce59e2755d2a0b49d39a532262156b903ef92d51cfd769f3e66eceb932c"
diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml
index 8dcfc6a6..c4234875 100644
--- a/libs/libcommon/pyproject.toml
+++ b/libs/libcommon/pyproject.toml
@@ -10 +10 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index e138e9d3..bea23aef 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -579 +579 @@ name = "datasets"
-version = "2.13.0"
+version = "2.13.1"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
- {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
@@ -1084 +1084 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
@@ -2661 +2660,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index 77c2de7d..d688e743 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -621 +621 @@ name = "datasets"
-version = "2.13.0"
+version = "2.13.1"
@@ -627,2 +627,2 @@ files = [
- {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
- {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
@@ -1146 +1146 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
@@ -2852 +2851,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 5f59e6b5..0ece7030 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -910 +910 @@ name = "datasets"
-version = "2.13.0"
+version = "2.13.1"
@@ -916,2 +916,2 @@ files = [
- {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
- {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
+ {file = "datasets-2.13.1-py3-none-any.whl", hash = "sha256:844d8dbc1759e0b6b8e5063af019dc95d6af07ea075002b03323a280bf8d53f6"},
+ {file = "datasets-2.13.1.tar.gz", hash = "sha256:bacb7750b1a434417312b4281a55225a3f7e0163abdd12a2a3e2d700310d5221"},
@@ -1780 +1780 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.1", extras = ["audio", "vision"]}
@@ -2865 +2864,0 @@ files = [
- {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"},
@@ -4401 +4399,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
@@ -4694,0 +4693,2 @@ files = [
+ {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:9c9b14fbb73ec4cb0f209722a1489020fd8614c92ae22589f2309c48cefdf21f"},
+ {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_x86_64.whl", hash = "sha256:6a54539bd076746f69ae8bef7282f981674fe4dbf59c3a84c4af86ae6bae9d5c"},
|
|
5584b9cf9e72e4b6b6b22fa436a5f0936a4b3d96
|
Quentin Lhoest
| 2023-06-22T12:51:40 |
keep image and audio untruncated (#1417)
|
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
index 64112c28..a4f64ece 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
@@ -7 +7 @@ from typing import List
-from datasets import Features
+from datasets import Audio, Features, Image
@@ -142,0 +143 @@ def compute_first_rows_response(
+ columns_to_keep_untruncated = [col for col, feature in features.items() if isinstance(feature, (Image, Audio))]
@@ -147,0 +149 @@ def compute_first_rows_response(
+ columns_to_keep_untruncated=columns_to_keep_untruncated,
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
index a49cc8ea..3dc01f41 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
@@ -8 +8,8 @@ from typing import List, Optional, Union
-from datasets import Features, IterableDataset, get_dataset_config_info, load_dataset
+from datasets import (
+ Audio,
+ Features,
+ Image,
+ IterableDataset,
+ get_dataset_config_info,
+ load_dataset,
+)
@@ -246,0 +254 @@ def compute_first_rows_response(
+ columns_to_keep_untruncated = [col for col, feature in features.items() if isinstance(feature, (Image, Audio))]
@@ -251,0 +260 @@ def compute_first_rows_response(
+ columns_to_keep_untruncated=columns_to_keep_untruncated,
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index b4869d67..982da72d 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -201 +201 @@ def utf8_byte_truncate(text: str, max_bytes: int) -> str:
-def truncate_row_item(row_item: RowItem, min_cell_bytes: int) -> RowItem:
+def truncate_row_item(row_item: RowItem, min_cell_bytes: int, columns_to_keep_untruncated: List[str]) -> RowItem:
@@ -208 +208 @@ def truncate_row_item(row_item: RowItem, min_cell_bytes: int) -> RowItem:
- if len(cell_json) <= min_cell_bytes:
+ if len(cell_json) <= min_cell_bytes or column_name in columns_to_keep_untruncated:
@@ -224 +224,3 @@ COMMA_SIZE = 1 # the comma "," is encoded with one byte in utf-8
-def truncate_row_items(row_items: List[RowItem], min_cell_bytes: int, rows_max_bytes: int) -> List[RowItem]:
+def truncate_row_items(
+ row_items: List[RowItem], min_cell_bytes: int, rows_max_bytes: int, columns_to_keep_untruncated: List[str]
+) -> List[RowItem]:
@@ -233 +235,3 @@ def truncate_row_items(row_items: List[RowItem], min_cell_bytes: int, rows_max_b
- row_item = truncate_row_item(row_item=row_item, min_cell_bytes=min_cell_bytes)
+ row_item = truncate_row_item(
+ row_item=row_item, min_cell_bytes=min_cell_bytes, columns_to_keep_untruncated=columns_to_keep_untruncated
+ )
@@ -251,0 +256 @@ def create_truncated_row_items(
+ columns_to_keep_untruncated: List[str],
@@ -277 +282,6 @@ def create_truncated_row_items(
- return truncate_row_items(row_items=row_items, min_cell_bytes=min_cell_bytes, rows_max_bytes=rows_max_bytes)
+ return truncate_row_items(
+ row_items=row_items,
+ min_cell_bytes=min_cell_bytes,
+ rows_max_bytes=rows_max_bytes,
+ columns_to_keep_untruncated=columns_to_keep_untruncated,
+ )
|
|
8763a497aa8f40970cc9008c9e68072f2c964794
|
Rémy
| 2023-06-22T07:51:06 |
feat: use external-secrets to read secrets from AWS (#1419)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index e7420e4e..96acf9f7 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -51,0 +52,10 @@ secrets:
+ externalSecret:
+ enabled: true
+ secretName: "datasets-server-prod-secrets"
+ secretStoreName: "datasets-server-prod-secretstore"
+ parameters:
+ MONGO_URL: "hub-prod-datasets-server-mongo-url"
+ HF_TOKEN: "hub-prod-datasets-server-hf-token"
+ PARQUET_CONVERTER_HF_TOKEN: "hub-prod-datasets-server-parquet-converter-hf-token"
+ WEBHOOK_SECRET: "hub-prod-datasets-server-webhook-secret"
+ SPAWNING_TOKEN: "hub-prod-datasets-server-spawning-token"
@@ -54,2 +64 @@ secrets:
- secretName: "mongo-url"
- value: mongo://
+ secretName: "datasets-server-prod-secrets"
@@ -58 +67 @@ secrets:
- secretName: "hf-token"
+ secretName: "datasets-server-prod-secrets"
@@ -61 +70 @@ secrets:
- secretName: "parquet-converter-hf-token"
+ secretName: "datasets-server-prod-secrets"
@@ -64 +73 @@ secrets:
- secretName: "webhook-secret"
+ secretName: "datasets-server-prod-secrets"
@@ -67 +76 @@ secrets:
- secretName: "spawning-token"
+ secretName: "datasets-server-prod-secrets"
diff --git a/chart/env/staging.yaml b/chart/env/staging.yaml
index 54263cb8..3014a6ef 100644
--- a/chart/env/staging.yaml
+++ b/chart/env/staging.yaml
@@ -49,0 +50,10 @@ secrets:
+ externalSecret:
+ enabled: true
+ secretName: "datasets-server-staging-secrets"
+ secretStoreName: "datasets-server-ephemeral-secretstore"
+ parameters:
+ MONGO_URL: "hub-ephemeral-datasets-server-mongo-url"
+ HF_TOKEN: "hub-ephemeral-datasets-server-hf-token"
+ PARQUET_CONVERTER_HF_TOKEN: "hub-ephemeral-datasets-server-parquet-converter-hf-token"
+ WEBHOOK_SECRET: "hub-ephemeral-datasets-server-webhook-secret"
+ SPAWNING_TOKEN: "hub-ephemeral-datasets-server-spawning-token"
@@ -52,2 +62 @@ secrets:
- secretName: "mongo-url"
- value: mongo://
+ secretName: "datasets-server-staging-secrets"
@@ -56 +65 @@ secrets:
- secretName: "datasets-server-hf-token"
+ secretName: "datasets-server-staging-secrets"
@@ -59 +68 @@ secrets:
- secretName: "parquet-converter-hf-token"
+ secretName: "datasets-server-staging-secrets"
@@ -62,2 +71 @@ secrets:
- secretName: "webhook-secret"
- value: ""
+ secretName: "datasets-server-staging-secrets"
@@ -66 +74 @@ secrets:
- secretName: "spawning-token"
+ secretName: "datasets-server-staging-secrets"
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
index 3d9765fc..6f2d1411 100644
--- a/chart/templates/_envWorker.tpl
+++ b/chart/templates/_envWorker.tpl
@@ -49 +49 @@
- key: HF_TOKEN
+ key: PARQUET_CONVERTER_HF_TOKEN
diff --git a/chart/templates/secrets.yaml b/chart/templates/secrets.yaml
new file mode 100644
index 00000000..6e371747
--- /dev/null
+++ b/chart/templates/secrets.yaml
@@ -0,0 +1,20 @@
+{{- if .Values.secrets.externalSecret.enabled }}
+apiVersion: "external-secrets.io/v1beta1"
+kind: ExternalSecret
+metadata:
+ name: {{ include "name" $ }}-external-secret
+ namespace: {{ $.Release.Namespace }}
+spec:
+ refreshInterval: 1h
+ secretStoreRef:
+ name: {{ .Values.secrets.externalSecret.secretStoreName }}
+ kind: SecretStore
+ target:
+ name: {{ .Values.secrets.externalSecret.secretName }}
+ data:
+ {{- range $key, $value := .Values.secrets.externalSecret.parameters }}
+ - secretKey: {{ $key | quote }}
+ remoteRef:
+ key: {{ $value | quote }}
+ {{- end }}
+{{- end }}
\ No newline at end of file
diff --git a/chart/values.yaml b/chart/values.yaml
index 0f99fb5d..4a57fc7a 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -68,0 +69,5 @@ secrets:
+ externalSecret:
+ enabled: false
+ secretName: ""
+ secretStoreName: ""
+ parameters: { }
|
|
31424f197038003216e23ebae9d68de91a3b1a3b
|
Sylvain Lesage
| 2023-06-21T15:06:14 |
Test concurrency in parquet and info (#1415)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index a0c0f528..1e1c4d9b 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -1005,6 +1004,0 @@ def compute_config_parquet_and_info_response(
- # check if the repo exists and get the list of refs
- try:
- refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE)
- except RepositoryNotFoundError as err:
- raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err
-
@@ -1021,0 +1016,2 @@ def compute_config_parquet_and_info_response(
+ except FileNotFoundError as err:
+ raise DatasetNotFoundError("The dataset, or the revision, does not exist on the Hub.") from err
@@ -1041,11 +1036,0 @@ def compute_config_parquet_and_info_response(
- # create the target revision if we managed to get the parquet files and it does not exist yet
- # (clone from initial commit to avoid cloning all repo's files)
- try:
- if all(ref.ref != target_revision for ref in refs.converts):
- initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id
- committer_hf_api.create_branch(
- repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit, exist_ok=True
- )
- except RepositoryNotFoundError as err:
- raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err
-
@@ -1053 +1038,2 @@ def compute_config_parquet_and_info_response(
- sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300]
+ sleeps = [1, 1, 1, 1, 1, 10, 10, 10, 10, 100] * 3
+ # ^ timeouts after ~7 minutes
@@ -1054,0 +1041,15 @@ def compute_config_parquet_and_info_response(
+ # create the target revision if we managed to get the parquet files and it does not exist yet
+ # (clone from initial commit to avoid cloning all repo's files)
+ refs = retry(on=[requests.exceptions.ConnectionError], sleeps=[1, 1, 1, 10, 10])(hf_api.list_repo_refs)(
+ repo_id=dataset, repo_type=DATASET_TYPE
+ )
+ if all(ref.ref != target_revision for ref in refs.converts):
+ initial_commit = hf_api.list_repo_commits(repo_id=dataset, repo_type=DATASET_TYPE)[-1].commit_id
+ committer_hf_api.create_branch(
+ repo_id=dataset,
+ branch=target_revision,
+ repo_type=DATASET_TYPE,
+ revision=initial_commit,
+ exist_ok=True,
+ )
+ # commit the parquet files
@@ -1064,0 +1066,2 @@ def compute_config_parquet_and_info_response(
+ # call the API again to get the list of parquet files
+ target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True)
@@ -1066,0 +1070,2 @@ def compute_config_parquet_and_info_response(
+ except RepositoryNotFoundError as err:
+ raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err
@@ -1068,2 +1072,0 @@ def compute_config_parquet_and_info_response(
- # call the API again to get the list of parquet files
- target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=True)
diff --git a/services/worker/tests/fixtures/files.py b/services/worker/tests/fixtures/files.py
index f2530413..1bc2a1f3 100644
--- a/services/worker/tests/fixtures/files.py
+++ b/services/worker/tests/fixtures/files.py
@@ -151,0 +152,34 @@ def dataset_script_with_two_configs_path(tmp_path_factory: pytest.TempPathFactor
+# N = 15
+DATASET_SCRIPT_WITH_N_CONFIGS = """
+import os
+
+import datasets
+from datasets import DatasetInfo, BuilderConfig, Features, Split, SplitGenerator, Value
+
+
+class DummyDataset(datasets.GeneratorBasedBuilder):
+
+ BUILDER_CONFIGS = [BuilderConfig(name="config"+str(i)) for i in range(15)]
+
+ def _info(self) -> DatasetInfo:
+ return DatasetInfo(features=Features({"text": Value("string")}))
+
+ def _split_generators(self, dl_manager):
+ return [
+ SplitGenerator(Split.TRAIN, gen_kwargs={"text": self.config.name}),
+ ]
+
+ def _generate_examples(self, text, **kwargs):
+ for i in range(1000):
+ yield i, {"text": text}
+"""
+
+
[email protected](scope="session")
+def dataset_script_with_n_configs_path(tmp_path_factory: pytest.TempPathFactory) -> str:
+ path = str(tmp_path_factory.mktemp("data") / "{dataset_name}.py")
+ with open(path, "w", newline="") as f:
+ f.write(DATASET_SCRIPT_WITH_N_CONFIGS)
+ return path
+
+
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 34941810..f34ca587 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -262,0 +263,7 @@ def hub_public_legacy_configs(dataset_script_with_two_configs_path: str) -> Iter
[email protected](scope="session")
+def hub_public_n_configs(dataset_script_with_n_configs_path: str) -> Iterator[str]:
+ repo_id = create_hub_dataset_repo(prefix="n_configs", file_paths=[dataset_script_with_n_configs_path])
+ yield repo_id
+ delete_hub_dataset_repo(repo_id=repo_id)
+
+
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index 41533372..9de4ac77 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -9 +9,3 @@ from http import HTTPStatus
-from typing import Any, Callable, Iterator, List, Optional
+from multiprocessing import Pool
+from pathlib import Path
+from typing import Any, Callable, Iterator, List, Optional, TypedDict
@@ -28 +30,2 @@ from libcommon.exceptions import (
-from libcommon.processing_graph import ProcessingGraph
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
+from libcommon.queue import Queue
@@ -31 +34 @@ from libcommon.simple_cache import CachedArtifactError, upsert_response
-from libcommon.utils import Priority
+from libcommon.utils import JobInfo, JobParams, Priority
@@ -33,0 +37 @@ from worker.config import AppConfig
+from worker.job_manager import JobManager
@@ -44,0 +49 @@ from worker.job_runners.config.parquet_and_info import (
+from worker.job_runners.dataset.config_names import DatasetConfigNamesJobRunner
@@ -45,0 +51 @@ from worker.resources import LibrariesResource
+from worker.utils import CompleteJobResult
@@ -651,0 +658,139 @@ def test_create_commits(
+
+
+GetDatasetConfigNamesJobRunner = Callable[[str, AppConfig], DatasetConfigNamesJobRunner]
+
+
[email protected]
+def get_dataset_config_names_job_runner(
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetDatasetConfigNamesJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ app_config: AppConfig,
+ ) -> DatasetConfigNamesJobRunner:
+ processing_step_name = DatasetConfigNamesJobRunner.get_job_type()
+ processing_graph = ProcessingGraph(
+ {
+ processing_step_name: {
+ "input_type": "dataset",
+ "job_runner_version": DatasetConfigNamesJobRunner.get_job_runner_version(),
+ }
+ }
+ )
+ return DatasetConfigNamesJobRunner(
+ job_info={
+ "type": DatasetConfigNamesJobRunner.get_job_type(),
+ "params": {
+ "dataset": dataset,
+ "revision": "revision",
+ "config": None,
+ "split": None,
+ },
+ "job_id": "job_id",
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=processing_graph.get_processing_step(processing_step_name),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_job_runner
+
+
+class JobRunnerArgs(TypedDict):
+ dataset: str
+ revision: str
+ config: str
+ app_config: AppConfig
+ tmp_path: Path
+
+
+def launch_job_runner(job_runner_args: JobRunnerArgs) -> CompleteJobResult:
+ config = job_runner_args["config"]
+ dataset = job_runner_args["dataset"]
+ revision = job_runner_args["revision"]
+ app_config = job_runner_args["app_config"]
+ tmp_path = job_runner_args["tmp_path"]
+ job_runner = ConfigParquetAndInfoJobRunner(
+ job_info=JobInfo(
+ job_id=f"job_{config}",
+ type="config-parquet-and-info",
+ params=JobParams(dataset=dataset, revision=revision, config=config, split=None),
+ priority=Priority.NORMAL,
+ ),
+ app_config=app_config,
+ processing_step=ProcessingStep(
+ name="config-parquet-and-info",
+ input_type="config",
+ job_runner_version=ConfigParquetAndInfoJobRunner.get_job_runner_version(),
+ ),
+ hf_datasets_cache=tmp_path,
+ )
+ return job_runner.compute()
+
+
+def test_concurrency(
+ hub_public_n_configs: str,
+ app_config: AppConfig,
+ tmp_path: Path,
+ get_dataset_config_names_job_runner: GetDatasetConfigNamesJobRunner,
+ queue_mongo_resource: QueueMongoResource,
+ cache_mongo_resource: CacheMongoResource,
+) -> None:
+ """
+ Test that multiple job runners (to compute config-parquet-and-info) can run in parallel,
+ without having conflicts when sending commits to the Hub.
+ For this test, we need a lot of configs for the same dataset (say 20) and one job runner for each.
+ Ideally we would try for both quick and slow jobs.
+ """
+ repo_id = hub_public_n_configs
+ hf_api = HfApi(endpoint=CI_HUB_ENDPOINT, token=CI_USER_TOKEN)
+ revision = hf_api.dataset_info(repo_id=repo_id, files_metadata=False).sha
+ if revision is None:
+ raise ValueError(f"Could not find revision for dataset {repo_id}")
+
+ # fill the cache for the step dataset-config-names, required by the job_runner
+ # it's a lot of code 😅
+ job_info = JobInfo(
+ job_id="not_used",
+ type="dataset-config-names",
+ params=JobParams(dataset=repo_id, revision=revision, config=None, split=None),
+ priority=Priority.NORMAL,
+ )
+ queue = Queue()
+ queue.create_jobs([job_info])
+ job_info = queue.start_job(job_types_only=["dataset-config-names"])
+ job_manager = JobManager(
+ job_info=job_info,
+ app_config=app_config,
+ processing_graph=ProcessingGraph(
+ {
+ "dataset-config-names": {
+ "input_type": "dataset",
+ "provides_dataset_config_names": True,
+ "job_runner_version": DatasetConfigNamesJobRunner.get_job_runner_version(),
+ }
+ }
+ ),
+ job_runner=get_dataset_config_names_job_runner(repo_id, app_config),
+ )
+ job_result = job_manager.run_job()
+ job_manager.finish(job_result=job_result)
+ if not job_result["output"]:
+ raise ValueError("Could not get config names")
+ configs = [str(config_name["config"]) for config_name in job_result["output"]["content"]["config_names"]]
+
+ # launch the job runners
+ NUM_JOB_RUNNERS = 10
+ with Pool(NUM_JOB_RUNNERS) as pool:
+ pool.map(
+ launch_job_runner,
+ [
+ JobRunnerArgs(
+ dataset=repo_id, revision=revision, config=config, app_config=app_config, tmp_path=tmp_path
+ )
+ for config in configs
+ ],
+ )
|
|
58c55b64c093929e7f08c527951c50b88879d47b
|
Sylvain Lesage
| 2023-06-21T14:25:15 |
Fix the lock (#1414)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 1371ee6a..550b1a52 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -254,6 +253,0 @@ class lock(contextlib.AbstractContextManager["lock"]):
- try:
- Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save(
- write_concern={"w": "majority", "fsync": True}
- )
- except NotUniqueError:
- pass
@@ -261 +255 @@ class lock(contextlib.AbstractContextManager["lock"]):
- acquired = (
+ try:
@@ -263 +257,5 @@ class lock(contextlib.AbstractContextManager["lock"]):
- job_id=self.job_id, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True}
+ upsert=True,
+ write_concern={"w": "majority", "fsync": True},
+ read_concern={"level": "majority"},
+ job_id=self.job_id,
+ updated_at=get_datetime(),
@@ -265,3 +262,0 @@ class lock(contextlib.AbstractContextManager["lock"]):
- > 0
- )
- if acquired:
@@ -269,2 +264,3 @@ class lock(contextlib.AbstractContextManager["lock"]):
- logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'")
- time.sleep(sleep)
+ except NotUniqueError:
+ logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'")
+ time.sleep(sleep)
@@ -275 +271,5 @@ class lock(contextlib.AbstractContextManager["lock"]):
- job_id=None, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True}
+ upsert=True,
+ write_concern={"w": "majority", "fsync": True},
+ read_concern={"level": "majority"},
+ job_id=None,
+ updated_at=get_datetime(),
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 18f3e259..82b930fb 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -5,0 +6,2 @@ import os
+import random
+import time
@@ -391,0 +394,5 @@ def test_has_ttl_index_on_finished_at_field() -> None:
+def random_sleep() -> None:
+ MAX_SLEEP_MS = 40
+ time.sleep(MAX_SLEEP_MS / 1000 * random.random())
+
+
@@ -392,0 +400 @@ def increment(tmp_file: Path) -> None:
+ random_sleep()
@@ -394,0 +403 @@ def increment(tmp_file: Path) -> None:
+ random_sleep()
@@ -396,0 +406 @@ def increment(tmp_file: Path) -> None:
+ random_sleep()
@@ -400 +410,2 @@ def locked_increment(tmp_file: Path) -> None:
- with lock(key="test_lock", job_id=str(os.getpid())):
+ sleeps = [0.05, 0.05, 0.05, 1, 1, 1, 1, 1, 1, 5, 5, 5, 5]
+ with lock(key="test_lock", job_id=str(os.getpid()), sleeps=sleeps):
@@ -420 +431 @@ def git_branch_locked_increment(tmp_file: Path) -> None:
- sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300]
+ sleeps = [0.05, 0.05, 0.05, 1, 1, 1, 1, 1, 1, 5, 5, 5, 5]
|
|
2848ab4660367d9a74a01ae2f4566c0e243d3009
|
Albert Villanova del Moral
| 2023-06-21T14:22:44 |
Remove unused code from /rows API endpoint (#1394)
|
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index 702fea88..618a6e02 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -4 +3,0 @@
-import asyncio
@@ -14 +12,0 @@ from datasets import Features
-from fsspec.implementations.http import HTTPFileSystem
@@ -41,4 +38,0 @@ logger = logging.getLogger(__name__)
-httpfs = HTTPFileSystem()
-session = asyncio.run(httpfs.set_session())
-
-
|
|
178ed7b0ff6199d85a5503bd8804e70cbf02c1f2
|
Quentin Lhoest
| 2023-06-21T14:21:56 |
unblock DFKI-SLT/few-nerd (#1412)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 06b172c5..e7420e4e 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -91 +91 @@ parquetAndInfo:
- blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
+ blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
|
|
1bb3b274b750e551fb943a729706755d5519467f
|
Quentin Lhoest
| 2023-06-21T09:54:11 |
Retry get_parquet_file_and_size (#1404)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 73dfb0e8..a0c0f528 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -16,0 +17 @@ import numpy as np
+import pyarrow as pa
@@ -618,0 +620,6 @@ def copy_parquet_files(builder: DatasetBuilder) -> List[CommitOperationCopy]:
+class NotAParquetFileError(ValueError):
+ """When a remote parquet file can't be parsed"""
+
+ pass
+
+
@@ -624,0 +632,14 @@ def get_parquet_file_and_size(url: str, fs: HTTPFileSystem, hf_token: Optional[s
+def retry_get_parquet_file_and_size(
+ url: str, fs: HTTPFileSystem, hf_token: Optional[str]
+) -> Tuple[pq.ParquetFile, int]:
+ try:
+ sleeps = [1, 1, 1, 10, 10, 10]
+ pf, size = retry(on=[pa.ArrowInvalid], sleeps=sleeps)(get_parquet_file_and_size)(url, fs, hf_token)
+ return pf, size
+ except RuntimeError as err:
+ if err.__cause__ and isinstance(err.__cause__, pa.ArrowInvalid):
+ raise NotAParquetFileError(f"Not a parquet file: '{url}'") from err.__cause__
+ else:
+ raise err
+
+
@@ -639 +660 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None:
- partial(get_parquet_file_and_size, fs=fs, hf_token=hf_token),
+ partial(retry_get_parquet_file_and_size, fs=fs, hf_token=hf_token),
|
|
dd8f9fc58c9f652578e796d6bd1ffaa887dfc527
|
Sylvain Lesage
| 2023-06-20T21:26:59 |
fix: 🐛 try to ensure only one job can get the lock (#1410)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 27128ccb..1371ee6a 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -255 +255,3 @@ class lock(contextlib.AbstractContextManager["lock"]):
- Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save()
+ Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save(
+ write_concern={"w": "majority", "fsync": True}
+ )
@@ -261 +263 @@ class lock(contextlib.AbstractContextManager["lock"]):
- job_id=self.job_id, updated_at=get_datetime()
+ job_id=self.job_id, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True}
@@ -272 +274,3 @@ class lock(contextlib.AbstractContextManager["lock"]):
- Lock.objects(key=self.key, job_id=self.job_id).update(job_id=None, updated_at=get_datetime())
+ Lock.objects(key=self.key, job_id=self.job_id).update(
+ job_id=None, updated_at=get_datetime(), write_concern={"w": "majority", "fsync": True}
+ )
|
|
86bd831913706c4beecd197b54015e4de2ddb0ab
|
Sylvain Lesage
| 2023-06-20T20:35:11 |
test: 💍 add a test on lock.git_branch (#1409)
|
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 02b9578e..18f3e259 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -3,0 +4 @@
+import json
@@ -415,0 +417,26 @@ def test_lock(tmp_path_factory: pytest.TempPathFactory, queue_mongo_resource: Qu
+
+
+def git_branch_locked_increment(tmp_file: Path) -> None:
+ sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300]
+ dataset = "dataset"
+ branch = "refs/convert/parquet"
+ with lock.git_branch(dataset=dataset, branch=branch, job_id=str(os.getpid()), sleeps=sleeps):
+ increment(tmp_file)
+
+
+def test_lock_git_branch(tmp_path_factory: pytest.TempPathFactory, queue_mongo_resource: QueueMongoResource) -> None:
+ tmp_file = Path(tmp_path_factory.mktemp("test_lock") / "tmp.txt")
+ tmp_file.touch()
+ max_parallel_jobs = 5
+ num_jobs = 43
+
+ with Pool(max_parallel_jobs, initializer=queue_mongo_resource.allocate) as pool:
+ pool.map(git_branch_locked_increment, [tmp_file] * num_jobs)
+
+ expected = num_jobs
+ with open(tmp_file, "r") as f:
+ assert int(f.read()) == expected
+ assert Lock.objects().count() == 1
+ assert Lock.objects().get().key == json.dumps({"dataset": "dataset", "branch": "refs/convert/parquet"})
+ assert Lock.objects().get().job_id is None
+ Lock.objects().delete()
|
|
f38b99d2d49a0e5deb08014705c61ace13282f88
|
Sylvain Lesage
| 2023-06-20T19:08:39 |
Add tests on create_commits (#1408)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 2769bdd8..73dfb0e8 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -731 +731 @@ def create_commits(
- The ma number of operations per commit, to avoid time out errors from the Hub. Defaults to 500.
+ The max number of operations per commit, to avoid time out errors from the Hub. Defaults to 500.
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index da77c3fc..34941810 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -621 +621 @@ def hub_reponses_empty(hub_public_empty: str) -> HubDatasetTest:
-def hub_reponses_public(hub_public_csv: str) -> HubDatasetTest:
+def hub_responses_public(hub_public_csv: str) -> HubDatasetTest:
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index 2121d53d..41533372 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -17 +17 @@ from datasets import Audio, Features, Image, Value, load_dataset_builder
-from huggingface_hub.hf_api import HfApi
+from huggingface_hub.hf_api import CommitOperationAdd, HfApi
@@ -35,0 +36 @@ from worker.job_runners.config.parquet_and_info import (
+ create_commits,
@@ -118 +119 @@ def test_compute(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -120,2 +121,2 @@ def test_compute(
- dataset = hub_reponses_public["name"]
- config = hub_reponses_public["config_names_response"]["config_names"][0]["config"]
+ dataset = hub_responses_public["name"]
+ config = hub_responses_public["config_names_response"]["config_names"][0]["config"]
@@ -126 +127 @@ def test_compute(
- content=hub_reponses_public["config_names_response"],
+ content=hub_responses_public["config_names_response"],
@@ -134 +135 @@ def test_compute(
- assert_content_is_equal(content, hub_reponses_public["parquet_and_info_response"])
+ assert_content_is_equal(content, hub_responses_public["parquet_and_info_response"])
@@ -446 +447 @@ def test_compute_splits_response_simple_csv_ok(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -454 +455 @@ def test_compute_splits_response_simple_csv_ok(
- hub_datasets = {"public": hub_reponses_public, "audio": hub_reponses_audio, "gated": hub_reponses_gated}
+ hub_datasets = {"public": hub_responses_public, "audio": hub_reponses_audio, "gated": hub_reponses_gated}
@@ -531 +532 @@ def test_compute_splits_response_simple_csv_error_2(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -538,2 +539,2 @@ def test_compute_splits_response_simple_csv_error_2(
- dataset = hub_reponses_public["name"]
- config_names_response = hub_reponses_public["config_names_response"]
+ dataset = hub_responses_public["name"]
+ config_names_response = hub_responses_public["config_names_response"]
@@ -559 +560 @@ def test_previous_step_error(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -562,2 +563,2 @@ def test_previous_step_error(
- dataset = hub_reponses_public["name"]
- config = hub_reponses_public["config_names_response"]["config_names"][0]["config"]
+ dataset = hub_responses_public["name"]
+ config = hub_responses_public["config_names_response"]["config_names"][0]["config"]
@@ -615,0 +617,35 @@ def test_get_writer_batch_size(ds_info: datasets.info.DatasetInfo, has_big_chunk
+
+
[email protected](
+ "max_operations_per_commit,use_parent_commit,expected_num_commits",
+ [(2, False, 1), (1, False, 2), (2, True, 1), (1, True, 2)],
+)
+def test_create_commits(
+ hub_public_legacy_configs: str, max_operations_per_commit: int, use_parent_commit: bool, expected_num_commits: int
+) -> None:
+ NUM_FILES = 2
+ repo_id = hub_public_legacy_configs
+ hf_api = HfApi(endpoint=CI_HUB_ENDPOINT, token=CI_USER_TOKEN)
+ if use_parent_commit:
+ target_dataset_info = hf_api.dataset_info(repo_id=repo_id, files_metadata=False)
+ parent_commit = target_dataset_info.sha
+ else:
+ parent_commit = None
+ directory = f".test_create_commits_{max_operations_per_commit}_{use_parent_commit}"
+ operations: List[CommitOperationAdd] = [
+ CommitOperationAdd(path_in_repo=f"{directory}/file{i}.txt", path_or_fileobj=f"content{i}".encode("UTF-8"))
+ for i in range(NUM_FILES)
+ ]
+ commit_infos = create_commits(
+ hf_api=hf_api,
+ repo_id=repo_id,
+ operations=operations,
+ commit_message="test",
+ max_operations_per_commit=max_operations_per_commit,
+ parent_commit=parent_commit,
+ )
+ assert len(commit_infos) == expected_num_commits
+ # check that the files were created
+ filenames = hf_api.list_repo_files(repo_id=repo_id, repo_type="dataset")
+ for i in range(NUM_FILES):
+ assert f"{directory}/file{i}.txt" in filenames
diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
index 56e28c7e..2e5f3215 100644
--- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
@@ -91 +91 @@ def test_compute_split_names_from_streaming_response(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -105 +105 @@ def test_compute_split_names_from_streaming_response(
- "public": hub_reponses_public,
+ "public": hub_responses_public,
diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py
index a94427f4..dc6aafe2 100644
--- a/services/worker/tests/job_runners/dataset/test_config_names.py
+++ b/services/worker/tests/job_runners/dataset/test_config_names.py
@@ -113 +113 @@ def test_compute_splits_response_simple_csv(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -127 +127 @@ def test_compute_splits_response_simple_csv(
- "public": hub_reponses_public,
+ "public": hub_responses_public,
diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
index fd2dd64b..9f255de8 100644
--- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
+++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
@@ -116 +116 @@ def test_number_rows(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -140 +140 @@ def test_number_rows(
- "public": hub_reponses_public,
+ "public": hub_responses_public,
diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
index e3fa31ac..8c8d3d74 100644
--- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
+++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
@@ -176 +176 @@ def test_compute(
- hub_reponses_public: HubDatasetTest,
+ hub_responses_public: HubDatasetTest,
@@ -185 +185 @@ def test_compute(
- hub_datasets = {"public": hub_reponses_public, "spawning_opt_in_out": hub_reponses_spawning_opt_in_out}
+ hub_datasets = {"public": hub_responses_public, "spawning_opt_in_out": hub_reponses_spawning_opt_in_out}
|
|
ca55a35ebd4c9b26af7ea19508bf773e3f953739
|
Andrea Francis Soria Jimenez
| 2023-06-20T18:43:07 |
Rename parent job runners (#1398)
|
diff --git a/services/worker/src/worker/job_runners/_cached_directory_job_runner.py b/services/worker/src/worker/job_runners/_job_runner_with_cache.py
similarity index 98%
rename from services/worker/src/worker/job_runners/_cached_directory_job_runner.py
rename to services/worker/src/worker/job_runners/_job_runner_with_cache.py
index 1ff02d16..1eca7d3b 100644
--- a/services/worker/src/worker/job_runners/_cached_directory_job_runner.py
+++ b/services/worker/src/worker/job_runners/_job_runner_with_cache.py
@@ -19 +19 @@ from worker.job_runner import JobRunner
-class CacheDirectoryJobRunner(JobRunner):
+class JobRunnerWithCache(JobRunner):
diff --git a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py b/services/worker/src/worker/job_runners/_job_runner_with_datasets_cache.py
similarity index 91%
rename from services/worker/src/worker/job_runners/_datasets_based_job_runner.py
rename to services/worker/src/worker/job_runners/_job_runner_with_datasets_cache.py
index f07c0da9..dc007a02 100644
--- a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
+++ b/services/worker/src/worker/job_runners/_job_runner_with_datasets_cache.py
@@ -13 +13 @@ from worker.config import AppConfig
-from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner
+from worker.job_runners._job_runner_with_cache import JobRunnerWithCache
@@ -16 +16 @@ from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRun
-class DatasetsBasedJobRunner(CacheDirectoryJobRunner):
+class JobRunnerWithDatasetsCache(JobRunnerWithCache):
diff --git a/services/worker/src/worker/job_runners/config/config_job_runner.py b/services/worker/src/worker/job_runners/config/config_job_runner.py
index 75723d36..b18d09c2 100644
--- a/services/worker/src/worker/job_runners/config/config_job_runner.py
+++ b/services/worker/src/worker/job_runners/config/config_job_runner.py
@@ -11 +11,3 @@ from worker.config import AppConfig
-from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.job_runners._job_runner_with_datasets_cache import (
+ JobRunnerWithDatasetsCache,
+)
@@ -30 +32 @@ class ConfigJobRunner(DatasetJobRunner):
-class ConfigCachedJobRunner(DatasetsBasedJobRunner, ConfigJobRunner):
+class ConfigJobRunnerWithDatasetsCache(JobRunnerWithDatasetsCache, ConfigJobRunner):
@@ -38 +40 @@ class ConfigCachedJobRunner(DatasetsBasedJobRunner, ConfigJobRunner):
- DatasetsBasedJobRunner.__init__(
+ JobRunnerWithDatasetsCache.__init__(
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index d75e9e55..2769bdd8 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -77 +77 @@ from worker.config import AppConfig, ParquetAndInfoConfig
-from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner
+from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatasetsCache
@@ -1074 +1074 @@ def compute_config_parquet_and_info_response(
-class ConfigParquetAndInfoJobRunner(ConfigCachedJobRunner):
+class ConfigParquetAndInfoJobRunner(ConfigJobRunnerWithDatasetsCache):
diff --git a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
index 8208dece..6c4ddc10 100644
--- a/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
+++ b/services/worker/src/worker/job_runners/config/split_names_from_streaming.py
@@ -20 +20 @@ from libcommon.exceptions import (
-from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner
+from worker.job_runners.config.config_job_runner import ConfigJobRunnerWithDatasetsCache
@@ -80 +80 @@ def compute_split_names_from_streaming_response(
-class ConfigSplitNamesFromStreamingJobRunner(ConfigCachedJobRunner):
+class ConfigSplitNamesFromStreamingJobRunner(ConfigJobRunnerWithDatasetsCache):
diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py
index a3a80354..faf74109 100644
--- a/services/worker/src/worker/job_runners/dataset/config_names.py
+++ b/services/worker/src/worker/job_runners/dataset/config_names.py
@@ -17 +17,3 @@ from libcommon.exceptions import (
-from worker.job_runners.dataset.dataset_job_runner import DatasetCachedJobRunner
+from worker.job_runners.dataset.dataset_job_runner import (
+ DatasetJobRunnerWithDatasetsCache,
+)
@@ -83 +85 @@ def compute_config_names_response(
-class DatasetConfigNamesJobRunner(DatasetCachedJobRunner):
+class DatasetConfigNamesJobRunner(DatasetJobRunnerWithDatasetsCache):
diff --git a/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py b/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py
index a21f9e14..693e01a3 100644
--- a/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py
+++ b/services/worker/src/worker/job_runners/dataset/dataset_job_runner.py
@@ -12 +12,3 @@ from worker.job_runner import JobRunner
-from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.job_runners._job_runner_with_datasets_cache import (
+ JobRunnerWithDatasetsCache,
+)
@@ -30 +32 @@ class DatasetJobRunner(JobRunner):
-class DatasetCachedJobRunner(DatasetsBasedJobRunner, DatasetJobRunner):
+class DatasetJobRunnerWithDatasetsCache(JobRunnerWithDatasetsCache, DatasetJobRunner):
@@ -38 +40 @@ class DatasetCachedJobRunner(DatasetsBasedJobRunner, DatasetJobRunner):
- DatasetsBasedJobRunner.__init__(
+ JobRunnerWithDatasetsCache.__init__(
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
index 43221906..a49cc8ea 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
@@ -29 +29 @@ from worker.config import AppConfig, FirstRowsConfig
-from worker.job_runners.split.split_job_runner import SplitCachedJobRunner
+from worker.job_runners.split.split_job_runner import SplitJobRunnerWithDatasetsCache
@@ -261 +261 @@ def compute_first_rows_response(
-class SplitFirstRowsFromStreamingJobRunner(SplitCachedJobRunner):
+class SplitFirstRowsFromStreamingJobRunner(SplitJobRunnerWithDatasetsCache):
diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
index 7544df77..dbb20277 100644
--- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
@@ -25 +25 @@ from worker.config import AppConfig, OptInOutUrlsScanConfig
-from worker.job_runners.split.split_job_runner import SplitCachedJobRunner
+from worker.job_runners.split.split_job_runner import SplitJobRunnerWithDatasetsCache
@@ -266 +266 @@ def compute_opt_in_out_urls_scan_response(
-class SplitOptInOutUrlsScanJobRunner(SplitCachedJobRunner):
+class SplitOptInOutUrlsScanJobRunner(SplitJobRunnerWithDatasetsCache):
diff --git a/services/worker/src/worker/job_runners/split/split_job_runner.py b/services/worker/src/worker/job_runners/split/split_job_runner.py
index 55a87b28..cca736bd 100644
--- a/services/worker/src/worker/job_runners/split/split_job_runner.py
+++ b/services/worker/src/worker/job_runners/split/split_job_runner.py
@@ -11 +11,4 @@ from worker.config import AppConfig
-from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.job_runners._job_runner_with_cache import JobRunnerWithCache
+from worker.job_runners._job_runner_with_datasets_cache import (
+ JobRunnerWithDatasetsCache,
+)
@@ -30 +33 @@ class SplitJobRunner(ConfigJobRunner):
-class SplitCachedJobRunner(DatasetsBasedJobRunner, SplitJobRunner):
+class SplitJobRunnerWithDatasetsCache(JobRunnerWithDatasetsCache, SplitJobRunner):
@@ -38 +41 @@ class SplitCachedJobRunner(DatasetsBasedJobRunner, SplitJobRunner):
- DatasetsBasedJobRunner.__init__(
+ JobRunnerWithDatasetsCache.__init__(
@@ -50,0 +54,23 @@ class SplitCachedJobRunner(DatasetsBasedJobRunner, SplitJobRunner):
+
+
+class SplitJobRunnerWithCache(JobRunnerWithCache, SplitJobRunner):
+ def __init__(
+ self,
+ job_info: JobInfo,
+ app_config: AppConfig,
+ processing_step: ProcessingStep,
+ cache_directory: Path,
+ ) -> None:
+ JobRunnerWithCache.__init__(
+ self,
+ job_info=job_info,
+ app_config=app_config,
+ processing_step=processing_step,
+ cache_directory=cache_directory,
+ )
+ SplitJobRunner.__init__(
+ self,
+ job_info=job_info,
+ app_config=app_config,
+ processing_step=processing_step,
+ )
diff --git a/services/worker/tests/job_runners/test__cached_directory_job_runner.py b/services/worker/tests/job_runners/test__job_runner_with_cache.py
similarity index 97%
rename from services/worker/tests/job_runners/test__cached_directory_job_runner.py
rename to services/worker/tests/job_runners/test__job_runner_with_cache.py
index a8a83a9b..db580ea6 100644
--- a/services/worker/tests/job_runners/test__cached_directory_job_runner.py
+++ b/services/worker/tests/job_runners/test__job_runner_with_cache.py
@@ -14 +14 @@ from worker.config import AppConfig
-from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner
+from worker.job_runners._job_runner_with_cache import JobRunnerWithCache
@@ -21 +21 @@ from ..fixtures.hub import get_default_config_split
-class DummyJobRunner(CacheDirectoryJobRunner):
+class DummyJobRunner(JobRunnerWithCache):
diff --git a/services/worker/tests/job_runners/test__datasets_based_job_runner.py b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
similarity index 96%
rename from services/worker/tests/job_runners/test__datasets_based_job_runner.py
rename to services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
index 207200d8..30ea2795 100644
--- a/services/worker/tests/job_runners/test__datasets_based_job_runner.py
+++ b/services/worker/tests/job_runners/test__job_runner_with_datasets_cache.py
@@ -14 +14,3 @@ from worker.config import AppConfig
-from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.job_runners._job_runner_with_datasets_cache import (
+ JobRunnerWithDatasetsCache,
+)
@@ -21 +23 @@ from ..fixtures.hub import get_default_config_split
-class DummyJobRunner(DatasetsBasedJobRunner):
+class DummyJobRunner(JobRunnerWithDatasetsCache):
|
|
9643c233e12765ed83373e8cf56a64336674196b
|
Andrea Francis Soria Jimenez
| 2023-06-20T15:19:59 |
Modify TTL index condition (#1405)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 06412c36..4b6dba13 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -224,0 +225,7 @@ class MigrationsCollector:
+ MigrationQueueDeleteTTLIndex(
+ version="202306201100",
+ description=(
+ "delete the TTL index on the 'finished_at' field in the queue database to update its TTL condition"
+ ),
+ field_name="finished_at",
+ ),
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index 2c1d062d..09cb6301 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -4 +3,0 @@
-import pytest
@@ -120 +118,0 @@ def test_metrics_deletion_migration(mongo_host: str) -> None:
[email protected](reason="temporaly disabled because of index removal")
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index b1c5e80c..27128ccb 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -26,0 +27 @@ from libcommon.constants import (
+ QUEUE_TTL_SECONDS,
@@ -137,0 +139,5 @@ class Job(Document):
+ {
+ "fields": ["finished_at"],
+ "expireAfterSeconds": QUEUE_TTL_SECONDS,
+ "partialFilterExpression": {"status": {"$in": [Status.SUCCESS, Status.ERROR, Status.CANCELLED]}},
+ },
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 5262acd1..02b9578e 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -379 +378,0 @@ def test_queue_get_zombies() -> None:
[email protected](reason="temporaly disabled because of index removal")
|
|
82f3b7393b0b57bd3cf5fc7599e1dad9141632e6
|
Andrea Francis Soria Jimenez
| 2023-06-20T14:59:07 |
Temporaly remove TTL index (#1403)
|
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index 09cb6301..2c1d062d 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -3,0 +4 @@
+import pytest
@@ -118,0 +120 @@ def test_metrics_deletion_migration(mongo_host: str) -> None:
[email protected](reason="temporaly disabled because of index removal")
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 6c233cab..b1c5e80c 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -27 +26,0 @@ from libcommon.constants import (
- QUEUE_TTL_SECONDS,
@@ -139 +137,0 @@ class Job(Document):
- {"fields": ["finished_at"], "expireAfterSeconds": QUEUE_TTL_SECONDS},
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 02b9578e..5262acd1 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -378,0 +379 @@ def test_queue_get_zombies() -> None:
[email protected](reason="temporaly disabled because of index removal")
|
|
bf7cc0b3b0c4315d299603217683b9fb1050de5e
|
Quentin Lhoest
| 2023-06-20T13:52:22 |
increase max job duration (#1395)
|
diff --git a/chart/values.yaml b/chart/values.yaml
index 5294d83c..0f99fb5d 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -131 +131 @@ worker:
- maxJobDurationSeconds: 1200
+ maxJobDurationSeconds: 2400
|
|
d017e7c79a4196535d38743b163293d2d296aa1c
|
Sylvain Lesage
| 2023-06-20T13:28:26 |
fix: 🐛 split the default value to get a list of strings (#1401)
|
diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py
index e1717543..1bbab28c 100644
--- a/libs/libcommon/src/libcommon/orchestrator.py
+++ b/libs/libcommon/src/libcommon/orchestrator.py
@@ -330 +330 @@ class DatasetBackfillPlan(Plan):
- self.error_codes_to_retry = list(ERROR_CODES_TO_RETRY)
+ self.error_codes_to_retry = ERROR_CODES_TO_RETRY.split(",")
|
|
7626cd0dfff2d5d70618893636a2e63e85b4ffa4
|
Sylvain Lesage
| 2023-06-20T12:36:44 |
Raise retryable error on hfhubhttperror (#1393)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 163a431d..06b172c5 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -135 +135 @@ backfill:
- error_codes_to_retry: ""
+ error_codes_to_retry: "CreateCommitError,LockedDatasetTimeoutError"
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
index fc0dd91e..cd41126f 100644
--- a/libs/libcommon/src/libcommon/constants.py
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -43,0 +44,2 @@ PARQUET_REVISION = "refs/convert/parquet"
+
+ERROR_CODES_TO_RETRY = "CreateCommitError,LockedDatasetTimeoutError"
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py
index 672111e1..f9de8f3f 100644
--- a/libs/libcommon/src/libcommon/exceptions.py
+++ b/libs/libcommon/src/libcommon/exceptions.py
@@ -76,0 +77 @@ CacheableErrorCode = Literal[
+ "CreateCommitError",
@@ -144,0 +146,7 @@ class ConfigNamesError(CacheableError):
+class CreateCommitError(CacheableError):
+ """Raised when a commit could not be created on the Hub."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.INTERNAL_SERVER_ERROR, "CreateCommitError", cause, False)
+
+
diff --git a/libs/libcommon/src/libcommon/orchestrator.py b/libs/libcommon/src/libcommon/orchestrator.py
index fb8518bd..e1717543 100644
--- a/libs/libcommon/src/libcommon/orchestrator.py
+++ b/libs/libcommon/src/libcommon/orchestrator.py
@@ -10,0 +11 @@ import pandas as pd
+from libcommon.constants import ERROR_CODES_TO_RETRY
@@ -327,0 +329,2 @@ class DatasetBackfillPlan(Plan):
+ if self.error_codes_to_retry is None:
+ self.error_codes_to_retry = list(ERROR_CODES_TO_RETRY)
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 22460dc6..d75e9e55 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -50,0 +51 @@ from libcommon.exceptions import (
+ CreateCommitError,
@@ -745,2 +746,2 @@ def create_commits(
- [`huggingface_hub.utils.HfHubHTTPError`]:
- If another commit happened after `parent_commit` or in between the commits.
+ [`libcommon.exceptions.CreateCommitError`]:
+ If one of the commits could not be created on the Hub.
@@ -748,0 +750 @@ def create_commits(
+ sleeps = [1, 1, 1, 10, 10, 10]
@@ -752,3 +754,3 @@ def create_commits(
- retry_create_commit = retry(on=[HfHubHTTPError], sleeps=[1, 1, 1, 10, 10, 10])(hf_api.create_commit)
- commit_infos.append(
- retry_create_commit(
+ retry_create_commit = retry(on=[HfHubHTTPError], sleeps=sleeps)(hf_api.create_commit)
+ try:
+ commit_info = retry_create_commit(
@@ -760 +762 @@ def create_commits(
- parent_commit=parent_commit if not commit_infos else commit_infos[-1].oid,
+ parent_commit=commit_infos[-1].oid if commit_infos else parent_commit,
@@ -762 +764,11 @@ def create_commits(
- )
+ except RuntimeError as e:
+ if e.__cause__ and isinstance(e.__cause__, HfHubHTTPError):
+ raise CreateCommitError(
+ message=(
+ f"Commit {commit_idx}/{len(offsets)} could not be created on the Hub (after"
+ f" {len(sleeps)} attempts)."
+ ),
+ cause=e.__cause__,
+ ) from e.__cause__
+ raise e
+ commit_infos.append(commit_info)
@@ -813,4 +825,2 @@ def commit_parquet_conversion(
- [`RuntimeError`]: If it fails after several retries.
- The retry fails can come from:
- [`huggingface_hub.utils.HfHubHTTPError`]:
- If another commit happened after `parent_commit` or in between the commits.
+ [`libcommon.exceptions.CreateCommitError`]:
+ If one of the commits could not be created on the Hub.
@@ -911,0 +922,2 @@ def compute_config_parquet_and_info_response(
+ - [`libcommon.exceptions.CreateCommitError`]:
+ If one of the commits could not be created on the Hub.
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index 7472e39b..b4869d67 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -322 +322 @@ class retry:
- raise RuntimeError(f"Give up after {attempt} attempts with {type(last_err)}") from last_err
+ raise RuntimeError(f"Give up after {attempt} attempts. The last one raised {type(last_err)}") from last_err
|
|
8d2f9ab14c3771e64d8341b879f36ef0ec96b7e3
|
Andrea Francis Soria Jimenez
| 2023-06-20T12:21:31 |
New parent job runner for cached data (#1388)
|
diff --git a/services/worker/src/worker/job_runners/_cached_directory_job_runner.py b/services/worker/src/worker/job_runners/_cached_directory_job_runner.py
new file mode 100644
index 00000000..1ff02d16
--- /dev/null
+++ b/services/worker/src/worker/job_runners/_cached_directory_job_runner.py
@@ -0,0 +1,63 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2023 The HuggingFace Authors.
+
+import json
+import random
+import re
+from hashlib import sha1
+from pathlib import Path
+from typing import Optional
+
+from libcommon.processing_graph import ProcessingStep
+from libcommon.storage import init_dir, remove_dir
+from libcommon.utils import JobInfo
+
+from worker.config import AppConfig
+from worker.job_runner import JobRunner
+
+
+class CacheDirectoryJobRunner(JobRunner):
+ """Base class for job runners that use a temporary cache directory."""
+
+ base_cache_directory: Path
+ cache_subdirectory: Optional[Path] = None
+
+ def __init__(
+ self,
+ job_info: JobInfo,
+ app_config: AppConfig,
+ processing_step: ProcessingStep,
+ cache_directory: Path,
+ ) -> None:
+ super().__init__(
+ job_info=job_info,
+ app_config=app_config,
+ processing_step=processing_step,
+ )
+ self.base_cache_directory = cache_directory
+
+ def get_cache_subdirectory(self, digits: int = 14) -> str:
+ random_str = f"{random.randrange(10**(digits - 1), 10**digits)}" # nosec B311
+ # TODO: Refactor, need a way to generate payload based only on provided params
+ payload = (
+ random_str,
+ self.get_job_type(),
+ self.job_info["params"]["dataset"],
+ self.job_info["params"]["config"],
+ self.job_info["params"]["split"],
+ )
+ hash_suffix = sha1(json.dumps(payload, sort_keys=True).encode(), usedforsecurity=False).hexdigest()[:8]
+ prefix = f"{random_str}-{self.get_job_type()}-{self.job_info['params']['dataset']}"[:64]
+ subdirectory = f"{prefix}-{hash_suffix}"
+ return "".join([c if re.match(r"[\w-]", c) else "-" for c in subdirectory])
+
+ def pre_compute(self) -> None:
+ new_directory = self.base_cache_directory / self.get_cache_subdirectory()
+ self.cache_subdirectory = Path(init_dir(new_directory))
+
+ def post_compute(self) -> None:
+ # empty the cache after the job to save storage space
+ previous_cache = self.cache_subdirectory
+ if previous_cache is not None:
+ remove_dir(previous_cache)
+ self.cache_subdirectory = None
diff --git a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
index 310b4bde..f07c0da9 100644
--- a/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
+++ b/services/worker/src/worker/job_runners/_datasets_based_job_runner.py
@@ -4 +3,0 @@
-import json
@@ -6,3 +4,0 @@ import logging
-import random
-import re
-from hashlib import sha1
@@ -14 +9,0 @@ from libcommon.processing_graph import ProcessingStep
-from libcommon.storage import init_dir, remove_dir
@@ -17,2 +12,2 @@ from libcommon.utils import JobInfo
-from worker.config import AppConfig, DatasetsBasedConfig
-from worker.job_runner import JobRunner
+from worker.config import AppConfig
+from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner
@@ -21 +16 @@ from worker.job_runner import JobRunner
-class DatasetsBasedJobRunner(JobRunner):
+class DatasetsBasedJobRunner(CacheDirectoryJobRunner):
@@ -24,8 +18,0 @@ class DatasetsBasedJobRunner(JobRunner):
- datasets_based_config: DatasetsBasedConfig
- base_datasets_cache: Path
-
- # the datasets library cache directories (for data, downloads, extraction, NOT for modules)
- # the job runner should have only one running job at the same time, then it should
- # be safe to use a global variable (and to set the datasets cache globally)
- datasets_cache: Optional[Path] = None
-
@@ -42,0 +30 @@ class DatasetsBasedJobRunner(JobRunner):
+ cache_directory=hf_datasets_cache,
@@ -44,17 +31,0 @@ class DatasetsBasedJobRunner(JobRunner):
- self.datasets_based_config = app_config.datasets_based
- self.base_datasets_cache = hf_datasets_cache
-
- def get_cache_subdirectory(self, digits: int = 14) -> str:
- random_str = f"{random.randrange(10**(digits - 1), 10**digits)}" # nosec B311
- # TODO: Refactor, need a way to generate payload based only on provided params
- payload = (
- random_str,
- self.get_job_type(),
- self.job_info["params"]["dataset"],
- self.job_info["params"]["config"],
- self.job_info["params"]["split"],
- )
- hash_suffix = sha1(json.dumps(payload, sort_keys=True).encode(), usedforsecurity=False).hexdigest()[:8]
- prefix = f"{random_str}-{self.get_job_type()}-{self.job_info['params']['dataset']}"[:64]
- subdirectory = f"{prefix}-{hash_suffix}"
- return "".join([c if re.match(r"[\w-]", c) else "-" for c in subdirectory])
@@ -62,3 +33,2 @@ class DatasetsBasedJobRunner(JobRunner):
- def set_datasets_cache(self, datasets_cache: Path) -> None:
- self.datasets_cache = Path(init_dir(datasets_cache))
- datasets.config.HF_DATASETS_CACHE = self.datasets_cache
+ def set_datasets_cache(self, cache_subdirectory: Optional[Path]) -> None:
+ datasets.config.HF_DATASETS_CACHE = cache_subdirectory
@@ -73,15 +42,0 @@ class DatasetsBasedJobRunner(JobRunner):
- def unset_datasets_cache(self) -> None:
- previous_datasets_cache = self.datasets_cache
- self.set_datasets_cache(self.base_datasets_cache)
- if previous_datasets_cache is not None and self.datasets_cache != previous_datasets_cache:
- remove_dir(previous_datasets_cache)
- logging.debug(f"temporary datasets data cache deleted: {previous_datasets_cache}")
- self.datasets_cache = None
-
- def set_cache(self) -> None:
- cache_subdirectory = self.get_cache_subdirectory()
- self.set_datasets_cache(self.base_datasets_cache / cache_subdirectory)
-
- def unset_cache(self) -> None:
- self.unset_datasets_cache()
-
@@ -89 +44,2 @@ class DatasetsBasedJobRunner(JobRunner):
- self.set_cache()
+ super().pre_compute()
+ self.set_datasets_cache(self.cache_subdirectory)
@@ -92,2 +48,2 @@ class DatasetsBasedJobRunner(JobRunner):
- # empty the cache after the job to save storage space
- self.unset_cache()
+ super().post_compute()
+ self.set_datasets_cache(self.base_cache_directory)
diff --git a/services/worker/tests/job_runners/test__datasets_based_worker.py b/services/worker/tests/job_runners/test__cached_directory_job_runner.py
similarity index 68%
rename from services/worker/tests/job_runners/test__datasets_based_worker.py
rename to services/worker/tests/job_runners/test__cached_directory_job_runner.py
index a57c7578..a8a83a9b 100644
--- a/services/worker/tests/job_runners/test__datasets_based_worker.py
+++ b/services/worker/tests/job_runners/test__cached_directory_job_runner.py
@@ -2 +2 @@
-# Copyright 2022 The HuggingFace Authors.
+# Copyright 2023 The HuggingFace Authors.
@@ -8 +7,0 @@ from typing import Callable, Optional
-import datasets.config
@@ -15 +14 @@ from worker.config import AppConfig
-from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.job_runners._cached_directory_job_runner import CacheDirectoryJobRunner
@@ -22 +21 @@ from ..fixtures.hub import get_default_config_split
-class DummyJobRunner(DatasetsBasedJobRunner):
+class DummyJobRunner(CacheDirectoryJobRunner):
@@ -26,2 +24,0 @@ class DummyJobRunner(DatasetsBasedJobRunner):
- # ^ borrowing the type, so that the processing step exists and the job runner can be initialized
- # refactoring libcommon.processing_graph might help avoiding this
@@ -75 +72 @@ def get_job_runner(
- hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ cache_directory=libraries_resource.hf_datasets_cache,
@@ -112,12 +109 @@ def test_get_cache_subdirectory(
-def test_set_and_unset_datasets_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
- dataset, config, split = get_default_config_split("dataset")
- job_runner = get_job_runner(dataset, config, split, app_config)
- base_path = job_runner.base_datasets_cache
- dummy_path = base_path / "dummy"
- job_runner.set_datasets_cache(dummy_path)
- assert_datasets_cache_path(path=dummy_path, exists=True)
- job_runner.unset_datasets_cache()
- assert_datasets_cache_path(path=base_path, exists=True)
-
-
-def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+def test_pre_compute_post_compute(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
@@ -126,5 +112,5 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner
- datasets_base_path = job_runner.base_datasets_cache
- job_runner.set_cache()
- assert str(datasets.config.HF_DATASETS_CACHE).startswith(str(datasets_base_path))
- assert "dummy-job-runner-user-dataset" in str(datasets.config.HF_DATASETS_CACHE)
- job_runner.unset_cache()
+ datasets_base_path = job_runner.base_cache_directory
+ job_runner.pre_compute()
+ datasets_cache_subdirectory = job_runner.cache_subdirectory
+ assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=True)
+ job_runner.post_compute()
@@ -131,0 +118 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner
+ assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=False)
@@ -134 +121,2 @@ def test_set_and_unset_cache(app_config: AppConfig, get_job_runner: GetJobRunner
-def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) -> None:
+def assert_datasets_cache_path(path: Optional[Path], exists: bool) -> None:
+ assert path is not None
@@ -136,3 +123,0 @@ def assert_datasets_cache_path(path: Path, exists: bool, equals: bool = True) ->
- assert (datasets.config.HF_DATASETS_CACHE == path) is equals
- assert (datasets.config.DOWNLOADED_DATASETS_PATH == path / datasets.config.DOWNLOADED_DATASETS_DIR) is equals
- assert (datasets.config.EXTRACTED_DATASETS_PATH == path / datasets.config.EXTRACTED_DATASETS_DIR) is equals
diff --git a/services/worker/tests/job_runners/test__datasets_based_job_runner.py b/services/worker/tests/job_runners/test__datasets_based_job_runner.py
new file mode 100644
index 00000000..207200d8
--- /dev/null
+++ b/services/worker/tests/job_runners/test__datasets_based_job_runner.py
@@ -0,0 +1,108 @@
+# SPDX-License-Identifier: Apache-2.0
+# Copyright 2022 The HuggingFace Authors.
+
+from pathlib import Path
+from typing import Callable, Optional
+
+import datasets.config
+import pytest
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.resources import CacheMongoResource, QueueMongoResource
+from libcommon.utils import Priority
+
+from worker.config import AppConfig
+from worker.job_runners._datasets_based_job_runner import DatasetsBasedJobRunner
+from worker.resources import LibrariesResource
+from worker.utils import CompleteJobResult
+
+from ..fixtures.hub import get_default_config_split
+
+
+class DummyJobRunner(DatasetsBasedJobRunner):
+ @staticmethod
+ def get_job_type() -> str:
+ return "dummy-job-runner"
+ # ^ borrowing the type, so that the processing step exists and the job runner can be initialized
+ # refactoring libcommon.processing_graph might help avoiding this
+
+ @staticmethod
+ def get_job_runner_version() -> int:
+ return 1
+
+ def compute(self) -> CompleteJobResult:
+ return CompleteJobResult({"col1": "a" * 200})
+
+
+GetJobRunner = Callable[[str, Optional[str], Optional[str], AppConfig], DummyJobRunner]
+
+
[email protected]
+def get_job_runner(
+ libraries_resource: LibrariesResource,
+ cache_mongo_resource: CacheMongoResource,
+ queue_mongo_resource: QueueMongoResource,
+) -> GetJobRunner:
+ def _get_job_runner(
+ dataset: str,
+ config: Optional[str],
+ split: Optional[str],
+ app_config: AppConfig,
+ ) -> DummyJobRunner:
+ processing_step_name = DummyJobRunner.get_job_type()
+ processing_graph = ProcessingGraph(
+ {
+ processing_step_name: {
+ "input_type": "dataset",
+ "job_runner_version": DummyJobRunner.get_job_runner_version(),
+ }
+ }
+ )
+ return DummyJobRunner(
+ job_info={
+ "type": DummyJobRunner.get_job_type(),
+ "params": {
+ "dataset": dataset,
+ "revision": "revision",
+ "config": config,
+ "split": split,
+ },
+ "job_id": "job_id",
+ "priority": Priority.NORMAL,
+ },
+ app_config=app_config,
+ processing_step=processing_graph.get_processing_step(processing_step_name),
+ hf_datasets_cache=libraries_resource.hf_datasets_cache,
+ )
+
+ return _get_job_runner
+
+
+def test_set_datasets_cache(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset, config, split = get_default_config_split("dataset")
+ job_runner = get_job_runner(dataset, config, split, app_config)
+ base_path = job_runner.base_cache_directory
+ dummy_path = base_path / "dummy"
+ job_runner.set_datasets_cache(dummy_path)
+ assert str(datasets.config.HF_DATASETS_CACHE).startswith(str(dummy_path))
+
+
+def test_pre_compute_post_compute(app_config: AppConfig, get_job_runner: GetJobRunner) -> None:
+ dataset, config, split = get_default_config_split("user/dataset")
+ job_runner = get_job_runner(dataset, config, split, app_config)
+ datasets_base_path = job_runner.base_cache_directory
+ job_runner.pre_compute()
+ datasets_cache_subdirectory = job_runner.cache_subdirectory
+ assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=True)
+ assert str(datasets.config.HF_DATASETS_CACHE).startswith(str(datasets_base_path))
+ assert "dummy-job-runner-user-dataset" in str(datasets.config.HF_DATASETS_CACHE)
+ job_runner.post_compute()
+ assert_datasets_cache_path(path=datasets_base_path, exists=True)
+ assert_datasets_cache_path(path=datasets_cache_subdirectory, exists=False, equals=False)
+
+
+def assert_datasets_cache_path(path: Optional[Path], exists: bool, equals: bool = True) -> None:
+ assert path is not None
+ assert path.exists() is exists
+ assert (datasets.config.HF_DATASETS_CACHE == path) is equals
+ assert (datasets.config.DOWNLOADED_DATASETS_PATH == path / datasets.config.DOWNLOADED_DATASETS_DIR) is equals
+ assert (datasets.config.EXTRACTED_DATASETS_PATH == path / datasets.config.EXTRACTED_DATASETS_DIR) is equals
|
|
b7f13c8ee5a1b774c75874095b368a79309a226f
|
Sylvain Lesage
| 2023-06-19T12:35:59 |
feat: 🎸 10x the size of supported images (#1392)
|
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index bc187b6f..7472e39b 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -37 +37 @@ from libcommon.utils import orjson_dumps
-MAX_IMAGE_PIXELS = 1_000_000_000
+MAX_IMAGE_PIXELS = 10_000_000_000
|
|
cb563d9451eb6898c7be87a34bd2f0262fd4330a
|
Sylvain Lesage
| 2023-06-19T12:12:18 |
Rename dev to staging, and use staging mongodb cluster (#1383)
|
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml
index 878e3e3b..d1394601 100644
--- a/.github/workflows/cd.yml
+++ b/.github/workflows/cd.yml
@@ -79,2 +79,2 @@ jobs:
- - name: Lint chart with dev values
- run: helm lint --values env/dev.yaml
+ - name: Lint chart with staging values
+ run: helm lint --values env/staging.yaml
@@ -86 +86 @@ jobs:
- deploy-dev-and-prod:
+ deploy-staging-and-prod:
diff --git a/.github/workflows/chart-pr.yml b/.github/workflows/chart-pr.yml
index cd5789bc..93880b23 100644
--- a/.github/workflows/chart-pr.yml
+++ b/.github/workflows/chart-pr.yml
@@ -22,2 +22,2 @@ jobs:
- - name: Lint chart with dev values
- run: helm lint --values env/dev.yaml
+ - name: Lint chart with staging values
+ run: helm lint --values env/staging.yaml
diff --git a/chart/Makefile b/chart/Makefile
index 555a89f3..57716f53 100644
--- a/chart/Makefile
+++ b/chart/Makefile
@@ -1,2 +0,0 @@
-K8S_NAMESPACE := datasets-server
-
@@ -7,36 +4,0 @@ init:
-.PHONY: uninstall
-uninstall:
- helm uninstall $(ENV) -n $(K8S_NAMESPACE)
-
-.PHONY: diff
-diff:
- helm diff upgrade --install $(ENV) . --values env/$(ENV).yaml -n $(K8S_NAMESPACE)
-
-.PHONY: upgrade
-upgrade:
- helm upgrade --install $(ENV) . --values env/$(ENV).yaml -n $(K8S_NAMESPACE)
-
-.PHONY: diff-dev
-diff-dev:
- @make diff ENV=dev
-
-.PHONY: uninstall-dev
-uninstall-dev:
- @make uninstall ENV=dev
-
-.PHONY: upgrade-dev
-upgrade-dev:
- @make upgrade ENV=dev
-
-.PHONY: diff-prod
-diff-prod:
- @make diff ENV=prod
-
-.PHONY: uninstall-prod
-uninstall-prod:
- @make uninstall ENV=prod
-
-.PHONY: upgrade-prod
-upgrade-prod:
- @make upgrade ENV=prod
-
diff --git a/chart/README.md b/chart/README.md
index 26865e85..aa378210 100644
--- a/chart/README.md
+++ b/chart/README.md
@@ -16,18 +16 @@ Note that this Helm chart is used to manage the deployment of the `datasets-serv
-To deploy to the `hub-ephemeral` Kubernetes cluster, ensure to first:
-
-- install the tools (aws, kubectl, helm)
-- authenticate with AWS
-- select the `hub-ephemeral` cluster
-
-Dry run:
-
-```shell
-make init
-make diff-dev
-```
-
-Deploy:
-
-```shell
-make upgrade-dev
-```
+To deploy, go to https://cd.internal.huggingface.tech/applications.
diff --git a/chart/env/dev.yaml b/chart/env/staging.yaml
similarity index 91%
rename from chart/env/dev.yaml
rename to chart/env/staging.yaml
index b66d3f88..54263cb8 100644
--- a/chart/env/dev.yaml
+++ b/chart/env/staging.yaml
@@ -7,4 +6,0 @@ global:
- imageRegistry: ""
- imagePullSecrets: []
- privateHub:
- enabled: false
@@ -12,0 +9 @@ global:
+ # ^ the domain contains "dev", not "staging". We don't change for now.
@@ -54 +51 @@ secrets:
- fromSecret: false
+ fromSecret: true
@@ -78 +75 @@ mongodb:
- enabled: true
+ enabled: false
@@ -87,0 +85,3 @@ log:
+firstRows:
+ maxBytes: "200_000"
+
@@ -104,0 +105,4 @@ cacheMaintenance:
+ log:
+ level: "debug"
+ backfill:
+ error_codes_to_retry: ""
@@ -117,2 +121,2 @@ metricsCollector:
- schedule: "*/5 * * * *"
- # every five minutes
+ schedule: "*/2 * * * *"
+ # every two minutes
@@ -122 +126 @@ metricsCollector:
- cpu: 0
+ cpu: 1
@@ -124 +128,2 @@ metricsCollector:
- cpu: 0
+ cpu: 1
+ memory: "512Mi"
@@ -163,2 +168,2 @@ ingress:
- alb.ingress.kubernetes.io/load-balancer-name: "hub-datasets-server-dev"
- alb.ingress.kubernetes.io/tags: "Env=dev,Project=datasets-server,Terraform=true"
+ alb.ingress.kubernetes.io/load-balancer-name: "hub-datasets-server-staging"
+ alb.ingress.kubernetes.io/tags: "Env=staging,Project=datasets-server,Terraform=true"
|
|
170b1c6be8bbfc0347056ac5b89e9e6b816ecc6c
|
Sylvain Lesage
| 2023-06-19T11:21:39 |
fix: 🐛 support bigger images (#1387)
|
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 8d817352..5f59e6b5 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -4695,2 +4694,0 @@ files = [
- {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:9c9b14fbb73ec4cb0f209722a1489020fd8614c92ae22589f2309c48cefdf21f"},
- {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_x86_64.whl", hash = "sha256:6a54539bd076746f69ae8bef7282f981674fe4dbf59c3a84c4af86ae6bae9d5c"},
@@ -5012,0 +5011,12 @@ test = ["black (>=22.3.0,<23.0.0)", "coverage (>=5.2,<6.0)", "isort (>=5.0.6,<6.
+[[package]]
+name = "types-pillow"
+version = "9.5.0.4"
+description = "Typing stubs for Pillow"
+category = "dev"
+optional = false
+python-versions = "*"
+files = [
+ {file = "types-Pillow-9.5.0.4.tar.gz", hash = "sha256:f1b6af47abd151847ee25911ffeba784899bc7dc7f9eba8ca6a5aac522b012ef"},
+ {file = "types_Pillow-9.5.0.4-py3-none-any.whl", hash = "sha256:69427d9fa4320ff6e30f00fb9c0dd71185dc0a16de4757774220104759483466"},
+]
+
@@ -5584 +5594 @@ python-versions = "3.9.15"
-content-hash = "ad3c5c34e9ea75e4cb4394930bb35c5afdff65fe03c5f51b8cbebbea37f62f1d"
+content-hash = "109af95b92d54671ee4da9ddd322e3b03b5f9882ec38c9ace064afcbe37b5a2b"
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index 9fabc531..f5feb938 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -22,0 +23 @@ lxml = "^4.9.2"
+mirakuru = "^2.4.2"
@@ -28 +28,0 @@ pdf2image = "^1.16.2"
-pyarrow = "^11.0.0"
@@ -29,0 +30 @@ py7zr = "^0.20.4"
+pyarrow = "^11.0.0"
@@ -47 +47,0 @@ wget = "^3.2"
-mirakuru = "^2.4.2"
@@ -59,0 +60 @@ pytest-cov = "^2.12.1"
+types-pillow = "^9.5.0.4"
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index 14d209a2..bc187b6f 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -24,0 +25 @@ from typing import (
+import PIL
@@ -35,0 +37,3 @@ from libcommon.utils import orjson_dumps
+MAX_IMAGE_PIXELS = 1_000_000_000
+# ^ see https://pillow.readthedocs.io/en/stable/reference/Image.html#PIL.Image.MAX_IMAGE_PIXELS
+
@@ -333,0 +338 @@ def get_rows(
+ PIL.Image.MAX_IMAGE_PIXELS = MAX_IMAGE_PIXELS
|
|
aa4a6c6442ce9f5f40ed664a0afb1f0c401be72d
|
Bas Krahmer
| 2023-06-19T10:39:35 |
Add docker internal to extra_hosts (#1390)
|
diff --git a/tools/docker-compose-dev-base.yml b/tools/docker-compose-dev-base.yml
index 90a1e9cb..b9ca8e9e 100644
--- a/tools/docker-compose-dev-base.yml
+++ b/tools/docker-compose-dev-base.yml
@@ -27,0 +28,2 @@ services:
+ extra_hosts:
+ - "host.docker.internal:host-gateway"
|
|
de349b5df053b8ea1dcc1ed5e626b8de3a223f07
|
Albert Villanova del Moral
| 2023-06-19T08:55:12 |
Fix typo in erro rmessage (#1391)
|
diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py
index b1045c31..a3a80354 100644
--- a/services/worker/src/worker/job_runners/dataset/config_names.py
+++ b/services/worker/src/worker/job_runners/dataset/config_names.py
@@ -77 +77 @@ def compute_config_names_response(
- f"The maximun number of configs allowed is {max_number}, dataset has {number_of_configs} configs."
+ f"The maximum number of configs allowed is {max_number}, dataset has {number_of_configs} configs."
|
|
03d633edfb5f5852c5b2f70e685860509ca9657d
|
Bas Krahmer
| 2023-06-19T08:51:24 |
Fix closing brackets and GH action link (#1389)
|
diff --git a/DEVELOPER_GUIDE.md b/DEVELOPER_GUIDE.md
index d446e6d0..e2101dc5 100644
--- a/DEVELOPER_GUIDE.md
+++ b/DEVELOPER_GUIDE.md
@@ -54 +54 @@ If you use VSCode, it might be useful to use the ["monorepo" workspace](./.vscod
-The repository is structured as a monorepo, with Python libraries and applications in [jobs](./jobs)), [libs](./libs) and [services](./services):
+The repository is structured as a monorepo, with Python libraries and applications in [jobs](./jobs), [libs](./libs) and [services](./services):
@@ -100 +100 @@ The following environments contain all the modules: reverse proxy, API server, a
-The CI checks the quality of the code through a [GitHub action](./.github/workflows/quality.yml). To manually format the code of a job, library, service or worker:
+The CI checks the quality of the code through a [GitHub action](./.github/workflows/_quality-python.yml). To manually format the code of a job, library, service or worker:
diff --git a/services/reverse-proxy/README.md b/services/reverse-proxy/README.md
index 2801bf44..d322fda2 100644
--- a/services/reverse-proxy/README.md
+++ b/services/reverse-proxy/README.md
@@ -11 +11 @@ The reverse proxy uses nginx:
-- it serves the static assets directly (the API also serves them if required, but it's unnecessary to go through starlette for this, and it generates errors in Safari, see [1](https://github.com/encode/starlette/issues/950) and [2](https://developer.apple.com/library/archive/documentation/AppleApplications/Reference/SafariWebContent/CreatingVideoforSafarioniPhone/CreatingVideoforSafarioniPhone.html#//apple_ref/doc/uid/TP40006514-SW6)
+- it serves the static assets directly (the API also serves them if required, but it's unnecessary to go through starlette for this, and it generates errors in Safari, see [1](https://github.com/encode/starlette/issues/950) and [2](https://developer.apple.com/library/archive/documentation/AppleApplications/Reference/SafariWebContent/CreatingVideoforSafarioniPhone/CreatingVideoforSafarioniPhone.html#//apple_ref/doc/uid/TP40006514-SW6))
|
|
dbd15643a377d9747c21000c364e231cfd6492df
|
Steven Liu
| 2023-06-16T16:10:04 |
[docs] Improvements (#1376)
|
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml
index 0767f5ab..b4dbd129 100644
--- a/docs/source/_toctree.yml
+++ b/docs/source/_toctree.yml
@@ -6,0 +7,2 @@
+ - local: analyze_data
+ title: Analyze a dataset on the Hub
diff --git a/docs/source/analyze_data.mdx b/docs/source/analyze_data.mdx
new file mode 100644
index 00000000..e6c1a673
--- /dev/null
+++ b/docs/source/analyze_data.mdx
@@ -0,0 +1,63 @@
+# Analyze a dataset on the Hub
+
+[[open-in-colab]]
+
+In the Quickstart, you were introduced to various endpoints for interacting with datasets on the Hub. One of the most useful ones is the `/parquet` endpoint, which allows you to get a dataset stored on the Hub and analyze it. This is a great way to explore the dataset, and get a better understanding of it's contents.
+
+To demonstrate, this guide will show you an end-to-end example of how to retrieve a dataset from the Hub and do some basic data analysis with the Pandas library.
+
+## Get a dataset
+
+The [Hub](https://huggingface.co/datasets) is home to more than 40,000 datasets across a wide variety of tasks, sizes, and languages. For this example, you'll use the [`codeparrot/codecomplex`](https://huggingface.co/datasets/codeparrot/codecomplex) dataset, but feel free to explore and find another dataset that interests you! The dataset contains Java code from programming competitions, and the time complexity of the code is labeled by a group of algorithm experts.
+
+Let's say you're interested in the average length of the submitted code as it relates to the time complexity. Here's how you can get started.
+
+Use the `/parquet` endpoint to convert the dataset to a Parquet file and return the URL to it:
+
+```py
+import requests
+API_URL = "https://datasets-server.huggingface.co/parquet?dataset=codeparrot/codecomplex"
+def query():
+ response = requests.get(API_URL)
+ return response.json()
+data = query()
+print(data)
+{'parquet_files':
+ [
+ {'dataset': 'codeparrot/codecomplex', 'config': 'codeparrot--codecomplex', 'split': 'train', 'url': 'https://huggingface.co/datasets/codeparrot/codecomplex/resolve/refs%2Fconvert%2Fparquet/codeparrot--codecomplex/json-train.parquet', 'filename': 'json-train.parquet', 'size': 4115908}
+ ],
+ 'pending': [], 'failed': []
+}
+```
+
+## Read dataset with Pandas
+
+With the URL, you can read the Parquet file into a Pandas DataFrame:
+
+```py
+import pandas as pd
+
+url = "https://huggingface.co/datasets/codeparrot/codecomplex/resolve/refs%2Fconvert%2Fparquet/codeparrot--codecomplex/json-train.parquet"
+df = pd.read_parquet(url)
+df.head(5)
+```
+
+| src | complexity | problem | from |
+|--------------------------------------------------:|-----------:|--------------------------------:|-----------:|
+| import java.io.*;\nimport java.math.BigInteger... | quadratic | 1179_B. Tolik and His Uncle | CODEFORCES |
+| import java.util.Scanner;\n \npublic class pil... | linear | 1197_B. Pillars | CODEFORCES |
+| import java.io.BufferedReader;\nimport java.io... | linear | 1059_C. Sequence Transformation | CODEFORCES |
+| import java.util.*;\n\nimport java.io.*;\npubl... | linear | 1011_A. Stages | CODEFORCES |
+| import java.io.OutputStream;\nimport java.io.I... | linear | 1190_C. Tokitsukaze and Duel | CODEFORCES |
+
+## Calculate mean code length by time complexity
+
+Pandas is a powerful library for data analysis; group the dataset by time complexity, apply a function to calculate the average length of the code snippet, and plot the results:
+
+```py
+df.groupby('complexity')['src'].apply(lambda x: x.str.len().mean()).sort_values(ascending=False).plot.barh(color="orange")
+```
+
+<div class="flex justify-center">
+ <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets-server/codecomplex.png"/>
+</div>
\ No newline at end of file
diff --git a/docs/source/quick_start.mdx b/docs/source/quick_start.mdx
index d784bb20..b798b8e8 100644
--- a/docs/source/quick_start.mdx
+++ b/docs/source/quick_start.mdx
@@ -2,0 +3,2 @@
+[[open-in-colab]]
+
@@ -89,0 +92,7 @@ curl https://datasets-server.huggingface.co/is-valid?dataset=rotten_tomatoes \
+You'll see the following error if you're trying to access a gated dataset without providing your user token:
+
+```py
+print(data)
+{'error': 'The dataset does not exist, or is not accessible without authentication (private or gated). Please check the spelling of the dataset name or retry with authentication.'}
+```
+
@@ -130,0 +140,16 @@ curl https://datasets-server.huggingface.co/valid \
+This returns a list of all the datasets that load without an error:
+
+```py
+print(data)
+{
+ "valid": [
+ "0n1xus/codexglue",
+ "0n1xus/pytorrent-standalone",
+ "0x7194633/rupile",
+ "51la5/keyword-extraction",
+ ...,
+ ...,
+ ]
+}
+```
+
@@ -169,0 +195,7 @@ curl https://datasets-server.huggingface.co/is-valid?dataset=rotten_tomatoes \
+This returns whether the `valid` key is `true` or `false`:
+
+```py
+print(data)
+{'valid': True}
+```
+
@@ -210,0 +243,15 @@ curl https://datasets-server.huggingface.co/splits?dataset=rotten_tomatoes \
+This returns the available configuration and splits in the dataset:
+
+```py
+print(data)
+{'splits':
+ [
+ {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'train'},
+ {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'validation'},
+ {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'test'}
+ ],
+ 'pending': [],
+ 'failed': []
+}
+```
+
@@ -251,0 +299,20 @@ curl https://datasets-server.huggingface.co/first-rows?dataset=rotten_tomatoes&c
+This returns the first 100 rows of the dataset:
+
+```py
+print(data)
+{'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'train',
+ 'features':
+ [
+ {'feature_idx': 0, 'name': 'text', 'type': {'dtype': 'string', '_type': 'Value'}},
+ {'feature_idx': 1, 'name': 'label', 'type': {'names': ['neg', 'pos'], '_type': 'ClassLabel'}}
+ ],
+ 'rows':
+ [
+ {'row_idx': 0, 'row': {'text': 'the rock is destined to be the 21st century\'s new " conan " and that he\'s going to make a splash even greater than arnold schwarzenegger , jean-claud van damme or steven segal .', 'label': 1}, 'truncated_cells': []},
+ {'row_idx': 1, 'row': {'text': 'the gorgeously elaborate continuation of " the lord of the rings " trilogy is so huge that a column of words cannot adequately describe co-writer/director peter jackson\'s expanded vision of j . r . r . tolkien\'s middle-earth .', 'label': 1}, 'truncated_cells': []}
+ ...,
+ ...,
+ ],
+}
+```
+
@@ -296,0 +364,21 @@ You can download slices of 100 rows maximum at a time.
+The response looks like:
+
+```py
+print(data)
+{'features':
+ [
+ {'feature_idx': 0, 'name': 'text', 'type': {'dtype': 'string', '_type': 'Value'}},
+ {'feature_idx': 1, 'name': 'label', 'type': {'names': ['neg', 'pos'], '_type': 'ClassLabel'}}],
+ 'rows':
+ [
+ {'row_idx': 150, 'row': {'text': 'enormously likable , partly because it is aware of its own grasp of the absurd .', 'label': 1}, 'truncated_cells': []},
+ {'row_idx': 151, 'row': {'text': "here's a british flick gleefully unconcerned with plausibility , yet just as determined to entertain you .", 'label': 1}, 'truncated_cells': []},
+ {'row_idx': 152, 'row': {'text': "it's an old story , but a lively script , sharp acting and partially animated interludes make just a kiss seem minty fresh .", 'label': 1}, 'truncated_cells': []},
+ {'row_idx': 153, 'row': {'text': 'must be seen to be believed .', 'label': 1}, 'truncated_cells': []},
+ {'row_idx': 154, 'row': {'text': "ray liotta and jason patric do some of their best work in their underwritten roles , but don't be fooled : nobody deserves any prizes here .", 'label': 1}, 'truncated_cells': []},
+ ...,
+ ...,
+ ]
+}
+```
+
@@ -336,0 +425,15 @@ curl https://datasets-server.huggingface.co/parquet?dataset=rotten_tomatoes \
+
+This returns a URL to the Parquet file for each split:
+
+```py
+print(data)
+{'parquet_files':
+ [
+ {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'test', 'url': 'https://huggingface.co/datasets/rotten_tomatoes/resolve/refs%2Fconvert%2Fparquet/default/rotten_tomatoes-test.parquet', 'filename': 'rotten_tomatoes-test.parquet', 'size': 92206},
+ {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'train', 'url': 'https://huggingface.co/datasets/rotten_tomatoes/resolve/refs%2Fconvert%2Fparquet/default/rotten_tomatoes-train.parquet', 'filename': 'rotten_tomatoes-train.parquet', 'size': 698845},
+ {'dataset': 'rotten_tomatoes', 'config': 'default', 'split': 'validation', 'url': 'https://huggingface.co/datasets/rotten_tomatoes/resolve/refs%2Fconvert%2Fparquet/default/rotten_tomatoes-validation.parquet', 'filename': 'rotten_tomatoes-validation.parquet', 'size': 90001}
+ ],
+ 'pending': [],
+ 'failed': []
+}
+```
\ No newline at end of file
|
|
a3da0bfc18e7990d36ddb6b0820868857303cdc9
|
Quentin Lhoest
| 2023-06-16T12:59:01 |
Multi commit Parquet copies (#1355)
|
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py
index f7048d5a..672111e1 100644
--- a/libs/libcommon/src/libcommon/exceptions.py
+++ b/libs/libcommon/src/libcommon/exceptions.py
@@ -101,0 +102 @@ CacheableErrorCode = Literal[
+ "LockedDatasetTimeoutError",
@@ -239,0 +241,7 @@ class DatasetWithTooManyParquetFilesError(CacheableError):
+class LockedDatasetTimeoutError(CacheableError):
+ """Raised when a dataset is locked by another job."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "LockedDatasetTimeoutError", cause, True)
+
+
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 14f54073..6c233cab 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -4,0 +5 @@ import contextlib
+import json
@@ -216 +217 @@ class Lock(Document):
-class lock:
+class lock(contextlib.AbstractContextManager["lock"]):
@@ -241 +242,3 @@ class lock:
- def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = (0.05, 0.05, 0.05, 1, 1, 1, 5)) -> None:
+ _default_sleeps = (0.05, 0.05, 0.05, 1, 1, 1, 5)
+
+ def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> None:
@@ -276,0 +280,14 @@ class lock:
+ @classmethod
+ def git_branch(cls, dataset: str, branch: str, job_id: str, sleeps: Sequence[float] = _default_sleeps) -> "lock":
+ """
+ Lock a git branch of a dataset on the hub for read/write
+
+ Args:
+ dataset (`str`): the dataset repository
+ branch (`str`): the branch to lock
+ job_id (`str`): the current job id that holds the lock
+ sleeps (`Sequence[float]`): the time in seconds to sleep between each attempt to acquire the lock
+ """
+ key = json.dumps({"dataset": dataset, "branch": branch})
+ return cls(key=key, job_id=job_id, sleeps=sleeps)
+
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index d9d10395..22460dc6 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -40,2 +40,2 @@ from huggingface_hub._commit_api import (
-from huggingface_hub.hf_api import DatasetInfo, HfApi, RepoFile
-from huggingface_hub.utils._errors import RepositoryNotFoundError
+from huggingface_hub.hf_api import CommitInfo, DatasetInfo, HfApi, RepoFile
+from huggingface_hub.utils._errors import HfHubHTTPError, RepositoryNotFoundError
@@ -64,0 +65 @@ from libcommon.exceptions import (
+ LockedDatasetTimeoutError,
@@ -68,0 +70 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.queue import lock
@@ -75 +77 @@ from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner
-from worker.utils import CompleteJobResult
+from worker.utils import CompleteJobResult, retry
@@ -93,0 +96 @@ MAX_FILES_PER_DIRECTORY = 10_000 # hf hub limitation
+MAX_OPERATIONS_PER_COMMIT = 500
@@ -680,0 +684,166 @@ def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]:
+def create_commits(
+ hf_api: HfApi,
+ repo_id: str,
+ operations: List[CommitOperation],
+ *,
+ commit_message: str,
+ revision: Optional[str] = None,
+ parent_commit: Optional[str] = None,
+ max_operations_per_commit: int = MAX_OPERATIONS_PER_COMMIT,
+) -> List[CommitInfo]:
+ """
+ Creates one or several commits in the given dataset repo, deleting & uploading files as needed.
+
+ Args:
+ hf_api (`huggingface_hub.HfApi`):
+ The HfApi to use to commit the operations.
+ repo_id (`str`):
+ The repository in which the commit will be created, for example:
+ `"username/my_dataset"`
+ operations (`Iterable` of [`huggingface_hub.hf_api.CommitOperation`]):
+ An iterable of operations to include in the commit, either:
+
+ - [`huggingface_hub.hf_api.CommitOperationAdd`] to upload a file
+ - [`huggingface_hub.hf_api.CommitOperationDelete`] to delete a file
+ - [`huggingface_hub.hf_api.CommitOperationCopy`] to copy a file
+ commit_message (`str`):
+ The summary (first line) of the commit that will be created.
+ commit_description (`str`, *optional*):
+ The description of the commit that will be created
+ token (`str`, *optional*):
+ Authentication token, obtained with `HfApi.login` method. Will
+ default to the stored token.
+ repo_type (`str`, *optional*):
+ Set to `"dataset"` or `"space"` if uploading to a dataset or
+ space, `None` or `"model"` if uploading to a model. Default is
+ `None`.
+ revision (`str`, *optional*):
+ The git revision to commit from. Defaults to the head of the `"main"` branch.
+ parent_commit (`str`, *optional*):
+ The OID / SHA of the parent commit, as a hexadecimal string.
+ Shorthands (7 first characters) are also supported. If specified and `create_pr` is `False`,
+ the commit will fail if `revision` does not point to `parent_commit`. If specified and `create_pr`
+ is `True`, the pull request will be created from `parent_commit`. Specifying `parent_commit`
+ ensures the repo has not changed before committing the changes, and can be especially useful
+ if the repo is updated / committed to concurrently.
+ max_operations_per_commit (`int`, *optional*):
+ The ma number of operations per commit, to avoid time out errors from the Hub. Defaults to 500.
+ Returns:
+ [`List[huggingface_hub.CommitInfo]`]:
+ List of [`CommitInfo`] containing information about the newly created commit (commit hash, commit
+ url, pr url, commit message,...).
+ Raises:
+ [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError)
+ If commit message is empty.
+ [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError)
+ If parent commit is not a valid commit OID.
+ [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError)
+ If the Hub API returns an HTTP 400 error (bad request)
+ [`huggingface_hub.utils.RepositoryNotFoundError`]:
+ If repository is not found (error 404): wrong repo_id/repo_type, private
+ but not authenticated or repo does not exist.
+ [`huggingface_hub.utils.HfHubHTTPError`]:
+ If another commit happened after `parent_commit` or in between the commits.
+ """
+ commit_infos: List[CommitInfo] = []
+ offsets = range(0, len(operations), max_operations_per_commit)
+ for commit_idx, offset in enumerate(offsets):
+ batch_msg = f" (step {commit_idx + 1} of {len(offsets)})" if len(offsets) > 1 else ""
+ retry_create_commit = retry(on=[HfHubHTTPError], sleeps=[1, 1, 1, 10, 10, 10])(hf_api.create_commit)
+ commit_infos.append(
+ retry_create_commit(
+ repo_id=repo_id,
+ repo_type=DATASET_TYPE,
+ revision=revision,
+ operations=operations[offset : offset + max_operations_per_commit], # noqa: E203
+ commit_message=commit_message + batch_msg,
+ parent_commit=parent_commit if not commit_infos else commit_infos[-1].oid,
+ )
+ )
+ return commit_infos
+
+
+def commit_parquet_conversion(
+ hf_api: HfApi,
+ committer_hf_api: HfApi,
+ dataset: str,
+ config: str,
+ config_names: Set[str],
+ parquet_operations: List[CommitOperation],
+ commit_message: str,
+ target_revision: Optional[str],
+) -> List[CommitInfo]:
+ """
+ Creates one or several commits in the given dataset repo, deleting & uploading files as needed.
+
+ Args:
+ hf_api (`huggingface_hub.HfApi`):
+ The HfApi to get the dataset info.
+ committer_hf_api (`huggingface_hub.HfApi`):
+ The HfApi to use to commit the operations.
+ dataset (`str`):
+ The dataset in which the commit will be created, for example:
+ `"username/my_dataset"`
+ config (`str`):
+ The dataset configuration.
+ config_names (`List[str]`):
+ The list of all the configurations of this dataset. This is used to clean
+ the other fiels and directories in the repo, if any.
+ parquet_operations (`List[huggingface_hub.hf_api.CommitOperation]`):
+ List of commit operation for the parquet conversion. It could be
+ file additions or file copies for example.
+ commit_message (`str`):
+ The summary (first line) of the commit that will be created.
+ target_revision (`str`, *optional*):
+ The git revision to commit from. Defaults to the head of the `"main"` branch.
+ Returns:
+ [`List[huggingface_hub.CommitInfo]`]:
+ List of [`CommitInfo`] containing information about the newly created commit (commit hash, commit
+ url, pr url, commit message,...).
+ Raises:
+ [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError)
+ If commit message is empty.
+ [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError)
+ If parent commit is not a valid commit OID.
+ [`ValueError`](https://docs.python.org/3/library/exceptions.html#ValueError)
+ If the Hub API returns an HTTP 400 error (bad request)
+ [`huggingface_hub.utils.RepositoryNotFoundError`]:
+ If repository is not found (error 404): wrong repo_id/repo_type, private
+ but not authenticated or repo does not exist.
+ [`RuntimeError`]: If it fails after several retries.
+ The retry fails can come from:
+ [`huggingface_hub.utils.HfHubHTTPError`]:
+ If another commit happened after `parent_commit` or in between the commits.
+ """
+ target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False)
+ # - get repo parquet files
+ all_repo_files: Set[str] = {f.rfilename for f in target_dataset_info.siblings}
+ repo_parquet_files: Set[str] = {file for file in all_repo_files if file.endswith(".parquet")}
+ # - get files that will be preserved in repo: files belonging to other configs and .gitattributes
+ # we exclude files of current config because otherwise outdated files might be preserved
+ files_to_ignore: Set[str] = {
+ file
+ for other_config in config_names.difference({config})
+ for file in repo_parquet_files
+ if file.startswith(f"{other_config}/")
+ }.union({".gitattributes"})
+ # - get files to be deleted - all files except for:
+ # - parquet files obtained for current config at this processing step,
+ # - parquet files belonging to other existing configs
+ # - .gitattributes
+ files_to_add = [operation.path_in_repo for operation in parquet_operations]
+ files_to_delete = all_repo_files - set(files_to_add).union(files_to_ignore)
+ delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete]
+ logging.debug(f"{delete_operations=}")
+
+ operations = delete_operations + parquet_operations
+ return create_commits(
+ committer_hf_api,
+ repo_id=dataset,
+ revision=target_revision,
+ operations=operations,
+ commit_message=commit_message,
+ parent_commit=target_dataset_info.sha,
+ )
+
+
@@ -681,0 +851 @@ def compute_config_parquet_and_info_response(
+ job_id: str,
@@ -699,0 +870,2 @@ def compute_config_parquet_and_info_response(
+ job_id (`str`):
+ The id of the current Job. It is used to lock the access to the parquet conversion branch on the Hub.
@@ -847,30 +1019,15 @@ def compute_config_parquet_and_info_response(
- builder_info = asdict(builder.info)
- target_dataset_info = hf_api.dataset_info(repo_id=dataset, revision=target_revision, files_metadata=False)
- # - get repo parquet files
- all_repo_files: Set[str] = {f.rfilename for f in target_dataset_info.siblings}
- repo_parquet_files: Set[str] = {file for file in all_repo_files if file.endswith(".parquet")}
- # - get files that will be preserved in repo: files belonging to other configs and .gitattributes
- # we exclude files of current config because otherwise outdated files might be preserved
- files_to_ignore: Set[str] = {
- file
- for other_config in config_names.difference({config})
- for file in repo_parquet_files
- if file.startswith(f"{other_config}/")
- }.union({".gitattributes"})
- # - get files to be deleted - all files except for:
- # - parquet files obtained for current config at this processing step,
- # - parquet files belonging to other existing configs
- # - .gitattributes
- config_files_to_add = [operation.path_in_repo for operation in parquet_operations]
- files_to_delete = all_repo_files - set(config_files_to_add).union(files_to_ignore)
- delete_operations: List[CommitOperation] = [CommitOperationDelete(path_in_repo=file) for file in files_to_delete]
- logging.debug(f"{delete_operations=}")
-
- committer_hf_api.create_commit(
- repo_id=dataset,
- repo_type=DATASET_TYPE,
- revision=target_revision,
- operations=delete_operations + parquet_operations,
- commit_message=commit_message,
- parent_commit=target_dataset_info.sha,
- )
+ try:
+ sleeps = [1, 1, 1, 10, 10, 100, 100, 100, 300]
+ with lock.git_branch(dataset=dataset, branch=target_revision, job_id=job_id, sleeps=sleeps):
+ commit_parquet_conversion(
+ hf_api=hf_api,
+ committer_hf_api=committer_hf_api,
+ dataset=dataset,
+ config=config,
+ parquet_operations=parquet_operations,
+ config_names=config_names,
+ target_revision=target_revision,
+ commit_message=commit_message,
+ )
+ except TimeoutError as err:
+ raise LockedDatasetTimeoutError("the dataset is currently locked, please try again later.") from err
@@ -901 +1058 @@ def compute_config_parquet_and_info_response(
- dataset_info=builder_info,
+ dataset_info=asdict(builder.info),
@@ -933,0 +1091 @@ class ConfigParquetAndInfoJobRunner(ConfigCachedJobRunner):
+ job_id=self.job_info["job_id"],
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index 4399bd6e..14d209a2 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -15,0 +16,3 @@ from typing import (
+ Sequence,
+ Tuple,
+ Type,
@@ -287,0 +291,2 @@ FuncT = TypeVar("FuncT", bound=Callable[..., Any])
+RETRY_SLEEPS = (1, 1, 1, 10, 10, 10, 60, 60, 60, 10 * 60)
+RETRY_ON: Tuple[Type[Exception]] = (Exception,)
@@ -290 +295 @@ FuncT = TypeVar("FuncT", bound=Callable[..., Any])
-def retry(func: FuncT) -> FuncT:
+class retry:
@@ -292,24 +297,27 @@ def retry(func: FuncT) -> FuncT:
- SLEEPS = [1, 7, 70, 7 * 60, 70 * 60]
- MAX_ATTEMPTS = len(SLEEPS)
-
- @functools.wraps(func)
- def decorator(*args: Any, **kwargs: Any) -> Any:
- attempt = 0
- last_err = None
- while attempt < MAX_ATTEMPTS:
- try:
- """always sleep before calling the function. It will prevent rate limiting in the first place"""
- duration = SLEEPS[attempt]
- logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.")
- time.sleep(duration)
- return func(*args, **kwargs)
- except ConnectionError as err:
- logging.info("Got a ConnectionError, possibly due to rate limiting. Let's retry.")
- last_err = err
- attempt += 1
- raise RuntimeError(f"Give up after {attempt} attempts with ConnectionError") from last_err
-
- return cast(FuncT, decorator)
-
-
-@retry
+
+ def __init__(self, sleeps: Sequence[int] = RETRY_SLEEPS, on: Sequence[Type[Exception]] = RETRY_ON) -> None:
+ self.sleeps = sleeps
+ self.on = on
+
+ def __call__(self, func: FuncT) -> FuncT:
+ @functools.wraps(func)
+ def decorator(*args: Any, **kwargs: Any) -> Any:
+ attempt = 0
+ last_err = None
+ while attempt < len(self.sleeps):
+ try:
+ """always sleep before calling the function. It will prevent rate limiting in the first place"""
+ duration = self.sleeps[attempt]
+ logging.info(f"Sleep during {duration} seconds to preventively mitigate rate limiting.")
+ time.sleep(duration)
+ return func(*args, **kwargs)
+ except tuple(self.on) as err:
+ logging.info(f"Got a {type(err)}. Let's retry.")
+ last_err = err
+ attempt += 1
+ raise RuntimeError(f"Give up after {attempt} attempts with {type(last_err)}") from last_err
+
+ return cast(FuncT, decorator)
+
+
+@retry(on=[ConnectionError])
|
|
031daa5a5beee4e44573fb487f094d1429f3292d
|
Sylvain Lesage
| 2023-06-16T07:29:54 |
Revert "Delete ttl index from queue.py code (#1378)" (#1380)
|
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index 2c1d062d..09cb6301 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -4 +3,0 @@
-import pytest
@@ -120 +118,0 @@ def test_metrics_deletion_migration(mongo_host: str) -> None:
[email protected](reason="temporaly disabled because of index removal")
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index bec930c1..14f54073 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -25,0 +26 @@ from libcommon.constants import (
+ QUEUE_TTL_SECONDS,
@@ -136,0 +138 @@ class Job(Document):
+ {"fields": ["finished_at"], "expireAfterSeconds": QUEUE_TTL_SECONDS},
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 5262acd1..02b9578e 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -379 +378,0 @@ def test_queue_get_zombies() -> None:
[email protected](reason="temporaly disabled because of index removal")
|
|
c96497924667ce699ef4cb77fc04377f276337d0
|
Andrea Francis Soria Jimenez
| 2023-06-15T23:20:24 |
Rollback index (#1379)
|
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index 09cb6301..2c1d062d 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -3,0 +4 @@
+import pytest
@@ -118,0 +120 @@ def test_metrics_deletion_migration(mongo_host: str) -> None:
[email protected](reason="temporaly disabled because of index removal")
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index e226b886..bec930c1 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -26 +25,0 @@ from libcommon.constants import (
- QUEUE_TTL_SECONDS,
@@ -138,5 +136,0 @@ class Job(Document):
- {
- "fields": ["finished_at"],
- "expireAfterSeconds": QUEUE_TTL_SECONDS,
- "partialFilterExpression": {"status": {"$in": [Status.SUCCESS, Status.ERROR, Status.CANCELLED]}},
- },
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 02b9578e..5262acd1 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -378,0 +379 @@ def test_queue_get_zombies() -> None:
[email protected](reason="temporaly disabled because of index removal")
|
|
335754dd4284b19ccfa3b75e0f96171858ecfd2a
|
Andrea Francis Soria Jimenez
| 2023-06-15T22:32:21 |
Adding condition to Jobs Collection - ttl index (#1325)
|
diff --git a/jobs/mongodb_migration/src/mongodb_migration/collector.py b/jobs/mongodb_migration/src/mongodb_migration/collector.py
index 88e61e94..06412c36 100644
--- a/jobs/mongodb_migration/src/mongodb_migration/collector.py
+++ b/jobs/mongodb_migration/src/mongodb_migration/collector.py
@@ -217,0 +218,7 @@ class MigrationsCollector:
+ MigrationQueueDeleteTTLIndex(
+ version="20230607154800",
+ description=(
+ "delete the TTL index on the 'finished_at' field in the queue database to update its TTL condition"
+ ),
+ field_name="finished_at",
+ ),
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index 2c1d062d..09cb6301 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -4 +3,0 @@
-import pytest
@@ -120 +118,0 @@ def test_metrics_deletion_migration(mongo_host: str) -> None:
[email protected](reason="temporaly disabled because of index removal")
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index bec930c1..e226b886 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -25,0 +26 @@ from libcommon.constants import (
+ QUEUE_TTL_SECONDS,
@@ -136,0 +138,5 @@ class Job(Document):
+ {
+ "fields": ["finished_at"],
+ "expireAfterSeconds": QUEUE_TTL_SECONDS,
+ "partialFilterExpression": {"status": {"$in": [Status.SUCCESS, Status.ERROR, Status.CANCELLED]}},
+ },
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 5262acd1..02b9578e 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -379 +378,0 @@ def test_queue_get_zombies() -> None:
[email protected](reason="temporaly disabled because of index removal")
|
|
47ea65b2567db4482579cd7000393cf0a15b412e
|
Andrea Francis Soria Jimenez
| 2023-06-15T22:08:06 |
Delete ttl index from queue.py code (#1378)
|
diff --git a/jobs/mongodb_migration/tests/test_deletion_migrations.py b/jobs/mongodb_migration/tests/test_deletion_migrations.py
index 09cb6301..2c1d062d 100644
--- a/jobs/mongodb_migration/tests/test_deletion_migrations.py
+++ b/jobs/mongodb_migration/tests/test_deletion_migrations.py
@@ -3,0 +4 @@
+import pytest
@@ -118,0 +120 @@ def test_metrics_deletion_migration(mongo_host: str) -> None:
[email protected](reason="temporaly disabled because of index removal")
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 14f54073..bec930c1 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -26 +25,0 @@ from libcommon.constants import (
- QUEUE_TTL_SECONDS,
@@ -138 +136,0 @@ class Job(Document):
- {"fields": ["finished_at"], "expireAfterSeconds": QUEUE_TTL_SECONDS},
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 02b9578e..5262acd1 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -378,0 +379 @@ def test_queue_get_zombies() -> None:
[email protected](reason="temporaly disabled because of index removal")
|
|
32c833873068ac89428df7ac0a8802de6ee33414
|
Quentin Lhoest
| 2023-06-15T20:53:42 |
Refac hub_datasets fixture (#1373)
|
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 110dba75..da77c3fc 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -576,16 +576,2 @@ SPAWNING_OPT_IN_OUT_rows = ["http://testurl.test/test_image.jpg", "http://testur
[email protected](scope="session")
-def hub_datasets(
- hub_public_empty: str,
- hub_public_csv: str,
- hub_private_csv: str,
- hub_gated_csv: str,
- hub_public_jsonl: str,
- hub_public_audio: str,
- hub_public_image: str,
- hub_public_images_list: str,
- hub_public_big: str,
- hub_public_big_no_info: str,
- hub_public_big_csv: str,
- hub_public_external_files: str,
- hub_public_spawning_opt_in_out: str,
-) -> HubDatasets:
[email protected]
+def hub_reponses_does_not_exist() -> HubDatasetTest:
@@ -593,124 +579,180 @@ def hub_datasets(
- "does_not_exist": {
- "name": "does_not_exist",
- "config_names_response": None,
- "splits_response": None,
- "first_rows_response": None,
- "parquet_and_info_response": None,
- },
- "does_not_exist_config": {
- "name": "does_not_exist_config",
- "config_names_response": None,
- "splits_response": None,
- "first_rows_response": None,
- "parquet_and_info_response": None,
- },
- "does_not_exist_split": {
- "name": "does_not_exist_split",
- "config_names_response": None,
- "splits_response": None,
- "first_rows_response": None,
- "parquet_and_info_response": None,
- },
- "empty": {
- "name": hub_public_empty,
- "config_names_response": None,
- "splits_response": None,
- "first_rows_response": None,
- "parquet_and_info_response": None,
- },
- "public": {
- "name": hub_public_csv,
- "config_names_response": create_config_names_response(hub_public_csv),
- "splits_response": create_splits_response(hub_public_csv),
- "first_rows_response": create_first_rows_response(hub_public_csv, DATA_cols, DATA_rows),
- "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_csv, data_type="csv"),
- },
- "private": {
- "name": hub_private_csv,
- "config_names_response": create_config_names_response(hub_private_csv),
- "splits_response": create_splits_response(hub_private_csv),
- "first_rows_response": create_first_rows_response(hub_private_csv, DATA_cols, DATA_rows),
- "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_private_csv, data_type="csv"),
- },
- "gated": {
- "name": hub_gated_csv,
- "config_names_response": create_config_names_response(hub_gated_csv),
- "splits_response": create_splits_response(hub_gated_csv),
- "first_rows_response": create_first_rows_response(hub_gated_csv, DATA_cols, DATA_rows),
- "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_gated_csv, data_type="csv"),
- },
- "jsonl": {
- "name": hub_public_jsonl,
- "config_names_response": create_config_names_response(hub_public_jsonl),
- "splits_response": create_splits_response(hub_public_jsonl),
- "first_rows_response": create_first_rows_response(hub_public_jsonl, JSONL_cols, JSONL_rows),
- "parquet_and_info_response": None,
- },
- "audio": {
- "name": hub_public_audio,
- "config_names_response": create_config_names_response(hub_public_audio),
- "splits_response": create_splits_response(hub_public_audio),
- "first_rows_response": create_first_rows_response(
- hub_public_audio, AUDIO_cols, get_AUDIO_rows(hub_public_audio)
- ),
- "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_audio, data_type="audio"),
- },
- "image": {
- "name": hub_public_image,
- "config_names_response": create_config_names_response(hub_public_image),
- "splits_response": create_splits_response(hub_public_image),
- "first_rows_response": create_first_rows_response(
- hub_public_image, IMAGE_cols, get_IMAGE_rows(hub_public_image)
- ),
- "parquet_and_info_response": None,
- },
- "images_list": {
- "name": hub_public_images_list,
- "config_names_response": create_config_names_response(hub_public_images_list),
- "splits_response": create_splits_response(hub_public_images_list),
- "first_rows_response": create_first_rows_response(
- hub_public_images_list, IMAGES_LIST_cols, get_IMAGES_LIST_rows(hub_public_images_list)
- ),
- "parquet_and_info_response": None,
- },
- "big": {
- "name": hub_public_big,
- "config_names_response": create_config_names_response(hub_public_big),
- "splits_response": create_splits_response(hub_public_big),
- "first_rows_response": create_first_rows_response(hub_public_big, BIG_cols, BIG_rows),
- "parquet_and_info_response": create_parquet_and_info_response(
- dataset=hub_public_big, data_type="big_parquet"
- ),
- },
- "big-no-info": {
- "name": hub_public_big_no_info,
- "config_names_response": create_config_names_response(hub_public_big_no_info),
- "splits_response": create_splits_response(hub_public_big_no_info),
- "first_rows_response": create_first_rows_response(hub_public_big_no_info, BIG_cols, BIG_rows),
- "parquet_and_info_response": create_parquet_and_info_response(
- dataset=hub_public_big_no_info, data_type="big_parquet_no_info"
- ),
- },
- "big-csv": {
- "name": hub_public_big_csv,
- "config_names_response": create_config_names_response(hub_public_big_csv),
- "splits_response": create_splits_response(hub_public_big_csv),
- "first_rows_response": create_first_rows_response(hub_public_big_csv, BIG_cols, BIG_rows),
- "parquet_and_info_response": None,
- },
- "external_files": {
- "name": hub_public_external_files,
- "config_names_response": create_config_names_response(hub_public_external_files),
- "splits_response": create_splits_response(hub_public_external_files),
- "first_rows_response": create_first_rows_response(hub_public_external_files, TEXT_cols, TEXT_rows),
- "parquet_and_info_response": None,
- },
- "spawning_opt_in_out": {
- "name": hub_public_spawning_opt_in_out,
- "config_names_response": create_config_names_response(hub_public_spawning_opt_in_out),
- "splits_response": create_splits_response(hub_public_spawning_opt_in_out),
- "first_rows_response": create_first_rows_response(
- hub_public_spawning_opt_in_out, SPAWNING_OPT_IN_OUT_cols, SPAWNING_OPT_IN_OUT_rows
- ),
- "parquet_and_info_response": None,
- },
+ "name": "does_not_exist",
+ "config_names_response": None,
+ "splits_response": None,
+ "first_rows_response": None,
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_does_not_exist_config() -> HubDatasetTest:
+ return {
+ "name": "does_not_exist_config",
+ "config_names_response": None,
+ "splits_response": None,
+ "first_rows_response": None,
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_does_not_exist_split() -> HubDatasetTest:
+ return {
+ "name": "does_not_exist_split",
+ "config_names_response": None,
+ "splits_response": None,
+ "first_rows_response": None,
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_empty(hub_public_empty: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_empty,
+ "config_names_response": None,
+ "splits_response": None,
+ "first_rows_response": None,
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_public(hub_public_csv: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_csv,
+ "config_names_response": create_config_names_response(hub_public_csv),
+ "splits_response": create_splits_response(hub_public_csv),
+ "first_rows_response": create_first_rows_response(hub_public_csv, DATA_cols, DATA_rows),
+ "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_csv, data_type="csv"),
+ }
+
+
[email protected]
+def hub_reponses_private(hub_private_csv: str) -> HubDatasetTest:
+ return {
+ "name": hub_private_csv,
+ "config_names_response": create_config_names_response(hub_private_csv),
+ "splits_response": create_splits_response(hub_private_csv),
+ "first_rows_response": create_first_rows_response(hub_private_csv, DATA_cols, DATA_rows),
+ "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_private_csv, data_type="csv"),
+ }
+
+
[email protected]
+def hub_reponses_gated(hub_gated_csv: str) -> HubDatasetTest:
+ return {
+ "name": hub_gated_csv,
+ "config_names_response": create_config_names_response(hub_gated_csv),
+ "splits_response": create_splits_response(hub_gated_csv),
+ "first_rows_response": create_first_rows_response(hub_gated_csv, DATA_cols, DATA_rows),
+ "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_gated_csv, data_type="csv"),
+ }
+
+
[email protected]
+def hub_reponses_jsonl(hub_public_jsonl: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_jsonl,
+ "config_names_response": create_config_names_response(hub_public_jsonl),
+ "splits_response": create_splits_response(hub_public_jsonl),
+ "first_rows_response": create_first_rows_response(hub_public_jsonl, JSONL_cols, JSONL_rows),
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_audio(hub_public_audio: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_audio,
+ "config_names_response": create_config_names_response(hub_public_audio),
+ "splits_response": create_splits_response(hub_public_audio),
+ "first_rows_response": create_first_rows_response(
+ hub_public_audio, AUDIO_cols, get_AUDIO_rows(hub_public_audio)
+ ),
+ "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_audio, data_type="audio"),
+ }
+
+
[email protected]
+def hub_reponses_image(hub_public_image: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_image,
+ "config_names_response": create_config_names_response(hub_public_image),
+ "splits_response": create_splits_response(hub_public_image),
+ "first_rows_response": create_first_rows_response(
+ hub_public_image, IMAGE_cols, get_IMAGE_rows(hub_public_image)
+ ),
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_images_list(hub_public_images_list: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_images_list,
+ "config_names_response": create_config_names_response(hub_public_images_list),
+ "splits_response": create_splits_response(hub_public_images_list),
+ "first_rows_response": create_first_rows_response(
+ hub_public_images_list, IMAGES_LIST_cols, get_IMAGES_LIST_rows(hub_public_images_list)
+ ),
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_big(hub_public_big: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_big,
+ "config_names_response": create_config_names_response(hub_public_big),
+ "splits_response": create_splits_response(hub_public_big),
+ "first_rows_response": create_first_rows_response(hub_public_big, BIG_cols, BIG_rows),
+ "parquet_and_info_response": create_parquet_and_info_response(dataset=hub_public_big, data_type="big_parquet"),
+ }
+
+
[email protected]
+def hub_reponses_big_no_info(hub_public_big_no_info: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_big_no_info,
+ "config_names_response": create_config_names_response(hub_public_big_no_info),
+ "splits_response": create_splits_response(hub_public_big_no_info),
+ "first_rows_response": create_first_rows_response(hub_public_big_no_info, BIG_cols, BIG_rows),
+ "parquet_and_info_response": create_parquet_and_info_response(
+ dataset=hub_public_big_no_info, data_type="big_parquet_no_info"
+ ),
+ }
+
+
[email protected]
+def hub_reponses_big_csv(hub_public_big_csv: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_big_csv,
+ "config_names_response": create_config_names_response(hub_public_big_csv),
+ "splits_response": create_splits_response(hub_public_big_csv),
+ "first_rows_response": create_first_rows_response(hub_public_big_csv, BIG_cols, BIG_rows),
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_external_files(hub_public_external_files: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_external_files,
+ "config_names_response": create_config_names_response(hub_public_external_files),
+ "splits_response": create_splits_response(hub_public_external_files),
+ "first_rows_response": create_first_rows_response(hub_public_external_files, TEXT_cols, TEXT_rows),
+ "parquet_and_info_response": None,
+ }
+
+
[email protected]
+def hub_reponses_spawning_opt_in_out(hub_public_spawning_opt_in_out: str) -> HubDatasetTest:
+ return {
+ "name": hub_public_spawning_opt_in_out,
+ "config_names_response": create_config_names_response(hub_public_spawning_opt_in_out),
+ "splits_response": create_splits_response(hub_public_spawning_opt_in_out),
+ "first_rows_response": create_first_rows_response(
+ hub_public_spawning_opt_in_out, SPAWNING_OPT_IN_OUT_cols, SPAWNING_OPT_IN_OUT_rows
+ ),
+ "parquet_and_info_response": None,
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index 66944344..2121d53d 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -4,0 +5 @@ import io
+from contextlib import contextmanager
@@ -46 +47 @@ from ...constants import CI_HUB_ENDPOINT, CI_USER_TOKEN
-from ...fixtures.hub import HubDatasets
+from ...fixtures.hub import HubDatasetTest
@@ -49,14 +50,5 @@ from ...fixtures.hub import HubDatasets
-# see https://github.com/pytest-dev/pytest/issues/363#issuecomment-406536200
[email protected](scope="module", autouse=True)
-def set_supported_datasets(hub_datasets: HubDatasets) -> Iterator[pytest.MonkeyPatch]:
- mp = pytest.MonkeyPatch()
- mp.setenv(
- "PARQUET_AND_INFO_BLOCKED_DATASETS",
- ",".join(value["name"] for value in hub_datasets.values() if "jsonl" in value["name"]),
- )
- mp.setenv(
- "PARQUET_AND_INFO_SUPPORTED_DATASETS",
- ",".join(value["name"] for value in hub_datasets.values() if "big" not in value["name"]),
- )
- yield mp
- mp.undo()
+@contextmanager
+def blocked(app_config: AppConfig, repo_id: str) -> Iterator[None]:
+ app_config.parquet_and_info.blocked_datasets.append(repo_id)
+ yield
+ app_config.parquet_and_info.blocked_datasets.remove(repo_id)
@@ -126 +118 @@ def test_compute(
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
@@ -128,2 +120,2 @@ def test_compute(
- dataset = hub_datasets["public"]["name"]
- config = hub_datasets["public"]["config_names_response"]["config_names"][0]["config"]
+ dataset = hub_reponses_public["name"]
+ config = hub_reponses_public["config_names_response"]["config_names"][0]["config"]
@@ -134 +126 @@ def test_compute(
- content=hub_datasets["public"]["config_names_response"],
+ content=hub_reponses_public["config_names_response"],
@@ -142 +134 @@ def test_compute(
- assert_content_is_equal(content, hub_datasets["public"]["parquet_and_info_response"])
+ assert_content_is_equal(content, hub_reponses_public["parquet_and_info_response"])
@@ -245 +237,2 @@ def test_raise_if_too_big_from_hub(
- hub_datasets: HubDatasets,
+ hub_public_csv: str,
+ hub_public_big: str,
@@ -250 +243 @@ def test_raise_if_too_big_from_hub(
- dataset = hub_datasets[name]["name"]
+ dataset = hub_public_csv if name == "public" else hub_public_big
@@ -274 +267,2 @@ def test_raise_if_too_big_from_datasets(
- hub_datasets: HubDatasets,
+ hub_public_csv: str,
+ hub_public_big: str,
@@ -279 +273 @@ def test_raise_if_too_big_from_datasets(
- dataset = hub_datasets[name]["name"]
+ dataset = hub_public_csv if name == "public" else hub_public_big
@@ -363 +357 @@ def test_supported_if_big_parquet(
- hub_datasets: HubDatasets,
+ hub_reponses_big: HubDatasetTest,
@@ -368,2 +362,2 @@ def test_supported_if_big_parquet(
- dataset = hub_datasets["big"]["name"]
- config = hub_datasets["big"]["config_names_response"]["config_names"][0]["config"]
+ dataset = hub_reponses_big["name"]
+ config = hub_reponses_big["config_names_response"]["config_names"][0]["config"]
@@ -374 +368 @@ def test_supported_if_big_parquet(
- content=hub_datasets["big"]["config_names_response"],
+ content=hub_reponses_big["config_names_response"],
@@ -382 +376 @@ def test_supported_if_big_parquet(
- assert_content_is_equal(content, hub_datasets["big"]["parquet_and_info_response"])
+ assert_content_is_equal(content, hub_reponses_big["parquet_and_info_response"])
@@ -388 +382 @@ def test_not_supported_if_big_non_parquet(
- hub_datasets: HubDatasets,
+ hub_reponses_big_csv: HubDatasetTest,
@@ -392,2 +386,2 @@ def test_not_supported_if_big_non_parquet(
- dataset = hub_datasets["big-csv"]["name"]
- config = hub_datasets["big-csv"]["config_names_response"]["config_names"][0]["config"]
+ dataset = hub_reponses_big_csv["name"]
+ config = hub_reponses_big_csv["config_names_response"]["config_names"][0]["config"]
@@ -398 +392 @@ def test_not_supported_if_big_non_parquet(
- content=hub_datasets["big-csv"]["config_names_response"],
+ content=hub_reponses_big_csv["config_names_response"],
@@ -409 +403 @@ def test_supported_if_gated(
- hub_datasets: HubDatasets,
+ hub_reponses_gated: HubDatasetTest,
@@ -412,2 +406,2 @@ def test_supported_if_gated(
- dataset = hub_datasets["gated"]["name"]
- config = hub_datasets["gated"]["config_names_response"]["config_names"][0]["config"]
+ dataset = hub_reponses_gated["name"]
+ config = hub_reponses_gated["config_names_response"]["config_names"][0]["config"]
@@ -418 +412 @@ def test_supported_if_gated(
- content=hub_datasets["gated"]["config_names_response"],
+ content=hub_reponses_gated["config_names_response"],
@@ -429 +423 @@ def test_blocked(
- hub_datasets: HubDatasets,
+ hub_reponses_jsonl: HubDatasetTest,
@@ -432,12 +426,13 @@ def test_blocked(
- dataset = hub_datasets["jsonl"]["name"]
- config = hub_datasets["jsonl"]["config_names_response"]["config_names"][0]["config"]
- upsert_response(
- kind="dataset-config-names",
- dataset=dataset,
- http_status=HTTPStatus.OK,
- content=hub_datasets["jsonl"]["config_names_response"],
- )
- job_runner = get_job_runner(dataset, config, app_config)
- with pytest.raises(CustomError) as e:
- job_runner.compute()
- assert e.typename == "DatasetInBlockListError"
+ with blocked(app_config, repo_id=hub_reponses_jsonl["name"]):
+ dataset = hub_reponses_jsonl["name"]
+ config = hub_reponses_jsonl["config_names_response"]["config_names"][0]["config"]
+ upsert_response(
+ kind="dataset-config-names",
+ dataset=dataset,
+ http_status=HTTPStatus.OK,
+ content=hub_reponses_jsonl["config_names_response"],
+ )
+ job_runner = get_job_runner(dataset, config, app_config)
+ with pytest.raises(CustomError) as e:
+ job_runner.compute()
+ assert e.typename == "DatasetInBlockListError"
@@ -451 +446,3 @@ def test_compute_splits_response_simple_csv_ok(
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
+ hub_reponses_audio: HubDatasetTest,
+ hub_reponses_gated: HubDatasetTest,
@@ -456,0 +454 @@ def test_compute_splits_response_simple_csv_ok(
+ hub_datasets = {"public": hub_reponses_public, "audio": hub_reponses_audio, "gated": hub_reponses_gated}
@@ -495 +493 @@ def test_compute_splits_response_simple_csv_error(
- hub_datasets: HubDatasets,
+ hub_reponses_private: HubDatasetTest,
@@ -502,2 +500,2 @@ def test_compute_splits_response_simple_csv_error(
- dataset = hub_datasets[name]["name"]
- config_names_response = hub_datasets[name]["config_names_response"]
+ dataset = hub_reponses_private["name"]
+ config_names_response = hub_reponses_private["config_names_response"]
@@ -509 +507 @@ def test_compute_splits_response_simple_csv_error(
- content=hub_datasets[name]["config_names_response"],
+ content=hub_reponses_private["config_names_response"],
@@ -533 +531 @@ def test_compute_splits_response_simple_csv_error_2(
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
@@ -540,2 +538,2 @@ def test_compute_splits_response_simple_csv_error_2(
- dataset = hub_datasets[name]["name"]
- config_names_response = hub_datasets[name]["config_names_response"]
+ dataset = hub_reponses_public["name"]
+ config_names_response = hub_reponses_public["config_names_response"]
@@ -561,2 +559 @@ def test_previous_step_error(
- hub_public_csv: str,
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
@@ -565,2 +562,2 @@ def test_previous_step_error(
- dataset = hub_datasets["public"]["name"]
- config = hub_datasets["public"]["config_names_response"]["config_names"][0]["config"]
+ dataset = hub_reponses_public["name"]
+ config = hub_reponses_public["config_names_response"]["config_names"][0]["config"]
diff --git a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
index 2d60370b..56e28c7e 100644
--- a/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
+++ b/services/worker/tests/job_runners/config/test_split_names_from_streaming.py
@@ -20 +20 @@ from worker.resources import LibrariesResource
-from ...fixtures.hub import HubDatasets, get_default_config_split
+from ...fixtures.hub import HubDatasetTest, get_default_config_split
@@ -91 +91,6 @@ def test_compute_split_names_from_streaming_response(
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
+ hub_reponses_audio: HubDatasetTest,
+ hub_reponses_gated: HubDatasetTest,
+ hub_reponses_private: HubDatasetTest,
+ hub_reponses_empty: HubDatasetTest,
+ hub_reponses_does_not_exist: HubDatasetTest,
@@ -98,0 +104,8 @@ def test_compute_split_names_from_streaming_response(
+ hub_datasets = {
+ "public": hub_reponses_public,
+ "audio": hub_reponses_audio,
+ "gated": hub_reponses_gated,
+ "private": hub_reponses_private,
+ "empty": hub_reponses_empty,
+ "does_not_exist": hub_reponses_does_not_exist,
+ }
diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py
index 3f0ed076..a94427f4 100644
--- a/services/worker/tests/job_runners/dataset/test_config_names.py
+++ b/services/worker/tests/job_runners/dataset/test_config_names.py
@@ -18 +18 @@ from worker.resources import LibrariesResource
-from ...fixtures.hub import HubDatasets
+from ...fixtures.hub import HubDatasetTest
@@ -113 +113,6 @@ def test_compute_splits_response_simple_csv(
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
+ hub_reponses_audio: HubDatasetTest,
+ hub_reponses_gated: HubDatasetTest,
+ hub_reponses_private: HubDatasetTest,
+ hub_reponses_empty: HubDatasetTest,
+ hub_reponses_does_not_exist: HubDatasetTest,
@@ -120,0 +126,8 @@ def test_compute_splits_response_simple_csv(
+ hub_datasets = {
+ "public": hub_reponses_public,
+ "audio": hub_reponses_audio,
+ "gated": hub_reponses_gated,
+ "private": hub_reponses_private,
+ "empty": hub_reponses_empty,
+ "does_not_exist": hub_reponses_does_not_exist,
+ }
diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
index e5ed7cfa..fd2dd64b 100644
--- a/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
+++ b/services/worker/tests/job_runners/split/test_first_rows_from_streaming.py
@@ -24 +24 @@ from worker.utils import get_json_size
-from ...fixtures.hub import HubDatasets, get_default_config_split
+from ...fixtures.hub import HubDatasetTest, get_default_config_split
@@ -116 +116,10 @@ def test_number_rows(
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
+ hub_reponses_audio: HubDatasetTest,
+ hub_reponses_image: HubDatasetTest,
+ hub_reponses_images_list: HubDatasetTest,
+ hub_reponses_jsonl: HubDatasetTest,
+ hub_reponses_gated: HubDatasetTest,
+ hub_reponses_private: HubDatasetTest,
+ hub_reponses_empty: HubDatasetTest,
+ hub_reponses_does_not_exist_config: HubDatasetTest,
+ hub_reponses_does_not_exist_split: HubDatasetTest,
@@ -129,0 +139,12 @@ def test_number_rows(
+ hub_datasets = {
+ "public": hub_reponses_public,
+ "audio": hub_reponses_audio,
+ "image": hub_reponses_image,
+ "images_list": hub_reponses_images_list,
+ "jsonl": hub_reponses_jsonl,
+ "gated": hub_reponses_gated,
+ "private": hub_reponses_private,
+ "empty": hub_reponses_empty,
+ "does_not_exist_config": hub_reponses_does_not_exist_config,
+ "does_not_exist_split": hub_reponses_does_not_exist_split,
+ }
@@ -187 +208,2 @@ def test_truncation(
- hub_datasets: HubDatasets,
+ hub_public_csv: str,
+ hub_public_big: str,
@@ -195 +217,2 @@ def test_truncation(
- dataset, config, split = get_default_config_split(hub_datasets[name]["name"])
+ dataset = hub_public_csv if name == "public" else hub_public_big
+ dataset, config, split = get_default_config_split(dataset)
diff --git a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
index f56e08b4..e3fa31ac 100644
--- a/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
+++ b/services/worker/tests/job_runners/split/test_opt_in_out_urls_scan_from_streaming.py
@@ -32 +32 @@ from ...constants import CI_SPAWNING_TOKEN
-from ...fixtures.hub import HubDatasets, get_default_config_split
+from ...fixtures.hub import HubDatasetTest, get_default_config_split
@@ -176 +176,2 @@ def test_compute(
- hub_datasets: HubDatasets,
+ hub_reponses_public: HubDatasetTest,
+ hub_reponses_spawning_opt_in_out: HubDatasetTest,
@@ -183,0 +185 @@ def test_compute(
+ hub_datasets = {"public": hub_reponses_public, "spawning_opt_in_out": hub_reponses_spawning_opt_in_out}
@@ -238 +240 @@ def test_compute_failed(
- hub_datasets: HubDatasets,
+ hub_reponses_spawning_opt_in_out: HubDatasetTest,
@@ -247 +249 @@ def test_compute_failed(
- dataset = hub_datasets["spawning_opt_in_out"]["name"]
+ dataset = hub_reponses_spawning_opt_in_out["name"]
|
|
ae6ffbe68ec50aa144e2b8986a7cd85c161d1faa
|
Quentin Lhoest
| 2023-06-15T20:37:34 |
fix fill_builder_info (#1375)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 8f2f0f10..d9d10395 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -631,0 +632 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None:
+ split = str(split) # in case it's a NamedSplit
|
|
0ce4b5b508ac660eba658ac77a8a90883541516c
|
Steven Liu
| 2023-06-15T20:22:50 |
add build notebook (#1377)
|
diff --git a/.github/workflows/doc-build.yml b/.github/workflows/doc-build.yml
index 8dc7ac99..6238f7c7 100644
--- a/.github/workflows/doc-build.yml
+++ b/.github/workflows/doc-build.yml
@@ -19,0 +20 @@ jobs:
+ notebook_folder: datasets-server_doc
|
|
6865d8fab8c9476ab1efccd713479c0b432bd22d
|
Albert Villanova del Moral
| 2023-06-15T15:59:03 |
Update datasets dependency to 2.13.0 version (#1372)
|
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index cd3fcb36..d43c651e 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -529 +529 @@ name = "datasets"
-version = "2.12.0"
+version = "2.13.0"
@@ -535,2 +535,2 @@ files = [
- {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"},
- {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"},
+ {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
+ {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
@@ -553 +552,0 @@ requests = ">=2.19.0"
-responses = "<0.19"
@@ -562 +561 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)",
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -565 +564 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
@@ -570 +569 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -1249 +1248 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.12.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
@@ -2605,19 +2603,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"]
-[[package]]
-name = "responses"
-version = "0.18.0"
-description = "A utility library for mocking out the `requests` Python library."
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"},
- {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"},
-]
-
-[package.dependencies]
-requests = ">=2.0,<3.0"
-urllib3 = ">=1.25.10"
-
-[package.extras]
-tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"]
-
diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock
index bd34f45d..6b1a8a05 100644
--- a/jobs/cache_maintenance/poetry.lock
+++ b/jobs/cache_maintenance/poetry.lock
@@ -579 +579 @@ name = "datasets"
-version = "2.12.0"
+version = "2.13.0"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"},
- {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"},
+ {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
+ {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
@@ -603 +602,0 @@ requests = ">=2.19.0"
-responses = "<0.19"
@@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)",
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
@@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -1027 +1026 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.12.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
@@ -2390,19 +2388,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"]
-[[package]]
-name = "responses"
-version = "0.18.0"
-description = "A utility library for mocking out the `requests` Python library."
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"},
- {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"},
-]
-
-[package.dependencies]
-requests = ">=2.0,<3.0"
-urllib3 = ">=1.25.10"
-
-[package.extras]
-tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"]
-
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index bd34f45d..6b1a8a05 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -579 +579 @@ name = "datasets"
-version = "2.12.0"
+version = "2.13.0"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"},
- {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"},
+ {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
+ {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
@@ -603 +602,0 @@ requests = ">=2.19.0"
-responses = "<0.19"
@@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)",
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
@@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -1027 +1026 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.12.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
@@ -2390,19 +2388,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"]
-[[package]]
-name = "responses"
-version = "0.18.0"
-description = "A utility library for mocking out the `requests` Python library."
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"},
- {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"},
-]
-
-[package.dependencies]
-requests = ">=2.0,<3.0"
-urllib3 = ">=1.25.10"
-
-[package.extras]
-tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"]
-
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index 1ddc15bc..99b16a6a 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -579 +579 @@ name = "datasets"
-version = "2.12.0"
+version = "2.13.0"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"},
- {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"},
+ {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
+ {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
@@ -603 +602,0 @@ requests = ">=2.19.0"
-responses = "<0.19"
@@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)",
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
@@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -2371,19 +2369,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"]
-[[package]]
-name = "responses"
-version = "0.18.0"
-description = "A utility library for mocking out the `requests` Python library."
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"},
- {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"},
-]
-
-[package.dependencies]
-requests = ">=2.0,<3.0"
-urllib3 = ">=1.25.10"
-
-[package.extras]
-tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"]
-
@@ -3035 +3015 @@ python-versions = "3.9.15"
-content-hash = "3eab4cefa7dd3f05dcf9657d8a1e81f7d7df494f07b6eb0f85fbc97e8a6aa56e"
+content-hash = "bfdd77727f041d88e26a4da8c23a6b8434f20c6747f068f1ca3c11d841e38aa1"
diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml
index 1bc1d006..8dcfc6a6 100644
--- a/libs/libcommon/pyproject.toml
+++ b/libs/libcommon/pyproject.toml
@@ -10 +10 @@ appdirs = "^1.4.4"
-datasets = { extras = ["audio", "vision"], version = "^2.12.0" }
+datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index 928e56ca..e138e9d3 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -579 +579 @@ name = "datasets"
-version = "2.12.0"
+version = "2.13.0"
@@ -585,2 +585,2 @@ files = [
- {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"},
- {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"},
+ {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
+ {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
@@ -603 +602,0 @@ requests = ">=2.19.0"
-responses = "<0.19"
@@ -612 +611 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)",
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -615 +614 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
@@ -620 +619 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -1085 +1084 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.12.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
@@ -2452 +2451 @@ description = "A utility library for mocking out the `requests` Python library."
-category = "main"
+category = "dev"
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index 70a86b1c..77c2de7d 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -621 +621 @@ name = "datasets"
-version = "2.12.0"
+version = "2.13.0"
@@ -627,2 +627,2 @@ files = [
- {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"},
- {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"},
+ {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
+ {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
@@ -645 +644,0 @@ requests = ">=2.19.0"
-responses = "<0.19"
@@ -654 +653 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)",
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -657 +656 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
@@ -662 +661 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -1147 +1146 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.12.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
@@ -2658,19 +2656,0 @@ use-chardet-on-py3 = ["chardet (>=3.0.2,<6)"]
-[[package]]
-name = "responses"
-version = "0.18.0"
-description = "A utility library for mocking out the `requests` Python library."
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"},
- {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"},
-]
-
-[package.dependencies]
-requests = ">=2.0,<3.0"
-urllib3 = ">=1.25.10"
-
-[package.extras]
-tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"]
-
@@ -3433 +3413 @@ python-versions = "3.9.15"
-content-hash = "8f47c3a3309578c035ed9ed97d3b5b965b9efaabca2c187e5e501303ab3310aa"
+content-hash = "781ea6fdf22b868dd68ea1f37735c621a9064e2ee616b3e2538fa8924697b1b2"
diff --git a/services/api/pyproject.toml b/services/api/pyproject.toml
index d1055331..7fc589ed 100644
--- a/services/api/pyproject.toml
+++ b/services/api/pyproject.toml
@@ -10 +9,0 @@ cryptography = "^41.0.1"
-datasets = { extras = ["audio", "vision"], version = "^2.11.0" }
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 6ae2e954..8d817352 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -910 +910 @@ name = "datasets"
-version = "2.12.0"
+version = "2.13.0"
@@ -916,2 +916,2 @@ files = [
- {file = "datasets-2.12.0-py3-none-any.whl", hash = "sha256:0a23bdf1fc28d82dd496375289d72f7917d149a95062ab2647cf621d67ed74ca"},
- {file = "datasets-2.12.0.tar.gz", hash = "sha256:faf164c18a41bea51df3f369e872f8be5b84c12ea5f6393c3896f56038af1ea3"},
+ {file = "datasets-2.13.0-py3-none-any.whl", hash = "sha256:26671d474990ad8fd7388e8c67cde4d72f6c1f0e87af685fc09af5d9a5992274"},
+ {file = "datasets-2.13.0.tar.gz", hash = "sha256:b8c3bcf9c3d0c74f101c7645e42231de9f45206a2e742df15799da9bfa625608"},
@@ -934 +933,0 @@ requests = ">=2.19.0"
-responses = "<0.19"
@@ -943 +942 @@ benchmarks = ["numpy (==1.18.5)", "protobuf (==3.20.3)", "tensorflow (==2.3.0)",
-dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+dev = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "black (>=23.1,<24.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "pyyaml (>=5.3.1)", "rarfile (>=4.0)", "ruff (>=0.0.241)", "s3fs", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -946 +945 @@ jax = ["jax (>=0.2.8,!=0.3.2,<=0.3.25)", "jaxlib (>=0.1.65,<=0.3.25)"]
-metrics-tests = ["Werkzeug (>=1.0.1)", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
+metrics-tests = ["Werkzeug (>=1.0.1)", "accelerate", "bert-score (>=0.3.6)", "jiwer", "langdetect", "mauve-text", "nltk", "requests-file (>=1.5.1)", "rouge-score", "sacrebleu", "sacremoses", "scikit-learn", "scipy", "sentencepiece", "seqeval", "six (>=1.15.0,<1.16.0)", "spacy (>=3.0.0)", "texttable (>=1.6.3)", "tldextract", "tldextract (>=3.1.0)", "toml (>=0.10.1)", "typer (<0.5.0)"]
@@ -951 +950 @@ tensorflow-gpu = ["tensorflow-gpu (>=2.2.0,!=2.6.0,!=2.6.1)"]
-tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
+tests = ["Pillow (>=6.2.1)", "absl-py", "apache-beam (>=2.26.0,<2.44.0)", "elasticsearch (<8.0.0)", "faiss-cpu (>=1.6.4)", "joblibspark", "librosa", "lz4", "py7zr", "pyspark (>=3.4)", "pytest", "pytest-datadir", "pytest-xdist", "rarfile (>=4.0)", "s3fs (>=2021.11.1)", "soundfile (>=0.12.1)", "sqlalchemy (<2.0.0)", "tensorflow (>=2.3,!=2.6.0,!=2.6.1)", "tensorflow-macos", "tiktoken", "torch", "transformers", "zstandard"]
@@ -1781 +1780 @@ appdirs = "^1.4.4"
-datasets = {version = "^2.12.0", extras = ["audio", "vision"]}
+datasets = {version = "^2.13.0", extras = ["audio", "vision"]}
@@ -4149,19 +4147,0 @@ rsa = ["oauthlib[signedtoken] (>=3.0.0)"]
-[[package]]
-name = "responses"
-version = "0.18.0"
-description = "A utility library for mocking out the `requests` Python library."
-category = "main"
-optional = false
-python-versions = ">=3.7"
-files = [
- {file = "responses-0.18.0-py3-none-any.whl", hash = "sha256:15c63ad16de13ee8e7182d99c9334f64fd81f1ee79f90748d527c28f7ca9dd51"},
- {file = "responses-0.18.0.tar.gz", hash = "sha256:380cad4c1c1dc942e5e8a8eaae0b4d4edf708f4f010db8b7bcfafad1fcd254ff"},
-]
-
-[package.dependencies]
-requests = ">=2.0,<3.0"
-urllib3 = ">=1.25.10"
-
-[package.extras]
-tests = ["coverage (>=6.0.0)", "flake8", "mypy", "pytest (>=4.6)", "pytest-cov", "pytest-localserver", "types-mock", "types-requests"]
-
|
|
8396a0d518d446b24334c0e4daf4841327fdf56f
|
Andrea Francis Soria Jimenez
| 2023-06-15T14:58:42 |
Adding limit for number of configs (#1371)
|
diff --git a/chart/templates/_envWorker.tpl b/chart/templates/_envWorker.tpl
index ffbec5ae..3d9765fc 100644
--- a/chart/templates/_envWorker.tpl
+++ b/chart/templates/_envWorker.tpl
@@ -86,0 +87,2 @@
+- name: CONFIG_NAMES_MAX_NUMBER
+ value: {{ .Values.configNames.maxNumber | quote }}
diff --git a/chart/values.yaml b/chart/values.yaml
index aaae2c89..5294d83c 100644
--- a/chart/values.yaml
+++ b/chart/values.yaml
@@ -182,0 +183,4 @@ optInOutUrlsScan:
+configNames:
+ # the max number of configs per dataset
+ maxNumber: 3_000
+
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py
index eaf8cc95..f7048d5a 100644
--- a/libs/libcommon/src/libcommon/exceptions.py
+++ b/libs/libcommon/src/libcommon/exceptions.py
@@ -87,0 +88 @@ CacheableErrorCode = Literal[
+ "DatasetWithTooManyConfigsError",
@@ -481,0 +483,7 @@ class UnsupportedExternalFilesError(CacheableError):
+
+
+class DatasetWithTooManyConfigsError(CacheableError):
+ """Raised when the number of configs of a dataset exceeded the limit."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyConfigsError", cause, True)
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index 3419e9aa..bc3fd397 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -219,0 +220,16 @@ class NumbaConfig:
+CONFIG_NAMES_MAX_NUMBER = 3_000
+
+
+@dataclass(frozen=True)
+class ConfigNamesConfig:
+ max_number: int = CONFIG_NAMES_MAX_NUMBER
+
+ @classmethod
+ def from_env(cls) -> "ConfigNamesConfig":
+ env = Env(expand_vars=True)
+ with env.prefixed("CONFIG_NAMES_"):
+ return cls(
+ max_number=env.int(name="MAX_NUMBER", default=CONFIG_NAMES_MAX_NUMBER),
+ )
+
+
@@ -224,0 +241 @@ class AppConfig:
+ config_names: ConfigNamesConfig = field(default_factory=ConfigNamesConfig)
@@ -240,0 +258 @@ class AppConfig:
+ config_names=ConfigNamesConfig.from_env(),
diff --git a/services/worker/src/worker/job_runners/dataset/config_names.py b/services/worker/src/worker/job_runners/dataset/config_names.py
index b9c447be..b1045c31 100644
--- a/services/worker/src/worker/job_runners/dataset/config_names.py
+++ b/services/worker/src/worker/job_runners/dataset/config_names.py
@@ -12,0 +13 @@ from libcommon.exceptions import (
+ DatasetWithTooManyConfigsError,
@@ -30,0 +32 @@ def compute_config_names_response(
+ max_number: int,
@@ -70,0 +73,7 @@ def compute_config_names_response(
+
+ number_of_configs = len(config_name_items)
+ if number_of_configs > max_number:
+ raise DatasetWithTooManyConfigsError(
+ f"The maximun number of configs allowed is {max_number}, dataset has {number_of_configs} configs."
+ )
+
@@ -85 +94,5 @@ class DatasetConfigNamesJobRunner(DatasetCachedJobRunner):
- compute_config_names_response(dataset=self.dataset, hf_token=self.app_config.common.hf_token)
+ compute_config_names_response(
+ dataset=self.dataset,
+ hf_token=self.app_config.common.hf_token,
+ max_number=self.app_config.config_names.max_number,
+ )
diff --git a/services/worker/tests/job_runners/dataset/test_config_names.py b/services/worker/tests/job_runners/dataset/test_config_names.py
index 9116bb5d..3f0ed076 100644
--- a/services/worker/tests/job_runners/dataset/test_config_names.py
+++ b/services/worker/tests/job_runners/dataset/test_config_names.py
@@ -5,0 +6 @@ from typing import Callable
+from unittest.mock import patch
@@ -68,0 +70,27 @@ def test_compute(app_config: AppConfig, hub_public_csv: str, get_job_runner: Get
[email protected](
+ "max_number_of_configs,error_code",
+ [
+ (1, "DatasetWithTooManyConfigsError"),
+ (2, None),
+ (3, None),
+ ],
+)
+def test_compute_too_many_configs(
+ app_config: AppConfig, get_job_runner: GetJobRunner, max_number_of_configs: int, error_code: str
+) -> None:
+ dataset = "dataset"
+ configs = ["config_1", "config_2"]
+ job_runner = get_job_runner(
+ dataset,
+ replace(app_config, config_names=replace(app_config.config_names, max_number=max_number_of_configs)),
+ )
+
+ with patch("worker.job_runners.dataset.config_names.get_dataset_config_names", return_value=configs):
+ if error_code:
+ with pytest.raises(CustomError) as exc_info:
+ job_runner.compute()
+ assert exc_info.value.code == error_code
+ else:
+ assert job_runner.compute() is not None
+
+
diff --git a/tools/docker-compose-datasets-server.yml b/tools/docker-compose-datasets-server.yml
index 4f1cf859..d6b2bd92 100644
--- a/tools/docker-compose-datasets-server.yml
+++ b/tools/docker-compose-datasets-server.yml
@@ -99,0 +100 @@ services:
+ CONFIG_NAMES_MAX_NUMBER: ${CONFIG_NAMES_MAX_NUMBER-3_000}
diff --git a/tools/docker-compose-dev-datasets-server.yml b/tools/docker-compose-dev-datasets-server.yml
index 6c7e39af..c50d781f 100644
--- a/tools/docker-compose-dev-datasets-server.yml
+++ b/tools/docker-compose-dev-datasets-server.yml
@@ -103,0 +104 @@ services:
+ CONFIG_NAMES_MAX_NUMBER: ${CONFIG_NAMES_MAX_NUMBER-3_000}
|
|
364900464d575c160bd74b80e85857b3a84bb03b
|
Sylvain Lesage
| 2023-06-14T15:25:09 |
feat: 🎸 reduce the resources (#1368)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 4d70bcc0..163a431d 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -263 +263 @@ workers:
- replicas: 160
+ replicas: 20
@@ -278 +278 @@ workers:
- replicas: 12
+ replicas: 4
|
|
3dad30355148f2bf0a886014bafbff87a12719c1
|
Quentin Lhoest
| 2023-06-14T14:53:45 |
Add lock for queue based on mongodb (#1357)
|
diff --git a/libs/libcommon/src/libcommon/constants.py b/libs/libcommon/src/libcommon/constants.py
index 0a3a5494..fc0dd91e 100644
--- a/libs/libcommon/src/libcommon/constants.py
+++ b/libs/libcommon/src/libcommon/constants.py
@@ -12,0 +13 @@ QUEUE_COLLECTION_JOBS = "jobsBlue"
+QUEUE_COLLECTION_LOCKS = "locks"
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index 9b636995..14f54073 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -5,0 +6 @@ import logging
+import time
@@ -11 +12,2 @@ from operator import itemgetter
-from typing import Generic, List, Optional, Type, TypedDict, TypeVar
+from types import TracebackType
+from typing import Generic, List, Literal, Optional, Sequence, Type, TypedDict, TypeVar
@@ -15,0 +18 @@ from mongoengine import Document, DoesNotExist
+from mongoengine.errors import NotUniqueError
@@ -20,0 +24 @@ from libcommon.constants import (
+ QUEUE_COLLECTION_LOCKS,
@@ -200,0 +205,73 @@ class Job(Document):
+class Lock(Document):
+ meta = {"collection": QUEUE_COLLECTION_LOCKS, "db_alias": QUEUE_MONGOENGINE_ALIAS, "indexes": [("key", "job_id")]}
+ key = StringField(primary_key=True)
+ job_id = StringField()
+
+ created_at = DateTimeField(required=True)
+ updated_at = DateTimeField()
+
+ objects = QuerySetManager["Lock"]()
+
+
+class lock:
+ """
+ Provides a simple way of inter-worker communication using a MongoDB lock.
+ A lock is used to indicate another worker of your application that a resource
+ or working directory is currently used in a job.
+
+ Example of usage:
+
+ ```python
+ key = json.dumps({"type": job.type, "dataset": job.dataset})
+ with lock(key=key, job_id=job.pk):
+ ...
+ ```
+
+ Or using a try/except:
+
+ ```python
+ try:
+ key = json.dumps({"type": job.type, "dataset": job.dataset})
+ lock(key=key, job_id=job.pk).acquire()
+ except TimeoutError:
+ ...
+ ```
+ """
+
+ def __init__(self, key: str, job_id: str, sleeps: Sequence[float] = (0.05, 0.05, 0.05, 1, 1, 1, 5)) -> None:
+ self.key = key
+ self.job_id = job_id
+ self.sleeps = sleeps
+
+ def acquire(self) -> None:
+ try:
+ Lock(key=self.key, job_id=self.job_id, created_at=get_datetime()).save()
+ except NotUniqueError:
+ pass
+ for sleep in self.sleeps:
+ acquired = (
+ Lock.objects(key=self.key, job_id__in=[None, self.job_id]).update(
+ job_id=self.job_id, updated_at=get_datetime()
+ )
+ > 0
+ )
+ if acquired:
+ return
+ logging.debug(f"Sleep {sleep}s to acquire lock '{self.key}' for job_id='{self.job_id}'")
+ time.sleep(sleep)
+ raise TimeoutError("lock couldn't be acquired")
+
+ def release(self) -> None:
+ Lock.objects(key=self.key, job_id=self.job_id).update(job_id=None, updated_at=get_datetime())
+
+ def __enter__(self) -> "lock":
+ self.acquire()
+ return self
+
+ def __exit__(
+ self, exctype: Optional[Type[BaseException]], excinst: Optional[BaseException], exctb: Optional[TracebackType]
+ ) -> Literal[False]:
+ self.release()
+ return False
+
+
@@ -817,0 +895 @@ def _clean_queue_database() -> None:
+ Lock.drop_collection() # type: ignore
diff --git a/libs/libcommon/src/libcommon/resources.py b/libs/libcommon/src/libcommon/resources.py
index e89121ce..b91e851d 100644
--- a/libs/libcommon/src/libcommon/resources.py
+++ b/libs/libcommon/src/libcommon/resources.py
@@ -6 +6 @@ from types import TracebackType
-from typing import Optional, Type, TypeVar
+from typing import Any, Optional, Type, TypeVar
@@ -106,0 +107,5 @@ class MongoResource(Resource):
+ def __reduce__(self) -> tuple[Any, ...]:
+ # Needed to be able to use the resource in subprocesses in tests (e.g. tests/test_queue.py::test_lock).
+ # This is because the _client in not picklable.
+ return (MongoResource, (self.database, self.host, self.mongoengine_alias, self.server_selection_timeout_ms))
+
diff --git a/libs/libcommon/tests/test_queue.py b/libs/libcommon/tests/test_queue.py
index 36226b22..02b9578e 100644
--- a/libs/libcommon/tests/test_queue.py
+++ b/libs/libcommon/tests/test_queue.py
@@ -3,0 +4 @@
+import os
@@ -4,0 +6,2 @@ from datetime import datetime, timedelta
+from multiprocessing import Pool
+from pathlib import Path
@@ -12 +15 @@ from libcommon.constants import QUEUE_TTL_SECONDS
-from libcommon.queue import EmptyQueueError, Job, Queue
+from libcommon.queue import EmptyQueueError, Job, Lock, Queue, lock
@@ -385,0 +389,27 @@ def test_has_ttl_index_on_finished_at_field() -> None:
+
+
+def increment(tmp_file: Path) -> None:
+ with open(tmp_file, "r") as f:
+ current = int(f.read() or 0)
+ with open(tmp_file, "w") as f:
+ f.write(str(current + 1))
+
+
+def locked_increment(tmp_file: Path) -> None:
+ with lock(key="test_lock", job_id=str(os.getpid())):
+ increment(tmp_file)
+
+
+def test_lock(tmp_path_factory: pytest.TempPathFactory, queue_mongo_resource: QueueMongoResource) -> None:
+ tmp_file = Path(tmp_path_factory.mktemp("test_lock") / "tmp.txt")
+ tmp_file.touch()
+ max_parallel_jobs = 4
+ num_jobs = 42
+
+ with Pool(max_parallel_jobs, initializer=queue_mongo_resource.allocate) as pool:
+ pool.map(locked_increment, [tmp_file] * num_jobs)
+
+ expected = num_jobs
+ with open(tmp_file, "r") as f:
+ assert int(f.read()) == expected
+ Lock.objects(key="test_lock").delete()
|
|
ca1d155af54b1aa372d9677115d12ce94fd1021b
|
Sylvain Lesage
| 2023-06-14T11:49:04 |
feat: 🎸 run jobs killing routines with some random delay (#1364)
|
diff --git a/libs/libcommon/src/libcommon/queue.py b/libs/libcommon/src/libcommon/queue.py
index af6bcc0d..9b636995 100644
--- a/libs/libcommon/src/libcommon/queue.py
+++ b/libs/libcommon/src/libcommon/queue.py
@@ -786 +786 @@ class Queue:
- def get_zombies(self, max_seconds_without_heartbeat: int) -> List[JobInfo]:
+ def get_zombies(self, max_seconds_without_heartbeat: float) -> List[JobInfo]:
diff --git a/services/worker/src/worker/config.py b/services/worker/src/worker/config.py
index b2c2ecf9..3419e9aa 100644
--- a/services/worker/src/worker/config.py
+++ b/services/worker/src/worker/config.py
@@ -39 +39 @@ class WorkerConfig:
- heartbeat_interval_seconds: int = WORKER_HEARTBEAT_INTERVAL_SECONDS
+ heartbeat_interval_seconds: float = WORKER_HEARTBEAT_INTERVAL_SECONDS
@@ -42,2 +42,2 @@ class WorkerConfig:
- kill_long_job_interval_seconds: int = WORKER_KILL_LONG_JOB_INTERVAL_SECONDS
- kill_zombies_interval_seconds: int = WORKER_KILL_ZOMBIES_INTERVAL_SECONDS
+ kill_long_job_interval_seconds: float = WORKER_KILL_LONG_JOB_INTERVAL_SECONDS
+ kill_zombies_interval_seconds: float = WORKER_KILL_ZOMBIES_INTERVAL_SECONDS
@@ -45 +45 @@ class WorkerConfig:
- max_job_duration_seconds: int = WORKER_MAX_JOB_DURATION_SECONDS
+ max_job_duration_seconds: float = WORKER_MAX_JOB_DURATION_SECONDS
@@ -49 +49 @@ class WorkerConfig:
- sleep_seconds: int = WORKER_SLEEP_SECONDS
+ sleep_seconds: float = WORKER_SLEEP_SECONDS
@@ -59 +59 @@ class WorkerConfig:
- heartbeat_interval_seconds=env.int(
+ heartbeat_interval_seconds=env.float(
@@ -64 +64 @@ class WorkerConfig:
- kill_long_job_interval_seconds=env.int(
+ kill_long_job_interval_seconds=env.float(
@@ -67 +67 @@ class WorkerConfig:
- kill_zombies_interval_seconds=env.int(
+ kill_zombies_interval_seconds=env.float(
@@ -71 +71 @@ class WorkerConfig:
- max_job_duration_seconds=env.int(
+ max_job_duration_seconds=env.float(
@@ -77 +77 @@ class WorkerConfig:
- sleep_seconds=env.int(name="SLEEP_SECONDS", default=WORKER_SLEEP_SECONDS),
+ sleep_seconds=env.float(name="SLEEP_SECONDS", default=WORKER_SLEEP_SECONDS),
diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py
index a0acf747..d663a775 100644
--- a/services/worker/src/worker/executor.py
+++ b/services/worker/src/worker/executor.py
@@ -8 +8,2 @@ from datetime import datetime, timedelta
-from typing import Any, Callable, Optional
+from random import random
+from typing import Any, Callable, Optional, Tuple, Union
@@ -27 +28,5 @@ async def every(
- func: Callable[..., Optional[Any]], *args: Any, seconds: int, stop_on: Optional[Any] = None, **kwargs: Any
+ func: Callable[..., Optional[Any]],
+ *args: Any,
+ seconds: Union[float, Tuple[float, float]],
+ stop_on: Optional[Any] = None,
+ **kwargs: Any,
@@ -33 +38,4 @@ async def every(
- await asyncio.sleep(seconds)
+ delay = (
+ seconds[0] + (seconds[1] - seconds[0]) * random() if isinstance(seconds, tuple) else seconds # nosec B311
+ )
+ await asyncio.sleep(delay)
@@ -81 +89,9 @@ class WorkerExecutor:
- loop.create_task(every(self.kill_zombies, seconds=self.app_config.worker.kill_zombies_interval_seconds))
+ loop.create_task(
+ every(
+ self.kill_zombies,
+ seconds=(
+ self.app_config.worker.kill_zombies_interval_seconds * 0.5,
+ self.app_config.worker.kill_zombies_interval_seconds * 1.5,
+ ),
+ )
+ )
@@ -86 +102,4 @@ class WorkerExecutor:
- seconds=self.app_config.worker.kill_long_job_interval_seconds,
+ seconds=(
+ self.app_config.worker.kill_long_job_interval_seconds * 0.5,
+ self.app_config.worker.kill_long_job_interval_seconds * 1.5,
+ ),
@@ -90 +109 @@ class WorkerExecutor:
- every(self.is_worker_alive, worker_loop_executor=worker_loop_executor, seconds=1, stop_on=False)
+ every(self.is_worker_alive, worker_loop_executor=worker_loop_executor, seconds=1.0, stop_on=False)
|
|
7bebee49ebf214eefb6410241fd544dbb4454aa4
|
Sylvain Lesage
| 2023-06-14T09:40:52 |
ci: 🎡 don't run doc CI if no doc has been changed (#1363)
|
diff --git a/.github/workflows/doc-pr-build.yml b/.github/workflows/doc-pr-build.yml
index 9c920c4f..6ee6a305 100644
--- a/.github/workflows/doc-pr-build.yml
+++ b/.github/workflows/doc-pr-build.yml
@@ -7,0 +8,3 @@ on:
+ paths:
+ - "docs/**"
+ - ".github/workflows/doc-pr-build.yml"
diff --git a/.github/workflows/doc-pr-delete-trigger.yml b/.github/workflows/doc-pr-delete-trigger.yml
index ebc0d0ad..89037f89 100644
--- a/.github/workflows/doc-pr-delete-trigger.yml
+++ b/.github/workflows/doc-pr-delete-trigger.yml
@@ -8,0 +9,3 @@ on:
+ paths:
+ - "docs/**"
+ - ".github/workflows/doc-pr-delete-trigger.yml"
|
|
7b266d24ee81c2f5b86fbd352466e97b34d46c99
|
Sylvain Lesage
| 2023-06-14T09:25:58 |
style: 💄 fix style for github action files (#1362)
|
diff --git a/.github/workflows/_e2e_tests.yml b/.github/workflows/_e2e_tests.yml
index 3e0a01f5..0126945c 100644
--- a/.github/workflows/_e2e_tests.yml
+++ b/.github/workflows/_e2e_tests.yml
@@ -21,2 +21 @@ jobs:
- -
- name: Checkout
+ - name: Checkout
@@ -24,2 +23 @@ jobs:
- -
- name: Build and launch the services (no cache)
+ - name: Build and launch the services (no cache)
@@ -48,2 +46 @@ jobs:
- -
- name: Install poetry
+ - name: Install poetry
@@ -51,2 +48 @@ jobs:
- -
- name: Use Python
+ - name: Use Python
@@ -56 +52 @@ jobs:
- cache: 'poetry'
+ cache: "poetry"
@@ -59,2 +55 @@ jobs:
- -
- name: Install dependencies
+ - name: Install dependencies
@@ -64,2 +59 @@ jobs:
- -
- name: End-to-end tests
+ - name: End-to-end tests
@@ -85 +79 @@ jobs:
- WORKER_SLEEP_SECONDS : "1"
+ WORKER_SLEEP_SECONDS: "1"
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml
index 9f9638e0..c285df42 100644
--- a/.github/workflows/_quality-python.yml
+++ b/.github/workflows/_quality-python.yml
@@ -5 +5 @@ name: Check Python code quality
-on:
+on:
@@ -31 +31 @@ jobs:
- cache: 'poetry'
+ cache: "poetry"
diff --git a/.github/workflows/_unit-tests-python.yml b/.github/workflows/_unit-tests-python.yml
index 78d2daef..763748bf 100644
--- a/.github/workflows/_unit-tests-python.yml
+++ b/.github/workflows/_unit-tests-python.yml
@@ -5 +5 @@ name: Launch Python unit tests
-on:
+on:
@@ -32 +32 @@ jobs:
- cache: 'poetry'
+ cache: "poetry"
@@ -48 +48 @@ jobs:
- METRICS_MONGO_URL: mongodb://localhost:${{ env.mongo-port }}
+ METRICS_MONGO_URL: mongodb://localhost:${{ env.mongo-port }}
diff --git a/.github/workflows/cd.yml b/.github/workflows/cd.yml
index d938a7f6..878e3e3b 100644
--- a/.github/workflows/cd.yml
+++ b/.github/workflows/cd.yml
@@ -89 +89 @@ jobs:
- needs: [ build-and-push-images ]
+ needs: [build-and-push-images]
diff --git a/.github/workflows/chart-pr.yml b/.github/workflows/chart-pr.yml
index 6e33e252..cd5789bc 100644
--- a/.github/workflows/chart-pr.yml
+++ b/.github/workflows/chart-pr.yml
@@ -8,2 +8,2 @@ on:
- - 'chart/**'
- - '.github/workflows/chart-pr.yml'
+ - "chart/**"
+ - ".github/workflows/chart-pr.yml"
diff --git a/.github/workflows/doc-build.yml b/.github/workflows/doc-build.yml
index 7378ed14..8dc7ac99 100644
--- a/.github/workflows/doc-build.yml
+++ b/.github/workflows/doc-build.yml
@@ -11,2 +11,2 @@ on:
- - 'docs/**'
- - '.github/workflows/doc-build.yml'
+ - "docs/**"
+ - ".github/workflows/doc-build.yml"
@@ -15 +15 @@ jobs:
- build:
+ build:
diff --git a/.github/workflows/doc-pr-build.yml b/.github/workflows/doc-pr-build.yml
index 7dc41015..9c920c4f 100644
--- a/.github/workflows/doc-pr-build.yml
+++ b/.github/workflows/doc-pr-build.yml
@@ -8 +7,0 @@ on:
-
diff --git a/.github/workflows/doc-pr-delete-trigger.yml b/.github/workflows/doc-pr-delete-trigger.yml
index b23e2591..ebc0d0ad 100644
--- a/.github/workflows/doc-pr-delete-trigger.yml
+++ b/.github/workflows/doc-pr-delete-trigger.yml
@@ -8,2 +8 @@ on:
- types: [ closed ]
-
+ types: [closed]
diff --git a/.github/workflows/doc-pr-upload.yml b/.github/workflows/doc-pr-upload.yml
index 517a69ca..99c3a4cf 100644
--- a/.github/workflows/doc-pr-upload.yml
+++ b/.github/workflows/doc-pr-upload.yml
@@ -19 +19 @@ jobs:
- comment_bot_token: ${{ secrets.COMMENT_BOT_TOKEN }}
\ No newline at end of file
+ comment_bot_token: ${{ secrets.COMMENT_BOT_TOKEN }}
diff --git a/.github/workflows/e2e.yml b/.github/workflows/e2e.yml
index 81a6ad18..3fbd1662 100644
--- a/.github/workflows/e2e.yml
+++ b/.github/workflows/e2e.yml
@@ -11,9 +11,9 @@ on:
- - 'e2e/**'
- - 'libs/**'
- - 'services/**'
- - 'chart/static-files/openapi.json'
- - '.github/workflows/_e2e_tests.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/e2e.yml'
- - 'tools/Python.mk'
- - 'tools/docker-compose-datasets-server.yml'
+ - "e2e/**"
+ - "libs/**"
+ - "services/**"
+ - "chart/static-files/openapi.json"
+ - ".github/workflows/_e2e_tests.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/e2e.yml"
+ - "tools/Python.mk"
+ - "tools/docker-compose-datasets-server.yml"
@@ -22,9 +22,9 @@ on:
- - 'e2e/**'
- - 'libs/**'
- - 'services/**'
- - 'chart/static-files/openapi.json'
- - '.github/workflows/_e2e_tests.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/e2e.yml'
- - 'tools/Python.mk'
- - 'tools/docker-compose-datasets-server.yml'
+ - "e2e/**"
+ - "libs/**"
+ - "services/**"
+ - "chart/static-files/openapi.json"
+ - ".github/workflows/_e2e_tests.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/e2e.yml"
+ - "tools/Python.mk"
+ - "tools/docker-compose-datasets-server.yml"
diff --git a/.github/workflows/j-cache-maintenance.yml b/.github/workflows/j-cache-maintenance.yml
index 74052dcf..f0ad6474 100644
--- a/.github/workflows/j-cache-maintenance.yml
+++ b/.github/workflows/j-cache-maintenance.yml
@@ -11,6 +11,6 @@ on:
- - 'jobs/cache_maintenance/**'
- - 'libs/libcommon/**'
- - '.github/workflows/j-cache-maintenance.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "jobs/cache_maintenance/**"
+ - "libs/libcommon/**"
+ - ".github/workflows/j-cache-maintenance.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
@@ -19,6 +19,6 @@ on:
- - 'jobs/cache_maintenance/**'
- - 'libs/libcommon/**'
- - '.github/workflows/j-cache-maintenance.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "jobs/cache_maintenance/**"
+ - "libs/libcommon/**"
+ - ".github/workflows/j-cache-maintenance.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
diff --git a/.github/workflows/j-migration.yml b/.github/workflows/j-migration.yml
index 201eaea8..40ffa4ed 100644
--- a/.github/workflows/j-migration.yml
+++ b/.github/workflows/j-migration.yml
@@ -11,6 +11,6 @@ on:
- - 'jobs/mongodb_migration/**'
- - 'libs/libcommon/**'
- - '.github/workflows/j-migration.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "jobs/mongodb_migration/**"
+ - "libs/libcommon/**"
+ - ".github/workflows/j-migration.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
@@ -19,6 +19,6 @@ on:
- - 'jobs/mongodb_migration/**'
- - 'libs/libcommon/**'
- - '.github/workflows/j-migration.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "jobs/mongodb_migration/**"
+ - "libs/libcommon/**"
+ - ".github/workflows/j-migration.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
diff --git a/.github/workflows/l-libcommon.yml b/.github/workflows/l-libcommon.yml
index 3ba6a069..96a8c1ae 100644
--- a/.github/workflows/l-libcommon.yml
+++ b/.github/workflows/l-libcommon.yml
@@ -11,5 +11,5 @@ on:
- - 'libs/libcommon/**'
- - '.github/workflows/l-libcommon.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "libs/libcommon/**"
+ - ".github/workflows/l-libcommon.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
@@ -18,5 +18,5 @@ on:
- - 'libs/libcommon/**'
- - '.github/workflows/l-libcommon.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "libs/libcommon/**"
+ - ".github/workflows/l-libcommon.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
diff --git a/.github/workflows/openapi-spec.yml b/.github/workflows/openapi-spec.yml
index a8062cfc..bfae8dcf 100644
--- a/.github/workflows/openapi-spec.yml
+++ b/.github/workflows/openapi-spec.yml
@@ -11,2 +11,2 @@ on:
- - 'chart/static-files/opanapi.json'
- - '.github/workflows/openapi.yml'
+ - "chart/static-files/opanapi.json"
+ - ".github/workflows/openapi.yml"
@@ -15,2 +15,2 @@ on:
- - 'chart/static-files/opanapi.json'
- - '.github/workflows/openapi.yml'
+ - "chart/static-files/opanapi.json"
+ - ".github/workflows/openapi.yml"
@@ -38 +38 @@ jobs:
- cache: 'poetry'
+ cache: "poetry"
diff --git a/.github/workflows/publish-helm.yml b/.github/workflows/publish-helm.yml
index fe9b8997..0c176e80 100644
--- a/.github/workflows/publish-helm.yml
+++ b/.github/workflows/publish-helm.yml
@@ -9 +9 @@ on:
- - 'chart/**'
+ - "chart/**"
diff --git a/.github/workflows/s-admin.yml b/.github/workflows/s-admin.yml
index 7dd4c411..0be67500 100644
--- a/.github/workflows/s-admin.yml
+++ b/.github/workflows/s-admin.yml
@@ -11,6 +11,6 @@ on:
- - 'libs/libcommon/**'
- - 'services/admin/**'
- - '.github/workflows/s-admin.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "libs/libcommon/**"
+ - "services/admin/**"
+ - ".github/workflows/s-admin.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
@@ -19,6 +19,6 @@ on:
- - 'libs/libcommon/**'
- - 'services/admin/**'
- - '.github/workflows/s-admin.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "libs/libcommon/**"
+ - "services/admin/**"
+ - ".github/workflows/s-admin.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
diff --git a/.github/workflows/s-api.yml b/.github/workflows/s-api.yml
index 972d6639..63c440cc 100644
--- a/.github/workflows/s-api.yml
+++ b/.github/workflows/s-api.yml
@@ -11,6 +11,6 @@ on:
- - 'libs/libcommon/**'
- - 'services/api/**'
- - '.github/workflows/s-api.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "libs/libcommon/**"
+ - "services/api/**"
+ - ".github/workflows/s-api.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
@@ -19,6 +19,6 @@ on:
- - 'libs/libcommon/**'
- - 'services/api/**'
- - '.github/workflows/s-api.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
+ - "libs/libcommon/**"
+ - "services/api/**"
+ - ".github/workflows/s-api.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
diff --git a/.github/workflows/s-worker.yml b/.github/workflows/s-worker.yml
index ffb3be37..146a6bcf 100644
--- a/.github/workflows/s-worker.yml
+++ b/.github/workflows/s-worker.yml
@@ -11,7 +11,7 @@ on:
- - 'libs/libcommon/**'
- - 'services/worker/**'
- - '.github/workflows/s-worker.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
- - 'vendors/'
+ - "libs/libcommon/**"
+ - "services/worker/**"
+ - ".github/workflows/s-worker.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
+ - "vendors/"
@@ -20,7 +20,7 @@ on:
- - 'libs/libcommon/**'
- - 'services/worker/**'
- - '.github/workflows/s-worker.yml'
- - '.github/workflows/_quality-python.yml'
- - '.github/workflows/_unit-tests-python.yml'
- - 'tools/docker-compose-mongo.yml'
- - 'vendors/'
+ - "libs/libcommon/**"
+ - "services/worker/**"
+ - ".github/workflows/s-worker.yml"
+ - ".github/workflows/_quality-python.yml"
+ - ".github/workflows/_unit-tests-python.yml"
+ - "tools/docker-compose-mongo.yml"
+ - "vendors/"
diff --git a/.github/workflows/stale.yml b/.github/workflows/stale.yml
index ede171dc..9586e941 100644
--- a/.github/workflows/stale.yml
+++ b/.github/workflows/stale.yml
@@ -18 +18 @@ jobs:
- - uses: actions/checkout@v3
+ - uses: actions/checkout@v3
@@ -20,4 +20,4 @@ jobs:
- - name: Setup Python
- uses: actions/setup-python@v4
- with:
- python-version: 3.7
+ - name: Setup Python
+ uses: actions/setup-python@v4
+ with:
+ python-version: 3.7
@@ -25,6 +25,6 @@ jobs:
- - name: Install requirements
- run: |
- pip install PyGithub
- - name: Close stale issues
- run: |
- python tools/stale.py
+ - name: Install requirements
+ run: |
+ pip install PyGithub
+ - name: Close stale issues
+ run: |
+ python tools/stale.py
|
|
7b4b78f04473d8f00e7c2226381982a9a84ab48c
|
Sylvain Lesage
| 2023-06-13T22:27:15 |
feat: 🎸 add resources for temporary load (#1360)
|
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 163a431d..4d70bcc0 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -263 +263 @@ workers:
- replicas: 20
+ replicas: 160
@@ -278 +278 @@ workers:
- replicas: 4
+ replicas: 12
|
|
2990bbc93b6622d536bb8bb50ea7fdfa72b17112
|
Steven Liu
| 2023-06-12T18:40:42 |
separate page for each library (#1354)
|
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml
index dcd3ef45..0767f5ab 100644
--- a/docs/source/_toctree.yml
+++ b/docs/source/_toctree.yml
@@ -19,2 +19,10 @@
- - local: parquet_process
- title: Query datasets from Datasets Server
+ - title: Query datasets from Datasets Server
+ sections:
+ - local: parquet_process
+ title: Overview
+ - local: duckdb
+ title: DuckDB
+ - local: polars
+ title: Polars
+ - local: pandas
+ title: Pandas
diff --git a/docs/source/duckdb.mdx b/docs/source/duckdb.mdx
new file mode 100644
index 00000000..b55fcd6c
--- /dev/null
+++ b/docs/source/duckdb.mdx
@@ -0,0 +1,90 @@
+# DuckDB
+
+[DuckDB](https://duckdb.org/docs/) is a database that supports reading and querying Parquet files really fast. Begin by creating a connection to DuckDB, and then install and load the [`httpfs`](https://duckdb.org/docs/extensions/httpfs.html) extension to read and write remote files:
+
+<inferencesnippet>
+<python>
+```py
+import duckdb
+
+url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet"
+
+con = duckdb.connect()
+con.execute("INSTALL httpfs;")
+con.execute("LOAD httpfs;")
+```
+</python>
+<js>
+```js
+var duckdb = require('duckdb');
+var db = new duckdb.Database(':memory:');
+var con = db.connect();
+con.exec('INSTALL httpfs');
+con.exec('LOAD httpfs');
+
+const url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet"
+```
+</js>
+</inferencesnippet>
+
+Now you can write and execute your SQL query on the Parquet file:
+
+<inferencesnippet>
+<python>
+```py
+con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '{url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)")
+┌───────────┬──────────────┬────────────────────┐
+│ horoscope │ count_star() │ avg_blog_length │
+│ varchar │ int64 │ double │
+├───────────┼──────────────┼────────────────────┤
+│ Aquarius │ 34062 │ 1129.218836239798 │
+│ Cancer │ 41509 │ 1098.366812016671 │
+│ Capricorn │ 33961 │ 1073.2002002296751 │
+│ Libra │ 40302 │ 1072.0718326633914 │
+│ Leo │ 40587 │ 1064.0536871412028 │
+└───────────┴──────────────┴────────────────────┘
+```
+</python>
+<js>
+```js
+con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '${url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) {
+ if (err) {
+ throw err;
+ }
+ console.log(res)
+});
+```
+</js>
+</inferencesnippet>
+
+To query multiple files - for example, if the dataset is sharded:
+
+<inferencesnippet>
+<python>
+```py
+con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet({urls[:2]}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)")
+┌─────────────┬──────────────┬────────────────────┐
+│ horoscope │ count_star() │ avg_blog_length │
+│ varchar │ int64 │ double │
+├─────────────┼──────────────┼────────────────────┤
+│ Aquarius │ 49568 │ 1125.8306770497095 │
+│ Cancer │ 63512 │ 1097.95608703867 │
+│ Libra │ 60304 │ 1060.6110539931017 │
+│ Capricorn │ 49402 │ 1059.5552609206104 │
+│ Sagittarius │ 50431 │ 1057.4589835616982 │
+└─────────────┴──────────────┴────────────────────┘
+```
+</python>
+<js>
+```js
+con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet(${JSON.stringify(urls)}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) {
+ if (err) {
+ throw err;
+ }
+ console.log(res)
+});
+```
+</js>
+</inferencesnippet>
+
+[DuckDB-Wasm](https://duckdb.org/docs/api/wasm), a package powered by [WebAssembly](https://webassembly.org/), is also available for running DuckDB in any browser. This could be useful, for instance, if you want to create a web app to query Parquet files from the browser!
diff --git a/docs/source/pandas.mdx b/docs/source/pandas.mdx
new file mode 100644
index 00000000..f15c4d29
--- /dev/null
+++ b/docs/source/pandas.mdx
@@ -0,0 +1,29 @@
+# Pandas
+
+[Pandas](https://pandas.pydata.org/docs/index.html) is a popular DataFrame library for data analysis.
+
+To read from a single Parquet file, use the [`read_parquet`](https://pandas.pydata.org/docs/reference/api/pandas.read_parquet.html) function to read it into a DataFrame:
+
+```py
+import pandas as pd
+
+df = (
+ pd.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet")
+ .groupby('horoscope')['text']
+ .apply(lambda x: x.str.len().mean())
+ .sort_values(ascending=False)
+ .head(5)
+)
+```
+
+To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pandas.pydata.org/docs/reference/api/pandas.concat.html) function to concatenate the files into a single DataFrame:
+
+```py
+df = (
+ pd.concat([pd.read_parquet(url) for url in urls])
+ .groupby('horoscope')['text']
+ .apply(lambda x: x.str.len().mean())
+ .sort_values(ascending=False)
+ .head(5)
+)
+```
\ No newline at end of file
diff --git a/docs/source/parquet_process.mdx b/docs/source/parquet_process.mdx
index 40798731..1ace2523 100644
--- a/docs/source/parquet_process.mdx
+++ b/docs/source/parquet_process.mdx
@@ -1 +1 @@
-# Query datasets from Datasets Server
+# Overview
@@ -3 +3 @@
-Datasets Server automatically converts and publishes datasets on the Hub as Parquet files. [Parquet](https://parquet.apache.org/docs/) files are column-based and they shine when you're working with big data. There are several ways you can work with Parquet files, and this guide will show you how to:
+Datasets Server automatically converts and publishes public datasets less than 5GB on the Hub as Parquet files. [Parquet](https://parquet.apache.org/docs/) files are column-based and they shine when you're working with big data. There are several different libraries you can use to work with the published Parquet files:
@@ -5,231 +5,3 @@ Datasets Server automatically converts and publishes datasets on the Hub as Parq
-- read and query Parquet files with Pandas and Polars
-- connect, read and query Parquet files with DuckDB and DuckDB-Wasm
-
-## Polars
-
-[Polars](https://pola-rs.github.io/polars-book/user-guide/introduction.html) is a fast DataFrame library written in Rust with Arrow as its foundation.
-
-<Tip>
-
-💡 Learn more about how to get the dataset URLs in the [List Parquet files](parquet) guide.
-
-</Tip>
-
-Let's start by grabbing the URLs to the `train` split of the [`blog_authorship_corpus`](https://huggingface.co/datasets/blog_authorship_corpus) dataset from Datasets Server:
-
-```py
-r = requests.get("https://datasets-server.huggingface.co/parquet?dataset=blog_authorship_corpus")
-j = r.json()
-urls = [f['url'] for f in j['parquet_files'] if f['split'] == 'train']
-urls
-['https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet',
- 'https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00001-of-00002.parquet']
-```
-
-To read from a single Parquet file, use the [`read_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.read_parquet.html) function to read it into a DataFrame and then execute your query:
-
-```py
-import polars as pl
-
-df = (
- pl.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet")
- .groupby("horoscope")
- .agg(
- [
- pl.count(),
- pl.col("text").str.n_chars().mean().alias("avg_blog_length")
- ]
- )
- .sort("avg_blog_length", descending=True)
- .limit(5)
-)
-print(df)
-shape: (5, 3)
-┌───────────┬───────┬─────────────────┐
-│ horoscope ┆ count ┆ avg_blog_length │
-│ --- ┆ --- ┆ --- │
-│ str ┆ u32 ┆ f64 │
-╞═══════════╪═══════╪═════════════════╡
-│ Aquarius ┆ 34062 ┆ 1129.218836 │
-│ Cancer ┆ 41509 ┆ 1098.366812 │
-│ Capricorn ┆ 33961 ┆ 1073.2002 │
-│ Libra ┆ 40302 ┆ 1072.071833 │
-│ Leo ┆ 40587 ┆ 1064.053687 │
-└───────────┴───────┴─────────────────┘
-```
-
-To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.concat.html) function to concatenate the files into a single DataFrame:
-
-```py
-import polars as pl
-df = (
- pl.concat([pl.read_parquet(url) for url in urls])
- .groupby("horoscope")
- .agg(
- [
- pl.count(),
- pl.col("text").str.n_chars().mean().alias("avg_blog_length")
- ]
- )
- .sort("avg_blog_length", descending=True)
- .limit(5)
-)
-print(df)
-shape: (5, 3)
-┌─────────────┬───────┬─────────────────┐
-│ horoscope ┆ count ┆ avg_blog_length │
-│ --- ┆ --- ┆ --- │
-│ str ┆ u32 ┆ f64 │
-╞═════════════╪═══════╪═════════════════╡
-│ Aquarius ┆ 49568 ┆ 1125.830677 │
-│ Cancer ┆ 63512 ┆ 1097.956087 │
-│ Libra ┆ 60304 ┆ 1060.611054 │
-│ Capricorn ┆ 49402 ┆ 1059.555261 │
-│ Sagittarius ┆ 50431 ┆ 1057.458984 │
-└─────────────┴───────┴─────────────────┘
-```
-
-### Lazy API
-
-Polars offers a [lazy API](https://pola-rs.github.io/polars-book/user-guide/lazy-api/intro.html) that is more performant and memory-efficient for large Parquet files. The LazyFrame API keeps track of what you want to do, and it'll only execute the entire query when you're ready. This way, the lazy API doesn't load everything into RAM beforehand, and it allows you to work with datasets larger than your available RAM.
-
-To lazily read a Parquet file, use the [`scan_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.scan_parquet.html) function instead. Then, execute the entire query with the [`collect`](https://pola-rs.github.io/polars/py-polars/html/reference/lazyframe/api/polars.LazyFrame.collect.html) function:
-
-```py
-import polars as pl
-
-q = (
- pl.scan_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet")
- .groupby("horoscope")
- .agg(
- [
- pl.count(),
- pl.col("text").str.n_chars().mean().alias("avg_blog_length")
- ]
- )
- .sort("avg_blog_length", descending=True)
- .limit(5)
-)
-df = q.collect()
-```
-
-## Pandas
-
-You can also use the popular Pandas DataFrame library to read Parquet files.
-
-To read from a single Parquet file, use the [`read_parquet`](https://pandas.pydata.org/docs/reference/api/pandas.read_parquet.html) function to read it into a DataFrame:
-
-```py
-import pandas as pd
-
-df = (
- pd.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet")
- .groupby('horoscope')['text']
- .apply(lambda x: x.str.len().mean())
- .sort_values(ascending=False)
- .head(5)
-)
-```
-
-To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pandas.pydata.org/docs/reference/api/pandas.concat.html) function to concatenate the files into a single DataFrame:
-
-```py
-df = (
- pd.concat([pd.read_parquet(url) for url in urls])
- .groupby('horoscope')['text']
- .apply(lambda x: x.str.len().mean())
- .sort_values(ascending=False)
- .head(5)
-)
-```
-
-## DuckDB
-
-[DuckDB](https://duckdb.org/docs/) is a database that supports reading and querying Parquet files really fast. Begin by creating a connection to DuckDB, and then install and load the [`httpfs`](https://duckdb.org/docs/extensions/httpfs.html) extension to read and write remote files:
-
-<inferencesnippet>
-<python>
-```py
-import duckdb
-
-url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet"
-
-con = duckdb.connect()
-con.execute("INSTALL httpfs;")
-con.execute("LOAD httpfs;")
-```
-</python>
-<js>
-```js
-var duckdb = require('duckdb');
-var db = new duckdb.Database(':memory:');
-var con = db.connect();
-con.exec('INSTALL httpfs');
-con.exec('LOAD httpfs');
-
-const url = "https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet"
-```
-</js>
-</inferencesnippet>
-
-Now you can write and execute your SQL query on the Parquet file:
-
-<inferencesnippet>
-<python>
-```py
-con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '{url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)")
-┌───────────┬──────────────┬────────────────────┐
-│ horoscope │ count_star() │ avg_blog_length │
-│ varchar │ int64 │ double │
-├───────────┼──────────────┼────────────────────┤
-│ Aquarius │ 34062 │ 1129.218836239798 │
-│ Cancer │ 41509 │ 1098.366812016671 │
-│ Capricorn │ 33961 │ 1073.2002002296751 │
-│ Libra │ 40302 │ 1072.0718326633914 │
-│ Leo │ 40587 │ 1064.0536871412028 │
-└───────────┴──────────────┴────────────────────┘
-```
-</python>
-<js>
-```js
-con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM '${url}' GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) {
- if (err) {
- throw err;
- }
- console.log(res)
-});
-```
-</js>
-</inferencesnippet>
-
-To query multiple files - for example, if the dataset is sharded:
-
-<inferencesnippet>
-<python>
-```py
-con.sql(f"SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet({urls[:2]}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)")
-┌─────────────┬──────────────┬────────────────────┐
-│ horoscope │ count_star() │ avg_blog_length │
-│ varchar │ int64 │ double │
-├─────────────┼──────────────┼────────────────────┤
-│ Aquarius │ 49568 │ 1125.8306770497095 │
-│ Cancer │ 63512 │ 1097.95608703867 │
-│ Libra │ 60304 │ 1060.6110539931017 │
-│ Capricorn │ 49402 │ 1059.5552609206104 │
-│ Sagittarius │ 50431 │ 1057.4589835616982 │
-└─────────────┴──────────────┴────────────────────┘
-```
-</python>
-<js>
-```js
-con.all(`SELECT horoscope, count(*), AVG(LENGTH(text)) AS avg_blog_length FROM read_parquet(${JSON.stringify(urls)}) GROUP BY horoscope ORDER BY avg_blog_length DESC LIMIT(5)`, function(err, res) {
- if (err) {
- throw err;
- }
- console.log(res)
-});
-```
-</js>
-</inferencesnippet>
-
-[DuckDB-Wasm](https://duckdb.org/docs/api/wasm), a package powered by [WebAssembly](https://webassembly.org/), is also available for running DuckDB in any browser. This could be useful, for instance, if you want to create a web app to query Parquet files from the browser!
+- [Polars](https://pola-rs.github.io/polars-book/user-guide/), a Rust based DataFrame library
+- [Pandas](https://pandas.pydata.org/docs/index.html), a data analysis tool for working with data structures
+- [DuckDB](https://duckdb.org/docs/), a high-performance SQL database for analytical queries
\ No newline at end of file
diff --git a/docs/source/polars.mdx b/docs/source/polars.mdx
new file mode 100644
index 00000000..a02c2167
--- /dev/null
+++ b/docs/source/polars.mdx
@@ -0,0 +1,107 @@
+# Polars
+
+[Polars](https://pola-rs.github.io/polars-book/user-guide/) is a fast DataFrame library written in Rust with Arrow as its foundation.
+
+<Tip>
+
+💡 Learn more about how to get the dataset URLs in the [List Parquet files](parquet) guide.
+
+</Tip>
+
+Let's start by grabbing the URLs to the `train` split of the [`blog_authorship_corpus`](https://huggingface.co/datasets/blog_authorship_corpus) dataset from Datasets Server:
+
+```py
+r = requests.get("https://datasets-server.huggingface.co/parquet?dataset=blog_authorship_corpus")
+j = r.json()
+urls = [f['url'] for f in j['parquet_files'] if f['split'] == 'train']
+urls
+['https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet',
+ 'https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00001-of-00002.parquet']
+```
+
+To read from a single Parquet file, use the [`read_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.read_parquet.html) function to read it into a DataFrame and then execute your query:
+
+```py
+import polars as pl
+
+df = (
+ pl.read_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet")
+ .groupby("horoscope")
+ .agg(
+ [
+ pl.count(),
+ pl.col("text").str.n_chars().mean().alias("avg_blog_length")
+ ]
+ )
+ .sort("avg_blog_length", descending=True)
+ .limit(5)
+)
+print(df)
+shape: (5, 3)
+┌───────────┬───────┬─────────────────┐
+│ horoscope ┆ count ┆ avg_blog_length │
+│ --- ┆ --- ┆ --- │
+│ str ┆ u32 ┆ f64 │
+╞═══════════╪═══════╪═════════════════╡
+│ Aquarius ┆ 34062 ┆ 1129.218836 │
+│ Cancer ┆ 41509 ┆ 1098.366812 │
+│ Capricorn ┆ 33961 ┆ 1073.2002 │
+│ Libra ┆ 40302 ┆ 1072.071833 │
+│ Leo ┆ 40587 ┆ 1064.053687 │
+└───────────┴───────┴─────────────────┘
+```
+
+To read multiple Parquet files - for example, if the dataset is sharded - you'll need to use the [`concat`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.concat.html) function to concatenate the files into a single DataFrame:
+
+```py
+import polars as pl
+df = (
+ pl.concat([pl.read_parquet(url) for url in urls])
+ .groupby("horoscope")
+ .agg(
+ [
+ pl.count(),
+ pl.col("text").str.n_chars().mean().alias("avg_blog_length")
+ ]
+ )
+ .sort("avg_blog_length", descending=True)
+ .limit(5)
+)
+print(df)
+shape: (5, 3)
+┌─────────────┬───────┬─────────────────┐
+│ horoscope ┆ count ┆ avg_blog_length │
+│ --- ┆ --- ┆ --- │
+│ str ┆ u32 ┆ f64 │
+╞═════════════╪═══════╪═════════════════╡
+│ Aquarius ┆ 49568 ┆ 1125.830677 │
+│ Cancer ┆ 63512 ┆ 1097.956087 │
+│ Libra ┆ 60304 ┆ 1060.611054 │
+│ Capricorn ┆ 49402 ┆ 1059.555261 │
+│ Sagittarius ┆ 50431 ┆ 1057.458984 │
+└─────────────┴───────┴─────────────────┘
+```
+
+## Lazy API
+
+Polars offers a [lazy API](https://pola-rs.github.io/polars-book/user-guide/lazy/using/) that is more performant and memory-efficient for large Parquet files. The LazyFrame API keeps track of what you want to do, and it'll only execute the entire query when you're ready. This way, the lazy API doesn't load everything into RAM beforehand, and it allows you to work with datasets larger than your available RAM.
+
+To lazily read a Parquet file, use the [`scan_parquet`](https://pola-rs.github.io/polars/py-polars/html/reference/api/polars.scan_parquet.html) function instead. Then, execute the entire query with the [`collect`](https://pola-rs.github.io/polars/py-polars/html/reference/lazyframe/api/polars.LazyFrame.collect.html) function:
+
+```py
+import polars as pl
+
+q = (
+ pl.scan_parquet("https://huggingface.co/datasets/blog_authorship_corpus/resolve/refs%2Fconvert%2Fparquet/blog_authorship_corpus/blog_authorship_corpus-train-00000-of-00002.parquet")
+ .groupby("horoscope")
+ .agg(
+ [
+ pl.count(),
+ pl.col("text").str.n_chars().mean().alias("avg_blog_length")
+ ]
+ )
+ .sort("avg_blog_length", descending=True)
+ .limit(5)
+)
+df = q.collect()
+```
\ No newline at end of file
|
|
8c53732d93921822a96c9b256fd26f449519b069
|
Sylvain Lesage
| 2023-06-12T16:29:05 |
refactor: 💡 rename required_by_dataset_viewer: enables_preview (#1353)
|
diff --git a/chart/static-files/openapi.json b/chart/static-files/openapi.json
index 296ba170..11b6b1fe 100644
--- a/chart/static-files/openapi.json
+++ b/chart/static-files/openapi.json
@@ -293,4 +293 @@
- "enum": [
- "RowsPostProcessingError",
- "UnexpectedError"
- ]
+ "enum": ["RowsPostProcessingError", "UnexpectedError"]
@@ -899 +896 @@
- "required": ["valid"],
+ "required": ["preview", "viewer", "valid"],
@@ -901 +898,5 @@
- "valid": {
+ "preview": {
+ "type": "array",
+ "items": { "type": "string" }
+ },
+ "viewer": {
@@ -903,0 +905,5 @@
+ },
+ "valid": {
+ "type": "array",
+ "items": { "type": "string" },
+ "deprecated": true
@@ -2216 +2222 @@
- "maximum": 100,
+ "maximum": 100,
@@ -3149 +3155 @@
- "description": "The list of the Hub datasets that work without an error (for /splits and /first-rows).",
+ "description": "The lists of the Hub datasets that work without an error, by type. It lists the datasets that have a Dataset Viewer (i.e. have been converted to Parquet format, and can be paginated) and the datasets that only have the Dataset Preview (the first 100 rows).",
@@ -3151 +3157 @@
- "description": "See Valid datasets (Hub docs)",
+ "description": "See Valid datasets (Hub docs)",
@@ -3175,0 +3182,9 @@
+ "preview": [
+ "0n1xus/codexglue",
+ "0x7194633/rupile",
+ "AHussain0418/day2_data"
+ ],
+ "viewer": [
+ "0n1xus/pytorrent-standalone",
+ "51la5/keyword-extraction"
+ ],
diff --git a/libs/libcommon/src/libcommon/config.py b/libs/libcommon/src/libcommon/config.py
index 8a074b41..e4d63e70 100644
--- a/libs/libcommon/src/libcommon/config.py
+++ b/libs/libcommon/src/libcommon/config.py
@@ -217 +217 @@ class ProcessingGraphConfig:
- "required_by_dataset_viewer": True,
+ "enables_preview": True,
@@ -240 +240 @@ class ProcessingGraphConfig:
- "required_by_dataset_viewer": True,
+ "enables_preview": True,
@@ -266,0 +267 @@ class ProcessingGraphConfig:
+ "enables_viewer": True,
diff --git a/libs/libcommon/src/libcommon/processing_graph.py b/libs/libcommon/src/libcommon/processing_graph.py
index 1b50b779..f6f1e636 100644
--- a/libs/libcommon/src/libcommon/processing_graph.py
+++ b/libs/libcommon/src/libcommon/processing_graph.py
@@ -50 +50,2 @@ class ProcessingStepSpecification(TypedDict, total=False):
- required_by_dataset_viewer: Literal[True]
+ enables_preview: Literal[True]
+ enables_viewer: Literal[True]
@@ -137 +138,2 @@ class ProcessingGraph:
- _processing_steps_required_by_dataset_viewer: List[ProcessingStep] = field(init=False)
+ _processing_steps_enables_preview: List[ProcessingStep] = field(init=False)
+ _processing_steps_enables_viewer: List[ProcessingStep] = field(init=False)
@@ -182 +184,2 @@ class ProcessingGraph:
- required_by_dataset_viewer=specification.get("required_by_dataset_viewer", False),
+ enables_preview=specification.get("enables_preview", False),
+ enables_viewer=specification.get("enables_viewer", False),
@@ -216 +219 @@ class ProcessingGraph:
- self._processing_steps_required_by_dataset_viewer = [
+ self._processing_steps_enables_preview = [
@@ -218 +221,6 @@ class ProcessingGraph:
- for (processing_step_name, required) in _nx_graph.nodes(data="required_by_dataset_viewer")
+ for (processing_step_name, required) in _nx_graph.nodes(data="enables_preview")
+ if required
+ ]
+ self._processing_steps_enables_viewer = [
+ self._processing_steps[processing_step_name]
+ for (processing_step_name, required) in _nx_graph.nodes(data="enables_viewer")
@@ -373 +381,13 @@ class ProcessingGraph:
- def get_processing_steps_required_by_dataset_viewer(self) -> List[ProcessingStep]:
+ def get_processing_steps_enables_preview(self) -> List[ProcessingStep]:
+ """
+ Get the processing steps that enable the dataset preview (first rows).
+
+ The returned processing steps are copies of the original ones, so that they can be modified without affecting
+ the original ones.
+
+ Returns:
+ List[ProcessingStep]: The list of processing steps that enable the dataset preview
+ """
+ return copy_processing_steps_list(self._processing_steps_enables_preview)
+
+ def get_processing_steps_enables_viewer(self) -> List[ProcessingStep]:
@@ -375 +395 @@ class ProcessingGraph:
- Get the processing steps required by the dataset viewer.
+ Get the processing steps that enable the dataset viewer (all rows).
@@ -381 +401 @@ class ProcessingGraph:
- List[ProcessingStep]: The list of processing steps required by the dataset viewer
+ List[ProcessingStep]: The list of processing steps that enable the dataset viewer
@@ -383 +403 @@ class ProcessingGraph:
- return copy_processing_steps_list(self._processing_steps_required_by_dataset_viewer)
+ return copy_processing_steps_list(self._processing_steps_enables_viewer)
diff --git a/libs/libcommon/tests/test_processing_graph.py b/libs/libcommon/tests/test_processing_graph.py
index a4861594..49bd37fb 100644
--- a/libs/libcommon/tests/test_processing_graph.py
+++ b/libs/libcommon/tests/test_processing_graph.py
@@ -303,3 +303,8 @@ def test_default_graph_first_steps(graph: ProcessingGraph) -> None:
-def test_default_graph_required_by_dataset_viewer(graph: ProcessingGraph) -> None:
- required_by_dataset_viewer = ["split-first-rows-from-streaming", "split-first-rows-from-parquet"]
- assert_lists_are_equal(graph.get_processing_steps_required_by_dataset_viewer(), required_by_dataset_viewer)
+def test_default_graph_enables_preview(graph: ProcessingGraph) -> None:
+ enables_preview = ["split-first-rows-from-streaming", "split-first-rows-from-parquet"]
+ assert_lists_are_equal(graph.get_processing_steps_enables_preview(), enables_preview)
+
+
+def test_default_graph_enables_viewer(graph: ProcessingGraph) -> None:
+ enables_viewer = ["config-size"]
+ assert_lists_are_equal(graph.get_processing_steps_enables_viewer(), enables_viewer)
diff --git a/services/api/src/api/routes/valid.py b/services/api/src/api/routes/valid.py
index 4cf69d0b..606353e4 100644
--- a/services/api/src/api/routes/valid.py
+++ b/services/api/src/api/routes/valid.py
@@ -5 +5,2 @@ import logging
-from typing import List
+from dataclasses import dataclass, field
+from typing import List, Set, TypedDict
@@ -7 +8 @@ from typing import List
-from libcommon.processing_graph import ProcessingGraph
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
@@ -21,11 +22,42 @@ from api.utils import (
-def get_valid(processing_graph: ProcessingGraph) -> List[str]:
- # a dataset is considered valid if at least one response of any of the
- # "required_by_dataset_viewer" steps is valid.
- processing_steps = processing_graph.get_processing_steps_required_by_dataset_viewer()
- if not processing_steps:
- return []
- datasets = set.union(
- *[get_valid_datasets(kind=processing_step.cache_kind) for processing_step in processing_steps]
- )
- # note that the list is sorted alphabetically for consistency
- return sorted(datasets)
+class ValidContent(TypedDict):
+ valid: List[str]
+ preview: List[str]
+ viewer: List[str]
+
+
+@dataclass
+class ValidDatasets:
+ processing_graph: ProcessingGraph
+ content: ValidContent = field(init=False)
+
+ def __post_init__(self) -> None:
+ _viewer_set: Set[str] = self._get_valid_set(
+ processing_steps=self.processing_graph.get_processing_steps_enables_viewer()
+ )
+ _preview_set: Set[str] = self._get_valid_set(
+ processing_steps=self.processing_graph.get_processing_steps_enables_preview()
+ ).difference(_viewer_set)
+ _valid_set = set.union(_viewer_set, _preview_set)
+ self.content = ValidContent(
+ valid=sorted(_valid_set),
+ preview=sorted(_preview_set),
+ viewer=sorted(_viewer_set),
+ )
+
+ def _get_valid_set(self, processing_steps: List[ProcessingStep]) -> Set[str]:
+ """Returns the list of the valid datasets for the list of steps
+
+ A dataset is considered valid if at least one response of any of the artifacts for any of the
+ steps is valid.
+
+ Args:
+ processing_steps (List[ProcessingStep]): The list of processing steps
+
+ Returns:
+ List[str]: The list of valid datasets for the steps
+ """
+ if not processing_steps:
+ return set()
+ return set.union(
+ *[get_valid_datasets(kind=processing_step.cache_kind) for processing_step in processing_steps]
+ )
@@ -45 +77 @@ def create_valid_endpoint(
- content = {"valid": get_valid(processing_graph=processing_graph)}
+ content = ValidDatasets(processing_graph=processing_graph).content
diff --git a/services/api/tests/routes/test_valid.py b/services/api/tests/routes/test_valid.py
index 317b0946..38fb758f 100644
--- a/services/api/tests/routes/test_valid.py
+++ b/services/api/tests/routes/test_valid.py
@@ -9 +9 @@ from api.config import AppConfig
-from api.routes.valid import get_valid
+from api.routes.valid import ValidDatasets
@@ -29 +29 @@ def clean_mongo_databases(app_config: AppConfig) -> None:
- {step_1: {"required_by_dataset_viewer": True}},
+ {step_1: {"enables_preview": True}},
@@ -34 +34,2 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) ->
- assert get_valid(processing_graph=processing_graph) == []
+ valid_datasets = ValidDatasets(processing_graph=processing_graph)
+ assert valid_datasets.content == {"valid": [], "preview": [], "viewer": []}
@@ -38 +39 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) ->
- "processing_graph_specification,expected_valid",
+ "processing_graph_specification,expected_preview,expected_viewer,expected_valid",
@@ -40,4 +41,6 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) ->
- ({step_1: {}}, []),
- ({step_1: {"required_by_dataset_viewer": True}}, ["dataset"]),
- ({step_1: {}, step_2: {"required_by_dataset_viewer": True}}, []),
- ({step_1: {"required_by_dataset_viewer": True}, step_2: {"required_by_dataset_viewer": True}}, ["dataset"]),
+ ({step_1: {}, step_2: {}}, [], [], []),
+ ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset"], [], ["dataset"]),
+ ({step_1: {}, step_2: {"enables_preview": True}}, [], [], []),
+ ({step_1: {"enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]),
+ ({step_1: {}, step_2: {"enables_viewer": True}}, [], [], []),
+ ({step_1: {"enables_preview": True, "enables_viewer": True}, step_2: {}}, [], ["dataset"], ["dataset"]),
@@ -46 +49,6 @@ def test_empty(processing_graph_specification: ProcessingGraphSpecification) ->
-def test_one_dataset(processing_graph_specification: ProcessingGraphSpecification, expected_valid: List[str]) -> None:
+def test_one_dataset(
+ processing_graph_specification: ProcessingGraphSpecification,
+ expected_preview: List[str],
+ expected_viewer: List[str],
+ expected_valid: List[str],
+) -> None:
@@ -51 +59,6 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio
- assert get_valid(processing_graph=processing_graph) == expected_valid
+ valid_datasets = ValidDatasets(processing_graph=processing_graph)
+ assert valid_datasets.content == {
+ "valid": expected_valid,
+ "preview": expected_preview,
+ "viewer": expected_viewer,
+ }
@@ -55 +68 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio
- "processing_graph_specification,expected_valid",
+ "processing_graph_specification,expected_preview,expected_viewer,expected_valid",
@@ -57,3 +70,3 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio
- ({step_1: {}, step_2: {}}, []),
- ({step_1: {"required_by_dataset_viewer": True}, step_2: {}}, ["dataset1"]),
- ({step_1: {}, step_2: {"required_by_dataset_viewer": True}}, ["dataset2"]),
+ ({step_1: {}, step_2: {}}, [], [], []),
+ ({step_1: {"enables_preview": True}, step_2: {}}, ["dataset1"], [], ["dataset1"]),
+ ({step_1: {}, step_2: {"enables_preview": True}}, ["dataset2"], [], ["dataset2"]),
@@ -61 +74,9 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio
- {step_1: {"required_by_dataset_viewer": True}, step_2: {"required_by_dataset_viewer": True}},
+ {step_1: {"enables_preview": True}, step_2: {"enables_preview": True}},
+ ["dataset1", "dataset2"],
+ [],
+ ["dataset1", "dataset2"],
+ ),
+ (
+ {step_1: {"enables_preview": True}, step_2: {"enables_viewer": True}},
+ ["dataset1"],
+ ["dataset2"],
@@ -66 +87,6 @@ def test_one_dataset(processing_graph_specification: ProcessingGraphSpecificatio
-def test_two_datasets(processing_graph_specification: ProcessingGraphSpecification, expected_valid: List[str]) -> None:
+def test_two_datasets(
+ processing_graph_specification: ProcessingGraphSpecification,
+ expected_preview: List[str],
+ expected_viewer: List[str],
+ expected_valid: List[str],
+) -> None:
@@ -80 +106,6 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- assert get_valid(processing_graph=processing_graph) == expected_valid
+ valid_datasets = ValidDatasets(processing_graph=processing_graph)
+ assert valid_datasets.content == {
+ "valid": expected_valid,
+ "preview": expected_preview,
+ "viewer": expected_viewer,
+ }
@@ -84 +115 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- "processing_graph_specification,expected_valid",
+ "processing_graph_specification,expected_preview,expected_viewer,expected_valid",
@@ -92,0 +124,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
+ [],
+ [],
@@ -96 +129 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- dataset_step: {"required_by_dataset_viewer": True},
+ dataset_step: {"enables_preview": True},
@@ -100,0 +134,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
+ [],
+ ["dataset"],
@@ -108 +143 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- "required_by_dataset_viewer": True,
+ "enables_preview": True,
@@ -112,0 +148,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
+ [],
+ ["dataset"],
@@ -118 +155 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- split_step: {"input_type": "split", "triggered_by": config_step, "required_by_dataset_viewer": True},
+ split_step: {"input_type": "split", "triggered_by": config_step, "enables_preview": True},
@@ -120,0 +158,2 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
+ [],
+ ["dataset"],
@@ -124 +163 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- dataset_step: {"required_by_dataset_viewer": True},
+ dataset_step: {"enables_preview": True},
@@ -128 +167 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- "required_by_dataset_viewer": True,
+ "enables_preview": True,
@@ -130 +169 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
- split_step: {"input_type": "split", "triggered_by": config_step, "required_by_dataset_viewer": True},
+ split_step: {"input_type": "split", "triggered_by": config_step, "enables_preview": True},
@@ -132,0 +172,16 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
+ [],
+ ["dataset"],
+ ),
+ (
+ {
+ dataset_step: {},
+ config_step: {
+ "input_type": "config",
+ "triggered_by": dataset_step,
+ "enables_viewer": True,
+ },
+ split_step: {"input_type": "split", "triggered_by": config_step, "enables_preview": True},
+ },
+ [],
+ ["dataset"],
+ ["dataset"],
@@ -136 +191,6 @@ def test_two_datasets(processing_graph_specification: ProcessingGraphSpecificati
-def test_three_steps(processing_graph_specification: ProcessingGraphSpecification, expected_valid: List[str]) -> None:
+def test_three_steps(
+ processing_graph_specification: ProcessingGraphSpecification,
+ expected_preview: List[str],
+ expected_viewer: List[str],
+ expected_valid: List[str],
+) -> None:
@@ -162 +222,6 @@ def test_three_steps(processing_graph_specification: ProcessingGraphSpecificatio
- assert get_valid(processing_graph=processing_graph) == expected_valid
+ valid_datasets = ValidDatasets(processing_graph=processing_graph)
+ assert valid_datasets.content == {
+ "valid": expected_valid,
+ "preview": expected_preview,
+ "viewer": expected_viewer,
+ }
@@ -166 +231 @@ def test_errors() -> None:
- processing_graph = ProcessingGraph({dataset_step: {"required_by_dataset_viewer": True}})
+ processing_graph = ProcessingGraph({dataset_step: {"enables_preview": True}})
@@ -174 +239,12 @@ def test_errors() -> None:
- assert get_valid(processing_graph=processing_graph) == [dataset_a, dataset_b]
+ valid_datasets = ValidDatasets(processing_graph=processing_graph)
+ assert valid_datasets.content == {
+ "valid": [
+ dataset_a,
+ dataset_b,
+ ],
+ "preview": [
+ dataset_a,
+ dataset_b,
+ ],
+ "viewer": [],
+ }
diff --git a/services/api/tests/test_app.py b/services/api/tests/test_app.py
index 32c83a97..a883d39b 100644
--- a/services/api/tests/test_app.py
+++ b/services/api/tests/test_app.py
@@ -51,0 +52,2 @@ def test_get_valid_datasets(client: TestClient) -> None:
+ assert "preview" in response.json()
+ assert "viewer" in response.json()
|
|
db7913058dc3a940c862dc4649d260c5bde33147
|
Sylvain Lesage
| 2023-06-12T12:56:29 |
feat: 🎸 upgrade duckdb and gradio (#1351)
|
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index 19956d9d..cd3fcb36 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -619 +619 @@ name = "duckdb"
-version = "0.6.1"
+version = "0.8.0"
@@ -625,47 +625,47 @@ files = [
- {file = "duckdb-0.6.1-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e566514f9327f89264e98ac14ee7a84fbd9857328028258422c3e8375ee19d25"},
- {file = "duckdb-0.6.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b31c2883de5b19591a2852165e6b3f9821f77af649835f27bc146b26e4aa30cb"},
- {file = "duckdb-0.6.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:998165b2fb1f1d2b0ad742096015ea70878f7d40304643c7424c3ed3ddf07bfc"},
- {file = "duckdb-0.6.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3941b3a1e8a1cdb7b90ab3917b87af816e71f9692e5ada7f19b6b60969f731e5"},
- {file = "duckdb-0.6.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:143611bd1b7c13343f087d4d423a7a8a4f33a114c5326171e867febf3f0fcfe1"},
- {file = "duckdb-0.6.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:125ba45e8b08f28858f918ec9cbd3a19975e5d8d9e8275ef4ad924028a616e14"},
- {file = "duckdb-0.6.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:e609a65b31c92f2f7166831f74b56f5ed54b33d8c2c4b4c3974c26fdc50464c5"},
- {file = "duckdb-0.6.1-cp310-cp310-win32.whl", hash = "sha256:b39045074fb9a3f068496475a5d627ad4fa572fa3b4980e3b479c11d0b706f2d"},
- {file = "duckdb-0.6.1-cp310-cp310-win_amd64.whl", hash = "sha256:16fa96ffaa3d842a9355a633fb8bc092d119be08d4bc02013946d8594417bc14"},
- {file = "duckdb-0.6.1-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b4bbe2f6c1b109c626f9318eee80934ad2a5b81a51409c6b5083c6c5f9bdb125"},
- {file = "duckdb-0.6.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:cfea36b58928ce778d17280d4fb3bf0a2d7cff407667baedd69c5b41463ac0fd"},
- {file = "duckdb-0.6.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:0b64eb53d0d0695814bf1b65c0f91ab7ed66b515f89c88038f65ad5e0762571c"},
- {file = "duckdb-0.6.1-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:35b01bc724e1933293f4c34f410d2833bfbb56d5743b515d805bbfed0651476e"},
- {file = "duckdb-0.6.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fec2c2466654ce786843bda2bfba71e0e4719106b41d36b17ceb1901e130aa71"},
- {file = "duckdb-0.6.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:82cd30f5cf368658ef879b1c60276bc8650cf67cfe3dc3e3009438ba39251333"},
- {file = "duckdb-0.6.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:a782bbfb7f5e97d4a9c834c9e78f023fb8b3f6687c22ca99841e6ed944b724da"},
- {file = "duckdb-0.6.1-cp311-cp311-win32.whl", hash = "sha256:e3702d4a9ade54c6403f6615a98bbec2020a76a60f5db7fcf085df1bd270e66e"},
- {file = "duckdb-0.6.1-cp311-cp311-win_amd64.whl", hash = "sha256:93b074f473d68c944b0eeb2edcafd91ad11da8432b484836efaaab4e26351d48"},
- {file = "duckdb-0.6.1-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:adae183924d6d479202c39072e37d440b511326e84525bcb7432bca85f86caba"},
- {file = "duckdb-0.6.1-cp36-cp36m-win32.whl", hash = "sha256:546a1cd17595bd1dd009daf6f36705aa6f95337154360ce44932157d353dcd80"},
- {file = "duckdb-0.6.1-cp36-cp36m-win_amd64.whl", hash = "sha256:87b0d00eb9d1a7ebe437276203e0cdc93b4a2154ba9688c65e8d2a8735839ec6"},
- {file = "duckdb-0.6.1-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:8442e074de6e1969c3d2b24363a5a6d7f866d5ac3f4e358e357495b389eff6c1"},
- {file = "duckdb-0.6.1-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0a6bf2ae7bec803352dade14561cb0b461b2422e70f75d9f09b36ba2dad2613b"},
- {file = "duckdb-0.6.1-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5054792f22733f89d9cbbced2bafd8772d72d0fe77f159310221cefcf981c680"},
- {file = "duckdb-0.6.1-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:21cc503dffc2c68bb825e4eb3098e82f40e910b3d09e1b3b7f090d39ad53fbea"},
- {file = "duckdb-0.6.1-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:54b3da77ad893e99c073087ff7f75a8c98154ac5139d317149f12b74367211db"},
- {file = "duckdb-0.6.1-cp37-cp37m-win32.whl", hash = "sha256:f1d709aa6a26172a3eab804b57763d5cdc1a4b785ac1fc2b09568578e52032ee"},
- {file = "duckdb-0.6.1-cp37-cp37m-win_amd64.whl", hash = "sha256:f4edcaa471d791393e37f63e3c7c728fa6324e3ac7e768b9dc2ea49065cd37cc"},
- {file = "duckdb-0.6.1-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:d218c2dd3bda51fb79e622b7b2266183ac9493834b55010aa01273fa5b7a7105"},
- {file = "duckdb-0.6.1-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:0c7155cb93ab432eca44b651256c359281d26d927ff43badaf1d2276dd770832"},
- {file = "duckdb-0.6.1-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:0925778200090d3d5d8b6bb42b4d05d24db1e8912484ba3b7e7b7f8569f17dcb"},
- {file = "duckdb-0.6.1-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8b544dd04bb851d08bc68b317a7683cec6091547ae75555d075f8c8a7edb626e"},
- {file = "duckdb-0.6.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f2c37d5a0391cf3a3a66e63215968ffb78e6b84f659529fa4bd10478f6203071"},
- {file = "duckdb-0.6.1-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ce376966260eb5c351fcc6af627a979dbbcae3efeb2e70f85b23aa45a21e289d"},
- {file = "duckdb-0.6.1-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:73c974b09dd08dff5e8bdedba11c7d0aa0fc46ca93954ee7d19e1e18c9883ac1"},
- {file = "duckdb-0.6.1-cp38-cp38-win32.whl", hash = "sha256:bfe39ed3a03e8b1ed764f58f513b37b24afe110d245803a41655d16d391ad9f1"},
- {file = "duckdb-0.6.1-cp38-cp38-win_amd64.whl", hash = "sha256:afa97d982dbe6b125631a17e222142e79bee88f7a13fc4cee92d09285e31ec83"},
- {file = "duckdb-0.6.1-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:c35ff4b1117096ef72d101524df0079da36c3735d52fcf1d907ccffa63bd6202"},
- {file = "duckdb-0.6.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:5c54910fbb6de0f21d562e18a5c91540c19876db61b862fc9ffc8e31be8b3f03"},
- {file = "duckdb-0.6.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:99a7172563a3ae67d867572ce27cf3962f58e76f491cb7f602f08c2af39213b3"},
- {file = "duckdb-0.6.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7363ffe857d00216b659116647fbf1e925cb3895699015d4a4e50b746de13041"},
- {file = "duckdb-0.6.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:06c1cef25f896b2284ba048108f645c72fab5c54aa5a6f62f95663f44ff8a79b"},
- {file = "duckdb-0.6.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:e92dd6aad7e8c29d002947376b6f5ce28cae29eb3b6b58a64a46cdbfc5cb7943"},
- {file = "duckdb-0.6.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:4b280b2d8a01ecd4fe2feab041df70233c534fafbe33a38565b52c1e017529c7"},
- {file = "duckdb-0.6.1-cp39-cp39-win32.whl", hash = "sha256:d9212d76e90b8469743924a4d22bef845be310d0d193d54ae17d9ef1f753cfa7"},
- {file = "duckdb-0.6.1-cp39-cp39-win_amd64.whl", hash = "sha256:00b7be8f67ec1a8edaa8844f521267baa1a795f4c482bfad56c72c26e1862ab2"},
- {file = "duckdb-0.6.1.tar.gz", hash = "sha256:6d26e9f1afcb924a6057785e506810d48332d4764ddc4a5b414d0f2bf0cacfb4"},
+ {file = "duckdb-0.8.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:6455aee00af30770c20f4a8c5e4347918cf59b578f49ee996a13807b12911871"},
+ {file = "duckdb-0.8.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b8cf0622ae7f86d4ce72791f8928af4357a46824aadf1b6879c7936b3db65344"},
+ {file = "duckdb-0.8.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:6132e8183ca3ae08a593e43c97cb189794077dedd48546e27ce43bd6a51a9c33"},
+ {file = "duckdb-0.8.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:fe29e5343fa2a95f2cde4519a4f4533f4fd551a48d2d9a8ab5220d40ebf53610"},
+ {file = "duckdb-0.8.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:945165987ca87c097dc0e578dcf47a100cad77e1c29f5dd8443d53ce159dc22e"},
+ {file = "duckdb-0.8.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:673c60daf7ada1d9a8518286a6893ec45efabb64602954af5f3d98f42912fda6"},
+ {file = "duckdb-0.8.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:d5075fe1ff97ae62331ca5c61e3597e6e9f7682a6fdd418c23ba5c4873ed5cd1"},
+ {file = "duckdb-0.8.0-cp310-cp310-win32.whl", hash = "sha256:001f5102f45d3d67f389fa8520046c8f55a99e2c6d43b8e68b38ea93261c5395"},
+ {file = "duckdb-0.8.0-cp310-cp310-win_amd64.whl", hash = "sha256:cb00800f2e1e865584b13221e0121fce9341bb3a39a93e569d563eaed281f528"},
+ {file = "duckdb-0.8.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:b2707096d6df4321044fcde2c9f04da632d11a8be60957fd09d49a42fae71a29"},
+ {file = "duckdb-0.8.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b27df1b70ae74d2c88efb5ffca8490954fdc678099509a9c4404ca30acc53426"},
+ {file = "duckdb-0.8.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:75a97c800271b52dd0f37696d074c50576dcb4b2750b6115932a98696a268070"},
+ {file = "duckdb-0.8.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:804cac261a5e016506a6d67838a65d19b06a237f7949f1704f0e800eb708286a"},
+ {file = "duckdb-0.8.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c6b9abca7fa6713e1d031c18485343b4de99742c7e1b85c10718aa2f31a4e2c6"},
+ {file = "duckdb-0.8.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:51aa6d606d49072abcfeb3be209eb559ac94c1b5e70f58ac3adbb94aca9cd69f"},
+ {file = "duckdb-0.8.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:7c8dc769aaf2be0a1c57995ca657e5b92c1c56fc8437edb720ca6cab571adf14"},
+ {file = "duckdb-0.8.0-cp311-cp311-win32.whl", hash = "sha256:c4207d18b42387c4a035846d8878eb967070198be8ac26fd77797ce320d1a400"},
+ {file = "duckdb-0.8.0-cp311-cp311-win_amd64.whl", hash = "sha256:0c392257547c20794c3072fcbca99a49ef0a49974005d755e93893e2b4875267"},
+ {file = "duckdb-0.8.0-cp36-cp36m-macosx_10_9_x86_64.whl", hash = "sha256:2832379e122020814dbe869af7b9ddf3c9f21474cf345531145b099c63ffe17e"},
+ {file = "duckdb-0.8.0-cp36-cp36m-win32.whl", hash = "sha256:914896526f7caba86b170f2c4f17f11fd06540325deeb0000cb4fb24ec732966"},
+ {file = "duckdb-0.8.0-cp36-cp36m-win_amd64.whl", hash = "sha256:022ebda86d0e3204cdc206e4af45aa9f0ae0668b34c2c68cf88e08355af4a372"},
+ {file = "duckdb-0.8.0-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:96a31c0f3f4ccbf0f5b18f94319f37691205d82f80aae48c6fe04860d743eb2c"},
+ {file = "duckdb-0.8.0-cp37-cp37m-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a07c73c6e6a8cf4ce1a634625e0d1b17e5b817242a8a530d26ed84508dfbdc26"},
+ {file = "duckdb-0.8.0-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:424acbd6e857531b06448d757d7c2557938dbddbff0632092090efbf413b4699"},
+ {file = "duckdb-0.8.0-cp37-cp37m-musllinux_1_1_i686.whl", hash = "sha256:c83cfd2a868f1acb0692b9c3fd5ef1d7da8faa1348c6eabf421fbf5d8c2f3eb8"},
+ {file = "duckdb-0.8.0-cp37-cp37m-musllinux_1_1_x86_64.whl", hash = "sha256:5c6f6b2d8db56936f662c649539df81856b5a8cb769a31f9544edf18af2a11ff"},
+ {file = "duckdb-0.8.0-cp37-cp37m-win32.whl", hash = "sha256:0bd6376b40a512172eaf4aa816813b1b9d68994292ca436ce626ccd5f77f8184"},
+ {file = "duckdb-0.8.0-cp37-cp37m-win_amd64.whl", hash = "sha256:931221885bcf1e7dfce2400f11fd048a7beef566b775f1453bb1db89b828e810"},
+ {file = "duckdb-0.8.0-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:42e7853d963d68e72403ea208bcf806b0f28c7b44db0aa85ce49bb124d56c133"},
+ {file = "duckdb-0.8.0-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:fcc338399175be3d43366576600aef7d72e82114d415992a7a95aded98a0f3fd"},
+ {file = "duckdb-0.8.0-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:03dd08a4624d6b581a59f9f9dbfd34902416398d16795ad19f92361cf21fd9b5"},
+ {file = "duckdb-0.8.0-cp38-cp38-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0c7c24ea0c9d8563dbd5ad49ccb54b7a9a3c7b8c2833d35e5d32a08549cacea5"},
+ {file = "duckdb-0.8.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cb58f6505cc0f34b4e976154302d26563d2e5d16b206758daaa04b65e55d9dd8"},
+ {file = "duckdb-0.8.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:ef37ac7880100c4b3f913c8483a29a13f8289313b9a07df019fadfa8e7427544"},
+ {file = "duckdb-0.8.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:c2a4f5ee913ca8a6a069c78f8944b9934ffdbc71fd935f9576fdcea2a6f476f1"},
+ {file = "duckdb-0.8.0-cp38-cp38-win32.whl", hash = "sha256:73831c6d7aefcb5f4072cd677b9efebecbf6c578946d21710791e10a1fc41b9a"},
+ {file = "duckdb-0.8.0-cp38-cp38-win_amd64.whl", hash = "sha256:faa36d2854734364d234f37d7ef4f3d763b73cd6b0f799cbc2a0e3b7e2575450"},
+ {file = "duckdb-0.8.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:50a31ec237ed619e50f9ab79eb0ec5111eb9697d4475da6e0ab22c08495ce26b"},
+ {file = "duckdb-0.8.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:351abb4cc2d229d043920c4bc2a4c29ca31a79fef7d7ef8f6011cf4331f297bf"},
+ {file = "duckdb-0.8.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:568550a163aca6a787bef8313e358590254de3f4019025a8d68c3a61253fedc1"},
+ {file = "duckdb-0.8.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2b82617f0e7f9fc080eda217090d82b42d4fad083bc9f6d58dfda9cecb7e3b29"},
+ {file = "duckdb-0.8.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d01c9be34d272532b75e8faedda0ff77fa76d1034cde60b8f5768ae85680d6d3"},
+ {file = "duckdb-0.8.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:8549d6a6bf5f00c012b6916f605416226507e733a3ffc57451682afd6e674d1b"},
+ {file = "duckdb-0.8.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:8d145c6d51e55743c3ed1a74cffa109d9e72f82b07e203b436cfa453c925313a"},
+ {file = "duckdb-0.8.0-cp39-cp39-win32.whl", hash = "sha256:f8610dfd21e90d7b04e8598b244bf3ad68599fd6ba0daad3428c03cbfd74dced"},
+ {file = "duckdb-0.8.0-cp39-cp39-win_amd64.whl", hash = "sha256:d0f0f104d30418808bafbe9bccdcd238588a07bd246b3cff13842d60bfd8e8ba"},
+ {file = "duckdb-0.8.0.tar.gz", hash = "sha256:c68da35bab5072a64ada2646a5b343da620ddc75a7a6e84aa4a1e0628a7ec18f"},
@@ -674,3 +673,0 @@ files = [
-[package.dependencies]
-numpy = ">=1.14"
-
@@ -915 +912 @@ name = "gradio"
-version = "3.28.1"
+version = "3.34.0"
@@ -921,2 +918,2 @@ files = [
- {file = "gradio-3.28.1-py3-none-any.whl", hash = "sha256:0811f04be88ee789d921fee0089c69b7da3b12dd7c937f383e4c0998663b4f21"},
- {file = "gradio-3.28.1.tar.gz", hash = "sha256:15d959e73f77bb6980334a159ff177867f6636a2a1007a7154d55a1d6e6d8648"},
+ {file = "gradio-3.34.0-py3-none-any.whl", hash = "sha256:1cd8b25b598d983561d64f0a039af819382f1376c676aa9f84972c46b6875741"},
+ {file = "gradio-3.34.0.tar.gz", hash = "sha256:fd7fa7257ffc749f9dc7c297eba554eaa1e5acd1a5f9c973250b2080932d6a41"},
@@ -931 +928 @@ ffmpy = "*"
-gradio-client = ">=0.1.3"
+gradio-client = ">=0.2.6"
@@ -933 +930 @@ httpx = "*"
-huggingface-hub = ">=0.13.0"
+huggingface-hub = ">=0.14.0"
@@ -944,0 +942 @@ pydub = "*"
+pygments = ">=2.12.0"
@@ -950 +948 @@ typing-extensions = "*"
-uvicorn = "*"
+uvicorn = ">=0.14.0"
@@ -955 +953 @@ name = "gradio-client"
-version = "0.2.5"
+version = "0.2.6"
@@ -961,2 +959,2 @@ files = [
- {file = "gradio_client-0.2.5-py3-none-any.whl", hash = "sha256:922a5188c93797adce023b4caa655318b9c09834095d31763443c1e7a707e301"},
- {file = "gradio_client-0.2.5.tar.gz", hash = "sha256:1a24dd7a09976db70fd32cdb5512b6f7b15c1a084f310b5f6b3021c9794d90a4"},
+ {file = "gradio_client-0.2.6-py3-none-any.whl", hash = "sha256:daa3e2f7697cc07821ccefa7486b923e71c1e459d5b3c6e35318f343b566d789"},
+ {file = "gradio_client-0.2.6.tar.gz", hash = "sha256:a5d5c5799ce33ae3107e1d30992c27050f50506f15dd70481a39b13ac47e2613"},
@@ -2267,0 +2266,15 @@ files = [
+[[package]]
+name = "pygments"
+version = "2.15.1"
+description = "Pygments is a syntax highlighting package written in Python."
+category = "main"
+optional = false
+python-versions = ">=3.7"
+files = [
+ {file = "Pygments-2.15.1-py3-none-any.whl", hash = "sha256:db2db3deb4b4179f399a09054b023b6a586b76499d36965813c71aa8ed7b5fd1"},
+ {file = "Pygments-2.15.1.tar.gz", hash = "sha256:8ace4d3c1dd481894b2005f560ead0f9f19ee64fe983366be1a21e171d12775c"},
+]
+
+[package.extras]
+plugins = ["importlib-metadata"]
+
@@ -3270 +3283 @@ python-versions = "3.9.15"
-content-hash = "9e29354c885f6c0d8aba1c5e79d6087a8b1419b69418e4db09caa28e32d771ae"
+content-hash = "4c403bae2a09145ce4fb1217296ad92a9a01b184bf2d913aceb98a777b55c133"
diff --git a/front/admin_ui/pyproject.toml b/front/admin_ui/pyproject.toml
index 8947fe80..cf4bebc5 100644
--- a/front/admin_ui/pyproject.toml
+++ b/front/admin_ui/pyproject.toml
@@ -8 +8 @@ authors = ["Quentin Lhoest <[email protected]>"]
-gradio = "3.28.1"
+gradio = "^3.34.0"
@@ -14 +14 @@ huggingface-hub = "^0.15.1"
-duckdb = "^0.6.1"
+duckdb = "^0.8.0"
diff --git a/front/admin_ui/requirements.txt b/front/admin_ui/requirements.txt
index e178a30b..3f903ace 100644
--- a/front/admin_ui/requirements.txt
+++ b/front/admin_ui/requirements.txt
@@ -1 +1 @@
-gradio==3.28.1
+gradio==3.34.0
@@ -7 +7 @@ huggingface-hub~=0.15.1
-duckdb~=0.6.1
\ No newline at end of file
+duckdb~=0.8.0
|
|
ba115a9145e2c24f01f6d58dabdd6cf78fc8b4da
|
Sylvain Lesage
| 2023-06-12T11:19:00 |
Fix admin app (#1348)
|
diff --git a/front/admin_ui/app.py b/front/admin_ui/app.py
index 770ff8f3..a39a8e20 100644
--- a/front/admin_ui/app.py
+++ b/front/admin_ui/app.py
@@ -27,0 +28,2 @@ DSS_ENDPOINT = DEV_DSS_ENDPOINT if DEV else PROD_DSS_ENDPOINT
+PROCESSING_GRAPH = ProcessingGraph(ProcessingGraphConfig().specification)
+
@@ -43,2 +45 @@ def draw_graph(width, height):
- config = ProcessingGraphConfig()
- graph = ProcessingGraph(config.specification)._nx_graph
+ graph = PROCESSING_GRAPH._nx_graph
@@ -81 +82,2 @@ with gr.Blocks() as demo:
- refresh_type = gr.Textbox(label="Processing step type", placeholder="split-first-rows-from-streaming")
+ job_types = [processing_step.job_type for processing_step in PROCESSING_GRAPH.get_topologically_ordered_processing_steps()]
+ refresh_type = gr.Dropdown(job_types, multiselect=False, label="job type", value=job_types[0])
@@ -171 +173 @@ with gr.Blocks() as demo:
- "type": job_type,
+ "kind": cached_response["kind"],
@@ -187 +189 @@ with gr.Blocks() as demo:
- "type": job_type,
+ "type": job["type"],
@@ -188,0 +191 @@ with gr.Blocks() as demo:
+ "revision": job["revision"],
|
|
deba0fd4c5a538922f15c723838b6c0bd056d7a1
|
Quentin Lhoest
| 2023-06-12T10:10:51 |
fix parquet-and-info download_size (#1347)
|
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 91de381e..8f2f0f10 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -628,0 +629 @@ def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None:
+ builder.info.download_size = 0
|
|
42b364b24e03912d9e70148057c7468eb3f8a047
|
Albert Villanova del Moral
| 2023-06-12T09:25:24 |
Update locked cachecontrol version (#1344)
|
diff --git a/front/admin_ui/poetry.lock b/front/admin_ui/poetry.lock
index 54ed9fd8..19956d9d 100644
--- a/front/admin_ui/poetry.lock
+++ b/front/admin_ui/poetry.lock
@@ -1266,0 +1267 @@ starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
diff --git a/jobs/cache_maintenance/poetry.lock b/jobs/cache_maintenance/poetry.lock
index b67c295d..bd34f45d 100644
--- a/jobs/cache_maintenance/poetry.lock
+++ b/jobs/cache_maintenance/poetry.lock
@@ -265 +265 @@ name = "cachecontrol"
-version = "0.13.0"
+version = "0.13.1"
@@ -269 +269 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -271,2 +271,2 @@ files = [
- {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"},
- {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"},
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
@@ -280,0 +281 @@ requests = ">=2.16.0"
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
@@ -1041,0 +1043 @@ starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
diff --git a/jobs/mongodb_migration/poetry.lock b/jobs/mongodb_migration/poetry.lock
index b67c295d..bd34f45d 100644
--- a/jobs/mongodb_migration/poetry.lock
+++ b/jobs/mongodb_migration/poetry.lock
@@ -265 +265 @@ name = "cachecontrol"
-version = "0.13.0"
+version = "0.13.1"
@@ -269 +269 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -271,2 +271,2 @@ files = [
- {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"},
- {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"},
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
@@ -280,0 +281 @@ requests = ">=2.16.0"
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
@@ -1041,0 +1043 @@ starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
diff --git a/libs/libcommon/poetry.lock b/libs/libcommon/poetry.lock
index 534f3756..1ddc15bc 100644
--- a/libs/libcommon/poetry.lock
+++ b/libs/libcommon/poetry.lock
@@ -265 +265 @@ name = "cachecontrol"
-version = "0.13.0"
+version = "0.13.1"
@@ -269 +269 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -271,2 +271,2 @@ files = [
- {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"},
- {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"},
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
@@ -280,0 +281 @@ requests = ">=2.16.0"
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
@@ -3034 +3035 @@ python-versions = "3.9.15"
-content-hash = "859a5062a386f1bc1cec980e4006f3f7f6c7c5b418902232d16994cb9ffb6947"
+content-hash = "3eab4cefa7dd3f05dcf9657d8a1e81f7d7df494f07b6eb0f85fbc97e8a6aa56e"
diff --git a/services/admin/poetry.lock b/services/admin/poetry.lock
index 8ff004b9..928e56ca 100644
--- a/services/admin/poetry.lock
+++ b/services/admin/poetry.lock
@@ -265 +265 @@ name = "cachecontrol"
-version = "0.13.0"
+version = "0.13.1"
@@ -269 +269 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -271,2 +271,2 @@ files = [
- {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"},
- {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"},
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
@@ -280,0 +281 @@ requests = ">=2.16.0"
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
@@ -1099,0 +1101 @@ starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
diff --git a/services/api/poetry.lock b/services/api/poetry.lock
index a4b0eaf9..70a86b1c 100644
--- a/services/api/poetry.lock
+++ b/services/api/poetry.lock
@@ -265 +265 @@ name = "cachecontrol"
-version = "0.13.0"
+version = "0.13.1"
@@ -269 +269 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -271,2 +271,2 @@ files = [
- {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"},
- {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"},
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
@@ -280,0 +281 @@ requests = ">=2.16.0"
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
@@ -1161,0 +1163 @@ starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index 92064110..6ae2e954 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
+# This file is automatically @generated by Poetry and should not be changed by hand.
@@ -473 +473 @@ name = "cachecontrol"
-version = "0.13.0"
+version = "0.13.1"
@@ -477 +477 @@ optional = false
-python-versions = ">=3.6"
+python-versions = ">=3.7"
@@ -479,2 +479,2 @@ files = [
- {file = "CacheControl-0.13.0-py3-none-any.whl", hash = "sha256:4544a012a25cf0a73c53cd986f68b4f9c9f6b1df01d741c2923c3d56c66c7bda"},
- {file = "CacheControl-0.13.0.tar.gz", hash = "sha256:fd3fd2cb0ca66b9a6c1d56cc9709e7e49c63dbd19b1b1bcbd8d3f94cedfe8ce5"},
+ {file = "cachecontrol-0.13.1-py3-none-any.whl", hash = "sha256:95dedbec849f46dda3137866dc28b9d133fc9af55f5b805ab1291833e4457aa4"},
+ {file = "cachecontrol-0.13.1.tar.gz", hash = "sha256:f012366b79d2243a6118309ce73151bf52a38d4a5dac8ea57f09bd29087e506b"},
@@ -488,0 +489 @@ requests = ">=2.16.0"
+dev = ["CacheControl[filecache,redis]", "black", "build", "cherrypy", "mypy", "pytest", "pytest-cov", "sphinx", "tox", "types-redis", "types-requests"]
@@ -1795,0 +1797 @@ starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
@@ -2863,0 +2866 @@ files = [
+ {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"},
@@ -4417,0 +4421 @@ files = [
+ {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
|
|
e0eed9c9ad09d57a857e6498c44b76d6bd71ff76
|
Sylvain Lesage
| 2023-06-10T11:54:23 |
feat: 🎸 give 10x more time for the jobs for cerebras/SlimPajama (#1342)
|
diff --git a/services/worker/src/worker/executor.py b/services/worker/src/worker/executor.py
index e369028d..a0acf747 100644
--- a/services/worker/src/worker/executor.py
+++ b/services/worker/src/worker/executor.py
@@ -135 +135,5 @@ class WorkerExecutor:
- if last_updated + timedelta(seconds=self.app_config.worker.max_job_duration_seconds) <= get_datetime():
+ coefficient = 10 if long_job["params"]["dataset"] == "cerebras/SlimPajama-627B" else 1
+ if (
+ last_updated + timedelta(seconds=coefficient * self.app_config.worker.max_job_duration_seconds)
+ <= get_datetime()
+ ):
|
|
da29c537bc969d2dc0973e010bb94222384b2f62
|
Andrea Francis Soria Jimenez
| 2023-06-09T18:05:29 |
refactor: first-rows-from-parquet use same code as in /rows (#1287)
|
diff --git a/libs/libcommon/pyproject.toml b/libs/libcommon/pyproject.toml
index 24ca6c2a..1bc1d006 100644
--- a/libs/libcommon/pyproject.toml
+++ b/libs/libcommon/pyproject.toml
@@ -26,0 +27 @@ starlette-prometheus = "^0.9.0"
+tqdm = "^4.65.0"
@@ -71,0 +73,2 @@ module = [
+ "tqdm.*",
+ "fsspec.*",
diff --git a/libs/libcommon/src/libcommon/parquet_utils.py b/libs/libcommon/src/libcommon/parquet_utils.py
new file mode 100644
index 00000000..3a498888
--- /dev/null
+++ b/libs/libcommon/src/libcommon/parquet_utils.py
@@ -0,0 +1,496 @@
+import asyncio
+import logging
+import os
+from dataclasses import dataclass
+from functools import lru_cache, partial
+from os import PathLike
+from typing import Callable, List, Literal, Optional, Tuple, TypedDict, Union, cast
+
+import numpy as np
+import numpy.typing as npt
+import pyarrow as pa
+import pyarrow.parquet as pq
+from datasets import Features
+from fsspec.implementations.http import HTTPFile, HTTPFileSystem
+from huggingface_hub import HfFileSystem
+from huggingface_hub.hf_file_system import safe_quote
+from tqdm.contrib.concurrent import thread_map
+
+from libcommon.constants import PARQUET_REVISION
+from libcommon.exceptions import UnexpectedError
+from libcommon.processing_graph import ProcessingGraph
+from libcommon.prometheus import StepProfiler
+from libcommon.simple_cache import get_previous_step_or_raise
+
+StrPath = Union[str, PathLike[str]]
+
+
+class ParquetResponseEmptyError(Exception):
+ pass
+
+
+class ParquetResponseFormatError(Exception):
+ pass
+
+
+class FileSystemError(Exception):
+ pass
+
+
+class ParquetFileItem(TypedDict):
+ dataset: str
+ config: str
+ split: str
+ url: str
+ filename: str
+ size: int
+
+
+class ParquetFileMetadataItem(TypedDict):
+ dataset: str
+ config: str
+ split: str
+ url: str
+ filename: str
+ size: int
+ num_rows: int
+ parquet_metadata_subpath: str
+
+
+class HTTPFileSystemSession(object):
+ _singleton: Optional["HTTPFileSystemSession"] = None
+
+ @classmethod
+ def get_instance(cls) -> "HTTPFileSystemSession":
+ if cls._singleton is None:
+ HTTPFileSystemSession()
+ return cast("HTTPFileSystemSession", HTTPFileSystemSession._singleton)
+
+ def __new__(cls) -> "HTTPFileSystemSession":
+ if HTTPFileSystemSession._singleton is not None:
+ raise RuntimeError(
+ "cannot initialize another instance of HTTPFileSystemSession, use .get_instance() instead"
+ )
+ return super(HTTPFileSystemSession, cls).__new__(cls)
+
+ def __init__(self) -> None:
+ HTTPFileSystemSession._singleton = self
+ self.httpfs = HTTPFileSystem()
+ self.session = asyncio.run(self.httpfs.set_session())
+
+
+def get_supported_unsupported_columns(
+ features: Features,
+ unsupported_features_magic_strings: List[str] = [],
+) -> Tuple[List[str], List[str]]:
+ supported_columns, unsupported_columns = [], []
+
+ for column, feature in features.items():
+ str_feature = str(feature)
+ str_column = str(column)
+ if any(magic_string in str_feature for magic_string in unsupported_features_magic_strings):
+ unsupported_columns.append(str_column)
+ else:
+ supported_columns.append(str_column)
+ return supported_columns, unsupported_columns
+
+
+# TODO: how to invalidate the cache when the parquet branch is created or deleted?
+@lru_cache(maxsize=128)
+def get_hf_fs(hf_token: Optional[str]) -> HfFileSystem:
+ """Get the Hugging Face filesystem.
+
+ Args:
+ hf_token (Optional[str]): The token to access the filesystem.
+ Returns:
+ HfFileSystem: The Hugging Face filesystem.
+ """
+ return HfFileSystem(token=hf_token)
+
+
+def get_hf_parquet_uris(paths: List[str], dataset: str) -> List[str]:
+ """Get the Hugging Face URIs from the Parquet branch of the dataset repository (see PARQUET_REVISION).
+
+ Args:
+ paths (List[str]): List of paths.
+ dataset (str): The dataset name.
+ Returns:
+ List[str]: List of Parquet URIs.
+ """
+ return [f"hf://datasets/{dataset}@{safe_quote(PARQUET_REVISION)}/{path}" for path in paths]
+
+
+@dataclass
+class ParquetIndexWithoutMetadata:
+ features: Features
+ supported_columns: List[str]
+ unsupported_columns: List[str]
+ row_group_offsets: npt.NDArray[np.int64]
+ row_group_readers: List[Callable[[], pa.Table]]
+
+ def query(self, offset: int, length: int) -> pa.Table:
+ """Query the parquet files
+
+ Note that this implementation will always read at least one row group, to get the list of columns and always
+ have the same schema, even if the requested rows are invalid (out of range).
+
+ Args:
+ offset (int): The first row to read.
+ length (int): The number of rows to read.
+
+ Returns:
+ pa.Table: The requested rows.
+ """
+ if (len(self.row_group_offsets) == 0) or (len(self.row_group_readers) == 0):
+ raise ParquetResponseEmptyError("No parquet files found.")
+ last_row_in_parquet = self.row_group_offsets[-1] - 1
+ first_row = min(offset, last_row_in_parquet)
+ last_row = min(offset + length - 1, last_row_in_parquet)
+ first_row_group_id, last_row_group_id = np.searchsorted(
+ self.row_group_offsets, [first_row, last_row], side="right"
+ )
+ pa_table = pa.concat_tables(
+ [self.row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)]
+ )
+ first_row_in_pa_table = self.row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0
+ return pa_table.slice(offset - first_row_in_pa_table, length)
+
+ @staticmethod
+ def from_parquet_file_items(
+ parquet_file_items: List[ParquetFileItem],
+ dataset: str,
+ config: str,
+ split: str,
+ hf_token: Optional[str],
+ unsupported_features_magic_strings: List[str] = [],
+ ) -> "ParquetIndexWithoutMetadata":
+ try:
+ sources = sorted(f"{config}/{parquet_file['filename']}" for parquet_file in parquet_file_items)
+ except Exception as e:
+ raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e
+ logging.debug(
+ f"Found {len(sources)} parquet files for dataset={dataset}, config={config}, split={split}: {sources}"
+ )
+ if not sources:
+ raise ParquetResponseEmptyError("No parquet files found.")
+ with StepProfiler(
+ method="parquet_index_without_metadata.from_parquet_file_items", step="get the Hub's dataset filesystem"
+ ):
+ fs = get_hf_fs(hf_token=hf_token)
+ with StepProfiler(method="parquet_index_without_metadata.from_parquet_file_items", step="get the source URIs"):
+ source_uris = get_hf_parquet_uris(sources, dataset=dataset)
+ with StepProfiler(
+ method="parquet_index_without_metadata.from_parquet_file_items",
+ step="get one parquet reader per parquet file",
+ ):
+ desc = f"{dataset}/{config}/{split}"
+ try:
+ parquet_files: List[pq.ParquetFile] = thread_map(
+ partial(pq.ParquetFile, filesystem=fs), source_uris, desc=desc, unit="pq", disable=True
+ )
+ except Exception as e:
+ raise FileSystemError(f"Could not read the parquet files: {e}") from e
+ with StepProfiler(
+ method="parquet_index_without_metadata.from_parquet_file_items", step="get the dataset's features"
+ ):
+ features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema())
+ supported_columns, unsupported_columns = get_supported_unsupported_columns(
+ features, unsupported_features_magic_strings=unsupported_features_magic_strings
+ )
+
+ with StepProfiler(
+ method="parquet_index_without_metadata.from_parquet_file_items", step="create the row group offsets"
+ ):
+ row_group_offsets = np.cumsum(
+ [
+ parquet_file.metadata.row_group(group_id).num_rows
+ for parquet_file in parquet_files
+ for group_id in range(parquet_file.metadata.num_row_groups)
+ ]
+ )
+ with StepProfiler(
+ method="parquet_index_without_metadata.from_parquet_file_items", step="create the row group readers"
+ ):
+ row_group_readers: List[Callable[[], pa.Table]] = [
+ partial(parquet_file.read_row_group, i=group_id, columns=supported_columns)
+ for parquet_file in parquet_files
+ for group_id in range(parquet_file.metadata.num_row_groups)
+ ]
+ return ParquetIndexWithoutMetadata(
+ features=features,
+ supported_columns=supported_columns,
+ unsupported_columns=unsupported_columns,
+ row_group_offsets=row_group_offsets,
+ row_group_readers=row_group_readers,
+ )
+
+
+@dataclass
+class ParquetIndexWithMetadata:
+ features: Features
+ supported_columns: List[str]
+ unsupported_columns: List[str]
+ parquet_files_urls: List[str]
+ metadata_paths: List[str]
+ num_bytes: List[int]
+ num_rows: List[int]
+ hf_token: Optional[str]
+
+ def query(self, offset: int, length: int) -> pa.Table:
+ """Query the parquet files
+
+ Note that this implementation will always read at least one row group, to get the list of columns and always
+ have the same schema, even if the requested rows are invalid (out of range).
+
+ Args:
+ offset (int): The first row to read.
+ length (int): The number of rows to read.
+
+ Returns:
+ pa.Table: The requested rows.
+ """
+ with StepProfiler(
+ method="parquet_index_with_metadata.query", step="get the parquet files than contain the requested rows"
+ ):
+ parquet_file_offsets = np.cumsum(self.num_rows)
+
+ last_row_in_parquet = parquet_file_offsets[-1] - 1
+ first_row = min(offset, last_row_in_parquet)
+ last_row = min(offset + length - 1, last_row_in_parquet)
+ first_parquet_file_id, last_parquet_file_id = np.searchsorted(
+ parquet_file_offsets, [first_row, last_row], side="right"
+ )
+ parquet_offset = (
+ offset - parquet_file_offsets[first_parquet_file_id - 1] if first_parquet_file_id > 0 else offset
+ )
+ urls = self.parquet_files_urls[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203
+ metadata_paths = self.metadata_paths[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203
+ num_bytes = self.num_bytes[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203
+
+ with StepProfiler(
+ method="parquet_index_with_metadata.query", step="load the remote parquet files using metadata from disk"
+ ):
+ httpFileSystemSession = HTTPFileSystemSession.get_instance()
+ session = httpFileSystemSession.session
+ httpfs = httpFileSystemSession.httpfs
+ parquet_files = [
+ pq.ParquetFile(
+ HTTPFile(httpfs, url, session=session, size=size, loop=httpfs.loop, cache_type=None),
+ metadata=pq.read_metadata(metadata_path),
+ pre_buffer=True,
+ )
+ for url, metadata_path, size in zip(urls, metadata_paths, num_bytes)
+ ]
+
+ with StepProfiler(
+ method="parquet_index_with_metadata.query", step="get the row groups than contain the requested rows"
+ ):
+ row_group_offsets = np.cumsum(
+ [
+ parquet_file.metadata.row_group(group_id).num_rows
+ for parquet_file in parquet_files
+ for group_id in range(parquet_file.metadata.num_row_groups)
+ ]
+ )
+ row_group_readers: List[Callable[[], pa.Table]] = [
+ partial(parquet_file.read_row_group, i=group_id, columns=self.supported_columns)
+ for parquet_file in parquet_files
+ for group_id in range(parquet_file.metadata.num_row_groups)
+ ]
+
+ last_row_in_parquet = row_group_offsets[-1] - 1
+ first_row = min(parquet_offset, last_row_in_parquet)
+ last_row = min(parquet_offset + length - 1, last_row_in_parquet)
+
+ first_row_group_id, last_row_group_id = np.searchsorted(
+ row_group_offsets, [first_row, last_row], side="right"
+ )
+
+ with StepProfiler(method="parquet_index_with_metadata.query", step="read the row groups"):
+ pa_table = pa.concat_tables(
+ [row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)]
+ )
+ first_row_in_pa_table = row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0
+ return pa_table.slice(parquet_offset - first_row_in_pa_table, length)
+
+ @staticmethod
+ def from_parquet_metadata_items(
+ parquet_file_metadata_items: List[ParquetFileMetadataItem],
+ parquet_metadata_directory: StrPath,
+ hf_token: Optional[str],
+ unsupported_features_magic_strings: List[str] = [],
+ ) -> "ParquetIndexWithMetadata":
+ if not parquet_file_metadata_items:
+ raise ParquetResponseEmptyError("No parquet files found.")
+
+ with StepProfiler(
+ method="parquet_index_with_metadata.from_parquet_metadata_items",
+ step="get the index from parquet metadata",
+ ):
+ try:
+ parquet_files_metadata = sorted(
+ parquet_file_metadata_items, key=lambda parquet_file_metadata: parquet_file_metadata["filename"]
+ )
+ parquet_files_urls = [parquet_file_metadata["url"] for parquet_file_metadata in parquet_files_metadata]
+ metadata_paths = [
+ os.path.join(parquet_metadata_directory, parquet_file_metadata["parquet_metadata_subpath"])
+ for parquet_file_metadata in parquet_files_metadata
+ ]
+ num_bytes = [parquet_file_metadata["size"] for parquet_file_metadata in parquet_files_metadata]
+ num_rows = [parquet_file_metadata["num_rows"] for parquet_file_metadata in parquet_files_metadata]
+ except Exception as e:
+ raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e
+
+ with StepProfiler(
+ method="parquet_index_with_metadata.from_parquet_metadata_items", step="get the dataset's features"
+ ):
+ features = Features.from_arrow_schema(pq.read_schema(metadata_paths[0]))
+ supported_columns, unsupported_columns = get_supported_unsupported_columns(
+ features,
+ unsupported_features_magic_strings=unsupported_features_magic_strings,
+ )
+ return ParquetIndexWithMetadata(
+ features=features,
+ supported_columns=supported_columns,
+ unsupported_columns=unsupported_columns,
+ parquet_files_urls=parquet_files_urls,
+ metadata_paths=metadata_paths,
+ num_bytes=num_bytes,
+ num_rows=num_rows,
+ hf_token=hf_token,
+ )
+
+
+class RowsIndex:
+ def __init__(
+ self,
+ dataset: str,
+ config: str,
+ split: str,
+ processing_graph: ProcessingGraph,
+ hf_token: Optional[str],
+ parquet_metadata_directory: StrPath,
+ unsupported_features_magic_strings: List[str] = [],
+ ):
+ self.dataset = dataset
+ self.revision: Optional[str] = None
+ self.config = config
+ self.split = split
+ self.processing_graph = processing_graph
+ self.parquet_index = self._init_parquet_index(
+ hf_token=hf_token,
+ parquet_metadata_directory=parquet_metadata_directory,
+ unsupported_features_magic_strings=unsupported_features_magic_strings,
+ )
+
+ def _init_parquet_index(
+ self,
+ hf_token: Optional[str],
+ parquet_metadata_directory: StrPath,
+ unsupported_features_magic_strings: List[str] = [],
+ ) -> Union[ParquetIndexWithMetadata, ParquetIndexWithoutMetadata]:
+ with StepProfiler(method="rows_index._init_parquet_index", step="all"):
+ # get the list of parquet files
+ with StepProfiler(method="rows_index._init_parquet_index", step="get list of parquet files for split"):
+ config_parquet_processing_steps = self.processing_graph.get_config_parquet_processing_steps()
+ if not config_parquet_processing_steps:
+ raise RuntimeError("No processing steps are configured to provide a config's parquet response.")
+
+ config_parquet_metadata_processing_steps = (
+ self.processing_graph.get_config_parquet_metadata_processing_steps()
+ )
+
+ cache_kinds = [step.cache_kind for step in config_parquet_processing_steps]
+ cache_kinds.extend([step.cache_kind for step in config_parquet_metadata_processing_steps])
+
+ try:
+ result = get_previous_step_or_raise(
+ kinds=cache_kinds,
+ dataset=self.dataset,
+ config=self.config,
+ split=None,
+ )
+ self.revision = result.response["dataset_git_revision"]
+ content = result.response["content"]
+ except Exception as e:
+ raise UnexpectedError("Could not get the list of parquet files to fetch the rows from.") from e
+ # ^ TODO: improve the error, depending on the case
+ if content and "parquet_files" in content:
+ return ParquetIndexWithoutMetadata.from_parquet_file_items(
+ [
+ parquet_item
+ for parquet_item in content["parquet_files"]
+ if parquet_item["split"] == self.split and parquet_item["config"] == self.config
+ ],
+ dataset=self.dataset,
+ config=self.config,
+ split=self.split,
+ hf_token=hf_token,
+ unsupported_features_magic_strings=unsupported_features_magic_strings,
+ )
+ else:
+ return ParquetIndexWithMetadata.from_parquet_metadata_items(
+ [
+ parquet_item
+ for parquet_item in content["parquet_files_metadata"]
+ if parquet_item["split"] == self.split and parquet_item["config"] == self.config
+ ],
+ parquet_metadata_directory=parquet_metadata_directory,
+ hf_token=hf_token,
+ unsupported_features_magic_strings=unsupported_features_magic_strings,
+ )
+
+ # note that this cache size is global for the class, not per instance
+ @lru_cache(maxsize=1024)
+ def query(self, offset: int, length: int) -> pa.Table:
+ """Query the parquet files
+
+ Note that this implementation will always read at least one row group, to get the list of columns and always
+ have the same schema, even if the requested rows are invalid (out of range).
+
+ Args:
+ offset (int): The first row to read.
+ length (int): The number of rows to read.
+
+ Returns:
+ pa.Table: The requested rows.
+ """
+ return self.parquet_index.query(offset=offset, length=length)
+
+
+class Indexer:
+ def __init__(
+ self,
+ processing_graph: ProcessingGraph,
+ parquet_metadata_directory: StrPath,
+ unsupported_features_magic_strings: List[str] = [],
+ all_columns_supported_datasets_allow_list: Union[Literal["all"], List[str]] = "all",
+ hf_token: Optional[str] = None,
+ ):
+ self.processing_graph = processing_graph
+ self.parquet_metadata_directory = parquet_metadata_directory
+ self.hf_token = hf_token
+ self.unsupported_features_magic_strings = unsupported_features_magic_strings
+ self.all_columns_supported_datasets_allow_list = all_columns_supported_datasets_allow_list
+
+ @lru_cache(maxsize=128)
+ def get_rows_index(
+ self,
+ dataset: str,
+ config: str,
+ split: str,
+ ) -> RowsIndex:
+ filter_magic_strings = (
+ self.all_columns_supported_datasets_allow_list != "all"
+ and dataset not in self.all_columns_supported_datasets_allow_list
+ )
+ unsupported_features_magic_strings = self.unsupported_features_magic_strings if filter_magic_strings else []
+ return RowsIndex(
+ dataset=dataset,
+ config=config,
+ split=split,
+ processing_graph=self.processing_graph,
+ hf_token=self.hf_token,
+ parquet_metadata_directory=self.parquet_metadata_directory,
+ unsupported_features_magic_strings=unsupported_features_magic_strings,
+ )
diff --git a/libs/libcommon/src/libcommon/simple_cache.py b/libs/libcommon/src/libcommon/simple_cache.py
index d70a1112..f537eb28 100644
--- a/libs/libcommon/src/libcommon/simple_cache.py
+++ b/libs/libcommon/src/libcommon/simple_cache.py
@@ -402,0 +403,17 @@ def get_best_response(
+def get_previous_step_or_raise(
+ kinds: List[str], dataset: str, config: Optional[str] = None, split: Optional[str] = None
+) -> BestResponse:
+ """Get the previous step from the cache, or raise an exception if it failed."""
+ best_response = get_best_response(kinds=kinds, dataset=dataset, config=config, split=split)
+ if best_response.response["http_status"] != HTTPStatus.OK:
+ raise CachedArtifactError(
+ message="The previous step failed.",
+ kind=best_response.kind,
+ dataset=dataset,
+ config=config,
+ split=split,
+ cache_entry_with_details=best_response.response,
+ )
+ return best_response
+
+
diff --git a/services/api/src/api/app.py b/services/api/src/api/app.py
index 13479c33..34d68dc4 100644
--- a/services/api/src/api/app.py
+++ b/services/api/src/api/app.py
@@ -114 +113,0 @@ def create_app_with_config(app_config: AppConfig, endpoint_config: EndpointConfi
- hf_endpoint=app_config.common.hf_endpoint,
diff --git a/services/api/src/api/routes/rows.py b/services/api/src/api/routes/rows.py
index 7fca6036..702fea88 100644
--- a/services/api/src/api/routes/rows.py
+++ b/services/api/src/api/routes/rows.py
@@ -9,2 +8,0 @@ import shutil
-from dataclasses import dataclass
-from functools import lru_cache, partial
@@ -12,12 +10 @@ from itertools import islice
-from os import PathLike
-from typing import (
- Any,
- Callable,
- List,
- Literal,
- Mapping,
- Optional,
- Tuple,
- TypedDict,
- Union,
-)
+from typing import Any, List, Literal, Mapping, Optional, TypedDict, Union
@@ -25,2 +11,0 @@ from typing import (
-import numpy as np
-import numpy.typing as npt
@@ -28 +12,0 @@ import pyarrow as pa
-import pyarrow.parquet as pq
@@ -30,3 +14,2 @@ from datasets import Features
-from fsspec.implementations.http import HTTPFile, HTTPFileSystem
-from huggingface_hub import HfFileSystem
-from huggingface_hub.hf_file_system import safe_quote
+from fsspec.implementations.http import HTTPFileSystem
+from libcommon.parquet_utils import Indexer, StrPath
@@ -42 +24,0 @@ from starlette.responses import Response
-from tqdm.contrib.concurrent import thread_map
@@ -45 +26,0 @@ from api.authentication import auth_check
-from api.routes.endpoint import get_cache_entry_from_steps
@@ -66,17 +46,0 @@ MAX_ROWS = 100
-PARQUET_REVISION = "refs/convert/parquet"
-
-
-StrPath = Union[str, PathLike[str]]
-
-
-class FileSystemError(Exception):
- pass
-
-
-class ParquetResponseFormatError(Exception):
- pass
-
-
-class ParquetResponseEmptyError(Exception):
- pass
-
@@ -88,45 +51,0 @@ class ParquetDataProcessingError(Exception):
-class ParquetFileItem(TypedDict):
- dataset: str
- config: str
- split: str
- url: str
- filename: str
- size: int
-
-
-class ParquetFileMetadataItem(TypedDict):
- dataset: str
- config: str
- split: str
- url: str
- filename: str
- size: int
- num_rows: int
- parquet_metadata_subpath: str
-
-
-# TODO: how to invalidate the cache when the parquet branch is created or deleted?
-@lru_cache(maxsize=128)
-def get_hf_fs(hf_token: Optional[str]) -> HfFileSystem:
- """Get the Hugging Face filesystem.
-
- Args:
- hf_token (Optional[str]): The token to access the filesystem.
- Returns:
- HfFileSystem: The Hugging Face filesystem.
- """
- return HfFileSystem(token=hf_token)
-
-
-def get_hf_parquet_uris(paths: List[str], dataset: str) -> List[str]:
- """Get the Hugging Face URIs from the Parquet branch of the dataset repository (see PARQUET_REVISION).
-
- Args:
- paths (List[str]): List of paths.
- dataset (str): The dataset name.
- Returns:
- List[str]: List of Parquet URIs.
- """
- return [f"hf://datasets/{dataset}@{safe_quote(PARQUET_REVISION)}/{path}" for path in paths]
-
-
@@ -138,359 +56,0 @@ UNSUPPORTED_FEATURES_MAGIC_STRINGS = ["'binary'", "Audio("]
-
-def get_supported_unsupported_columns(features: Features, dataset_name: str) -> Tuple[List[str], List[str]]:
- supported_columns, unsupported_columns = [], []
- unsupported_features_magic_strings = (
- UNSUPPORTED_FEATURES_MAGIC_STRINGS
- if ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST != "all"
- and dataset_name not in ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST
- else []
- )
- for column, feature in features.items():
- str_feature = str(feature)
- str_column = str(column)
- if any(magic_string in str_feature for magic_string in unsupported_features_magic_strings):
- unsupported_columns.append(str_column)
- else:
- supported_columns.append(str_column)
- return supported_columns, unsupported_columns
-
-
-@dataclass
-class ParquetIndexWithoutMetadata:
- features: Features
- supported_columns: List[str]
- unsupported_columns: List[str]
- row_group_offsets: npt.NDArray[np.int64]
- row_group_readers: List[Callable[[], pa.Table]]
-
- def query(self, offset: int, length: int) -> pa.Table:
- """Query the parquet files
-
- Note that this implementation will always read at least one row group, to get the list of columns and always
- have the same schema, even if the requested rows are invalid (out of range).
-
- Args:
- offset (int): The first row to read.
- length (int): The number of rows to read.
-
- Returns:
- pa.Table: The requested rows.
- """
- if (len(self.row_group_offsets) == 0) or (len(self.row_group_readers) == 0):
- raise ParquetResponseEmptyError("No parquet files found.")
- last_row_in_parquet = self.row_group_offsets[-1] - 1
- first_row = min(offset, last_row_in_parquet)
- last_row = min(offset + length - 1, last_row_in_parquet)
- first_row_group_id, last_row_group_id = np.searchsorted(
- self.row_group_offsets, [first_row, last_row], side="right"
- )
- pa_table = pa.concat_tables(
- [self.row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)]
- )
- first_row_in_pa_table = self.row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0
- return pa_table.slice(offset - first_row_in_pa_table, length)
-
- @staticmethod
- def from_parquet_file_items(
- parquet_file_items: List[ParquetFileItem], dataset: str, config: str, split: str, hf_token: Optional[str]
- ) -> "ParquetIndexWithoutMetadata":
- try:
- sources = sorted(f"{config}/{parquet_file['filename']}" for parquet_file in parquet_file_items)
- except Exception as e:
- raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e
- logging.debug(
- f"Found {len(sources)} parquet files for dataset={dataset}, config={config}, split={split}: {sources}"
- )
- if not sources:
- raise ParquetResponseEmptyError("No parquet files found.")
- with StepProfiler(method="rows.index.without_metadata", step="get the Hub's dataset filesystem"):
- fs = get_hf_fs(hf_token=hf_token)
- with StepProfiler(method="rows.index.without_metadata", step="get the source URIs"):
- source_uris = get_hf_parquet_uris(sources, dataset=dataset)
- with StepProfiler(method="rows.index.without_metadata", step="get one parquet reader per parquet file"):
- desc = f"{dataset}/{config}/{split}"
- try:
- parquet_files: List[pq.ParquetFile] = thread_map(
- partial(pq.ParquetFile, filesystem=fs), source_uris, desc=desc, unit="pq", disable=True
- )
- except Exception as e:
- raise FileSystemError(f"Could not read the parquet files: {e}") from e
- with StepProfiler(method="rows.index.without_metadata", step="get the dataset's features"):
- features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema())
- supported_columns, unsupported_columns = get_supported_unsupported_columns(features, dataset_name=dataset)
-
- with StepProfiler(method="rows.index.without_metadata", step="create the row group offsets"):
- row_group_offsets = np.cumsum(
- [
- parquet_file.metadata.row_group(group_id).num_rows
- for parquet_file in parquet_files
- for group_id in range(parquet_file.metadata.num_row_groups)
- ]
- )
- with StepProfiler(method="rows.index.without_metadata", step="create the row group readers"):
- row_group_readers: List[Callable[[], pa.Table]] = [
- partial(parquet_file.read_row_group, i=group_id, columns=supported_columns)
- for parquet_file in parquet_files
- for group_id in range(parquet_file.metadata.num_row_groups)
- ]
- return ParquetIndexWithoutMetadata(
- features=features,
- supported_columns=supported_columns,
- unsupported_columns=unsupported_columns,
- row_group_offsets=row_group_offsets,
- row_group_readers=row_group_readers,
- )
-
-
-@dataclass
-class ParquetIndexWithMetadata:
- features: Features
- supported_columns: List[str]
- unsupported_columns: List[str]
- parquet_files_urls: List[str]
- metadata_paths: List[str]
- num_bytes: List[int]
- num_rows: List[int]
- hf_token: Optional[str]
-
- def query(self, offset: int, length: int) -> pa.Table:
- """Query the parquet files
-
- Note that this implementation will always read at least one row group, to get the list of columns and always
- have the same schema, even if the requested rows are invalid (out of range).
-
- Args:
- offset (int): The first row to read.
- length (int): The number of rows to read.
-
- Returns:
- pa.Table: The requested rows.
- """
- with StepProfiler(
- method="rows.query.with_metadata", step="get the parquet files than contain the requested rows"
- ):
- parquet_file_offsets = np.cumsum(self.num_rows)
-
- last_row_in_parquet = parquet_file_offsets[-1] - 1
- first_row = min(offset, last_row_in_parquet)
- last_row = min(offset + length - 1, last_row_in_parquet)
- first_parquet_file_id, last_parquet_file_id = np.searchsorted(
- parquet_file_offsets, [first_row, last_row], side="right"
- )
- parquet_offset = (
- offset - parquet_file_offsets[first_parquet_file_id - 1] if first_parquet_file_id > 0 else offset
- )
- urls = self.parquet_files_urls[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203
- metadata_paths = self.metadata_paths[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203
- num_bytes = self.num_bytes[first_parquet_file_id : last_parquet_file_id + 1] # noqa: E203
-
- with StepProfiler(
- method="rows.query.with_metadata", step="load the remote parquet files using metadata from disk"
- ):
- parquet_files = [
- pq.ParquetFile(
- HTTPFile(httpfs, url, session=session, size=size, loop=httpfs.loop, cache_type=None),
- metadata=pq.read_metadata(metadata_path),
- pre_buffer=True,
- )
- for url, metadata_path, size in zip(urls, metadata_paths, num_bytes)
- ]
-
- with StepProfiler(
- method="rows.query.with_metadata", step="get the row groups than contain the requested rows"
- ):
- row_group_offsets = np.cumsum(
- [
- parquet_file.metadata.row_group(group_id).num_rows
- for parquet_file in parquet_files
- for group_id in range(parquet_file.metadata.num_row_groups)
- ]
- )
- row_group_readers: List[Callable[[], pa.Table]] = [
- partial(parquet_file.read_row_group, i=group_id, columns=self.supported_columns)
- for parquet_file in parquet_files
- for group_id in range(parquet_file.metadata.num_row_groups)
- ]
-
- last_row_in_parquet = row_group_offsets[-1] - 1
- first_row = min(parquet_offset, last_row_in_parquet)
- last_row = min(parquet_offset + length - 1, last_row_in_parquet)
-
- first_row_group_id, last_row_group_id = np.searchsorted(
- row_group_offsets, [first_row, last_row], side="right"
- )
-
- with StepProfiler(method="rows.query.with_metadata", step="read the row groups"):
- pa_table = pa.concat_tables(
- [row_group_readers[i]() for i in range(first_row_group_id, last_row_group_id + 1)]
- )
- first_row_in_pa_table = row_group_offsets[first_row_group_id - 1] if first_row_group_id > 0 else 0
- return pa_table.slice(parquet_offset - first_row_in_pa_table, length)
-
- @staticmethod
- def from_parquet_metadata_items(
- parquet_file_metadata_items: List[ParquetFileMetadataItem],
- parquet_metadata_directory: StrPath,
- hf_token: Optional[str],
- ) -> "ParquetIndexWithMetadata":
- if not parquet_file_metadata_items:
- raise ParquetResponseEmptyError("No parquet files found.")
-
- with StepProfiler(method="rows.index", step="get the index from parquet metadata"):
- try:
- parquet_files_metadata = sorted(
- parquet_file_metadata_items, key=lambda parquet_file_metadata: parquet_file_metadata["filename"]
- )
- parquet_files_urls = [parquet_file_metadata["url"] for parquet_file_metadata in parquet_files_metadata]
- metadata_paths = [
- os.path.join(parquet_metadata_directory, parquet_file_metadata["parquet_metadata_subpath"])
- for parquet_file_metadata in parquet_files_metadata
- ]
- num_bytes = [parquet_file_metadata["size"] for parquet_file_metadata in parquet_files_metadata]
- num_rows = [parquet_file_metadata["num_rows"] for parquet_file_metadata in parquet_files_metadata]
- dataset_name = parquet_files_metadata[0]["dataset"]
- except Exception as e:
- raise ParquetResponseFormatError(f"Could not parse the list of parquet files: {e}") from e
-
- with StepProfiler(method="rows.index.with_metadata", step="get the dataset's features"):
- features = Features.from_arrow_schema(pq.read_schema(metadata_paths[0]))
- supported_columns, unsupported_columns = get_supported_unsupported_columns(
- features, dataset_name=dataset_name
- )
- return ParquetIndexWithMetadata(
- features=features,
- supported_columns=supported_columns,
- unsupported_columns=unsupported_columns,
- parquet_files_urls=parquet_files_urls,
- metadata_paths=metadata_paths,
- num_bytes=num_bytes,
- num_rows=num_rows,
- hf_token=hf_token,
- )
-
-
-class RowsIndex:
- def __init__(
- self,
- dataset: str,
- config: str,
- split: str,
- processing_graph: ProcessingGraph,
- hf_endpoint: str,
- hf_token: Optional[str],
- parquet_metadata_directory: StrPath,
- ):
- self.dataset = dataset
- self.revision: Optional[str] = None
- self.config = config
- self.split = split
- self.processing_graph = processing_graph
- self.parquet_index = self._init_parquet_index(
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- parquet_metadata_directory=parquet_metadata_directory,
- )
-
- def _init_parquet_index(
- self,
- hf_endpoint: str,
- hf_token: Optional[str],
- parquet_metadata_directory: StrPath,
- ) -> Union[ParquetIndexWithMetadata, ParquetIndexWithoutMetadata]:
- with StepProfiler(method="rows.index", step="all"):
- # get the list of parquet files
- with StepProfiler(method="rows.index", step="get list of parquet files for split"):
- config_parquet_processing_steps = self.processing_graph.get_config_parquet_processing_steps()
- config_parquet_metadata_processing_steps = (
- self.processing_graph.get_config_parquet_metadata_processing_steps()
- )
- if not config_parquet_processing_steps:
- raise RuntimeError("No processing steps are configured to provide a config's parquet response.")
- try:
- result = get_cache_entry_from_steps(
- processing_steps=config_parquet_metadata_processing_steps + config_parquet_processing_steps,
- dataset=self.dataset,
- config=self.config,
- split=None,
- processing_graph=self.processing_graph,
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- )
- self.revision = result["dataset_git_revision"]
- content = result["content"]
- except ApiCustomError as e:
- raise e
- except Exception as e:
- raise UnexpectedError("Could not get the list of parquet files to fetch the rows from.") from e
- # ^ TODO: improve the error, depending on the case
- if content and "parquet_files" in content:
- return ParquetIndexWithoutMetadata.from_parquet_file_items(
- [
- parquet_item
- for parquet_item in content["parquet_files"]
- if parquet_item["split"] == self.split and parquet_item["config"] == self.config
- ],
- dataset=self.dataset,
- config=self.config,
- split=self.split,
- hf_token=hf_token,
- )
- else:
- return ParquetIndexWithMetadata.from_parquet_metadata_items(
- [
- parquet_item
- for parquet_item in content["parquet_files_metadata"]
- if parquet_item["split"] == self.split and parquet_item["config"] == self.config
- ],
- parquet_metadata_directory=parquet_metadata_directory,
- hf_token=hf_token,
- )
-
- # note that this cache size is global for the class, not per instance
- @lru_cache(maxsize=1024)
- def query(self, offset: int, length: int) -> pa.Table:
- """Query the parquet files
-
- Note that this implementation will always read at least one row group, to get the list of columns and always
- have the same schema, even if the requested rows are invalid (out of range).
-
- Args:
- offset (int): The first row to read.
- length (int): The number of rows to read.
-
- Returns:
- pa.Table: The requested rows.
- """
- return self.parquet_index.query(offset=offset, length=length)
-
-
-class Indexer:
- def __init__(
- self,
- processing_graph: ProcessingGraph,
- parquet_metadata_directory: StrPath,
- hf_endpoint: str,
- hf_token: Optional[str] = None,
- ):
- self.processing_graph = processing_graph
- self.parquet_metadata_directory = parquet_metadata_directory
- self.hf_endpoint = hf_endpoint
- self.hf_token = hf_token
-
- @lru_cache(maxsize=128)
- def get_rows_index(
- self,
- dataset: str,
- config: str,
- split: str,
- ) -> RowsIndex:
- return RowsIndex(
- dataset=dataset,
- config=config,
- split=split,
- processing_graph=self.processing_graph,
- hf_endpoint=self.hf_endpoint,
- hf_token=self.hf_token,
- parquet_metadata_directory=self.parquet_metadata_directory,
- )
-
-
@@ -700 +259,0 @@ def create_rows_endpoint(
- hf_endpoint: str,
@@ -715 +273,0 @@ def create_rows_endpoint(
- hf_endpoint=hf_endpoint,
@@ -717,0 +276,2 @@ def create_rows_endpoint(
+ unsupported_features_magic_strings=UNSUPPORTED_FEATURES_MAGIC_STRINGS,
+ all_columns_supported_datasets_allow_list=ALL_COLUMNS_SUPPORTED_DATASETS_ALLOW_LIST,
@@ -752 +312,5 @@ def create_rows_endpoint(
- rows_index = indexer.get_rows_index(dataset=dataset, config=config, split=split)
+ rows_index = indexer.get_rows_index(
+ dataset=dataset,
+ config=config,
+ split=split,
+ )
diff --git a/services/api/tests/routes/test_rows.py b/services/api/tests/routes/test_rows.py
index 4b3f7f86..9dbad3ed 100644
--- a/services/api/tests/routes/test_rows.py
+++ b/services/api/tests/routes/test_rows.py
@@ -14,0 +15,6 @@ from fsspec import AbstractFileSystem
+from libcommon.parquet_utils import (
+ Indexer,
+ ParquetIndexWithMetadata,
+ ParquetIndexWithoutMetadata,
+ RowsIndex,
+)
@@ -21,8 +27 @@ from api.config import AppConfig
-from api.routes.rows import (
- Indexer,
- ParquetIndexWithMetadata,
- ParquetIndexWithoutMetadata,
- RowsIndex,
- clean_cached_assets,
- create_response,
-)
+from api.routes.rows import clean_cached_assets, create_response
@@ -251 +249,0 @@ def indexer(app_config: AppConfig, processing_graph: ProcessingGraph, parquet_me
- hf_endpoint=app_config.common.hf_endpoint,
@@ -268,2 +266,2 @@ def rows_index(
- with patch("api.routes.rows.get_hf_fs", return_value=ds_sharded_fs):
- with patch("api.routes.rows.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris):
+ with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_sharded_fs):
+ with patch("libcommon.parquet_utils.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris):
@@ -276,2 +274,2 @@ def test_indexer_get_rows_index(
- with patch("api.routes.rows.get_hf_fs", return_value=ds_fs):
- with patch("api.routes.rows.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris):
+ with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_fs):
+ with patch("libcommon.parquet_utils.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris):
@@ -295,2 +293,2 @@ def test_indexer_get_rows_index_sharded(
- with patch("api.routes.rows.get_hf_fs", return_value=ds_sharded_fs):
- with patch("api.routes.rows.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris):
+ with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_sharded_fs):
+ with patch("libcommon.parquet_utils.get_hf_parquet_uris", side_effect=mock_get_hf_parquet_uris):
@@ -325 +323 @@ def rows_index_with_parquet_metadata(
- with patch("api.routes.rows.HTTPFile", return_value=f):
+ with patch("libcommon.parquet_utils.HTTPFile", return_value=f):
@@ -333 +331 @@ def test_indexer_get_rows_index_with_parquet_metadata(
- with patch("api.routes.rows.HTTPFile", return_value=f):
+ with patch("libcommon.parquet_utils.HTTPFile", return_value=f):
@@ -354 +352 @@ def test_indexer_get_rows_index_sharded_with_parquet_metadata(
- with patch("api.routes.rows.HTTPFile", return_value=f):
+ with patch("libcommon.parquet_utils.HTTPFile", return_value=f):
diff --git a/services/worker/src/worker/job_runner_factory.py b/services/worker/src/worker/job_runner_factory.py
index 0e4bfe82..2352944c 100644
--- a/services/worker/src/worker/job_runner_factory.py
+++ b/services/worker/src/worker/job_runner_factory.py
@@ -174,0 +175 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ processing_graph=self.processing_graph,
@@ -175,0 +177 @@ class JobRunnerFactory(BaseJobRunnerFactory):
+ parquet_metadata_directory=self.parquet_metadata_directory,
diff --git a/services/worker/src/worker/job_runners/config/info.py b/services/worker/src/worker/job_runners/config/info.py
index 265ef38b..ae0e2760 100644
--- a/services/worker/src/worker/job_runners/config/info.py
+++ b/services/worker/src/worker/job_runners/config/info.py
@@ -5,0 +6 @@ from libcommon.exceptions import PreviousStepFormatError
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -8 +9 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import CompleteJobResult, get_previous_step_or_raise
+from worker.utils import CompleteJobResult
diff --git a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
index 54b18b0b..9c9306c4 100644
--- a/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/config/opt_in_out_urls_count.py
@@ -10,6 +10,2 @@ from libcommon.exceptions import PreviousStepFormatError
-from libcommon.simple_cache import DoesNotExist, get_response
-
-from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import (
- JobResult,
- OptInOutUrlsCountResponse,
+from libcommon.simple_cache import (
+ DoesNotExist,
@@ -16,0 +13 @@ from worker.utils import (
+ get_response,
@@ -18,0 +16,3 @@ from worker.utils import (
+from worker.job_runners.config.config_job_runner import ConfigJobRunner
+from worker.utils import JobResult, OptInOutUrlsCountResponse
+
diff --git a/services/worker/src/worker/job_runners/config/parquet.py b/services/worker/src/worker/job_runners/config/parquet.py
index ee88039b..eae1d780 100644
--- a/services/worker/src/worker/job_runners/config/parquet.py
+++ b/services/worker/src/worker/job_runners/config/parquet.py
@@ -8,0 +9 @@ from libcommon.exceptions import PreviousStepFormatError
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -12 +13 @@ from worker.job_runners.config.parquet_and_info import ParquetFileItem
-from worker.utils import CompleteJobResult, get_previous_step_or_raise
+from worker.utils import CompleteJobResult
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index e6c4b1b2..91de381e 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -68,0 +69 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -74 +75 @@ from worker.job_runners.config.config_job_runner import ConfigCachedJobRunner
-from worker.utils import CompleteJobResult, get_previous_step_or_raise
+from worker.utils import CompleteJobResult
diff --git a/services/worker/src/worker/job_runners/config/parquet_metadata.py b/services/worker/src/worker/job_runners/config/parquet_metadata.py
index 97600419..e1cb6ef5 100644
--- a/services/worker/src/worker/job_runners/config/parquet_metadata.py
+++ b/services/worker/src/worker/job_runners/config/parquet_metadata.py
@@ -16,0 +17 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -26 +27 @@ from worker.job_runners.config.parquet_and_info import ParquetFileItem
-from worker.utils import CompleteJobResult, get_previous_step_or_raise
+from worker.utils import CompleteJobResult
diff --git a/services/worker/src/worker/job_runners/config/size.py b/services/worker/src/worker/job_runners/config/size.py
index 4b2dfca4..e46b9108 100644
--- a/services/worker/src/worker/job_runners/config/size.py
+++ b/services/worker/src/worker/job_runners/config/size.py
@@ -8,0 +9 @@ from libcommon.exceptions import PreviousStepFormatError
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -11 +12 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import CompleteJobResult, get_previous_step_or_raise
+from worker.utils import CompleteJobResult
diff --git a/services/worker/src/worker/job_runners/config/split_names_from_info.py b/services/worker/src/worker/job_runners/config/split_names_from_info.py
index 7d6d9fb1..7037de05 100644
--- a/services/worker/src/worker/job_runners/config/split_names_from_info.py
+++ b/services/worker/src/worker/job_runners/config/split_names_from_info.py
@@ -11,0 +12 @@ from libcommon.exceptions import PreviousStepFormatError
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -14,7 +15 @@ from worker.job_runners.config.config_job_runner import ConfigJobRunner
-from worker.utils import (
- CompleteJobResult,
- JobRunnerInfo,
- SplitItem,
- SplitsList,
- get_previous_step_or_raise,
-)
+from worker.utils import CompleteJobResult, JobRunnerInfo, SplitItem, SplitsList
diff --git a/services/worker/src/worker/job_runners/dataset/info.py b/services/worker/src/worker/job_runners/dataset/info.py
index 2cedb82a..ad0c1408 100644
--- a/services/worker/src/worker/job_runners/dataset/info.py
+++ b/services/worker/src/worker/job_runners/dataset/info.py
@@ -10 +10,5 @@ from libcommon.exceptions import PreviousStepFormatError
-from libcommon.simple_cache import DoesNotExist, get_response
+from libcommon.simple_cache import (
+ DoesNotExist,
+ get_previous_step_or_raise,
+ get_response,
+)
@@ -13 +17 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult, PreviousJob, get_previous_step_or_raise
+from worker.utils import JobResult, PreviousJob
diff --git a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
index 8a36fbae..530d3145 100644
--- a/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/dataset/opt_in_out_urls_count.py
@@ -10,6 +10,2 @@ from libcommon.exceptions import PreviousStepFormatError
-from libcommon.simple_cache import DoesNotExist, get_response
-
-from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import (
- JobResult,
- OptInOutUrlsCountResponse,
+from libcommon.simple_cache import (
+ DoesNotExist,
@@ -16,0 +13 @@ from worker.utils import (
+ get_response,
@@ -18,0 +16,3 @@ from worker.utils import (
+from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
+from worker.utils import JobResult, OptInOutUrlsCountResponse
+
diff --git a/services/worker/src/worker/job_runners/dataset/parquet.py b/services/worker/src/worker/job_runners/dataset/parquet.py
index 6d12167f..4a360950 100644
--- a/services/worker/src/worker/job_runners/dataset/parquet.py
+++ b/services/worker/src/worker/job_runners/dataset/parquet.py
@@ -10 +10,5 @@ from libcommon.exceptions import PreviousStepFormatError
-from libcommon.simple_cache import DoesNotExist, get_response
+from libcommon.simple_cache import (
+ DoesNotExist,
+ get_previous_step_or_raise,
+ get_response,
+)
@@ -15 +19 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult, PreviousJob, get_previous_step_or_raise
+from worker.utils import JobResult, PreviousJob
diff --git a/services/worker/src/worker/job_runners/dataset/size.py b/services/worker/src/worker/job_runners/dataset/size.py
index 84db9f3c..3b69e948 100644
--- a/services/worker/src/worker/job_runners/dataset/size.py
+++ b/services/worker/src/worker/job_runners/dataset/size.py
@@ -10 +10,5 @@ from libcommon.exceptions import PreviousStepFormatError
-from libcommon.simple_cache import DoesNotExist, get_response
+from libcommon.simple_cache import (
+ DoesNotExist,
+ get_previous_step_or_raise,
+ get_response,
+)
@@ -14 +18 @@ from worker.job_runners.dataset.dataset_job_runner import DatasetJobRunner
-from worker.utils import JobResult, PreviousJob, get_previous_step_or_raise
+from worker.utils import JobResult, PreviousJob
diff --git a/services/worker/src/worker/job_runners/dataset/split_names.py b/services/worker/src/worker/job_runners/dataset/split_names.py
index eea5a259..259e66d4 100644
--- a/services/worker/src/worker/job_runners/dataset/split_names.py
+++ b/services/worker/src/worker/job_runners/dataset/split_names.py
@@ -10 +10 @@ from libcommon.exceptions import PreviousStepFormatError
-from libcommon.simple_cache import get_best_response
+from libcommon.simple_cache import get_best_response, get_previous_step_or_raise
@@ -19 +18,0 @@ from worker.utils import (
- get_previous_step_or_raise,
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
index 0d6862b0..64112c28 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_parquet.py
@@ -5,2 +5 @@ import logging
-from functools import lru_cache, partial
-from typing import List, Optional
+from typing import List
@@ -8 +6,0 @@ from typing import List, Optional
-import pyarrow as pa
@@ -10,2 +7,0 @@ from datasets import Features
-from huggingface_hub import HfFileSystem
-from huggingface_hub.hf_file_system import safe_quote
@@ -13 +8,0 @@ from libcommon.constants import (
- PARQUET_REVISION,
@@ -18,3 +12,0 @@ from libcommon.exceptions import (
- FileSystemError,
- ParquetResponseEmptyError,
- PreviousStepFormatError,
@@ -25 +17,2 @@ from libcommon.exceptions import (
-from libcommon.processing_graph import ProcessingStep
+from libcommon.parquet_utils import Indexer
+from libcommon.processing_graph import ProcessingGraph, ProcessingStep
@@ -29,2 +21,0 @@ from libcommon.viewer_utils.features import get_cell_value
-from pyarrow.parquet import ParquetFile
-from tqdm.contrib.concurrent import thread_map
@@ -42 +32,0 @@ from worker.utils import (
- get_previous_step_or_raise,
@@ -75,24 +64,0 @@ def transform_rows(
-@lru_cache(maxsize=128)
-def get_hf_fs(hf_token: Optional[str]) -> HfFileSystem:
- """Get the Hugging Face filesystem.
-
- Args:
- hf_token (Optional[str]): The token to access the filesystem.
- Returns:
- HfFileSystem: The Hugging Face filesystem.
- """
- return HfFileSystem(token=hf_token)
-
-
-def get_hf_parquet_uris(paths: List[str], dataset: str) -> List[str]:
- """Get the Hugging Face URIs from the Parquet branch of the dataset repository (see PARQUET_REVISION).
-
- Args:
- paths (List[str]): List of paths.
- dataset (str): The dataset name.
- Returns:
- List[str]: List of Parquet URIs.
- """
- return [f"hf://datasets/{dataset}@{safe_quote(PARQUET_REVISION)}/{path}" for path in paths]
-
-
@@ -104 +69,0 @@ def compute_first_rows_response(
- hf_token: Optional[str],
@@ -110,0 +76 @@ def compute_first_rows_response(
+ indexer: Indexer,
@@ -114,29 +80,5 @@ def compute_first_rows_response(
- # first ensure the tuple (dataset, config, split) exists on the Hub
-
- config_parquet_best_response = get_previous_step_or_raise(kinds=["config-parquet"], dataset=dataset, config=config)
- try:
- parquet_files_content = config_parquet_best_response.response["content"]["parquet_files"]
- sources = sorted(
- f"{config}/{parquet_file['filename']}"
- for parquet_file in parquet_files_content
- if parquet_file["split"] == split and parquet_file["config"] == config
- )
- if not sources:
- raise ParquetResponseEmptyError("No parquet files found.")
- except Exception as e:
- raise PreviousStepFormatError("Previous step did not return the expected content.") from e
-
- logging.debug(f"Found {len(sources)} parquet files for {dataset=}, {config=}, {split=}: {sources}")
-
- fs = get_hf_fs(hf_token=hf_token)
- source_uris = get_hf_parquet_uris(sources, dataset=dataset)
- desc = f"{dataset}/{config}/{split}"
- try:
- parquet_files: List[ParquetFile] = thread_map(
- partial(ParquetFile, filesystem=fs), source_uris, desc=desc, unit="pq", disable=True
- )
- except Exception as e:
- raise FileSystemError(f"Could not read the parquet files: {e}") from e
-
- # get the features
- features = Features.from_arrow_schema(parquet_files[0].schema.to_arrow_schema())
+ rows_index = indexer.get_rows_index(
+ dataset=dataset,
+ config=config,
+ split=split,
+ )
@@ -143,0 +86,2 @@ def compute_first_rows_response(
+ # validate the features
+ features = rows_index.parquet_index.features
@@ -169,22 +113 @@ def compute_first_rows_response(
- num_rows = 0
- last_row_group_id = 0
- row_group_readers = []
- for parquet_file in parquet_files:
- for group_id in range(parquet_file.metadata.num_row_groups):
- last_row_group_id = group_id
- row_group_readers.append(partial(parquet_file.read_row_group, i=group_id))
- if num_rows + parquet_file.metadata.row_group(group_id).num_rows >= rows_max_number:
- num_rows = rows_max_number
- break
- else:
- num_rows += parquet_file.metadata.row_group(group_id).num_rows
- else:
- continue
- break
-
- if len(row_group_readers) == 0:
- raise ParquetResponseEmptyError("No parquet files found.")
-
- pa_table = pa.concat_tables([row_group_readers[i]() for i in range(last_row_group_id + 1)])
- result = pa_table.slice(0, num_rows)
-
+ pa_table = rows_index.query(offset=0, length=rows_max_number)
@@ -199 +122 @@ def compute_first_rows_response(
- for idx, row in enumerate(result.to_pylist())
+ for idx, row in enumerate(pa_table.to_pylist())
@@ -234,0 +158 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner):
+ indexed: Indexer
@@ -255,0 +180 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner):
+ processing_graph: ProcessingGraph,
@@ -256,0 +182 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner):
+ parquet_metadata_directory: StrPath,
@@ -265,0 +192,8 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner):
+ self.parquet_metadata_directory = parquet_metadata_directory
+ self.indexer = Indexer(
+ processing_graph=processing_graph,
+ hf_token=self.app_config.common.hf_token,
+ parquet_metadata_directory=parquet_metadata_directory,
+ unsupported_features_magic_strings=[],
+ all_columns_supported_datasets_allow_list="all",
+ )
@@ -275 +208,0 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner):
- hf_token=self.app_config.common.hf_token,
@@ -280,0 +214 @@ class SplitFirstRowsFromParquetJobRunner(SplitJobRunner):
+ indexer=self.indexer,
diff --git a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
index d3a617cb..43221906 100644
--- a/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split/first_rows_from_streaming.py
@@ -22,0 +23 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -36 +36,0 @@ from worker.utils import (
- get_previous_step_or_raise,
diff --git a/services/worker/src/worker/job_runners/split/image_url_columns.py b/services/worker/src/worker/job_runners/split/image_url_columns.py
index 1686e800..c2b19607 100644
--- a/services/worker/src/worker/job_runners/split/image_url_columns.py
+++ b/services/worker/src/worker/job_runners/split/image_url_columns.py
@@ -7,0 +8 @@ from libcommon.exceptions import PreviousStepFormatError
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -15 +15,0 @@ from worker.utils import (
- get_previous_step_or_raise,
diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py
index cf83ec1f..af094266 100644
--- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py
+++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_count.py
@@ -7,0 +8 @@ from libcommon.exceptions import PreviousStepFormatError
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -10,5 +11 @@ from worker.job_runners.split.split_job_runner import SplitJobRunner
-from worker.utils import (
- CompleteJobResult,
- OptInOutUrlsCountResponse,
- get_previous_step_or_raise,
-)
+from worker.utils import CompleteJobResult, OptInOutUrlsCountResponse
diff --git a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
index 28fd6f0e..7544df77 100644
--- a/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
+++ b/services/worker/src/worker/job_runners/split/opt_in_out_urls_scan_from_streaming.py
@@ -20,0 +21 @@ from libcommon.processing_graph import ProcessingStep
+from libcommon.simple_cache import get_previous_step_or_raise
@@ -29 +29,0 @@ from worker.utils import (
- get_previous_step_or_raise,
diff --git a/services/worker/src/worker/utils.py b/services/worker/src/worker/utils.py
index 17ccd75b..4399bd6e 100644
--- a/services/worker/src/worker/utils.py
+++ b/services/worker/src/worker/utils.py
@@ -10 +9,0 @@ from dataclasses import dataclass, field
-from http import HTTPStatus
@@ -32 +30,0 @@ from libcommon.exceptions import NormalRowsError, StreamingRowsError
-from libcommon.simple_cache import BestResponse, CachedArtifactError, get_best_response
@@ -403,17 +400,0 @@ def get_rows_or_raise(
-
-
-def get_previous_step_or_raise(
- kinds: List[str], dataset: str, config: Optional[str] = None, split: Optional[str] = None
-) -> BestResponse:
- """Get the previous step from the cache, or raise an exception if it failed."""
- best_response = get_best_response(kinds=kinds, dataset=dataset, config=config, split=split)
- if best_response.response["http_status"] != HTTPStatus.OK:
- raise CachedArtifactError(
- message="The previous step failed.",
- kind=best_response.kind,
- dataset=dataset,
- config=config,
- split=split,
- cache_entry_with_details=best_response.response,
- )
- return best_response
diff --git a/services/worker/tests/conftest.py b/services/worker/tests/conftest.py
index 43fd1de7..851a99c6 100644
--- a/services/worker/tests/conftest.py
+++ b/services/worker/tests/conftest.py
@@ -143 +143 @@ def another_processing_step(test_processing_graph: ProcessingGraph) -> Processin
-pytest_plugins = ["tests.fixtures.datasets", "tests.fixtures.files", "tests.fixtures.hub"]
+pytest_plugins = ["tests.fixtures.datasets", "tests.fixtures.files", "tests.fixtures.hub", "tests.fixtures.fsspec"]
diff --git a/services/worker/tests/fixtures/fsspec.py b/services/worker/tests/fixtures/fsspec.py
new file mode 100644
index 00000000..848dceb5
--- /dev/null
+++ b/services/worker/tests/fixtures/fsspec.py
@@ -0,0 +1,119 @@
+# type: ignore
+import posixpath
+import shutil
+from pathlib import Path
+from unittest.mock import patch
+
+import fsspec
+import pytest
+from fsspec.implementations.local import (
+ AbstractFileSystem,
+ LocalFileSystem,
+ stringify_path,
+)
+
+
+class MockFileSystem(AbstractFileSystem):
+ protocol = "mock"
+
+ def __init__(self, *args, local_root_dir, **kwargs):
+ super().__init__()
+ self._fs = LocalFileSystem(*args, **kwargs)
+ self.local_root_dir = Path(local_root_dir).resolve().as_posix() + "/"
+
+ def mkdir(self, path, *args, **kwargs):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs.mkdir(path, *args, **kwargs)
+
+ def makedirs(self, path, *args, **kwargs):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs.makedirs(path, *args, **kwargs)
+
+ def rmdir(self, path):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs.rmdir(path)
+
+ def ls(self, path, detail=True, *args, **kwargs):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ out = self._fs.ls(path, detail=detail, *args, **kwargs)
+ if detail:
+ return [{**info, "name": info["name"][len(self.local_root_dir) :]} for info in out] # noqa: E203
+ else:
+ return [name[len(self.local_root_dir) :] for name in out] # noqa: E203
+
+ def info(self, path, *args, **kwargs):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ out = dict(self._fs.info(path, *args, **kwargs))
+ out["name"] = out["name"][len(self.local_root_dir) :] # noqa: E203
+ return out
+
+ def cp_file(self, path1, path2, *args, **kwargs):
+ path1 = posixpath.join(self.local_root_dir, self._strip_protocol(path1))
+ path2 = posixpath.join(self.local_root_dir, self._strip_protocol(path2))
+ return self._fs.cp_file(path1, path2, *args, **kwargs)
+
+ def rm_file(self, path, *args, **kwargs):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs.rm_file(path, *args, **kwargs)
+
+ def rm(self, path, *args, **kwargs):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs.rm(path, *args, **kwargs)
+
+ def _open(self, path, *args, **kwargs):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs._open(path, *args, **kwargs)
+
+ def created(self, path):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs.created(path)
+
+ def modified(self, path):
+ path = posixpath.join(self.local_root_dir, self._strip_protocol(path))
+ return self._fs.modified(path)
+
+ @classmethod
+ def _strip_protocol(cls, path):
+ path = stringify_path(path)
+ if path.startswith("mock://"):
+ path = path[7:]
+ return path
+
+
+class TmpDirFileSystem(MockFileSystem):
+ protocol = "tmp"
+ tmp_dir = None
+
+ def __init__(self, *args, **kwargs):
+ assert self.tmp_dir is not None, "TmpDirFileSystem.tmp_dir is not set"
+ super().__init__(*args, **kwargs, local_root_dir=self.tmp_dir, auto_mkdir=True)
+
+ @classmethod
+ def _strip_protocol(cls, path):
+ path = stringify_path(path)
+ if path.startswith("tmp://"):
+ path = path[6:]
+ return path
+
+
[email protected]
+def mock_fsspec():
+ original_registry = fsspec.registry.copy()
+ fsspec.register_implementation("mock", MockFileSystem)
+ fsspec.register_implementation("tmp", TmpDirFileSystem)
+ yield
+ fsspec.registry = original_registry
+
+
[email protected]
+def mockfs(tmp_path_factory, mock_fsspec):
+ local_fs_dir = tmp_path_factory.mktemp("mockfs")
+ return MockFileSystem(local_root_dir=local_fs_dir, auto_mkdir=True)
+
+
[email protected]
+def tmpfs(tmp_path_factory, mock_fsspec):
+ tmp_fs_dir = tmp_path_factory.mktemp("tmpfs")
+ with patch.object(TmpDirFileSystem, "tmp_dir", tmp_fs_dir):
+ yield TmpDirFileSystem()
+ shutil.rmtree(tmp_fs_dir)
diff --git a/services/worker/tests/job_runners/split/config/dataset-split.parquet b/services/worker/tests/job_runners/split/config/dataset-split.parquet
deleted file mode 100644
index 672ab11e..00000000
Binary files a/services/worker/tests/job_runners/split/config/dataset-split.parquet and /dev/null differ
diff --git a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py
index 46a92c80..3dcb37c1 100644
--- a/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py
+++ b/services/worker/tests/job_runners/split/test_first_rows_from_parquet.py
@@ -4 +3,0 @@
-import os
@@ -7 +6 @@ from http import HTTPStatus
-from typing import Callable, List
+from typing import Callable, Generator, List
@@ -10,0 +10,2 @@ import pytest
+from datasets import Dataset
+from fsspec import AbstractFileSystem
@@ -17 +17,0 @@ from libcommon.utils import Priority
-from pyarrow.fs import LocalFileSystem
@@ -30,0 +31 @@ def get_job_runner(
+ parquet_metadata_directory: StrPath,
@@ -44 +45,5 @@ def get_job_runner(
- "config-level": {"input_type": "dataset", "triggered_by": "dataset-level"},
+ "config-level": {
+ "input_type": "config",
+ "triggered_by": "dataset-level",
+ "provides_config_parquet": True,
+ },
@@ -65,0 +71 @@ def get_job_runner(
+ processing_graph=processing_graph,
@@ -66,0 +73 @@ def get_job_runner(
+ parquet_metadata_directory=parquet_metadata_directory,
@@ -71,0 +79,12 @@ def get_job_runner(
[email protected]
+def ds() -> Dataset:
+ return Dataset.from_dict({"col1": [1, 2, 3], "col2": ["a", "b", "c"]})
+
+
[email protected]
+def ds_fs(ds: Dataset, tmpfs: AbstractFileSystem) -> Generator[AbstractFileSystem, None, None]:
+ with tmpfs.open("config/dataset-split.parquet", "wb") as f:
+ ds.to_parquet(f)
+ yield tmpfs
+
+
@@ -84,0 +104 @@ def test_compute(
+ ds_fs: AbstractFileSystem,
@@ -91,0 +112,14 @@ def test_compute(
+
+ config_parquet_content = {
+ "parquet_files": [
+ {
+ "dataset": dataset,
+ "config": config,
+ "split": split,
+ "url": "https://fake.huggingface.co/datasets/dataset/resolve/refs%2Fconvert%2Fparquet/config/dataset-split.parquet", # noqa: E501
+ "filename": "dataset-split.parquet",
+ "size": 128,
+ }
+ ]
+ }
+
@@ -93 +127 @@ def test_compute(
- kind="config-parquet",
+ kind="config-level",
@@ -96,11 +130 @@ def test_compute(
- content={
- "parquet_files": [
- {
- "dataset": dataset,
- "config": config,
- "split": split,
- "filename": f"{dataset}-{split}.parquet",
- "size": 1000,
- }
- ]
- },
+ content=config_parquet_content,
@@ -110 +134 @@ def test_compute(
- with patch("worker.job_runners.split.first_rows_from_parquet.get_hf_fs") as mock_read:
+ with patch("libcommon.parquet_utils.get_hf_fs", return_value=ds_fs):
@@ -112 +136 @@ def test_compute(
- "worker.job_runners.split.first_rows_from_parquet.get_hf_parquet_uris",
+ "libcommon.parquet_utils.get_hf_parquet_uris",
@@ -115,6 +138,0 @@ def test_compute(
- initial_location = os.getcwd()
- os.chdir("tests/job_runners/split")
- # TODO: Make localsystem by relative path
- fs = LocalFileSystem()
- mock_read.return_value = fs
- # ^ Mocking file system with local file
@@ -156 +174 @@ def test_compute(
- assert response["features"][0]["type"]["dtype"] == "int32"
+ assert response["features"][0]["type"]["dtype"] == "int64"
@@ -170 +187,0 @@ def test_compute(
- os.chdir(initial_location)
|
|
e870cb24073f4c214734b155e6195b5a6d295291
|
Quentin Lhoest
| 2023-06-09T17:41:12 |
Copy parquet when possible (#1321)
|
diff --git a/.github/workflows/_quality-python.yml b/.github/workflows/_quality-python.yml
index 9cc6eb2a..9f9638e0 100644
--- a/.github/workflows/_quality-python.yml
+++ b/.github/workflows/_quality-python.yml
@@ -52 +52 @@ jobs:
- run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d')"
+ run: bash -c "poetry run pip-audit -r <(poetry export -f requirements.txt --with dev | sed '/^kenlm @/d' | sed '/^torch @/d' | sed '/^libcommon @/d' | sed '/^trec-car-tools @/d' | sed 's/^huggingface-hub @ git.*/huggingface-hub==0.15.1 ; python_full_version == \"3.9.15\" --hash=sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445 --hash=sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081/')"
diff --git a/chart/env/prod.yaml b/chart/env/prod.yaml
index 1bf2afcf..163a431d 100644
--- a/chart/env/prod.yaml
+++ b/chart/env/prod.yaml
@@ -91 +91 @@ parquetAndInfo:
- blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,bigcode/the-stack-dedup,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
+ blockedDatasets: "matallanas/linustechtips-transcript-audio-wav,KnutJaegersberg/Interpretable_word_embeddings_large_cskg,ashraf-ali/quran-data,cjvt/cc_gigafida,cmudrc/porous-microstructure-strain-fields,dlwh/MultiLegalPile_Wikipedia_Shuffled,izumaru/os2-datasets,joelito/MultiLegalPile_Wikipedia_Filtered,leviethoang/VBVLSP,nyanko7/yandere-images,severo/wit,texturedesign/td01_natural-ground-textures,Tristan/olm-october-2022-tokenized-1024-exact-dedup-only,Whispering-GPT/linustechtips-transcript-audio,beyond/chinese_clean_passages_80m,bigscience/xP3,dalle-mini/YFCC100M_OpenAI_subset,galman33/gal_yair_166000_256x256_fixed,matallanas/linustechtips-transcript-audio-mp3,mwitiderrick/arXiv,sjpmpzx/qm_ly_gy_soundn,tilos/ASR-CCANTCSC,matallanas/linustechtips-transcript-audio-ogg,VIMA/VIMA-Data,severo/wit,wmt/europarl,chrisjay/mnist-adversarial-dataset,mwitiderrick/arXiv,HuggingFaceM4/TextCaps,CristianaLazar/librispeech5k_train,texturedesign/td01_natural-ground-textures,cjvt/cc_gigafida,Yehor/ukrainian-tts-lada,YWjimmy/PeRFception-v1,SDbiaseval/dataset-dalle,Pinguin/images,DTU54DL/librispeech5k-augmentated-train-prepared,CristianaLazar/librispeech500,abdusahmbzuai/masc_dev,anonymousdeepcc/DeepCC,bigcode/the-stack-username-to-repo,bigscience/massive-probing-results,dgrnd4/stanford_dog_dataset,gigant/romanian_speech_synthesis_0_8_1,helena-balabin/sentences,icelab/ntrs_meta,joefox/Mozilla_Common_Voice_ru_test_noise,m-aliabbas/idrak_splitted_amy_1,marinone94/nst_sv,mbarnig/lb-de-fr-en-pt-12800-TTS-CORPUS,momilla/Ethereum_transacitons,nev/anime-giph,openclimatefix/nimrod-uk-1km-validation,raghav66/whisper-gpt,strombergnlp/broad_twitter_corpus,z-uo/female-LJSpeech-italian,Champion/vpc2020_clear_anon_speech,DelgadoPanadero/Pokemon,GEM/references,HuggingFaceM4/FairFace,Karavet/ILUR-news-text-classification-corpus,Voicemod/LibriTTS-100-preproc,YWjimmy/PeRFception-v1-1,albertvillanova/TextCaps,allenai/c4,dog/punks,chenghao/scielo_books,YWjimmy/PeRFception-v1-2,openclimatefix/era5,Carlisle/msmarco-passage-non-abs,SetFit/mnli,valurank/PoliticalBias_AllSides_Txt,Biomedical-TeMU/ProfNER_corpus_classification,LeoFeng/MLHW_6,pragnakalp/squad_v2_french_translated,textvqa,polinaeterna/vox_lingua,nishita/ade20k-sample,oyk100/ChaSES-data,YWjimmy/PeRFception-v1-3,YWjimmy/PeRFception-ScanNet,ChaiML/AnthropicRLHFPreferenceData,voidful/librispeech_asr_text,Isma/librispeech_1000_seed_42,Graphcore/vqa-lxmert,Tevatron/wikipedia-curated-corpus,adamlin/daily_dialog,cameronbc/synthtiger,clarin-pl/multiwiki_90k,echarlaix/vqa-lxmert,gigant/african_accented_french,Graphcore/vqa,echarlaix/vqa,jimregan/clarinpl_studio,GEM/xsum,Tevatron/wikipedia-squad-corpus,mulcyber/europarl-mono,nateraw/wit,bigscience/P3,tau/mrqa,uva-irlab/trec-cast-2019-multi-turn,vblagoje/wikipedia_snippets_streamed,Tevatron/wikipedia-wq-corpus,malteos/paperswithcode-aspects,Samip/Scotch,iluvvatar/RuREBus,nateraw/quickdraw,tau/scrolls,qanastek/MASSIVE,TalTechNLP/VoxLingua107,shanya/crd3,HugoLaurencon/libri_light,jerpint/imagenette,Leyo/TGIF,DFKI-SLT/few-nerd,crystina-z/msmarco-passage-dl20,HuggingFaceM4/epic_kitchens_100,HuggingFaceM4/yttemporal180m,andreagasparini/librispeech_train_other_only,allenai/nllb,biglam/nls_chapbook_illustrations,winvoker/lvis,Lacito/pangloss,indonesian-nlp/librivox-indonesia,Graphcore/gqa-lxmert,nanom/splittedspanish3bwc,cahya/librivox-indonesia,asapp/slue,sil-ai/audio-keyword-spotting,tner/wikiann,rogerdehe/xfund,arpelarpe/nota,mwhanna/ACT-Thor,sanchit-gandhi/librispeech_asr_clean,echarlaix/gqa-lxmert,shunk031/cocostuff,gigant/m-ailabs_speech_dataset_fr,jimregan/clarinpl_sejmsenat,1aurent/icdar-2011,marinone94/nst_no,jamescalam/unsplash-25k-images,stas/openwebtext-10k,florianbussmann/train_tickets-yu2020pick,benschill/brain-tumor-collection,imvladikon/paranames,PolyAI/evi,bengaliAI/cvbn,Sreyan88/librispeech_asr,superb,mozilla-foundation/common_voice_10_0,darkproger/librispeech_asr,kresnik/librispeech_asr_test,Lehrig/Monkey-Species-Collection,HuggingFaceM4/TGIF,crystina-z/miracl-bm25-negative,cats_vs_dogs,biglam/gallica_literary_fictions,common_language,competition_math,cornell_movie_dialog,evidence_infer_treatment,hebrew_projectbenyehuda,lj_speech,mc4,muchocine,opus_euconst,tab_fact,the_pile,tapaco,turkic_xwmt,web_nlg,vctk,mathaillah/BeritaHoaks-NonHoaks,universal_morphologies,LanceaKing/asvspoof2019,andreagasparini/librispeech_train_clean_only,nuprl/MultiPL-E,SLPL/naab-raw,mteb/results,SocialGrep/the-reddit-climate-change-dataset,bigscience-biomedical/anat_em,crystina-z/xor-tydi-corpus,qanastek/QUAERO,TomTBT/pmc_open_access_section,jamescalam/movielens-25m-ratings,HuggingFaceM4/charades,Tevatron/xor-tydi-corpus,khalidalt/tydiqa-primary,nvm472001/cvdataset-layoutlmv3,Lehrig/GTZAN-Collection,mteb/tatoeba-bitext-mining,sled-umich/Action-Effect,HamdiJr/Egyptian_hieroglyphs,joelito/lextreme,cooleel/xfund_de,oscar,mozilla-foundation/common_voice_7_0,KETI-AIR/vqa,Livingwithmachines/MapReader_Data_SIGSPATIAL_2022,NLPC-UOM/document_alignment_dataset-Sinhala-Tamil-English,miracl/miracl,Muennighoff/flores200,Murple/mmcrsc,mesolitica/dbp,CodedotAI/code_clippy,keshan/clean-si-mc4,yhavinga/ccmatrix,metashift,google/fleurs,HugoLaurencon/libri_light_bytes,biwi_kinect_head_pose,ami,bigscience-biomedical/ebm_pico,HuggingFaceM4/general-pmd-synthetic-testing,crystina-z/mmarco,robertmyers/pile_v2,bigbio/anat_em,biglam/early_printed_books_font_detection,nateraw/imagenet-sketch,jpwahle/dblp-discovery-dataset,andreagasparini/librispeech_test_only,crystina-z/mmarco-corpus,mozilla-foundation/common_voice_6_0,biglam/brill_iconclass,bigscience-biomedical/evidence_inference,HuggingFaceM4/cm4-synthetic-testing,SocialGrep/ten-million-reddit-answers,bnl_newspapers,multilingual_librispeech,openslr,GEM/BiSECT,Graphcore/gqa,SaulLu/Natural_Questions_HTML_reduced_all,ccdv/cnn_dailymail,mozilla-foundation/common_voice_1_0,huggan/anime-faces,Biomedical-TeMU/ProfNER_corpus_NER,MorVentura/TRBLLmaker,student/celebA,Rodion/uno_sustainable_development_goals,Nart/parallel-ab-ru,HuggingFaceM4/VQAv2,mesolitica/noisy-ms-en-augmentation,nateraw/rice-image-dataset,tensorcat/wikipedia-japanese,angelolab/ark_example,RAYZ/Mixed-Dia,ywchoi/mdpi_sept10,TomTBT/pmc_open_access_figure,society-ethics/lila_camera_traps,autoevaluator/shoes-vs-sandals-vs-boots,cjvt/slo_collocations,parambharat/mile_dataset,rossevine/tesis,ksaml/Stanford_dogs,nuprl/MultiPL-E-raw-data,ZihaoLin/zhlds,ACL-OCL/acl-anthology-corpus,mozilla-foundation/common_voice_2_0,Biomedical-TeMU/SPACCC_Sentence-Splitter,nateraw/rice-image-dataset-2,mesolitica/noisy-en-ms-augmentation,bigbio/ctebmsp,bigbio/distemist,nlphuji/vasr,parambharat/malayalam_asr_corpus,cjvt/sloleks,DavidVivancos/MindBigData2022_Imagenet_IN_Spct,KokeCacao/oracle,keremberke/nfl-object-detection,lafi23333/ds,Lykon/OnePiece,kaliansh/sdaia,sil-ai/audio-kw-in-context,andite/riyo-tag,ilhanemirhan/eee543,backslashlim/LoRA-Datasets,hr16/Miwano-Rag,ccdv/mediasum,mozilla-foundation/common_voice_3_0,mozilla-foundation/common_voice_4_0,bigbio/ebm_pico,parambharat/kannada_asr_corpus,parambharat/telugu_asr_corpus,Abuelnour/json_1000_Scientific_Paper,reazon-research/reazonspeech,shunk031/livedoor-news-corpus,mesolitica/translated-SQUAD,SamAct/medium_cleaned,EfaceD/ElysiumInspirations,cahya/fleurs,guangguang/azukijpg,genjib/LAVISHData,rohitp1/librispeech_asr_clean,azraahmadi/autotrain-data-xraydatasetp2,HuggingFaceM4/COCO,bio-datasets/e3c,nateraw/auto-cats-and-dogs,keremberke/smoke-object-detection,ds4sd/DocLayNet,nlphuji/utk_faces,corentinm7/MyoQuant-SDH-Data,xglue,grasshoff/lhc_sents,HugoLaurencon/IIIT-5K,alkzar90/CC6204-Hackaton-Cub-Dataset,RaphaelOlivier/whisper_adversarial_examples,bruno-cotrim/arch-max,keshan/multispeaker-tts-sinhala,Tevatron/beir-corpus,fcakyon/gun-object-detection,ccdv/arxiv-summarization,keremberke/protective-equipment-detection,mozilla-foundation/common_voice_5_0,nlphuji/winogavil,Poupou/Gitcoin-Grant-DataBuilder,orieg/elsevier-oa-cc-by,castorini/msmarco_v1_passage_doc2query-t5_expansions,inseq/divemt_attributions,crystina-z/msmarco-passage-dl19,mozilla-foundation/common_voice_5_1,matchbench/dbp15k-fr-en,keremberke/garbage-object-detection,crystina-z/no-nonself-mrtydi,ashraq/dhivehi-corpus,zyznull/dureader-retrieval-ranking,zyznull/msmarco-passage-corpus,zyznull/msmarco-passage-ranking,Tevatron/wikipedia-squad,Tevatron/wikipedia-trivia-corpus,NeuroSenko/senko_anime_full,plncmm/wl-disease,plncmm/wl-family-member"
diff --git a/libs/libcommon/src/libcommon/dataset.py b/libs/libcommon/src/libcommon/dataset.py
index 708f6257..7d5dfc42 100644
--- a/libs/libcommon/src/libcommon/dataset.py
+++ b/libs/libcommon/src/libcommon/dataset.py
@@ -63,0 +64,2 @@ def get_dataset_info_for_supported_datasets(
+ revision: Optional[str] = None,
+ files_metadata: bool = False,
@@ -76,0 +79,6 @@ def get_dataset_info_for_supported_datasets(
+ revision (`str`, *optional*, defaults to None):
+ The revision of the dataset repository from which to get the
+ information.
+ files_metadata (`bool`, *optional*, defaults to False):
+ Whether or not to retrieve metadata for files in the repository
+ (size, LFS metadata, etc). Defaults to `False`.
@@ -94 +102,5 @@ def get_dataset_info_for_supported_datasets(
- repo_id=dataset, token=hf_token, timeout=hf_timeout_seconds
+ repo_id=dataset,
+ token=hf_token,
+ timeout=hf_timeout_seconds,
+ revision=revision,
+ files_metadata=files_metadata,
diff --git a/libs/libcommon/src/libcommon/exceptions.py b/libs/libcommon/src/libcommon/exceptions.py
index 8ae12363..eaf8cc95 100644
--- a/libs/libcommon/src/libcommon/exceptions.py
+++ b/libs/libcommon/src/libcommon/exceptions.py
@@ -87,0 +88 @@ CacheableErrorCode = Literal[
+ "DatasetWithTooManyParquetFilesError",
@@ -225 +226 @@ class DatasetWithTooManyExternalFilesError(CacheableError):
- """Raised when the dataset size (sum of config sizes given by the datasets library) is too big."""
+ """Raised when the number of external data files of a dataset is too big."""
@@ -230,0 +232,7 @@ class DatasetWithTooManyExternalFilesError(CacheableError):
+class DatasetWithTooManyParquetFilesError(CacheableError):
+ """Raised when the number of parquet files of a dataset is too big."""
+
+ def __init__(self, message: str, cause: Optional[BaseException] = None):
+ super().__init__(message, HTTPStatus.NOT_IMPLEMENTED, "DatasetWithTooManyParquetFilesError", cause, True)
+
+
diff --git a/services/worker/poetry.lock b/services/worker/poetry.lock
index d233d5ff..92064110 100644
--- a/services/worker/poetry.lock
+++ b/services/worker/poetry.lock
@@ -1 +1 @@
-# This file is automatically @generated by Poetry and should not be changed by hand.
+# This file is automatically @generated by Poetry 1.4.0 and should not be changed by hand.
@@ -1457 +1457 @@ name = "huggingface-hub"
-version = "0.15.1"
+version = "0.16.0.dev0"
@@ -1462,4 +1462,2 @@ python-versions = ">=3.7.0"
-files = [
- {file = "huggingface_hub-0.15.1-py3-none-any.whl", hash = "sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445"},
- {file = "huggingface_hub-0.15.1.tar.gz", hash = "sha256:a61b7d1a7769fe10119e730277c72ab99d95c48d86a3d6da3e9f3d0f632a4081"},
-]
+files = []
+develop = false
@@ -1486,0 +1485,6 @@ typing = ["types-PyYAML", "types-requests", "types-simplejson", "types-toml", "t
+[package.source]
+type = "git"
+url = "https://github.com/huggingface/huggingface_hub"
+reference = "1055a56b2d2723b55ba4fdf1f3296e04cfd8d6db"
+resolved_reference = "1055a56b2d2723b55ba4fdf1f3296e04cfd8d6db"
+
@@ -2860 +2863,0 @@ files = [
- {file = "pdf2image-1.16.3.tar.gz", hash = "sha256:74208810c2cef4d9e347769b8e62a52303982ddb4f2dfd744c7ab4b940ae287e"},
@@ -4415 +4417,0 @@ files = [
- {file = "soundfile-0.12.1-py2.py3-none-manylinux_2_17_x86_64.whl", hash = "sha256:2dc3685bed7187c072a46ab4ffddd38cef7de9ae5eb05c03df2ad569cf4dacbc"},
@@ -4708,0 +4711,2 @@ files = [
+ {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_arm64.whl", hash = "sha256:9c9b14fbb73ec4cb0f209722a1489020fd8614c92ae22589f2309c48cefdf21f"},
+ {file = "tensorflow_macos-2.12.0-cp311-cp311-macosx_12_0_x86_64.whl", hash = "sha256:6a54539bd076746f69ae8bef7282f981674fe4dbf59c3a84c4af86ae6bae9d5c"},
@@ -5596 +5600 @@ python-versions = "3.9.15"
-content-hash = "f9c56aa59e9d0c6802245fea8dac8ae28a9b85c8c017d69a7085a9689328ca46"
+content-hash = "ad3c5c34e9ea75e4cb4394930bb35c5afdff65fe03c5f51b8cbebbea37f62f1d"
diff --git a/services/worker/pyproject.toml b/services/worker/pyproject.toml
index ecdc0f65..9fabc531 100644
--- a/services/worker/pyproject.toml
+++ b/services/worker/pyproject.toml
@@ -17 +17 @@ gdown = "^4.6.3"
-huggingface-hub = "^0.15.1"
+huggingface-hub = { git = "https://github.com/huggingface/huggingface_hub", rev = "1055a56b2d2723b55ba4fdf1f3296e04cfd8d6db" }
diff --git a/services/worker/src/worker/job_runners/config/parquet_and_info.py b/services/worker/src/worker/job_runners/config/parquet_and_info.py
index 3a0c75f6..e6c4b1b2 100644
--- a/services/worker/src/worker/job_runners/config/parquet_and_info.py
+++ b/services/worker/src/worker/job_runners/config/parquet_and_info.py
@@ -4 +3,0 @@
-import contextlib
@@ -17,0 +17 @@ import numpy as np
+import pyarrow.parquet as pq
@@ -19 +19 @@ import requests
-from datasets import DownloadConfig, get_dataset_config_info, load_dataset_builder
+from datasets import DownloadConfig, Features, load_dataset_builder
@@ -21,0 +22 @@ from datasets.data_files import EmptyDatasetError as _EmptyDatasetError
+from datasets.data_files import Url
@@ -22,0 +24,2 @@ from datasets.download import StreamingDownloadManager
+from datasets.packaged_modules.parquet.parquet import Parquet as ParquetBuilder
+from datasets.splits import SplitDict, SplitInfo
@@ -29,0 +33 @@ from datasets.utils.py_utils import asdict, map_nested
+from fsspec.implementations.http import HTTPFileSystem
@@ -32,0 +37 @@ from huggingface_hub._commit_api import (
+ CommitOperationCopy,
@@ -36 +41 @@ from huggingface_hub.hf_api import DatasetInfo, HfApi, RepoFile
-from huggingface_hub.utils._errors import RepositoryNotFoundError, RevisionNotFoundError
+from huggingface_hub.utils._errors import RepositoryNotFoundError
@@ -42,0 +48 @@ from libcommon.constants import (
+from libcommon.dataset import get_dataset_info_for_supported_datasets
@@ -48 +53,0 @@ from libcommon.exceptions import (
- DatasetRevisionNotFoundError,
@@ -52,0 +58 @@ from libcommon.exceptions import (
+ DatasetWithTooManyParquetFilesError,
@@ -57,0 +64 @@ from libcommon.exceptions import (
+ FileSystemError,
@@ -62,0 +70 @@ from libcommon.utils import JobInfo
+from tqdm.contrib.concurrent import thread_map
@@ -83,0 +92 @@ DATASET_TYPE = "dataset"
+MAX_FILES_PER_DIRECTORY = 10_000 # hf hub limitation
@@ -94 +103,2 @@ class ParquetFile:
- def repo_file(self) -> str:
+ @property
+ def path_in_repo(self) -> str:
@@ -172,40 +182,10 @@ def raise_if_blocked(
-def get_dataset_info_or_raise(
- dataset: str,
- hf_endpoint: str,
- hf_token: Optional[str],
- revision: str,
-) -> DatasetInfo:
- """
- Return the dataset info if possible.
- Raise an error if the dataset cannot be accessed (does not exist, gated with extra fields, private)
-
- Args:
- dataset (`str`):
- A namespace (user or an organization) and a repo name separated
- by a `/`.
- hf_endpoint (`str`):
- The Hub endpoint (for example: "https://huggingface.co")
- hf_token (`str`, `optional`):
- An app authentication token with read access to all the datasets.
- revision (`str`):
- The git revision (e.g. "main" or sha) of the dataset
- Returns:
- `DatasetInfo`: The dataset info
- Raises the following errors:
- - [`libcommon.exceptions.DatasetNotFoundError`]
- If the repository to download from cannot be found. This may be because it doesn't exist,
- or because it is set to `private` and you do not have access.
- - [`libcommon.exceptions.DatasetRevisionNotFoundError`]
- If the revision does not exist or cannot be accessed using the token.
- """
- try:
- dataset_info = HfApi(endpoint=hf_endpoint, token=hf_token).dataset_info(
- repo_id=dataset, revision=revision, files_metadata=True
- )
- except RepositoryNotFoundError as err:
- raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err
- except RevisionNotFoundError as err:
- raise DatasetRevisionNotFoundError("The dataset revision does not exist on the Hub.") from err
- if dataset_info.private:
- raise DatasetNotFoundError("The dataset does not exist on the Hub.")
- return dataset_info
+def is_parquet_builder_with_hub_files(builder: DatasetBuilder, hf_endpoint: str) -> bool:
+ if not isinstance(builder, ParquetBuilder) or not builder.config.data_files:
+ return False
+ for split in builder.config.data_files:
+ for data_file in builder.config.data_files[split]:
+ if not isinstance(data_file, Url):
+ return False
+ if not data_file.startswith(hf_endpoint + "/datasets/" + str(builder.repo_id) + "/"):
+ return False
+ return True
@@ -243,5 +223 @@ def raise_if_too_big_from_datasets(
- dataset: str,
- config: str,
- hf_endpoint: str,
- hf_token: Optional[str],
- revision: str,
+ info: datasets.DatasetInfo,
@@ -255,11 +231,2 @@ def raise_if_too_big_from_datasets(
- dataset (`str`):
- A namespace (user or an organization) and a repo name separated
- by a `/`.
- config (`str`):
- Dataset configuration name
- hf_endpoint (`str`):
- The Hub endpoint (for example: "https://huggingface.co")
- hf_token (`str`, `optional`):
- An app authentication token with read access to all the datasets.
- revision (`str`):
- The git revision (e.g. "main" or sha) of the dataset
+ info (`datasets.DatasetInfo`):
+ Dataset info from the datasets library
@@ -277,9 +244 @@ def raise_if_too_big_from_datasets(
- if datasets.config.HF_ENDPOINT != hf_endpoint:
- raise ValueError(
- f"Invalid datasets.config.HF_ENDPOINT value: '{datasets.config.HF_ENDPOINT}'. Please set it to:"
- f" '{hf_endpoint}'."
- )
- dataset_size = 0
- with contextlib.suppress(Exception):
- info = get_dataset_config_info(path=dataset, config_name=config, revision=revision, use_auth_token=hf_token)
- dataset_size = info.dataset_size if info.dataset_size is not None else 0
+ dataset_size = info.dataset_size if info.dataset_size is not None else 0
@@ -295,2 +254 @@ def raise_if_requires_manual_download(
- dataset: str,
- config: str,
+ builder: DatasetBuilder,
@@ -299 +256,0 @@ def raise_if_requires_manual_download(
- revision: str,
@@ -305,5 +262,2 @@ def raise_if_requires_manual_download(
- dataset (`str`):
- A namespace (user or an organization) and a repo name separated
- by a `/`.
- config (`str`):
- Dataset configuration name.
+ builder (`datasets.builder.DatasetBuilder`):
+ A dataset builder instance to check.
@@ -314,2 +267,0 @@ def raise_if_requires_manual_download(
- revision (`str`):
- The git revision (e.g. "main" or sha) of the dataset.
@@ -331,6 +282,0 @@ def raise_if_requires_manual_download(
- builder = load_dataset_builder(
- dataset,
- name=config,
- revision=revision,
- use_auth_token=hf_token,
- )
@@ -344 +290 @@ def raise_if_requires_manual_download(
- raise DatasetManualDownloadError(f"{dataset=} requires manual download.", cause=err) from err
+ raise DatasetManualDownloadError(f"dataset={builder.repo_id} requires manual download.", cause=err) from err
@@ -348,2 +294,2 @@ def raise_if_not_supported(
- dataset: str,
- config: str,
+ dataset_info: DatasetInfo,
+ builder: DatasetBuilder,
@@ -352,3 +297,0 @@ def raise_if_not_supported(
- revision: str,
- supported_datasets: List[str],
- blocked_datasets: List[str],
@@ -355,0 +299 @@ def raise_if_not_supported(
+ max_external_data_files: int,
@@ -364,5 +308,4 @@ def raise_if_not_supported(
- dataset (`str`):
- A namespace (user or an organization) and a repo name separated
- by a `/`.
- config (`str`):
- Dataset configuration name
+ dataset_info (`DatasetInfo`):
+ The dataset info
+ builder (`datasets.builder.DatasetBuilder`):
+ A dataset builder instance to check.
@@ -375,5 +317,0 @@ def raise_if_not_supported(
- supported_datasets (`List[str]`):
- The list of supported datasets, saving the blocked datasets. If empty, all datasets are supported
- (saving the blocked datasets).
- blocked_datasets (`List[str]`):
- The list of blocked datasets. If empty, no dataset is blocked.
@@ -382,0 +321,3 @@ def raise_if_not_supported(
+ max_external_data_files (`int`):
+ The maximum number of external data files (i.e. not hosted on HF).
+ If the dataset is under the limit (which means that the files can be fetched), it will be allowed.
@@ -387,6 +327,0 @@ def raise_if_not_supported(
- - [`libcommon.exceptions.DatasetNotFoundError`]:
- if the dataset does not exist, or if the token does not give the sufficient access to the dataset,
- - ['requests.exceptions.HTTPError'](https://requests.readthedocs.io/en/latest/api/#requests.HTTPError)
- any other error when asking access
- - [`libcommon.exceptions.DatasetInBlockListError`]
- If the dataset is in the list of blocked datasets.
@@ -402,0 +338,14 @@ def raise_if_not_supported(
+ - [`libcommon.exceptions.DatasetWithTooManyExternalFilesError`]
+ If the dataset has too many external files to be converted to parquet
+ - [`libcommon.exceptions.DatasetWithTooBigExternalFilesError`]
+ If the dataset is too big external files be converted to parquet
+ - [`libcommon.exceptions.UnsupportedExternalFilesError`]
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
+ - [`libcommon.exceptions.ExternalFilesSizeRequestHTTPError`]
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
+ - [`libcommon.exceptions.ExternalFilesSizeRequestConnectionError`]
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
+ - [`libcommon.exceptions.ExternalFilesSizeRequestTimeoutError`]
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
+ - [`libcommon.exceptions.ExternalFilesSizeRequestError`]
+ If we failed to get the external files sizes to make sure we can convert the dataset to parquet
@@ -404 +353 @@ def raise_if_not_supported(
- If the datasets.config.HF_ENDPOINT is not set to the expected value
+ If the datasets.config.HF_ENDPOINT is not set to the expected value
@@ -406,4 +355,5 @@ def raise_if_not_supported(
- raise_if_blocked(dataset=dataset, blocked_datasets=blocked_datasets)
- dataset_info = get_dataset_info_or_raise(
- dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, revision=revision
- )
+ if datasets.config.HF_ENDPOINT != hf_endpoint:
+ raise ValueError(
+ f"Invalid datasets.config.HF_ENDPOINT value: '{datasets.config.HF_ENDPOINT}'. Please set it to:"
+ f" '{hf_endpoint}'."
+ )
@@ -411,2 +361 @@ def raise_if_not_supported(
- dataset=dataset,
- config=config,
+ builder=builder,
@@ -415 +363,0 @@ def raise_if_not_supported(
- revision=revision,
@@ -417,6 +365,5 @@ def raise_if_not_supported(
- if dataset in supported_datasets:
- return
- raise_if_too_big_from_datasets(
- dataset=dataset,
- config=config,
- hf_endpoint=hf_endpoint,
+ raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size)
+ raise_if_too_big_from_external_data_files(
+ builder=builder,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
@@ -424 +371,3 @@ def raise_if_not_supported(
- revision=revision,
+ )
+ raise_if_too_big_from_datasets(
+ builder.info,
@@ -427 +375,0 @@ def raise_if_not_supported(
- raise_if_too_big_from_hub(dataset_info=dataset_info, max_dataset_size=max_dataset_size)
@@ -628,0 +577,101 @@ def get_writer_batch_size(ds_config_info: datasets.info.DatasetInfo) -> Optional
+def copy_parquet_files(builder: DatasetBuilder) -> List[CommitOperationCopy]:
+ """Copy parquet files by copying the git LFS pointer files"""
+ data_files = builder.config.data_files
+ if not data_files:
+ raise EmptyDatasetError("Empty parquet data_files")
+ parquet_operations = []
+ total_num_parquet_files = sum(len(data_files[split]) for split in data_files)
+ if total_num_parquet_files >= MAX_FILES_PER_DIRECTORY:
+ raise DatasetWithTooManyParquetFilesError(
+ f"The dataset has {total_num_parquet_files} parquet files and can't be linked in the parquet directory "
+ f"because it exceeds the maximum number of files per directory ({MAX_FILES_PER_DIRECTORY})."
+ )
+ for split in data_files:
+ num_shards = len(data_files[split])
+ for shard_idx, data_file in enumerate(data_files[split]):
+ src_revision, src_path_in_repo = data_file.split("/datasets/" + builder.repo_id + "/resolve/", 1)[1].split(
+ "/", 1
+ )
+
+ # for forward compatibility with https://github.com/huggingface/datasets/pull/5331
+ parquet_name = str(builder.dataset_name) if hasattr(builder, "dataset_name") else builder.name
+
+ if num_shards > 1:
+ path_in_repo = (
+ f"{builder.config.name}/{parquet_name}-{split}-{shard_idx:05d}-of-{num_shards:05d}.parquet"
+ )
+ else:
+ path_in_repo = f"{builder.config.name}/{parquet_name}-{split}.parquet"
+
+ parquet_operations.append(
+ CommitOperationCopy(
+ src_path_in_repo=src_path_in_repo, path_in_repo=path_in_repo, src_revision=src_revision
+ )
+ )
+ return parquet_operations
+
+
+def get_parquet_file_and_size(url: str, fs: HTTPFileSystem, hf_token: Optional[str]) -> Tuple[pq.ParquetFile, int]:
+ headers = get_authentication_headers_for_url(url, use_auth_token=hf_token)
+ f = fs.open(url, headers=headers)
+ return pq.ParquetFile(f), f.size
+
+
+def fill_builder_info(builder: DatasetBuilder, hf_token: Optional[str]) -> None:
+ """Fill the builder DatasetInfo from the copied parquet files"""
+ data_files = builder.config.data_files
+ if not data_files:
+ raise EmptyDatasetError("Empty parquet data_files")
+ fs = HTTPFileSystem()
+ if not builder.info.splits or not builder.info.download_size:
+ builder.info.splits = SplitDict()
+ builder.info.dataset_size = 0
+ for split in data_files:
+ try:
+ parquet_files_and_sizes: List[Tuple[pq.ParquetFile, int]] = thread_map(
+ partial(get_parquet_file_and_size, fs=fs, hf_token=hf_token),
+ data_files[split],
+ unit="pq",
+ disable=True,
+ )
+ parquet_files, sizes = zip(*parquet_files_and_sizes)
+ except Exception as e:
+ raise FileSystemError(f"Could not read the parquet files: {e}") from e
+ if parquet_files:
+ first_pf = parquet_files[0]
+ if builder.info.features is None:
+ builder.info.features = Features.from_arrow_schema(first_pf.schema_arrow)
+ first_row_group = first_pf.read_row_group(0)
+ compression_ratio = first_row_group.nbytes / first_row_group.num_rows
+ num_examples = sum(parquet_file.metadata.num_rows for parquet_file in parquet_files)
+ approx_num_bytes = int(compression_ratio * num_examples)
+ builder.info.splits.add(SplitInfo(split, num_bytes=approx_num_bytes, num_examples=num_examples))
+ builder.info.download_size += sum(sizes)
+ builder.info.dataset_size += approx_num_bytes
+
+
+def convert_to_parquet(builder: DatasetBuilder) -> List[CommitOperationAdd]:
+ """Download and prepare the dataset as parquet files and fills the builder info"""
+ # prepare the parquet files locally
+ writer_batch_size = get_writer_batch_size(builder.info)
+ if writer_batch_size is not None and (
+ builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size
+ ):
+ builder._writer_batch_size = writer_batch_size
+ builder.download_and_prepare(
+ file_format="parquet"
+ ) # the parquet files are stored in the cache dir and it fills the info
+ local_parquet_files = [
+ ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=builder.config.name)
+ for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet")
+ ]
+
+ # send the files to the target revision
+ parquet_operations: List[CommitOperationAdd] = [
+ CommitOperationAdd(path_in_repo=parquet_file.path_in_repo, path_or_fileobj=parquet_file.local_file)
+ for parquet_file in local_parquet_files
+ ]
+ logging.debug(f"{parquet_operations=}")
+ return parquet_operations
+
+
@@ -725,10 +774 @@ def compute_config_parquet_and_info_response(
- raise_if_not_supported(
- dataset=dataset,
- config=config,
- hf_endpoint=hf_endpoint,
- hf_token=hf_token,
- revision=source_revision,
- supported_datasets=supported_datasets,
- blocked_datasets=blocked_datasets,
- max_dataset_size=max_dataset_size,
- )
+ raise_if_blocked(dataset=dataset, blocked_datasets=blocked_datasets)
@@ -757,2 +797,6 @@ def compute_config_parquet_and_info_response(
- # prepare the parquet files locally
- local_parquet_files: List[ParquetFile] = []
+ # check if the repo exists and get the list of refs
+ try:
+ refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE)
+ except RepositoryNotFoundError as err:
+ raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err
+
@@ -770,17 +813,0 @@ def compute_config_parquet_and_info_response(
- writer_batch_size = get_writer_batch_size(builder.info)
- if writer_batch_size is not None and (
- builder._writer_batch_size is None or builder._writer_batch_size > writer_batch_size
- ):
- builder._writer_batch_size = writer_batch_size
- raise_if_too_big_from_external_data_files(
- builder=builder,
- max_dataset_size=max_dataset_size,
- max_external_data_files=max_external_data_files,
- hf_token=hf_token,
- )
- builder.download_and_prepare(file_format="parquet") # the parquet files are stored in the cache dir
- dataset_info = asdict(builder.info)
- local_parquet_files.extend(
- ParquetFile(local_file=local_file, local_dir=builder.cache_dir, config=config)
- for local_file in glob.glob(f"{builder.cache_dir}**/*.parquet")
- )
@@ -788 +815,20 @@ def compute_config_parquet_and_info_response(
- # create the target revision if it does not exist yet (clone from initial commit to avoid cloning all repo's files)
+ if is_parquet_builder_with_hub_files(builder, hf_endpoint=hf_endpoint):
+ parquet_operations = copy_parquet_files(builder)
+ fill_builder_info(builder, hf_token=hf_token)
+ else:
+ dataset_info = get_dataset_info_for_supported_datasets(
+ dataset=dataset, hf_endpoint=hf_endpoint, hf_token=hf_token, revision=source_revision, files_metadata=True
+ )
+ if dataset not in supported_datasets:
+ raise_if_not_supported(
+ dataset_info=dataset_info,
+ builder=builder,
+ hf_endpoint=hf_endpoint,
+ hf_token=hf_token,
+ max_dataset_size=max_dataset_size,
+ max_external_data_files=max_external_data_files,
+ )
+ parquet_operations = convert_to_parquet(builder)
+
+ # create the target revision if we managed to get the parquet files and it does not exist yet
+ # (clone from initial commit to avoid cloning all repo's files)
@@ -790 +835,0 @@ def compute_config_parquet_and_info_response(
- refs = hf_api.list_repo_refs(repo_id=dataset, repo_type=DATASET_TYPE)
@@ -794 +839 @@ def compute_config_parquet_and_info_response(
- repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit
+ repo_id=dataset, branch=target_revision, repo_type=DATASET_TYPE, revision=initial_commit, exist_ok=True
@@ -797 +842 @@ def compute_config_parquet_and_info_response(
- raise DatasetNotFoundError("The dataset does not exist on the Hub.") from err
+ raise DatasetNotFoundError("The dataset does not exist on the Hub (was deleted during job).") from err
@@ -798,0 +844 @@ def compute_config_parquet_and_info_response(
+ builder_info = asdict(builder.info)
@@ -803,4 +848,0 @@ def compute_config_parquet_and_info_response(
- # - get parquet files for current config
- config_files_to_add: Dict[str, str] = {
- parquet_file.repo_file(): parquet_file.local_file for parquet_file in local_parquet_files
- }
@@ -818,0 +861 @@ def compute_config_parquet_and_info_response(
+ config_files_to_add = [operation.path_in_repo for operation in parquet_operations]
@@ -823,7 +865,0 @@ def compute_config_parquet_and_info_response(
- # send the files to the target revision
- add_operations: List[CommitOperation] = [
- CommitOperationAdd(path_in_repo=file, path_or_fileobj=local_file)
- for file, local_file in config_files_to_add.items()
- ]
- logging.debug(f"{add_operations=}")
-
@@ -834 +870 @@ def compute_config_parquet_and_info_response(
- operations=delete_operations + add_operations,
+ operations=delete_operations + parquet_operations,
@@ -862 +898 @@ def compute_config_parquet_and_info_response(
- dataset_info=dataset_info,
+ dataset_info=builder_info,
diff --git a/services/worker/tests/fixtures/hub.py b/services/worker/tests/fixtures/hub.py
index 0c7287d2..110dba75 100644
--- a/services/worker/tests/fixtures/hub.py
+++ b/services/worker/tests/fixtures/hub.py
@@ -5,0 +6 @@
+import csv
@@ -225,0 +227,17 @@ def hub_public_big(datasets: Mapping[str, Dataset]) -> Iterator[str]:
[email protected](scope="session")
+def hub_public_big_no_info(datasets: Mapping[str, Dataset]) -> Iterator[str]:
+ repo_id = create_hub_dataset_repo(prefix="big-no-info", dataset=datasets["big"])
+ hf_api.delete_file(
+ "README.md", repo_id=repo_id, repo_type="dataset", commit_message="Delete README.md", token=CI_USER_TOKEN
+ )
+ yield repo_id
+ delete_hub_dataset_repo(repo_id=repo_id)
+
+
[email protected](scope="session")
+def hub_public_big_csv(big_csv_path: str) -> Iterator[str]:
+ repo_id = create_hub_dataset_repo(prefix="big-csv", file_paths=[big_csv_path])
+ yield repo_id
+ delete_hub_dataset_repo(repo_id=repo_id)
+
+
@@ -342,0 +361,30 @@ def create_dataset_info_response_for_csv(dataset: str, config: str) -> Any:
+def create_dataset_info_response_for_big_parquet() -> Any:
+ return {
+ "description": "",
+ "citation": "",
+ "homepage": "",
+ "license": "",
+ "features": BIG_cols,
+ "splits": {
+ "train": {"name": "train", "num_bytes": 5653946, "num_examples": len(BIG_rows), "dataset_name": None}
+ },
+ "download_size": BIG_PARQUET_FILE,
+ "dataset_size": 5653946,
+ }
+
+
+def create_dataset_info_response_for_big_parquet_no_info() -> Any:
+ return {
+ "description": "",
+ "citation": "",
+ "homepage": "",
+ "license": "",
+ "features": BIG_cols,
+ "splits": {
+ "train": {"name": "train", "num_bytes": 12345, "num_examples": len(BIG_rows), "dataset_name": None}
+ },
+ "download_size": BIG_PARQUET_FILE,
+ "dataset_size": 12345,
+ }
+
+
@@ -350,7 +398 @@ def create_dataset_info_response_for_audio() -> Any:
- "splits": {"train": {"name": "train", "num_bytes": 59, "num_examples": 1, "dataset_name": "parquet"}},
- "download_checksums": {
- "SOME_KEY": {
- "num_bytes": AUDIO_PARQUET_SIZE,
- "checksum": None,
- }
- },
+ "splits": {"train": {"name": "train", "num_bytes": 54, "num_examples": 1, "dataset_name": None}},
@@ -358,2 +400 @@ def create_dataset_info_response_for_audio() -> Any:
- "dataset_size": 59,
- "size_in_bytes": 1443,
+ "dataset_size": 54,
@@ -363 +404,3 @@ def create_dataset_info_response_for_audio() -> Any:
-def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "audio"]) -> Any:
+def create_parquet_and_info_response(
+ dataset: str, data_type: Literal["csv", "audio", "big_parquet", "big_parquet_no_info"]
+) -> Any:
@@ -367 +410 @@ def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "au
- size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE
+ size = CSV_PARQUET_SIZE if data_type == "csv" else AUDIO_PARQUET_SIZE if data_type == "audio" else BIG_PARQUET_FILE
@@ -371,0 +415,4 @@ def create_parquet_and_info_response(dataset: str, data_type: Literal["csv", "au
+ if data_type == "audio"
+ else create_dataset_info_response_for_big_parquet()
+ if data_type == "big_parquet"
+ else create_dataset_info_response_for_big_parquet_no_info()
@@ -391,0 +439 @@ AUDIO_PARQUET_SIZE = 1_384
+BIG_PARQUET_FILE = 38_896
@@ -488 +536 @@ BIG_cols = {
- "col": [{"_type": "Value", "dtype": "string"}],
+ "col": {"_type": "Value", "dtype": "string"},
@@ -491 +539,13 @@ BIG_cols = {
-BIG_rows = ["a" * 1_234 for _ in range(4_567)]
+BIG_rows = [{"col": "a" * 1_234} for _ in range(4_567)]
+
+
[email protected](scope="session")
+def big_csv_path(tmp_path_factory: pytest.TempPathFactory) -> str:
+ path = str(tmp_path_factory.mktemp("data") / "big_dataset.csv")
+ with open(path, "w", newline="") as f:
+ writer = csv.DictWriter(f, fieldnames=list(BIG_cols))
+ writer.writeheader()
+ for row in BIG_rows:
+ writer.writerow(row)
+ return path
+
@@ -526,0 +587,2 @@ def hub_datasets(
+ hub_public_big_no_info: str,
+ hub_public_big_csv: str,
@@ -618,0 +681,18 @@ def hub_datasets(
+ "parquet_and_info_response": create_parquet_and_info_response(
+ dataset=hub_public_big, data_type="big_parquet"
+ ),
+ },
+ "big-no-info": {
+ "name": hub_public_big_no_info,
+ "config_names_response": create_config_names_response(hub_public_big_no_info),
+ "splits_response": create_splits_response(hub_public_big_no_info),
+ "first_rows_response": create_first_rows_response(hub_public_big_no_info, BIG_cols, BIG_rows),
+ "parquet_and_info_response": create_parquet_and_info_response(
+ dataset=hub_public_big_no_info, data_type="big_parquet_no_info"
+ ),
+ },
+ "big-csv": {
+ "name": hub_public_big_csv,
+ "config_names_response": create_config_names_response(hub_public_big_csv),
+ "splits_response": create_splits_response(hub_public_big_csv),
+ "first_rows_response": create_first_rows_response(hub_public_big_csv, BIG_cols, BIG_rows),
diff --git a/services/worker/tests/job_runners/config/test_parquet_and_info.py b/services/worker/tests/job_runners/config/test_parquet_and_info.py
index 5ed9cd60..66944344 100644
--- a/services/worker/tests/job_runners/config/test_parquet_and_info.py
+++ b/services/worker/tests/job_runners/config/test_parquet_and_info.py
@@ -15 +15 @@ import requests
-from datasets import Audio, Features, Image, Value
+from datasets import Audio, Features, Image, Value, load_dataset_builder
@@ -16,0 +17 @@ from huggingface_hub.hf_api import HfApi
+from libcommon.dataset import get_dataset_info_for_supported_datasets
@@ -34 +34,0 @@ from worker.job_runners.config.parquet_and_info import (
- get_dataset_info_or_raise,
@@ -38 +37,0 @@ from worker.job_runners.config.parquet_and_info import (
- raise_if_not_supported,
@@ -122,4 +120,0 @@ def assert_content_is_equal(content: Any, expected: Any) -> None:
- assert len(content_value["download_checksums"]) == 1, content
- content_checksum = list(content_value["download_checksums"].values())[0]
- expected_checksum = list(expected_value["download_checksums"].values())[0]
- assert content_checksum == expected_checksum, content
@@ -235,0 +231 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_
+ builder = load_dataset_builder(hub_public_manual_download)
@@ -238,2 +234 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_
- hub_public_manual_download,
- "default",
+ builder=builder,
@@ -242 +236,0 @@ def test_raise_if_requires_manual_download(hub_public_manual_download: str, app_
- revision="main",
@@ -257 +251 @@ def test_raise_if_too_big_from_hub(
- dataset_info = get_dataset_info_or_raise(
+ dataset_info = get_dataset_info_for_supported_datasets(
@@ -261,0 +256 @@ def test_raise_if_too_big_from_hub(
+ files_metadata=True,
@@ -285 +280 @@ def test_raise_if_too_big_from_datasets(
- config = hub_datasets[name]["config_names_response"]["config_names"][0]["config"]
+ builder = load_dataset_builder(dataset)
@@ -289,5 +284 @@ def test_raise_if_too_big_from_datasets(
- dataset=dataset,
- config=config,
- hf_endpoint=app_config.common.hf_endpoint,
- hf_token=app_config.common.hf_token,
- revision="main",
+ info=builder.info,
@@ -298,5 +289 @@ def test_raise_if_too_big_from_datasets(
- dataset=dataset,
- config=config,
- hf_endpoint=app_config.common.hf_endpoint,
- hf_token=app_config.common.hf_token,
- revision="main",
+ info=builder.info,
@@ -373,9 +360 @@ def test_raise_if_too_many_external_files(
[email protected](
- "in_list,raises",
- [
- (True, False),
- (False, True),
- ],
-)
-def test_raise_if_not_supported(
- hub_datasets: HubDatasets,
+def test_supported_if_big_parquet(
@@ -383,2 +362,2 @@ def test_raise_if_not_supported(
- in_list: bool,
- raises: bool,
+ get_job_runner: GetJobRunner,
+ hub_datasets: HubDatasets,
@@ -385,0 +365,3 @@ def test_raise_if_not_supported(
+ # Not in the list of supported datasets and bigger than the maximum size
+ # but still supported since it's made of parquet files
+ # dataset = hub_public_big
@@ -388,23 +370,13 @@ def test_raise_if_not_supported(
- if raises:
- with pytest.raises(DatasetTooBigFromDatasetsError):
- raise_if_not_supported(
- dataset=dataset,
- config=config,
- hf_endpoint=app_config.common.hf_endpoint,
- hf_token=app_config.common.hf_token,
- revision="main",
- max_dataset_size=app_config.parquet_and_info.max_dataset_size,
- supported_datasets=[dataset] if in_list else ["another_dataset"],
- blocked_datasets=[],
- )
- else:
- raise_if_not_supported(
- dataset=dataset,
- config=config,
- hf_endpoint=app_config.common.hf_endpoint,
- hf_token=app_config.common.hf_token,
- revision="main",
- max_dataset_size=app_config.parquet_and_info.max_dataset_size,
- supported_datasets=[dataset] if in_list else ["another_dataset"],
- blocked_datasets=[],
- )
+ upsert_response(
+ kind="dataset-config-names",
+ dataset=dataset,
+ http_status=HTTPStatus.OK,
+ content=hub_datasets["big"]["config_names_response"],
+ )
+ job_runner = get_job_runner(dataset, config, app_config)
+ response = job_runner.compute()
+ assert response
+ content = response.content
+ assert content
+ assert len(content["parquet_files"]) == 1
+ assert_content_is_equal(content, hub_datasets["big"]["parquet_and_info_response"])
@@ -413 +385 @@ def test_raise_if_not_supported(
-def test_not_supported_if_big(
+def test_not_supported_if_big_non_parquet(
@@ -419,3 +391,3 @@ def test_not_supported_if_big(
- # dataset = hub_public_big
- dataset = hub_datasets["big"]["name"]
- config = hub_datasets["big"]["config_names_response"]["config_names"][0]["config"]
+ # dataset = hub_public_big_csv
+ dataset = hub_datasets["big-csv"]["name"]
+ config = hub_datasets["big-csv"]["config_names_response"]["config_names"][0]["config"]
@@ -426 +398 @@ def test_not_supported_if_big(
- content=hub_datasets["big"]["config_names_response"],
+ content=hub_datasets["big-csv"]["config_names_response"],
@@ -431 +403 @@ def test_not_supported_if_big(
- assert e.typename == "DatasetTooBigFromDatasetsError"
+ assert e.typename == "DatasetTooBigFromHubError"
@@ -532,0 +505,6 @@ def test_compute_splits_response_simple_csv_error(
+ upsert_response(
+ "dataset-config-names",
+ dataset=dataset,
+ http_status=HTTPStatus.OK,
+ content=hub_datasets[name]["config_names_response"],
+ )
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.