Datasets:
Modalities:
Text
Formats:
parquet
Languages:
English
Size:
1K - 10K
Tags:
biomedical-information-retrieval
citation-prediction-retrieval
passage-retrieval
news-retrieval
argument-retrieval
zero-shot-information-retrieval
License:
Convert to Parquet (#3)
Browse files- fix metadata (1bc6653564af544dbf15588077fceb056d475b98)
- Update README.md (e7ce91d25bbc4989f6bc879c5b962dab46d3fb73)
- Convert dataset to Parquet (bd00114f97da8b99bd310ef44a78aec1fd737f95)
- Add 'queries' config data files (bf00f5edaf006eefc706cbf09b10c2e854f0d72a)
- Delete loading script (80b5a63ff085ec6db829e1b6c6b9270b96e92cb6)
- Delete data file (72213e7a255d405dc7fd70d4f9f3b1fd00ca1387)
- Delete data file (266a5c58a26f7fff28cf9886d58e265cb1e039b6)
- Update README.md (9d1b0e188171749890135be5ef08256bce91e6ca)
Co-authored-by: Quentin Lhoest <[email protected]>
README.md
CHANGED
@@ -1,6 +1,4 @@
|
|
1 |
---
|
2 |
-
annotations_creators: []
|
3 |
-
language_creators: []
|
4 |
language:
|
5 |
- en
|
6 |
license:
|
@@ -9,54 +7,62 @@ multilinguality:
|
|
9 |
- monolingual
|
10 |
paperswithcode_id: beir
|
11 |
pretty_name: BEIR Benchmark
|
12 |
-
size_categories:
|
13 |
-
msmarco:
|
14 |
-
- 1M<n<10M
|
15 |
-
trec-covid:
|
16 |
-
- 100k<n<1M
|
17 |
-
nfcorpus:
|
18 |
-
- 1K<n<10K
|
19 |
-
nq:
|
20 |
-
- 1M<n<10M
|
21 |
-
hotpotqa:
|
22 |
-
- 1M<n<10M
|
23 |
-
fiqa:
|
24 |
-
- 10K<n<100K
|
25 |
-
arguana:
|
26 |
-
- 1K<n<10K
|
27 |
-
touche-2020:
|
28 |
-
- 100K<n<1M
|
29 |
-
cqadupstack:
|
30 |
-
- 100K<n<1M
|
31 |
-
quora:
|
32 |
-
- 100K<n<1M
|
33 |
-
dbpedia:
|
34 |
-
- 1M<n<10M
|
35 |
-
scidocs:
|
36 |
-
- 10K<n<100K
|
37 |
-
fever:
|
38 |
-
- 1M<n<10M
|
39 |
-
climate-fever:
|
40 |
-
- 1M<n<10M
|
41 |
-
scifact:
|
42 |
-
- 1K<n<10K
|
43 |
-
source_datasets: []
|
44 |
task_categories:
|
|
|
45 |
- text-retrieval
|
46 |
-
- zero-shot-retrieval
|
47 |
-
- information-retrieval
|
48 |
-
- zero-shot-information-retrieval
|
49 |
task_ids:
|
50 |
-
-
|
51 |
- entity-linking-retrieval
|
52 |
- fact-checking-retrieval
|
53 |
-
|
|
|
54 |
- citation-prediction-retrieval
|
55 |
-
-
|
56 |
-
- argument-retrieval
|
57 |
- news-retrieval
|
58 |
-
-
|
|
|
|
|
59 |
- question-answering-retrieval
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
60 |
---
|
61 |
|
62 |
# Dataset Card for BEIR Benchmark
|
|
|
1 |
---
|
|
|
|
|
2 |
language:
|
3 |
- en
|
4 |
license:
|
|
|
7 |
- monolingual
|
8 |
paperswithcode_id: beir
|
9 |
pretty_name: BEIR Benchmark
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
task_categories:
|
11 |
+
- zero-shot-classification
|
12 |
- text-retrieval
|
|
|
|
|
|
|
13 |
task_ids:
|
14 |
+
- document-retrieval
|
15 |
- entity-linking-retrieval
|
16 |
- fact-checking-retrieval
|
17 |
+
tags:
|
18 |
+
- biomedical-information-retrieval
|
19 |
- citation-prediction-retrieval
|
20 |
+
- passage-retrieval
|
|
|
21 |
- news-retrieval
|
22 |
+
- argument-retrieval
|
23 |
+
- zero-shot-information-retrieval
|
24 |
+
- tweet-retrieval
|
25 |
- question-answering-retrieval
|
26 |
+
- duplication-question-retrieval
|
27 |
+
- zero-shot-retrieval
|
28 |
+
configs:
|
29 |
+
- config_name: corpus
|
30 |
+
data_files:
|
31 |
+
- split: corpus
|
32 |
+
path: corpus/corpus-*
|
33 |
+
- config_name: queries
|
34 |
+
data_files:
|
35 |
+
- split: queries
|
36 |
+
path: queries/queries-*
|
37 |
+
dataset_info:
|
38 |
+
- config_name: corpus
|
39 |
+
features:
|
40 |
+
- name: _id
|
41 |
+
dtype: string
|
42 |
+
- name: title
|
43 |
+
dtype: string
|
44 |
+
- name: text
|
45 |
+
dtype: string
|
46 |
+
splits:
|
47 |
+
- name: corpus
|
48 |
+
num_bytes: 5856698
|
49 |
+
num_examples: 3633
|
50 |
+
download_size: 3193804
|
51 |
+
dataset_size: 5856698
|
52 |
+
- config_name: queries
|
53 |
+
features:
|
54 |
+
- name: _id
|
55 |
+
dtype: string
|
56 |
+
- name: title
|
57 |
+
dtype: string
|
58 |
+
- name: text
|
59 |
+
dtype: string
|
60 |
+
splits:
|
61 |
+
- name: queries
|
62 |
+
num_bytes: 141303
|
63 |
+
num_examples: 3237
|
64 |
+
download_size: 83904
|
65 |
+
dataset_size: 141303
|
66 |
---
|
67 |
|
68 |
# Dataset Card for BEIR Benchmark
|
corpus.jsonl.gz → corpus/corpus-00000-of-00001.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6d63810663c22a21d74f1689da586ecdc1aec941d4d99494b2a5f319f3078471
|
3 |
+
size 3193804
|
nfcorpus.py
DELETED
@@ -1,58 +0,0 @@
|
|
1 |
-
import json
|
2 |
-
import csv
|
3 |
-
import os
|
4 |
-
import datasets
|
5 |
-
|
6 |
-
logger = datasets.logging.get_logger(__name__)
|
7 |
-
|
8 |
-
_DESCRIPTION = "FIQA Dataset"
|
9 |
-
_SPLITS = ["corpus", "queries"]
|
10 |
-
|
11 |
-
URL = ""
|
12 |
-
_URLs = {subset: URL + f"{subset}.jsonl.gz" for subset in _SPLITS}
|
13 |
-
|
14 |
-
class BEIR(datasets.GeneratorBasedBuilder):
|
15 |
-
"""BEIR BenchmarkDataset."""
|
16 |
-
|
17 |
-
BUILDER_CONFIGS = [
|
18 |
-
datasets.BuilderConfig(
|
19 |
-
name=name,
|
20 |
-
description=f"This is the {name} in the FiQA dataset.",
|
21 |
-
) for name in _SPLITS
|
22 |
-
]
|
23 |
-
|
24 |
-
def _info(self):
|
25 |
-
|
26 |
-
return datasets.DatasetInfo(
|
27 |
-
description=_DESCRIPTION,
|
28 |
-
features=datasets.Features({
|
29 |
-
"_id": datasets.Value("string"),
|
30 |
-
"title": datasets.Value("string"),
|
31 |
-
"text": datasets.Value("string"),
|
32 |
-
}),
|
33 |
-
supervised_keys=None,
|
34 |
-
)
|
35 |
-
|
36 |
-
def _split_generators(self, dl_manager):
|
37 |
-
"""Returns SplitGenerators."""
|
38 |
-
|
39 |
-
my_urls = _URLs[self.config.name]
|
40 |
-
data_dir = dl_manager.download_and_extract(my_urls)
|
41 |
-
|
42 |
-
return [
|
43 |
-
datasets.SplitGenerator(
|
44 |
-
name=self.config.name,
|
45 |
-
# These kwargs will be passed to _generate_examples
|
46 |
-
gen_kwargs={"filepath": data_dir},
|
47 |
-
),
|
48 |
-
]
|
49 |
-
|
50 |
-
def _generate_examples(self, filepath):
|
51 |
-
"""Yields examples."""
|
52 |
-
with open(filepath, encoding="utf-8") as f:
|
53 |
-
texts = f.readlines()
|
54 |
-
for i, text in enumerate(texts):
|
55 |
-
text = json.loads(text)
|
56 |
-
if 'metadata' in text: del text['metadata']
|
57 |
-
if "title" not in text: text["title"] = ""
|
58 |
-
yield i, text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
queries.jsonl.gz → queries/queries-00000-of-00001.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:947d272a11835ab4690bb085928f951f26144bd4ecbc811dc48886297aa10b42
|
3 |
+
size 83904
|