nthakur lhoestq HF Staff commited on
Commit
9e6cd3a
·
verified ·
1 Parent(s): e28763e

Convert to Parquet (#3)

Browse files

- fix metadata (1bc6653564af544dbf15588077fceb056d475b98)
- Update README.md (e7ce91d25bbc4989f6bc879c5b962dab46d3fb73)
- Convert dataset to Parquet (bd00114f97da8b99bd310ef44a78aec1fd737f95)
- Add 'queries' config data files (bf00f5edaf006eefc706cbf09b10c2e854f0d72a)
- Delete loading script (80b5a63ff085ec6db829e1b6c6b9270b96e92cb6)
- Delete data file (72213e7a255d405dc7fd70d4f9f3b1fd00ca1387)
- Delete data file (266a5c58a26f7fff28cf9886d58e265cb1e039b6)
- Update README.md (9d1b0e188171749890135be5ef08256bce91e6ca)


Co-authored-by: Quentin Lhoest <[email protected]>

README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- annotations_creators: []
3
- language_creators: []
4
  language:
5
  - en
6
  license:
@@ -9,54 +7,62 @@ multilinguality:
9
  - monolingual
10
  paperswithcode_id: beir
11
  pretty_name: BEIR Benchmark
12
- size_categories:
13
- msmarco:
14
- - 1M<n<10M
15
- trec-covid:
16
- - 100k<n<1M
17
- nfcorpus:
18
- - 1K<n<10K
19
- nq:
20
- - 1M<n<10M
21
- hotpotqa:
22
- - 1M<n<10M
23
- fiqa:
24
- - 10K<n<100K
25
- arguana:
26
- - 1K<n<10K
27
- touche-2020:
28
- - 100K<n<1M
29
- cqadupstack:
30
- - 100K<n<1M
31
- quora:
32
- - 100K<n<1M
33
- dbpedia:
34
- - 1M<n<10M
35
- scidocs:
36
- - 10K<n<100K
37
- fever:
38
- - 1M<n<10M
39
- climate-fever:
40
- - 1M<n<10M
41
- scifact:
42
- - 1K<n<10K
43
- source_datasets: []
44
  task_categories:
 
45
  - text-retrieval
46
- - zero-shot-retrieval
47
- - information-retrieval
48
- - zero-shot-information-retrieval
49
  task_ids:
50
- - passage-retrieval
51
  - entity-linking-retrieval
52
  - fact-checking-retrieval
53
- - tweet-retrieval
 
54
  - citation-prediction-retrieval
55
- - duplication-question-retrieval
56
- - argument-retrieval
57
  - news-retrieval
58
- - biomedical-information-retrieval
 
 
59
  - question-answering-retrieval
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
60
  ---
61
 
62
  # Dataset Card for BEIR Benchmark
 
1
  ---
 
 
2
  language:
3
  - en
4
  license:
 
7
  - monolingual
8
  paperswithcode_id: beir
9
  pretty_name: BEIR Benchmark
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  task_categories:
11
+ - zero-shot-classification
12
  - text-retrieval
 
 
 
13
  task_ids:
14
+ - document-retrieval
15
  - entity-linking-retrieval
16
  - fact-checking-retrieval
17
+ tags:
18
+ - biomedical-information-retrieval
19
  - citation-prediction-retrieval
20
+ - passage-retrieval
 
21
  - news-retrieval
22
+ - argument-retrieval
23
+ - zero-shot-information-retrieval
24
+ - tweet-retrieval
25
  - question-answering-retrieval
26
+ - duplication-question-retrieval
27
+ - zero-shot-retrieval
28
+ configs:
29
+ - config_name: corpus
30
+ data_files:
31
+ - split: corpus
32
+ path: corpus/corpus-*
33
+ - config_name: queries
34
+ data_files:
35
+ - split: queries
36
+ path: queries/queries-*
37
+ dataset_info:
38
+ - config_name: corpus
39
+ features:
40
+ - name: _id
41
+ dtype: string
42
+ - name: title
43
+ dtype: string
44
+ - name: text
45
+ dtype: string
46
+ splits:
47
+ - name: corpus
48
+ num_bytes: 5856698
49
+ num_examples: 3633
50
+ download_size: 3193804
51
+ dataset_size: 5856698
52
+ - config_name: queries
53
+ features:
54
+ - name: _id
55
+ dtype: string
56
+ - name: title
57
+ dtype: string
58
+ - name: text
59
+ dtype: string
60
+ splits:
61
+ - name: queries
62
+ num_bytes: 141303
63
+ num_examples: 3237
64
+ download_size: 83904
65
+ dataset_size: 141303
66
  ---
67
 
68
  # Dataset Card for BEIR Benchmark
corpus.jsonl.gz → corpus/corpus-00000-of-00001.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae4d6ab2ab88d880e7e180776c956a09c19c43a703457742cd850d2422035802
3
- size 1978293
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d63810663c22a21d74f1689da586ecdc1aec941d4d99494b2a5f319f3078471
3
+ size 3193804
nfcorpus.py DELETED
@@ -1,58 +0,0 @@
1
- import json
2
- import csv
3
- import os
4
- import datasets
5
-
6
- logger = datasets.logging.get_logger(__name__)
7
-
8
- _DESCRIPTION = "FIQA Dataset"
9
- _SPLITS = ["corpus", "queries"]
10
-
11
- URL = ""
12
- _URLs = {subset: URL + f"{subset}.jsonl.gz" for subset in _SPLITS}
13
-
14
- class BEIR(datasets.GeneratorBasedBuilder):
15
- """BEIR BenchmarkDataset."""
16
-
17
- BUILDER_CONFIGS = [
18
- datasets.BuilderConfig(
19
- name=name,
20
- description=f"This is the {name} in the FiQA dataset.",
21
- ) for name in _SPLITS
22
- ]
23
-
24
- def _info(self):
25
-
26
- return datasets.DatasetInfo(
27
- description=_DESCRIPTION,
28
- features=datasets.Features({
29
- "_id": datasets.Value("string"),
30
- "title": datasets.Value("string"),
31
- "text": datasets.Value("string"),
32
- }),
33
- supervised_keys=None,
34
- )
35
-
36
- def _split_generators(self, dl_manager):
37
- """Returns SplitGenerators."""
38
-
39
- my_urls = _URLs[self.config.name]
40
- data_dir = dl_manager.download_and_extract(my_urls)
41
-
42
- return [
43
- datasets.SplitGenerator(
44
- name=self.config.name,
45
- # These kwargs will be passed to _generate_examples
46
- gen_kwargs={"filepath": data_dir},
47
- ),
48
- ]
49
-
50
- def _generate_examples(self, filepath):
51
- """Yields examples."""
52
- with open(filepath, encoding="utf-8") as f:
53
- texts = f.readlines()
54
- for i, text in enumerate(texts):
55
- text = json.loads(text)
56
- if 'metadata' in text: del text['metadata']
57
- if "title" not in text: text["title"] = ""
58
- yield i, text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
queries.jsonl.gz → queries/queries-00000-of-00001.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:baba3504a2736b0e32265962f04acc4fecfaf9649e0ea08e49ff441246f75364
3
- size 79080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:947d272a11835ab4690bb085928f951f26144bd4ecbc811dc48886297aa10b42
3
+ size 83904