parquet-converter commited on
Commit
7b1a54b
·
1 Parent(s): 5b250ef

Update parquet files

Browse files
corpus.jsonl.gz → default/msmarco-passage-corpus-train-00000-of-00007.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:634543cee77a992613bbf32d0aa4863659dffd064c405f062454400aa80aea3f
3
- size 1118496727
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73962b19c9777386696f7bc062ce5ecb5304ffa2b35a2af11c719bf5a96dbdbd
3
+ size 266662221
default/msmarco-passage-corpus-train-00001-of-00007.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea268eec345592711d48b7b08cfe739ab3aafaacb7a2078232363d743252ecc2
3
+ size 267171922
default/msmarco-passage-corpus-train-00002-of-00007.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1db78ad97274f9231318b8b2469b88475afb0f283998c1345b32930fda31f021
3
+ size 269469670
default/msmarco-passage-corpus-train-00003-of-00007.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f3081526a7c06146e0ce8b021dae335d5079c6a128647f1e7be820c22c015b8
3
+ size 269502551
default/msmarco-passage-corpus-train-00004-of-00007.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d71fe6a48eafa74cb3a4fa4d9db4a3f289932751c8a63bf691cd75a4e5523bc
3
+ size 258932986
default/msmarco-passage-corpus-train-00005-of-00007.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7cae415b497671301711673c771796f811b2c54c1a05c9b4cb869009b8cd44c
3
+ size 272091420
default/msmarco-passage-corpus-train-00006-of-00007.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d34a4c47e6cc1c4c22bf45af3fd99f53468b217f08bcb92c4b4a8768a470454a
3
+ size 202001221
msmarco-passage-corpus.py DELETED
@@ -1,91 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.Wikipedia
15
-
16
- # Lint as: python3
17
- """MsMarco Passage dataset."""
18
-
19
- import json
20
-
21
- import datasets
22
-
23
- _CITATION = """
24
- @misc{bajaj2018ms,
25
- title={MS MARCO: A Human Generated MAchine Reading COmprehension Dataset},
26
- author={Payal Bajaj and Daniel Campos and Nick Craswell and Li Deng and Jianfeng Gao and Xiaodong Liu
27
- and Rangan Majumder and Andrew McNamara and Bhaskar Mitra and Tri Nguyen and Mir Rosenberg and Xia Song
28
- and Alina Stoica and Saurabh Tiwary and Tong Wang},
29
- year={2018},
30
- eprint={1611.09268},
31
- archivePrefix={arXiv},
32
- primaryClass={cs.CL}
33
- }
34
- """
35
-
36
- _DESCRIPTION = "dataset load script for MSMARCO Passage Corpus"
37
-
38
- _DATASET_URLS = {
39
- 'train': "https://huggingface.co/datasets/Tevatron/msmarco-passage-corpus/resolve/main/corpus.jsonl.gz",
40
- }
41
-
42
-
43
- class MsMarcoPassageCorpus(datasets.GeneratorBasedBuilder):
44
- VERSION = datasets.Version("0.0.1")
45
-
46
- BUILDER_CONFIGS = [
47
- datasets.BuilderConfig(version=VERSION,
48
- description="MS MARCO passage Corpus"),
49
- ]
50
-
51
- def _info(self):
52
- features = datasets.Features(
53
- {'docid': datasets.Value('string'), 'title': datasets.Value('string'), 'text': datasets.Value('string')}
54
- )
55
-
56
- return datasets.DatasetInfo(
57
- # This is the description that will appear on the datasets page.
58
- description=_DESCRIPTION,
59
- # This defines the different columns of the dataset and their types
60
- features=features, # Here we define them above because they are different between the two configurations
61
- supervised_keys=None,
62
- # Homepage of the dataset for documentation
63
- homepage="",
64
- # License for the dataset if available
65
- license="",
66
- # Citation for the dataset
67
- citation=_CITATION,
68
- )
69
-
70
- def _split_generators(self, dl_manager):
71
- if self.config.data_files:
72
- downloaded_files = self.config.data_files
73
- else:
74
- downloaded_files = dl_manager.download_and_extract(_DATASET_URLS)
75
- splits = [
76
- datasets.SplitGenerator(
77
- name=split,
78
- gen_kwargs={
79
- "files": [downloaded_files[split]] if isinstance(downloaded_files[split], str) else downloaded_files[split],
80
- },
81
- ) for split in downloaded_files
82
- ]
83
- return splits
84
-
85
- def _generate_examples(self, files):
86
- """Yields examples."""
87
- for filepath in files:
88
- with open(filepath, encoding="utf-8") as f:
89
- for line in f:
90
- data = json.loads(line)
91
- yield data['docid'], data