YiDuo1999 commited on
Commit
8e98830
1 Parent(s): 21b8091

Upload 10 files

Browse files
Wiki/dataset_dict.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"splits": ["test", "train", "validation"]}
Wiki/test/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1811cc93f559e94692aec5a92d6daff8c837896ff241d61621c4d2327594ab8a
3
+ size 1291200
Wiki/test/dataset_info.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "wikitext_document_level",
3
+ "citation": "@misc{merity2016pointer,\n title={Pointer Sentinel Mixture Models},\n author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher},\n year={2016},\n eprint={1609.07843},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n",
4
+ "config_name": "wikitext-2-raw-v1",
5
+ "dataset_name": "wikitext_document_level",
6
+ "dataset_size": 13380433,
7
+ "description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n",
8
+ "download_checksums": {
9
+ "https://wikitext.smerity.com/wikitext-2-raw-v1.zip": {
10
+ "num_bytes": 4721645,
11
+ "checksum": null
12
+ }
13
+ },
14
+ "download_size": 4721645,
15
+ "features": {
16
+ "page": {
17
+ "dtype": "string",
18
+ "_type": "Value"
19
+ }
20
+ },
21
+ "homepage": "https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/",
22
+ "license": "Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)",
23
+ "size_in_bytes": 18102078,
24
+ "splits": {
25
+ "test": {
26
+ "name": "test",
27
+ "num_bytes": 1290775,
28
+ "num_examples": 62,
29
+ "dataset_name": "wikitext_document_level"
30
+ },
31
+ "train": {
32
+ "name": "train",
33
+ "num_bytes": 10942633,
34
+ "num_examples": 629,
35
+ "dataset_name": "wikitext_document_level"
36
+ },
37
+ "validation": {
38
+ "name": "validation",
39
+ "num_bytes": 1147025,
40
+ "num_examples": 60,
41
+ "dataset_name": "wikitext_document_level"
42
+ }
43
+ },
44
+ "version": {
45
+ "version_str": "1.0.0",
46
+ "major": 1,
47
+ "minor": 0,
48
+ "patch": 0
49
+ }
50
+ }
Wiki/test/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "55aaa3ceb3a80cdf",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": "test"
13
+ }
Wiki/train/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a61b4eceaf6d85585251938ceaf76a6726e852bd190f87a719c2f61753bc9c0
3
+ size 10943056
Wiki/train/dataset_info.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "wikitext_document_level",
3
+ "citation": "@misc{merity2016pointer,\n title={Pointer Sentinel Mixture Models},\n author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher},\n year={2016},\n eprint={1609.07843},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n",
4
+ "config_name": "wikitext-2-raw-v1",
5
+ "dataset_name": "wikitext_document_level",
6
+ "dataset_size": 13380433,
7
+ "description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n",
8
+ "download_checksums": {
9
+ "https://wikitext.smerity.com/wikitext-2-raw-v1.zip": {
10
+ "num_bytes": 4721645,
11
+ "checksum": null
12
+ }
13
+ },
14
+ "download_size": 4721645,
15
+ "features": {
16
+ "page": {
17
+ "dtype": "string",
18
+ "_type": "Value"
19
+ }
20
+ },
21
+ "homepage": "https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/",
22
+ "license": "Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)",
23
+ "size_in_bytes": 18102078,
24
+ "splits": {
25
+ "test": {
26
+ "name": "test",
27
+ "num_bytes": 1290775,
28
+ "num_examples": 62,
29
+ "dataset_name": "wikitext_document_level"
30
+ },
31
+ "train": {
32
+ "name": "train",
33
+ "num_bytes": 10942633,
34
+ "num_examples": 629,
35
+ "dataset_name": "wikitext_document_level"
36
+ },
37
+ "validation": {
38
+ "name": "validation",
39
+ "num_bytes": 1147025,
40
+ "num_examples": 60,
41
+ "dataset_name": "wikitext_document_level"
42
+ }
43
+ },
44
+ "version": {
45
+ "version_str": "1.0.0",
46
+ "major": 1,
47
+ "minor": 0,
48
+ "patch": 0
49
+ }
50
+ }
Wiki/train/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "95c0d7dd9f9809ed",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": "train"
13
+ }
Wiki/validation/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d868d420c1c3bad4d06aaa557bb1761d5096e5aee8078d58341a23f9180682f
3
+ size 1147456
Wiki/validation/dataset_info.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "wikitext_document_level",
3
+ "citation": "@misc{merity2016pointer,\n title={Pointer Sentinel Mixture Models},\n author={Stephen Merity and Caiming Xiong and James Bradbury and Richard Socher},\n year={2016},\n eprint={1609.07843},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n",
4
+ "config_name": "wikitext-2-raw-v1",
5
+ "dataset_name": "wikitext_document_level",
6
+ "dataset_size": 13380433,
7
+ "description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n",
8
+ "download_checksums": {
9
+ "https://wikitext.smerity.com/wikitext-2-raw-v1.zip": {
10
+ "num_bytes": 4721645,
11
+ "checksum": null
12
+ }
13
+ },
14
+ "download_size": 4721645,
15
+ "features": {
16
+ "page": {
17
+ "dtype": "string",
18
+ "_type": "Value"
19
+ }
20
+ },
21
+ "homepage": "https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/",
22
+ "license": "Creative Commons Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)",
23
+ "size_in_bytes": 18102078,
24
+ "splits": {
25
+ "test": {
26
+ "name": "test",
27
+ "num_bytes": 1290775,
28
+ "num_examples": 62,
29
+ "dataset_name": "wikitext_document_level"
30
+ },
31
+ "train": {
32
+ "name": "train",
33
+ "num_bytes": 10942633,
34
+ "num_examples": 629,
35
+ "dataset_name": "wikitext_document_level"
36
+ },
37
+ "validation": {
38
+ "name": "validation",
39
+ "num_bytes": 1147025,
40
+ "num_examples": 60,
41
+ "dataset_name": "wikitext_document_level"
42
+ }
43
+ },
44
+ "version": {
45
+ "version_str": "1.0.0",
46
+ "major": 1,
47
+ "minor": 0,
48
+ "patch": 0
49
+ }
50
+ }
Wiki/validation/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "2e82649119cc49d0",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": "validation"
13
+ }