albertvillanova HF staff commited on
Commit
0e634bb
1 Parent(s): 5dd2949

Convert dataset to Parquet

Browse files

Convert dataset to Parquet.

This dataset uses `tasks`, which are deprecated and will raise an error after the next release of `datasets`. See: https://github.com/huggingface/datasets/pull/6999

README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- pretty_name: Arabic Speech Corpus
3
  annotations_creators:
4
  - expert-generated
5
  language_creators:
@@ -10,7 +9,6 @@ license:
10
  - cc-by-4.0
11
  multilinguality:
12
  - monolingual
13
- paperswithcode_id: arabic-speech-corpus
14
  size_categories:
15
  - 1K<n<10K
16
  source_datasets:
@@ -18,22 +16,10 @@ source_datasets:
18
  task_categories:
19
  - automatic-speech-recognition
20
  task_ids: []
21
- train-eval-index:
22
- - config: clean
23
- task: automatic-speech-recognition
24
- task_id: speech_recognition
25
- splits:
26
- train_split: train
27
- eval_split: test
28
- col_mapping:
29
- file: path
30
- text: text
31
- metrics:
32
- - type: wer
33
- name: WER
34
- - type: cer
35
- name: CER
36
  dataset_info:
 
37
  features:
38
  - name: file
39
  dtype: string
@@ -47,16 +33,38 @@ dataset_info:
47
  dtype: string
48
  - name: orthographic
49
  dtype: string
50
- config_name: clean
51
  splits:
52
  - name: train
53
- num_bytes: 1002365
54
  num_examples: 1813
55
  - name: test
56
- num_bytes: 65784
57
  num_examples: 100
58
- download_size: 1192302846
59
- dataset_size: 1068149
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
60
  ---
61
 
62
  # Dataset Card for Arabic Speech Corpus
 
1
  ---
 
2
  annotations_creators:
3
  - expert-generated
4
  language_creators:
 
9
  - cc-by-4.0
10
  multilinguality:
11
  - monolingual
 
12
  size_categories:
13
  - 1K<n<10K
14
  source_datasets:
 
16
  task_categories:
17
  - automatic-speech-recognition
18
  task_ids: []
19
+ paperswithcode_id: arabic-speech-corpus
20
+ pretty_name: Arabic Speech Corpus
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  dataset_info:
22
+ config_name: clean
23
  features:
24
  - name: file
25
  dtype: string
 
33
  dtype: string
34
  - name: orthographic
35
  dtype: string
 
36
  splits:
37
  - name: train
38
+ num_bytes: 1527815416.966
39
  num_examples: 1813
40
  - name: test
41
+ num_bytes: 99851729.0
42
  num_examples: 100
43
+ download_size: 1347643373
44
+ dataset_size: 1627667145.966
45
+ configs:
46
+ - config_name: clean
47
+ data_files:
48
+ - split: train
49
+ path: clean/train-*
50
+ - split: test
51
+ path: clean/test-*
52
+ default: true
53
+ train-eval-index:
54
+ - config: clean
55
+ task: automatic-speech-recognition
56
+ task_id: speech_recognition
57
+ splits:
58
+ train_split: train
59
+ eval_split: test
60
+ col_mapping:
61
+ file: path
62
+ text: text
63
+ metrics:
64
+ - type: wer
65
+ name: WER
66
+ - type: cer
67
+ name: CER
68
  ---
69
 
70
  # Dataset Card for Arabic Speech Corpus
clean/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd5ea889532615c4ca9c63b5b83fc3bacb94e9fa156c26f5963b8da2c8e87768
3
+ size 90899095
clean/train-00000-of-00004.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c3f2931ab19224daf55126c1cf96ff068f3ad442d760c1f5db99805d5a290be
3
+ size 398895011
clean/train-00001-of-00004.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d02e7e7d080082d1d96929b83e19b924d7c10c8b59a39f190c373245559ea36d
3
+ size 322764456
clean/train-00002-of-00004.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ea3385f7d8496bf1e77d9b1a2696fb2bb3769e1ffa060e43fa4fc6c5e25cf06
3
+ size 291793854
clean/train-00003-of-00004.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91bbec3487d3ba745113c5869be40a6008ef815b9681fe683cf7ab46dd06efcf
3
+ size 243290957