Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -1,23 +1,126 @@
|
|
1 |
---
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
|
|
|
|
20 |
---
|
21 |
-
# Dataset Card for
|
22 |
-
|
23 |
-
[
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
pretty_name: Golos
|
3 |
+
annotations_creators:
|
4 |
+
- expert-generated
|
5 |
+
language_creators:
|
6 |
+
- crowdsourced
|
7 |
+
- expert-generated
|
8 |
+
language:
|
9 |
+
- ru
|
10 |
+
license:
|
11 |
+
- other
|
12 |
+
multilinguality:
|
13 |
+
- monolingual
|
14 |
+
paperswithcode_id: golos
|
15 |
+
size_categories:
|
16 |
+
- 10K<n<100k
|
17 |
+
source_datasets:
|
18 |
+
- extended
|
19 |
+
task_categories:
|
20 |
+
- automatic-speech-recognition
|
21 |
+
- audio-classification
|
22 |
---
|
23 |
+
# Dataset Card for sberdevices_golos_100h_farfield
|
24 |
+
## Table of Contents
|
25 |
+
- [Dataset Description](#dataset-description)
|
26 |
+
- [Dataset Summary](#dataset-summary)
|
27 |
+
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
|
28 |
+
- [Languages](#languages)
|
29 |
+
- [Dataset Structure](#dataset-structure)
|
30 |
+
- [Data Instances](#data-instances)
|
31 |
+
- [Data Fields](#data-fields)
|
32 |
+
- [Data Splits](#data-splits)
|
33 |
+
- [Dataset Creation](#dataset-creation)
|
34 |
+
- [Curation Rationale](#curation-rationale)
|
35 |
+
- [Source Data](#source-data)
|
36 |
+
- [Annotations](#annotations)
|
37 |
+
- [Personal and Sensitive Information](#personal-and-sensitive-information)
|
38 |
+
- [Considerations for Using the Data](#considerations-for-using-the-data)
|
39 |
+
- [Social Impact of Dataset](#social-impact-of-dataset)
|
40 |
+
- [Discussion of Biases](#discussion-of-biases)
|
41 |
+
- [Other Known Limitations](#other-known-limitations)
|
42 |
+
- [Additional Information](#additional-information)
|
43 |
+
- [Dataset Curators](#dataset-curators)
|
44 |
+
- [Licensing Information](#licensing-information)
|
45 |
+
- [Citation Information](#citation-information)
|
46 |
+
- [Contributions](#contributions)
|
47 |
+
## Dataset Description
|
48 |
+
- **Homepage:** [Golos ASR corpus](https://www.openslr.org/114)
|
49 |
+
- **Repository:** [Golos dataset](https://github.com/sberdevices/golos)
|
50 |
+
- **Paper:** [Golos: Russian Dataset for Speech Research](https://arxiv.org/pdf/2106.10161.pdf)
|
51 |
+
- **Leaderboard:** [The 🤗 Speech Bench](https://huggingface.co/spaces/huggingface/hf-speech-bench)
|
52 |
+
- **Point of Contact:** [Nikolay Karpov](mailto:karpnv@gmail.com)
|
53 |
+
### Dataset Summary
|
54 |
+
Sberdevices Golos is a corpus of approximately 1200 hours of 16kHz Russian speech from crowd (reading speech) and farfield (communication with smart devices) domains, prepared by SberDevices Team (Alexander Denisenko, Angelina Kovalenko, Fedor Minkin, and Nikolay Karpov). The data is derived from the crowd-sourcing platform, and has been manually annotated.
|
55 |
+
Authors divide all dataset into train and test subsets. The training subset includes approximately 1000 hours. For experiments with a limited number of records, authors identified training subsets of shorter length: 100 hours, 10 hours, 1 hour, 10 minutes.
|
56 |
+
This dataset is a simpler version of the above mentioned Golos:
|
57 |
+
- it includes the farfield domain only (without any sound from the crowd domain);
|
58 |
+
- validation split is built on the 10-hour training subset;
|
59 |
+
- training split corresponds to the 100-hour training subset without sounds from the 10-hour training subset;
|
60 |
+
- test split is a full original test split.
|
61 |
+
### Supported Tasks and Leaderboards
|
62 |
+
- `automatic-speech-recognition`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active Hugging Face leaderboard which can be found at https://huggingface.co/spaces/huggingface/hf-speech-bench. The leaderboard ranks models uploaded to the Hub based on their WER.
|
63 |
+
### Languages
|
64 |
+
The audio is in Russian.
|
65 |
+
## Dataset Structure
|
66 |
+
### Data Instances
|
67 |
+
A typical data point comprises the audio data, usually called `audio` and its transcription, called `transcription`. Any additional information about the speaker and the passage which contains the transcription is not provided.
|
68 |
+
```
|
69 |
+
{'audio': {'path': None,
|
70 |
+
'array': array([ 1.22070312e-04, 1.22070312e-04, 9.15527344e-05, ...,
|
71 |
+
6.10351562e-05, 6.10351562e-05, 3.05175781e-05]), dtype=float64),
|
72 |
+
'sampling_rate': 16000},
|
73 |
+
'transcription': 'джой источники истории турции'}
|
74 |
+
```
|
75 |
+
### Data Fields
|
76 |
+
- audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`.
|
77 |
+
- transcription: the transcription of the audio file.
|
78 |
+
### Data Splits
|
79 |
+
This dataset is a simpler version of the original Golos:
|
80 |
+
- it includes the farfield domain only (without any sound from the crowd domain);
|
81 |
+
- validation split is built on the 10-hour training subset;
|
82 |
+
- training split corresponds to the 100-hour training subset without sounds from the 10-hour training subset;
|
83 |
+
- test split is a full original test split.
|
84 |
+
| | Train | Validation | Test |
|
85 |
+
| ----- | ------ | ---------- | ----- |
|
86 |
+
| examples | 9570 | 933 | 1916 |
|
87 |
+
| hours | 10.3h | 1.0h | 1.4h |
|
88 |
+
## Dataset Creation
|
89 |
+
### Curation Rationale
|
90 |
+
[Needs More Information]
|
91 |
+
### Source Data
|
92 |
+
#### Initial Data Collection and Normalization
|
93 |
+
[Needs More Information]
|
94 |
+
#### Who are the source language producers?
|
95 |
+
[Needs More Information]
|
96 |
+
### Annotations
|
97 |
+
#### Annotation process
|
98 |
+
All recorded audio files were manually annotated on the crowd-sourcing platform.
|
99 |
+
#### Who are the annotators?
|
100 |
+
[Needs More Information]
|
101 |
+
### Personal and Sensitive Information
|
102 |
+
The dataset consists of people who have donated their voice. You agree to not attempt to determine the identity of speakers in this dataset.
|
103 |
+
## Considerations for Using the Data
|
104 |
+
### Social Impact of Dataset
|
105 |
+
[More Information Needed]
|
106 |
+
### Discussion of Biases
|
107 |
+
[More Information Needed]
|
108 |
+
### Other Known Limitations
|
109 |
+
[Needs More Information]
|
110 |
+
## Additional Information
|
111 |
+
### Dataset Curators
|
112 |
+
The dataset was initially created by Alexander Denisenko, Angelina Kovalenko, Fedor Minkin, and Nikolay Karpov.
|
113 |
+
### Licensing Information
|
114 |
+
[Public license with attribution and conditions reserved](https://github.com/sberdevices/golos/blob/master/license/en_us.pdf)
|
115 |
+
### Citation Information
|
116 |
+
```
|
117 |
+
@misc{karpov2021golos,
|
118 |
+
author = {Karpov, Nikolay and Denisenko, Alexander and Minkin, Fedor},
|
119 |
+
title = {Golos: Russian Dataset for Speech Research},
|
120 |
+
publisher = {arXiv},
|
121 |
+
year = {2021},
|
122 |
+
url = {https://arxiv.org/abs/2106.10161}
|
123 |
+
}
|
124 |
+
```
|
125 |
+
### Contributions
|
126 |
+
Thanks to [@bond005](https://github.com/bond005) for adding this dataset.
|