|
--- |
|
dataset_info: |
|
features: |
|
- name: name |
|
dtype: string |
|
- name: speaker_embeddings |
|
sequence: float32 |
|
splits: |
|
- name: validation |
|
num_bytes: 634175 |
|
num_examples: 305 |
|
download_size: 979354 |
|
dataset_size: 634175 |
|
license: mit |
|
language: |
|
- ar |
|
size_categories: |
|
- n<1K |
|
task_categories: |
|
- text-to-speech |
|
- audio-to-audio |
|
pretty_name: Arabic(M) Speaker Embeddings |
|
--- |
|
|
|
# Arabic Speaker Embeddings extracted from ASC and ClArTTS |
|
|
|
There is one speaker embedding for each utterance in the validation set of both datasets. The speaker embeddings are 512-element X-vectors. |
|
|
|
[Arabic Speech Corpus](https://huggingface.co/datasets/arabic_speech_corpus) has 100 files for a single male speaker and [ClArTTS](https://www.clartts.com/) has 205 files for a single male speaker. |
|
|
|
The X-vectors were extracted using [this script](https://huggingface.co/mechanicalsea/speecht5-vc/blob/main/manifest/utils/prep_cmu_arctic_spkemb.py), which uses the `speechbrain/spkrec-xvect-voxceleb` model. |
|
|
|
Usage: |
|
|
|
```python |
|
from datasets import load_dataset |
|
|
|
embeddings_dataset = load_dataset("herwoww/arabic_xvect_embeddings", split="validation") |
|
speaker_embedding = torch.tensor(embeddings_dataset[1]["speaker_embeddings"]).unsqueeze(0) |
|
``` |