Datasets:

Modalities:
Audio
Text
Formats:
parquet
Libraries:
Datasets
Dask
License:

You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Please enter your affiliation and agree to the following terms to use the dataset.

Log in or Sign Up to review the conditions and access this dataset content.

SpeechBSD Dataset

This is an extension of the BSD corpus, a Japanese--English dialogue translation corpus, with audio files and speaker attribute information. Although the primary intended usage is for speech-to-text translation, it can be viewed as text/speech Japanese/English/cross-language dialogue corpus and can be used for various tasks.

Dataset Statistics

Train Dev. Test
Scenarios 670 69 69
Sentences 20,000 2,051 2,120
En audio (h) 20.1 2.1 2.1
Ja audio (h) 25.3 2.7 2.7
En audio gender (male % / female %) 47.2 / 52.8 50.1 / 49.9 44.4 / 55.6
Ja audio gender (male % / female %) 68.0 / 32.0 62.3 / 37.7 69.0 / 31.0

Data Structure

We also provide the dataset at the GitHub repository. The data structure there is similar to the original BSD corpus. Here, the structure is changed to be represented in the JSONL format where one instance contains one audio file.

Data Instances

There are two types of instances, one that contains English wav file, and one that contains Japanese wav file.

A typical instance that contains an English wav file looks like this:

{
  'audio': {'path': '/path/to/speech-bsd-hf/data/dev/190315_E001_17_spk0_no10_en.wav', 'array': array([ 0.00000000e+00,  0.00000000e+00,  0.00000000e+00, ...,
       -6.10351562e-05, -1.83105469e-04,  6.10351562e-05]), 'sampling_rate': 16000},
  'no': 10,
  'en_speaker': 'Mr. Ben Sherman',
  'en_sentence': 'You can also check out major institutions like banks, accounting companies, and market research companies.',
  'en_spkid': '190315_E001_17_spk0_en',
  'en_wav': '190315_E001_17_spk0_no10_en.wav',
  'en_spk_gender': 'M',
  'en_spk_state': 'CA',
  'scenario_id': '190315_E001_17',
  'scenario_tag': 'training',
  'scenario_title': 'Training: How to do research',
  'scenario_original_language': 'en',
  'ja_speaker': None,
  'ja_sentence': None,
  'ja_spkid': None,
  'ja_wav': None,
  'ja_spk_gender': None,
  'ja_spk_prefecture': None
}

In the corresponding instance that contains a Japanese wav file, en_speaker, en_sentence, en_spkid, en_wav, en_spk_gender, and en_spk_state are None and corresponding Japanese ones are filled instead.

Data Fields

Each utterance is a part of a scenario.

The scenario information is shown with scenario_id, scenario_tag, scenario_title, and scenario_original_language, which corresponds respectively to id, tag, title, and original_language in the original BSD corpus.

The utterance-specific fields are the following:

  • no, ja_speaker, en_speaker, ja_sentence, en_sentence are identical to the ones of the BSD corpus.
  • ja_spkid and en_spkid show speaker IDs consistent throughout the conversation.
  • ja_wav and en_wav show the wavfile names.
  • ja_spk_gender and en_spk_gender show the gender of the speaker of the corresponding wav file (either "M" or "F").
  • ja_spk_prefecture and en_spk_state show the speaker's region of origin.
  • As any other huggingface audio datasets, audio contains full path to the audio file in your environment, audio array, and sampling rate (which is 16000). You will need librosa and soundfile packages to have access to these fields.

Here are some additional notes:

  • Speakers are different if speaker ID is different. For example, if a conversation is spoken by two speakers taking turns, there would be 4 speakers (2 Japanese speakers and 2 English speakers). However, it's possible that speakers with different speaker ID is actually spoken by the same person because of the way audio is collected.

  • Gender information of audio does not necessarily match with the one inferrable from text. For example, even if the en_speaker is "Mr. Sam Lee", the audio may contain female voice. This is because no explicit gender information is given in the original BSD corpus.

  • Japanese speech is collected from Japanese speakers who are from Japan.

    • ja_spk_prefecture is one of the 47 prefectures or "不明" (unknown). Japanese prefectures have four different ending characters, "県", "府", "都", and "道".
      • Prefectures that ends with "県" or "府" does not contain those characters (e.g., "神奈川", "京都").
      • Tokyo is "東京" without "都".
      • Hokkaido is "北海道".
  • English speech is collected from English speakers who are from the US.

    • en_spk is one of the 50 states, written in postal abbreviation.

Citation

If you find the dataset useful, please cite our ACL 2023 Findings paper: Towards Speech Dialogue Translation Mediating Speakers of Different Languages.

@inproceedings{shimizu-etal-2023-towards,
  title = "Towards Speech Dialogue Translation Mediating Speakers of Different Languages",
  author = "Shimizu, Shuichiro  and
    Chu, Chenhui  and
    Li, Sheng  and
    Kurohashi, Sadao",
  booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
  month = jul,
  year = "2023",
  address = "Toronto, Canada",
  publisher = "Association for Computational Linguistics",
  url = "https://aclanthology.org/2023.findings-acl.72",
  pages = "1122--1134",
  abstract = "We present a new task, speech dialogue translation mediating speakers of different languages. We construct the SpeechBSD dataset for the task and conduct baseline experiments. Furthermore, we consider context to be an important aspect that needs to be addressed in this task and propose two ways of utilizing context, namely monolingual context and bilingual context. We conduct cascaded speech translation experiments using Whisper and mBART, and show that bilingual context performs better in our settings.",
}

License

This dataset is licensed under CC-BY-NC-SA 4.0.

Downloads last month
30