metadata
library_name: transformers
license: apache-2.0
datasets:
- facebook/multilingual_librispeech
- speechcolab/gigaspeech
- mozilla-foundation/common_voice_15_0
- facebook/voxpopuli
- MLCommons/peoples_speech
language:
- en
base_model:
- mistralai/Mistral-7B-v0.1
pipeline_tag: audio-to-audio
Model Card
Unified Speech-Text Model (USTM), a speech-text cross-modal pretrained model obtained by further training Mistral-7B-v0.1 using a proposed unified speech-text pretraining methodology.
Paralinguistics-Aware Speech-Empowered LLMs for Natural Conversation [NeurIPS 2024]
- Repository: https://github.com/naver-ai/usdm
- Paper: https://openreview.net/forum?id=NjewXJUDYq
- Project Page: https://unifiedsdm.github.io/
BibTeX
@inproceedings{
kim2024paralinguisticsaware,
title={Paralinguistics-Aware Speech-Empowered Large Language Models for Natural Conversation},
author={Heeseung Kim and Soonshin Seo and Kyeongseok Jeong and Ohsung Kwon and Soyoon Kim and Jungwhan Kim and Jaehong Lee and Eunwoo Song and Myungwoo Oh and Jung-Woo Ha and Sungroh Yoon and Kang Min Yoo},
booktitle={The Thirty-eighth Annual Conference on Neural Information Processing Systems},
year={2024},
url={https://openreview.net/forum?id=NjewXJUDYq}
}