Beomseok-LEE's picture
Update example table front-end and logic, adding cache example
c6ba3b6 verified
raw
history blame
6.22 kB
import gradio as gr
import numpy as np
import librosa
import os
from asr.run_asr import run_asr_inference, load_asr_model
from nlu.run_nlu import run_nlu_inference, load_nlu_model
############### strings
mhubert_link = '[mHuBERT-147 model](https://huggingface.co/utter-project/mHuBERT-147)'
massive_link = '[Speech-MASSIVE dataset](https://huggingface.co/datasets/FBK-MT/Speech-MASSIVE)'
blog_post_link = '[blog post](https://huggingface.co/blog/mzboito/naver-demo-french-slu)'
title = "# DEMO: French Spoken Language Understanding using mHuBERT-147 and Speech-MASSIVE"
description=[
f"""
**Interspeech 2024 DEMO.** Cascaded SLU using {mhubert_link} and {massive_link} components.
""",
f"""This demo runs on CPU node. You may experience lagging due to network latency depending on the concurrent requests.""",
f"""For more details on the implementation, check our {blog_post_link}.""",
]
CACHE_EXAMPLES = os.getenv("CACHE_EXAMPLES") == "1"
examples = [
"resources/audios/utt_286.wav",
"resources/audios/utt_2414.wav",
"resources/audios/utt_16032.wav",
"resources/audios/utt_3060.wav",
"resources/audios/utt_1264.wav",
"resources/audios/utt_9912.wav",
"resources/audios/utt_14684.wav",
"resources/audios/utt_5410.wav",
]
transcriptions = [
"allume les lumières dans la cuisine",
"je veux commander une pizza chez michael's pizza",
"veuillez envoyer un e-mail à sally concernant la réunion de demain",
"quelles sont les nouvelles de newsource",
"mon réveil est-il réglé pour demain matin",
"olly combien de temps dois-je faire bouillir les oeufs",
"qui est le premier ministre de russie",
"lis moi les derniers gros titres du new york times"
]
intents = [
"iot_hue_lighton",
"takeaway_order",
"email_sendemail",
"news_query",
"alarm_query",
"cooking_recipe",
"qa_factoid",
"news_query"
]
slots = [
[ "Other", "Other", "Other", "Other", "Other", "house_place" ],
[ "Other", "Other", "Other", "Other", "food_type", "Other", "business_name", "business_name" ],
[ 'Other', 'Other', 'Other', 'Other', 'Other', 'Other', 'person', 'Other', 'Other', 'event_name', 'Other', 'date'],
[ 'Other', 'Other', 'Other', 'Other', 'Other', 'media_type'],
[ 'Other', 'Other', 'Other', 'Other', 'Other', 'Other', 'date', 'timeofday'],
[ 'Other', 'Other', 'Other', 'Other', 'Other', 'Other', 'Other', 'cooking_type', 'Other', 'food_type'],
[ 'Other', 'Other', 'Other', 'Other', 'Other', 'Other', 'place_name'],
[ 'Other', 'Other', 'Other', 'Other', 'Other', 'Other', 'Other', 'media_type', 'media_type', 'media_type']
]
example_list = [[example, transcription, slot, intent] for example, transcription, slot, intent in zip(examples, transcriptions, slots, intents)]
utter_ack_text = """This is an output of the European Project UTTER (Unified Transcription and Translation for Extended Reality) funded by European Union’s Horizon Europe Research and Innovation programme under grant agreement number 101070631.
For more information please visit https://he-utter.eu/"""
ack_authors = """This demo was made by [Beomseok Lee](https://mt.fbk.eu/author/blee/) and [Marcely Zanon Boito](https://sites.google.com/view/mzboito/marcely-zanon-boito)."""
eu_logo = """<img src="https://huggingface.co/spaces/naver/French-SLU-DEMO-Interspeech2024/resolve/main/resources/logos/EU_flag.jpg" width="100" height="100">"""
utter_logo = """<a href="https://he-utter.eu/" target="_blank"><img src="https://huggingface.co/spaces/naver/French-SLU-DEMO-Interspeech2024/resolve/main/resources/logos/Utter_logo.png" width="50" height="50"></a>"""
nle_logo = """<a href="https://europe.naverlabs.com/" target="_blank"><img src="https://huggingface.co/spaces/naver/French-SLU-DEMO-Interspeech2024/resolve/main/resources/logos/NAVERLABS_2_BLACK.png" width="100" height="100"></a>"""
fbk_logo = """<a href="https://mt.fbk.eu/" target="_blank"><img src="https://huggingface.co/spaces/naver/French-SLU-DEMO-Interspeech2024/resolve/main/resources/logos/FBK_logo.png" width="100" height="100"></a>"""
############### calls
def run_inference(audio_file):
print(audio_file)
audio_struct = librosa.load(audio_file, sr=16000)
print(audio_struct)
audio = {'sampling_rate': audio_struct[1], 'array': audio_struct[0]} #.astype(np.float32)
transcription = run_asr_inference(asr_model, processor, audio)
print(transcription)
structured_output = run_nlu_inference(nlu_model, tokenizer, transcription)
return structured_output
############### app
asr_model, processor = load_asr_model()
nlu_model, tokenizer = load_nlu_model()
demo = gr.Blocks(
title=title,
analytics_enabled=False,
theme=gr.themes.Base(),
)
with demo:
gr.Markdown(title)
for line in description:
gr.Markdown(line)
with gr.Row():
waveform_options = gr.WaveformOptions(sample_rate=16000)
audio_file = gr.Audio(
label="Audio file",
sources=['microphone','upload'],
type="filepath",
format='wav',
waveform_options=waveform_options,
show_download_button=False,
show_share_button=False,
max_length=20,
)
with gr.Row(visible=False):
_transcription = gr.Textbox(label="Transcription")
_slot = gr.Textbox(label="Slots")
_intent = gr.Textbox(label="Intent")
output = gr.HighlightedText(label="ASR result + NLU result")
gr.Button("Run Inference", variant='primary').click(
run_inference,
concurrency_limit=2,
inputs=audio_file,
outputs=output,
)
with gr.Row():
gr.Examples(
label="Examples(Speech-MASSIVE test utterances):",
examples=example_list,
inputs=[audio_file, _transcription, _slot, _intent],
cache_examples=CACHE_EXAMPLES,
)
gr.Markdown("# Aknowledgments")
gr.Markdown(utter_ack_text)
gr.Markdown(ack_authors)
with gr.Row():
gr.Markdown(eu_logo)
gr.Markdown(utter_logo)
gr.Markdown(nle_logo)
gr.Markdown(fbk_logo)
demo.queue()
demo.launch()