Marcus Edel commited on
Commit
b5073d7
·
1 Parent(s): a7b86dc

Add sample video to the README.

Browse files
Files changed (2) hide show
  1. README.md +5 -0
  2. README.qmd +6 -0
README.md CHANGED
@@ -1,5 +1,10 @@
1
  # WhisperFusion
2
 
 
 
 
 
 
3
 
4
  Welcome to WhisperFusion. WhisperFusion builds upon the capabilities of
5
  the [WhisperLive](https://github.com/collabora/WhisperLive) and
 
1
  # WhisperFusion
2
 
3
+ <h2 align="center">
4
+ <a href="https://www.youtube.com/watch?v=_PnaP0AQJnk"><img
5
+ src="https://img.youtube.com/vi/_PnaP0AQJnk/0.jpg" style="background-color:rgba(0,0,0,0);" height=300 alt="WhisperFusion"></a>
6
+ <br><br>Doing math with WhisperFusion: Ultra-low latency conversations with an AI chatbot<br><br>
7
+ </h2>
8
 
9
  Welcome to WhisperFusion. WhisperFusion builds upon the capabilities of
10
  the [WhisperLive](https://github.com/collabora/WhisperLive) and
README.qmd CHANGED
@@ -29,6 +29,12 @@ These steps are included in `{fname}`
29
 
30
  # WhisperFusion
31
 
 
 
 
 
 
 
32
  Welcome to WhisperFusion. WhisperFusion builds upon the capabilities of the [WhisperLive](https://github.com/collabora/WhisperLive) and [WhisperSpeech](https://github.com/collabora/WhisperSpeech) by integrating Mistral, a Large Language Model (LLM), on top of the real-time speech-to-text pipeline. WhisperLive relies on OpenAI Whisper, a powerful automatic speech recognition (ASR) system. Both Mistral and Whisper are optimized to run efficiently as TensorRT engines, maximizing performance and real-time processing capabilities.
33
 
34
  ## Features
 
29
 
30
  # WhisperFusion
31
 
32
+ <h2 align="center">
33
+ <a href="https://www.youtube.com/watch?v=_PnaP0AQJnk"><img
34
+ src="https://img.youtube.com/vi/_PnaP0AQJnk/0.jpg" style="background-color:rgba(0,0,0,0);" height=300 alt="WhisperFusion"></a>
35
+ <br><br>Doing math with WhisperFusion: Ultra-low latency conversations with an AI chatbot<br><br>
36
+ </h2>
37
+
38
  Welcome to WhisperFusion. WhisperFusion builds upon the capabilities of the [WhisperLive](https://github.com/collabora/WhisperLive) and [WhisperSpeech](https://github.com/collabora/WhisperSpeech) by integrating Mistral, a Large Language Model (LLM), on top of the real-time speech-to-text pipeline. WhisperLive relies on OpenAI Whisper, a powerful automatic speech recognition (ASR) system. Both Mistral and Whisper are optimized to run efficiently as TensorRT engines, maximizing performance and real-time processing capabilities.
39
 
40
  ## Features